WO2016066038A1 - 一种图像主体提取方法及系统 - Google Patents

一种图像主体提取方法及系统 Download PDF

Info

Publication number
WO2016066038A1
WO2016066038A1 PCT/CN2015/092505 CN2015092505W WO2016066038A1 WO 2016066038 A1 WO2016066038 A1 WO 2016066038A1 CN 2015092505 W CN2015092505 W CN 2015092505W WO 2016066038 A1 WO2016066038 A1 WO 2016066038A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
processed
area
super pixel
color
Prior art date
Application number
PCT/CN2015/092505
Other languages
English (en)
French (fr)
Inventor
安山
张洪明
刘彬
陈国成
Original Assignee
阿里巴巴集团控股有限公司
安山
张洪明
刘彬
陈国成
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 阿里巴巴集团控股有限公司, 安山, 张洪明, 刘彬, 陈国成 filed Critical 阿里巴巴集团控股有限公司
Priority to JP2017521554A priority Critical patent/JP6719457B2/ja
Publication of WO2016066038A1 publication Critical patent/WO2016066038A1/zh
Priority to US15/499,840 priority patent/US10497121B2/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/12Edge-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/143Segmentation; Edge detection involving probabilistic approaches, e.g. Markov random field [MRF] modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20076Probabilistic image processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person
    • G06T2207/30201Face

Definitions

  • the embodiments of the present application relate to the field of image processing technologies, and in particular, to an image subject extraction method.
  • the embodiment of the present application also relates to an image body extraction system.
  • online shopping has gradually become one of the main channels for people to shop, and the online shopping platform has made great progress.
  • the online shopping platform has accumulated a large amount of product image information, and how to more effectively realize the organization, analysis, retrieval and display to the consumer image information has become very important.
  • the content of the product image includes the product body and the background.
  • the traditional image subject extraction method is based on manual intervention, that is, it is necessary to artificially frame and set the divided regions. This method is inefficient and is not suitable for the amount of pictures on the Internet. Therefore, it is necessary to design an automatic extraction method of a product image body, which can accurately extract specific content in the image.
  • the embodiment of the present application provides an image subject extraction method for accurately determining and extracting content of an image body, and the method includes:
  • the image in the body region is subjected to foreground object extraction processing as a foreground image, and the extracted image is taken as a main portion of the image to be processed.
  • an image subject extraction system including:
  • a determining module configured to determine whether there is an area to be identified corresponding to the specified feature in the image to be processed
  • a rectangular frame obtaining module configured to determine, according to a preset body region feature parameter, a coordinate and a size of the to-be-identified region, a body region including the body portion of the image to be processed, when the region to be identified is present;
  • an extraction module configured to perform foreground object extraction processing on the image in the body region as a foreground image, and use the extracted image as a main portion of the image to be processed.
  • the main body area of the image to be processed is obtained in different manners, in the main body area.
  • a preset algorithm is executed in the image, and an image corresponding to the foreground region extracted by the algorithm is taken as a main portion of the image to be processed. Therefore, on the basis of realizing the automatic extraction of the image body, the extracted body is accurate and reliable, and the processing efficiency is improved.
  • FIG. 1 is a schematic flowchart diagram of an image subject extraction method according to the present application.
  • FIG. 2 is a schematic diagram of super pixel segmentation results of a product image in a specific embodiment of the present application
  • FIG. 3 is a schematic diagram of a head and shoulder detection result of a product image in a specific embodiment of the present application.
  • FIG. 4 is a schematic diagram showing the result of estimating the estimated rectangle according to the head and shoulders in the specific embodiment of the present application;
  • Figure 5a is a schematic view showing the original product of the head and shoulders not detected in the specific embodiment of the present application.
  • 6a is a schematic diagram of the saliency of a binarized product in a specific embodiment of the present application.
  • 6b is a schematic diagram of a rectangle enclosing each polygon in a specific embodiment of the present application.
  • 6c is a schematic diagram of a peripheral rectangular frame enclosing all rectangles in a specific embodiment of the present application.
  • FIG. 7a is a schematic diagram of a final subject extraction result of a product image in which a head and shoulder is detected in a specific embodiment of the present application;
  • 7b is a schematic diagram of the final body extraction result of the product image of the head and shoulder that is not detected in the specific embodiment of the present application;
  • FIG. 8 is a schematic diagram of an image subject extraction effect in a specific embodiment of the present application.
  • FIG. 9 is another schematic diagram of an image subject extraction effect in a specific embodiment of the present application.
  • FIG. 10 is another schematic diagram of an image subject extraction effect in a specific embodiment of the present application.
  • FIG. 11 is another schematic diagram of an image subject extraction effect in a specific embodiment of the present application.
  • FIG. 12 is a schematic structural diagram of an image extraction system according to the present application.
  • an image subject extraction method proposed by the embodiment of the present application determines that there is a corresponding to the specified feature in the image to be processed. After identifying the area, determining the body area including the body part according to the preset body area feature parameter and the coordinate information of the area to be identified, and performing foreground object extraction processing on the image in the body area to extract the body part, and realizing the image body automatically On the basis of the extraction, the extracted body is accurate and reliable, and the processing efficiency is improved.
  • a schematic flowchart of an image subject extraction method includes the following steps:
  • the embodiment of the present application may assign an associated feature to the body portion in advance, and perform a further fine extraction operation by detecting the to-be-identified region corresponding to the specified feature.
  • the embodiment of the present application preferably adopts a pre-processing mechanism for filtering a to-be-processed picture for a large number of original pictures, and filters the image by determining whether the image is already a solid color background. It is better to display the image of the subject and reduce the number of pictures that need to be extracted. Therefore, the following operations can be performed before this step:
  • a super pixel is a small area in a picture consisting of a series of pixels with similar positions, such as color, brightness, texture, etc., which mostly retain effective information for further image segmentation, and generally do not destroy objects in the image. Boundary information.
  • a color model (Lab) may be preferably used to measure the average of the colors of the respective superpixel blocks.
  • this step can further process the Lab color average of the super pixel blocks on the four sides of the image to be clustered according to preset parameters, and the number of data points including the cluster processing is included.
  • the color value represented by the cluster center is used as the background color of the image to be processed.
  • a ratio of the number of pixels included in the super pixel block that matches the color threshold of the background color to the total number of pixels of the image to be processed may be taken as the solid color background score.
  • After obtaining the solid color background score it may be further determined whether the solid color background score is higher than a preset threshold. If the judgment result is yes, it is confirmed that the image to be processed is a solid color background; otherwise, it may continue to determine whether the image to be processed exists and the specified feature. Corresponding area to be identified.
  • the present application uses SLIC (Simple linear iterative clustering) method for super pixel segmentation, first converting image data from initial color space to CIELAB color space, and normalizing pixel position to form 5-dimensional data.
  • SLIC Simple linear iterative clustering
  • K cluster centers C k [l k ,a k ,b k ,x k ,y k ] T that specify the grid distance S
  • m is the tightness factor.
  • the number of super pixels is set to 200, and the tightness factor is set to 10.
  • the product image can be divided into nearly 200 super pixel blocks, and the super pixel segmentation diagram is as shown in FIG. 2 .
  • is a preset color threshold. If the color average of the super pixel block is less than ⁇ from the background color of the image, it is determined that the super pixel block color is the same as the image background color.
  • the threshold is set to 100.
  • step B1 detecting whether there is an area corresponding to the specified feature in the image to be processed, if yes, proceeding to step B1); if not, ending the process.
  • the step is to determine the specified feature associated with the human head and shoulder area, and after determining the area corresponding to the face feature in the image to be processed in the head and shoulder area of the head and shoulder area of the human body, further Detecting, according to the upper body detection algorithm and the head and shoulder contour detector, an area corresponding to the shoulder area of the human body in the image to be processed, and verifying the detected area based on the head and shoulder apparent model, and confirming whether the area is The area to be identified corresponding to the head and shoulder area of the human body.
  • the head and shoulder contour detection in the above steps can be trained using the AdaBoost algorithm combined with the HoG feature, and the head and shoulder apparent model can be established based on the deep network model, and the significance detection uses the global unique method combined with the color space distribution.
  • the method is carried out.
  • those skilled in the art can further adopt other algorithms or optimization models, which are included in the protection scope of the present application.
  • a women's image is used as a training set, and a face detection algorithm can be used to detect a face in a women's image.
  • the upper body detection algorithm is used to detect the upper body head, neck, left shoulder, and right shoulder position of the human body in the picture. Inaccurate pictures can be detected by manual culling as needed.
  • the AdaBoost algorithm is used in combination with the HoG feature to train the head and shoulder detector. Compared with the Haar feature used in the face, the HoG feature is more focused on the description of the contour, and is more suitable for head and shoulder detection.
  • the 13-level AdaBoost classifier was trained with 5W positive samples and 5W negative samples, achieving a higher detection rate and a lower false alarm rate.
  • AdaBoost AdaBoost check
  • the detector Since the HoG feature only pays attention to the performance of the contour of the target object, the actual use of AdaBoost check When the detector is used, there will be a lot of false alarms with similar head and shoulder contours.
  • a head-shoulder model based on the deep network model can be added to detect AdaBoost according to the apparent features of the target. As a result, further verification is carried out.
  • the combination of the two has greatly reduced the false alarm rate without significantly reducing the detection rate. Specifically, the head and shoulder detection results are shown in FIG.
  • the area to be identified exists, determine a body area that includes the body part of the image to be processed according to a preset body area feature parameter, a coordinate and a size of the to-be-identified area.
  • the rectangular frame surrounding the human body area may be directly estimated based on the preset body area characteristic parameter, and the running time of the algorithm is saved.
  • the preset regional parameters can be flexibly set according to the previous empirical statistics, and the predicted effect can be achieved. The specific value does not affect the protection scope of the present application.
  • the coordinates (RectX, RectY) and the side length Length of the upper left vertex of the head and shoulders box are obtained by S101.
  • the best parameters of the rectangular frame surrounding the human body area are estimated by experimenting with a large number of clothing product pictures including the human head and shoulders.
  • a schematic diagram of estimating the estimated rectangle according to the head and shoulders is shown in FIG. 4, and the parameters of the rectangular frame are set as follows:
  • Top left corner vertex X coordinate RectX-0.5*Length(15)
  • width Length*2(17)
  • the embodiment of the present application may further perform saliency detection on the image to be processed, and determine the image to be processed according to the saliency value of each pixel in the image to be processed after the detection.
  • the main area may be further perform saliency detection on the image to be processed, and determine the image to be processed according to the saliency value of each pixel in the image to be processed after the detection.
  • the subject area can be determined by a significance value, as follows:
  • the Global Uniqueness method and the Color Spatial Distribution method can be used to detect the saliency of the product image.
  • the image color is clustered and expressed by a Gaussian Mixture Model (GMM).
  • GMM Gaussian Mixture Model
  • the color I s of each pixel is represented as a weighted combination of a plurality of GMM components, and the probability that it belongs to a certain component c is expressed as:
  • the Global Uniqueness method represents the uniqueness of the global component c i as a weighted color contrast relative to all other components:
  • D(c i , c j ) is the spatial distance between the centers of the two GMM components c i and c j .
  • the Color Spatial Distribution method first calculates the horizontal space variance of the cluster component C:
  • x h is the x coordinate of the pixel x
  • spatial variance of the cluster component C is:
  • V(C) V h (C)+V V (C) (13)
  • the vertical space variance V v (C) is defined similarly to the horizontal space variance V h (C).
  • the final color space distribution value is:
  • the saliency value of the image is detected using the above steps, and the saliency map of the saliency value of the binarized product image is set, and the binarization threshold is set to 0.1;
  • the outline of all the color patches in the image and the area of the enclosed area, the area smaller than the threshold is removed, and the smaller color block in the binarized image is cleared, where the threshold is set to 5000; and all the remaining color patches in the image are found.
  • Contour use polygon to approximate the outline of each color block; find the rectangle surrounding each polygon, and the outer rectangle surrounding all the rectangles, the relevant diagrams are shown in Figure 6a, Figure 6b and Figure 6c respectively.
  • S104 Perform foreground object extraction processing on the image in the body area as a foreground image, and use the extracted image as a main part of the image to be processed.
  • the embodiment of the present application proposes an optimization mechanism for the subject image, and the steps are as follows:
  • the obtained body region can be more accurate and reliable, and once the body region is determined, the image in the body region is set as the foreground image, and the GrabCut algorithm is initialized.
  • GrabCut is an effective interactive segmentation algorithm for extracting foreground targets from complex backgrounds. It uses Gaussian mixture model to describe the distribution of pixels, and iterative estimation method is used to minimize energy. It is one of the most excellent and practical algorithms. . Therefore, in the specific embodiment, the present application performs a GrabCut algorithm on the foreground image to obtain the main part of the image to be processed, but this is not the only way. On the basis of this, those skilled in the art can select other extraction algorithms to obtain the main part as well. All belong to the scope of protection of this application.
  • the present application first performs super pixel segmentation on the product image in this step.
  • the SLIC Simple linear iterative clustering
  • the number of super pixels is set to 20, and the tightness factor is set to 10.
  • the product image can be divided into nearly 20 super pixel blocks; the average significance value of all the pixels in each super pixel block is calculated; if the average significance value of the super pixel block is higher than the set threshold, the super pixel block is set
  • the mask value of each pixel image is non-zero, where the threshold is set to 0.6; a peripheral rectangular frame that surrounds the mask image other than 0 pixels is obtained.
  • the image in the rectangular region is then used as the foreground image, and the image outside the rectangular region is used as the background image to initialize the GrabCut algorithm.
  • the GrabCut algorithm utilizes the Gaussian mixture model to achieve image segmentation through continuous iteration of segmentation estimation and model parameter learning.
  • the foreground partial image obtained by the GrabCut algorithm is used as the commodity body.
  • the product image of the head and shoulders and the product image of the head and shoulder that are not detected are shown in Fig. 7a and Fig. 7b, respectively.
  • the schematic diagram of the image body extraction effect is shown in FIG. 8, FIG. 9, FIG. 10 and FIG.
  • the embodiment of the present application further provides an image subject extraction system, as shown in FIG. 12, including:
  • the determining module 1201 is configured to determine whether there is an area to be identified corresponding to the specified feature in the image to be processed;
  • a rectangular frame obtaining module 1202 configured to determine, according to a preset body region feature parameter, a coordinate and a size of the to-be-identified region, a body region including the body portion of the image to be processed, when the region to be identified is present;
  • the extracting module 1203 is configured to perform foreground object extraction processing on the image in the body region as a foreground image, and use the extracted image as a main portion of the image to be processed.
  • a segmentation module configured to perform superpixel segmentation processing on the image to be processed, and determine an average value of colors of each super pixel block after processing
  • a background color module configured to determine a background color of the image to be processed according to an average value of colors of the super pixel blocks
  • a solid color determining module configured to determine a solid color background score of the image to be processed, and determine, according to the solid color background score, whether the image to be processed is a solid color background, and the solid color background score is a color threshold corresponding to the background color
  • the determining module specifically includes:
  • a detecting submodule configured to detect whether an area corresponding to the specified feature exists in the image to be processed
  • a confirmation sub-module configured to perform contour detection on an area corresponding to the specified feature in the image to be processed when the detection sub-module confirms that an area corresponding to the specified feature exists, and according to the specified feature
  • the parameterized apparent model verifies the detected area and confirms whether the area is the area to be identified corresponding to the specified feature.
  • the specified feature is specifically a human head and shoulder area
  • the confirmation sub-module is specifically configured to: the image to be processed according to an upper body detection algorithm and a head and shoulder contour detector. The area corresponding to the head and shoulder area of the human body is detected, and the detected area is verified based on the head and shoulder apparent model.
  • the head and shoulder contour detector is trained by using an AdaBoost algorithm combined with a HoG feature, which is built based on a deep network model.
  • a saliency detection module configured to perform saliency detection on the image to be processed when the area to be identified is not present, and determine the to-be-processed according to the saliency value of each pixel in the image to be processed after the detection The body area of the image.
  • the saliency detection module is specifically configured to perform binarization processing on the saliency map formed by the saliency value, and to perform a smaller color on the binarized image.
  • the block performs a clearing process to find a contour line of the color block in the image after the clearing process, and approximates the contour line of each of the color blocks by using a polygon, and surrounds each of the polygons with a rectangle, and takes a peripheral rectangular frame surrounding each of the rectangles as a rectangle The body area.
  • a rectangular frame improvement module configured to perform super pixel segmentation processing on the image to be processed, and determine an average significance value of all pixels in each super pixel block after processing, and sequentially determine whether an average significance value of each of the super pixel blocks is Higher than a preset threshold, and setting an image mask value of each pixel in the super pixel block to be non-zero when the average significance value of the super pixel block is higher than a set threshold, and using the enclosing mask image to be non-zero pixels
  • the peripheral rectangular frame updates the body area.
  • the extraction module is specifically configured to set an image in the body region as a foreground image, and initialize a GrabCut algorithm, and execute the GrabCut algorithm on the foreground image.
  • the main area of the image to be processed is obtained in different manners according to whether there are different situations of the to-be-identified area corresponding to the specified feature in the image to be processed.
  • the preset algorithm is executed in the image of the domain and the body region, and the image corresponding to the foreground region extracted by the algorithm is taken as the main portion of the image to be processed. Therefore, on the basis of realizing the automatic extraction of the image body, the extracted body is accurate and reliable, and the processing efficiency is improved.
  • the present application can be implemented by hardware, or by software plus a necessary general hardware platform.
  • the technical solution of the present application may be embodied in the form of a software product, which may be stored in a non-volatile storage medium (which may be a CD-ROM, a USB flash drive, a mobile hard disk, etc.), including several The instructions are for causing a computer device (which may be a personal computer, server, or network device, etc.) to perform the methods described in various implementation scenarios of the present application.
  • modules in the apparatus in the implementation scenario may be distributed in the apparatus for implementing the scenario according to the implementation scenario description, or may be correspondingly changed in one or more devices different from the implementation scenario.
  • the modules of the above implementation scenarios may be combined into one module, or may be further split into multiple sub-modules.

Abstract

本申请公开了一种图像主体提取方法,根据待处理图像中是否存在与指定特征对应的待识别区域的不同情况,分别采取不同的方式获取待处理图像的主体区域,在主体区域的图像内执行预设的算法,将通过算法提取的前景区域对应的图像作为待处理图像的主体部分。从而在实现了图像主体自动提取的基础上,使提取出的主体精确、可靠,提高了处理效率。

Description

一种图像主体提取方法及系统 技术领域
本申请实施例涉及图像处理技术领域,特别涉及一种图像主体提取方法。本申请实施例同时还涉及一种图像主体提取系统。
背景技术
近年来随着技术的发展和人们观念的转变,在线购物逐渐成为人们购物的主要渠道之一,网购平台取得了长足的发展。在这种背景下,网购平台积累了大量的商品图像信息,如何更有效的实现对商品图像信息的组织、分析、检索和向消费者展示已经变得十分重要。
商品图像的内容包括商品主体和背景,当用户上传一幅商品图像并期望搜索与该图相同或相似的商品时,用户更关注于商品本身,背景信息的存在可能影响商品搜索的结果。因此,提取商品图像中的商品主体成为非常重要的一项工作。传统的图像主体提取方法基于人工干预,即需要人为对被分割的区域进行框选和设定,这种方法效率低下,不适于互联网上亿的图片量。因此,当下需要设计一种商品图像主体的自动提取方法,能够对图像中特定的内容进行准确的提取。
发明内容
本申请实施例提供了一种图像主体提取方法,用以实现对图像主体内容的精准判断以及提取,该方法包括:
判断待处理图像中是否存在与指定特征对应的待识别区域;
若存在所述待识别区域,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域;
将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
相应地,本申请实施例还提出了一种图像主体提取系统,包括:
判断模块,用于判断待处理图像中是否存在与指定特征对应的待识别区域;
矩形框获取模块,用于在存在所述待识别区域时,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域;
提取模块,用于将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
由此可见,通过应用本申请实施例的技术方案,根据待处理图像中是否存在与指定特征对应的待识别区域的不同情况,分别采取不同的方式获取待处理图像的主体区域,在主体区域的图像内执行预设的算法,将通过算法提取的前景区域对应的图像作为待处理图像的主体部分。从而在实现了图像主体自动提取的基础上,使提取出的主体精确、可靠,提高了处理效率。
附图说明
图1为本申请提出的一种图像主体提取方法的流程示意图;
图2为本申请具体实施例中商品图像的超像素分割结果示意图;
图3为本申请具体实施例中商品图像的头肩检测结果示意图;
图4为本申请具体实施例中根据头肩检测估计矩形的结果示意图;
图5a为本申请具体实施例中未检测到头肩的商品原图示意图;
图5b为本申请具体实施例基于图5a的商品显著性示意图;
图6a为本申请具体实施例中二值化的商品显著性示意图;
图6b为本申请具体实施例中包围各多边形的矩形示意图;
图6c为本申请具体实施例中包围所有矩形的外围矩形框示意图;
图7a为本申请具体实施例中检测到头肩的商品图像最终主体提取结果示意图;
图7b为本申请具体实施例中未检测到头肩的商品图像最终主体提取结果示意图;
图8为本申请具体实施例中图像主题提取效果示意图;
图9为本申请具体实施例中图像主题提取效果另一示意图;
图10为本申请具体实施例中图像主题提取效果另一示意图;
图11为本申请具体实施例中图像主题提取效果另一示意图;
图12为本申请提出的一种图像提取系统的结构示意图。
具体实施方式
如背景技术所述,现有的图像提取技术无法兼顾图像提取的效率以及准确度,为此本申请实施例提出的一种图像主体提取方法,在判断待处理图像中存在与指定特征对应的待识别区域之后,结合预设的主体区域特征参数以及待识别区域的坐标信息确定包含主体部分的主体区域,并针对主体区域中的图像进行前景目标提取处理以提取主体部分,在实现了图像主体自动提取的基础上,使提取出的主体精确、可靠,提高了处理效率。
如图1所示,为本申请实施例提出的一种图像主体提取方法的流程示意图,包括以下步骤:
S101,判断所述待处理图像中是否存在与指定特征对应的待识别区域。
基于所希望提取的图像主体部分,本申请实施例可以预先为该主体部分指定一个相关联的特征,通过检测与该指定特征对应的待识别区域来进行进一步的精细提取操作。
由于原始图片的数量一般都是非常庞大的,为了能够进一步地提高效率,本申请实施例针对大量的原始图片优选采取预处理机制来筛选待处理图片,通过判断图像是否已经为纯色背景,过滤已较好展现主体的图像,减少需提取主体的图片数目。因此,在本步骤之前可预先执行以下操作:
a).将所述待处理图像进行超像素分割处理,并确定处理后各超像素块的颜色的平均值。
超像素为图像中由一系列位置相邻且颜色、亮度、纹理等特征相似的像素点组成的小区域,这些小区域大多保留了进一步进行图像分割的有效信息,并且一般不会破坏图像中物体的边界信息。在通过超像素图像分割将图像分割为不含强边缘的超像素区域之后,可以优选采用颜色模型(Lab)来衡量各超像素块的颜色的平均值。
b).确定所述待处理图像的纯色背景分数,并根据所述纯色背景分数判断所述待处理图像是否为纯色背景。
基于上一步骤中所使用的Lab颜色,本步骤可进一步对待处理图像四边的超像素块的Lab颜色平均值根据预设的参数进行聚类处理,将包含聚类处理后数据点数目较多的聚类中心所代表的颜色值作为所述待处理图像的背景色。优选地,可将符合所述背景色的颜色阈值的超像素块中所包含的像素数与所述待处理图像的总像素数的比值作为所述纯色背景分数。在得到纯色背景分数之后,可进一步判断纯色背景分数是否高于预设的阈值,如果判断结果为是,则确认待处理图像为纯色背景,否则,可继续判断待处理图像中是否存在与指定特征对应的待识别区域。
在具体实施例中,本申请采用SLIC(Simple linear iterative clustering)方法进行超像素分割,首先将图像数据由初始色彩空间转为CIELAB色彩空间,加上归一化后像素点位置,形成5维数据空间;选择规定栅格距离S的K个聚类中心Ck=[lk,ak,bk,xk,yk]T;计算聚类中心周围2S×2S范围内数据点到聚类中 心的距离,将数据点划入最近的聚类中。
在以上过程中,距离Ds的计算公式为:
Figure PCTCN2015092505-appb-000001
Figure PCTCN2015092505-appb-000002
Figure PCTCN2015092505-appb-000003
其中,m为紧密因数。本实例中,设置超像素个数为200,紧密因数设为10,可将商品图像划分为接近200个超像素块,超像素分割示意图如图2所示。
之后继续计算各超像素块的Lab颜色平均值:
Figure PCTCN2015092505-appb-000004
取出靠近图像四边的超像素块的Lab颜色平均值,进行k=2,数据维度为3的kmeans聚类,取包含数据点数目较多的聚类中心代表的颜色值作为图像背景色(LB,aB,bB)。
计算符合下式的超像素块中包含的像素数
Figure PCTCN2015092505-appb-000005
其中,θ为预设的颜色阈值,若超像素块的颜色平均值与图像背景色的距离小于θ,则判断该超像素块颜色与图像背景色相同。此处阈值设为100。
最后计算图像的纯色背景分数:
pr=pθ/pall    (6)
其中,pall为图像的总像素数。根据实验确定,若纯色背景分数大于0.3,则商品图像为纯色背景。若为纯色背景则不用提取商品主体。
需要说明的是,以上通过Lab颜色以及聚类处理以获取纯色背景分数仅为本申请所提出的一种优选的实施方式,在上述处理完成之后,即可根据技术人员所指定的指定特征进行识别,步骤如下:
A1).检测所述待处理图像中是否存在与所述指定特征对应的区域,若存在,则转至步骤B1);若不存在,则结束处理。
B1).对所述待处理图像中与所述指定特征对应的区域进行轮廓检测,并根据与所述指定特征对应的参数化表观模型对已检测的区域进行验证,确认所述区域是否为与所述指定特征对应的待识别区域。
在大多数具体应用场景中,人体图像中着装主体的提取往往是最重要的。因此,在优选实施例中,本步骤将与之关联的指定特征确定为人体头肩区域,在人体头肩区域人体头肩区域判断待处理图像中存在与人脸特征对应的区域之后,可进一步根据上半身检测算法以及头肩轮廓检测器对所述待处理图像中与人体头肩区域对应的区域进行检测,并基于头肩表观模型对已检测的区域进行验证,确认所述区域是否为与人体头肩区域对应的待识别区域。
优选地,以上述步骤中的头肩轮廓检测可以使用AdaBoost算法结合HoG特征来训练得到,并且头肩表观模型则可基于深度网络模型建立,而显著性检测采用全局唯一性方法结合色彩空间分布方法来进行,当然本领域技术人员也可在此基础上进一步采取其他的算法或是优化模型,这些都包含在本申请的保护范围之内。
在本申请具体的实施例中,应用于服装类目的商品图片处理。例如,使用女装图像作为训练集,并可使用人脸检测算法,检测女装图像中的人脸。对于检测出单个人脸的图片,使用上半身检测算法,检测图片中的人体上半身头、颈、左肩、右肩位置。并可根据需要进一步通过人工剔除检测不准确的图片。之后继续使用AdaBoost算法结合HoG特征训练头肩检测器,相比人脸中使用的Haar特征,HoG特征更偏重于轮廓的描述,比较适用于头肩检测。采用5W正样本和5W负样本训练了13级AdaBoost分类器,取得了较高的检测率和比较低的虚警率。
由于HoG特征只是关注目标物体轮廓的表现,在实际使用AdaBoost检 测器的时候会有不少的和头肩轮廓相似的虚警被检测出来,为了进一步减少虚警,加入基于深度网络模型的头肩表观模型,可以根据目标的表观特征对AdaBoost的检测结果进行进一步的验证,在实际系统中,二者的结合在没有明显降低检测率的条件下,虚警率大大降低,具体地,头肩检测结果示意图如图3所示。
S102,若存在所述待识别区域,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域。
本步骤在S101中判断商品图像中存在人体头肩区域后,可基于预设的主体区域特征参数对检测到头肩的图片直接估计包围人体区域的矩形框,节省算法运行时间。其中预设的区域参数可以根据以往的经验统计数据灵活设置,以可达到预估效果为准,具体的取值的不同不影响本申请的保护范围。
在具体实施例中,由S101可获得头肩方框左上顶点的坐标(RectX,RectY)以及边长Length。通过对大量包含人体头肩的衣服商品图片实验,估计包围人体区域的矩形框的最佳参数。根据头肩检测估计矩形的示意图如图4所示,矩形框的参数设置为:
左上角顶点X坐标=RectX-0.5*Length(15)
左上角顶点Y坐标=RectY+0.7*Length(16)
矩形框的宽width=Length*2(17)
矩形框的高height=Length*10(18)。
在通过S101判断不存在与指定特征对应的待识别区域时,本申请实施例还可进一步对待处理图像进行显著性检测,并根据检测后待处理图像中各像素点的显著性值确定待处理图像的主体区域。
在优选的实施例中,可通过显著性值确定主体区域,具体步骤如下:
a).对由所述显著性值所构成的显著性图进行二值化处理;
b).对二值化处理后的图像中的较小的色块进行清除处理;
c).查找清除处理后的图像中色块的轮廓线,并使用多边形近似各所述色块的轮廓线;
d).利用矩形包围各所述多边形,将包围各所述矩形的外围矩形框作为所述主体区域。
在此需要指出的是,检测商品图像各像素点显著性值的方法有多种,以上通过全局唯一性方法(Global Uniqueness)结合色彩空间分布方法(Color Spatial Distribution)进行商品图像显著性检测为例进行说明。本申请对此不做限定,本领域技术人员也可使用其他可替代的方法包括基于直方图对比度的方法(HC),基于区域对比度的方法(RC),上下文感知方法(CA)和频域调谐方法(FT)等参考以上步骤实现主体区域的识别,这都在本申请的保护范围之内。
在具体的实施例中,可以采用全局唯一性方法(Global Uniqueness)结合色彩空间分布方法(Color Spatial Distribution)进行商品图像显著性检测,首先对图像颜色进行聚类并使用高斯混合模型(GMM)表示。每个像素的颜色Is表示为多个GMM分量的加权组合,其属于某个分量c的概率表示为:
Figure PCTCN2015092505-appb-000006
两个GMM分量ci和cj的相关性表示为:
Figure PCTCN2015092505-appb-000007
全局唯一性方法(Global Uniqueness)将全局分量ci的唯一性表示为相对所有其他分量的加权色彩对比度:
Figure PCTCN2015092505-appb-000008
其中,D(ci,cj)是两个GMM分量ci和cj中心的空间距离。
色彩空间分布方法(Color Spatial Distribution)首先计算聚类分量C的水平空间方差:
Figure PCTCN2015092505-appb-000009
Figure PCTCN2015092505-appb-000010
|X|c=∑xp(C|Ix)   (12)
其中,xh是像素x的x坐标,聚类分量C的空间方差为:
V(C)=Vh(C)+VV(C)    (13)
垂直空间方差Vv(C)定义与水平空间方差Vh(C)相似。最终的色彩空间分布值为:
S(C)=(1-V(C))×(1-D(C))   (14)
其中,D(C)=∑xp(C|Ix)dx,最终生成的显著性检测结果以及未检测到头肩的商品原图分别如图5b、图5a所示。
具体地,对于未检测出头肩的图像,使用以上步骤检测图像的显著性值,二值化商品图像的显著性值构成的显著性图,二值化阈值设为0.1;求出二值化后图像中所有色块的轮廓及包围的区域面积,去掉小于阈值的区域,用于清除二值化后图像中较小的色块,此处阈值设为5000;找出图像中剩余所有色块的轮廓线;使用多边形近似各个色块的轮廓线;求出包围各多边形的矩形,以及包围所有矩形的外围矩形框,相关示意图分别如图6a、图6b以及图6c所示。
S104,将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
为了使主体区域更加地准确,在执行本步骤之前,本申请实施例提出了针对主体图像的优化机制,步骤如下:
a).依次判断各所述超像素块的平均显著性值是否高于预设的阈值;
b).若所述超像素块的平均显著性值高于设定阈值,则设置所述超像素块 中各像素图像掩码值为非0;
c).根据包围掩码图像非0像素的外围矩形框,更新所述主体区域。
通过以上过程处理后,所得到的主体区域能够更加精确可靠,而一旦在确定主体区域之后,将所述主体区域中的图像设置为前景图像,并初始化GrabCut算法。其中,GrabCut是一种有效的从复杂背景中提取前景目标的交互式分割算法它利用高斯混合模型来描述像素的分布,并采用迭代估计法实现能量最小化,是目前比较优秀实用的算法之一。因此本申请在具体实施例中对前景图像执行GrabCut算法来得到待处理图像的主体部分,但这并非唯一的方式,在此基础上本领域技术人员可以选择其他的提取算法同样得到主体部分,这些都属于本申请的保护范围。
在具体的实施例中,本申请在本步骤首先对商品图像进行超像素分割,此处使用SLIC(Simple linear iterative clustering)方法进行超像素分割,超像素个数设为20,紧密因数设为10,可将商品图像划分为接近20个超像素块;计算每个超像素块中所有像素的平均显著性值;若超像素块的平均显著性值高于设定阈值,则设置该超像素块中各像素图像掩码值为非0,此处阈值设为0.6;求出包围掩码图像非0像素的外围矩形框。随后将所述矩形区域内的图像作为前景图像,所述矩形区域外的图像作为背景图像,初始化GrabCut算法。GrabCut算法利用高斯混合模型,通过不断进行分割估计和模型参数学习的交互迭代实现图像分割。将GrabCut算法求出的前景部分图像作为商品主体。检测到头肩的商品图像和未检测到头肩的商品图像最终主体提取结果分别如图7a、图7b所示。图像主体提取效果示意图如图8、图9、图10及图11所示。
与本申请实施例提供的对图像进行解码的方法相对应,本申请实施例还提供了一种图像主体提取系统,如图12所示,包括:
判断模块1201,用于判断待处理图像中是否存在与指定特征对应的待识别区域;
矩形框获取模块1202,用于在存在所述待识别区域时,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域;
提取模块1203,用于将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
在具体的应用场景中,还包括:
分割模块,用于将所述待处理图像进行超像素分割处理,并确定处理后各超像素块的颜色的平均值;
背景色模块,用于根据所述各超像素块的颜色的平均值确定所述待处理图像的背景色;
纯色判断模块,用于确定所述待处理图像的纯色背景分数,并根据所述纯色背景分数判断所述待处理图像是否为纯色背景,所述纯色背景分数为符合所述背景色的颜色阈值的超像素块中所包含的像素数与所述待处理图像的总像素数的比值。
在具体的应用场景中,所述判断模块,具体包括:
检测子模块,用于检测所述待处理图像中是否存在与所述指定特征对应的区域;
确认子模块,用于所述检测子模块确认存在与所述指定特征对应的区域时,对所述待处理图像中与所述指定特征对应的区域进行轮廓检测,并根据与所述指定特征对应的参数化表观模型对已检测的区域进行验证,确认所述区域是否为与所述指定特征对应的待识别区域。
在具体的应用场景中,所述指定特征具体为人体头肩区域,所述确认子模块,具体用于根据上半身检测算法以及头肩轮廓检测器对所述待处理图像 中与人体头肩区域对应的区域进行检测,并基于头肩表观模型对已检测的区域进行验证。
在具体的应用场景中,所述头肩轮廓检测器是通过使用AdaBoost算法结合HoG特征来训练得到的,所述头肩表观模型基于深度网络模型建立。
在具体的应用场景中,还包括:
显著性检测模块,用于在不存在所述待识别区域时,对所述待处理图像进行显著性检测,并根据检测后所述待处理图像中各像素点的显著性值确定所述待处理图像的主体区域。
在具体的应用场景中,所述显著性检测模块,具体用于对由所述显著性值所构成的显著性图进行二值化处理,对二值化处理后的图像中的较小的色块进行清除处理,查找清除处理后的图像中色块的轮廓线,并使用多边形近似各所述色块的轮廓线,利用矩形包围各所述多边形,将包围各所述矩形的外围矩形框作为所述主体区域。
在具体的应用场景中,还包括:
矩形框改进模块,用于对所述待处理图像进行超像素分割处理,并确定处理后各超像素块中所有像素的平均显著性值,依次判断各所述超像素块的平均显著性值是否高于预设的阈值,并在所述超像素块的平均显著性值高于设定阈值时设置所述超像素块中各像素图像掩码值为非0,利用包围掩码图像非0像素的外围矩形框更新所述主体区域。
在具体的应用场景中,所述提取模块,具体用于将所述主体区域中的图像设置为前景图像,并初始化GrabCut算法,对所述前景图像执行所述GrabCut算法。
通过应用本申请的技术方案,根据待处理图像中是否存在与指定特征对应的待识别区域的不同情况,分别采取不同的方式获取待处理图像的主体区 域,再主体区域的图像内执行预设的算法,将通过算法提取的前景区域对应的图像作为待处理图像的主体部分。从而在实现了图像主体自动提取的基础上,使提取出的主体精确、可靠,提高了处理效率。
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到本申请可以通过硬件实现,也可以借助软件加必要的通用硬件平台的方式来实现。基于这样的理解,本申请的技术方案可以以软件产品的形式体现出来,该软件产品可以存储在一个非易失性存储介质(可以是CD-ROM,U盘,移动硬盘等)中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)执行本申请各个实施场景所述的方法。
本领域技术人员可以理解附图只是一个优选实施场景的示意图,附图中的模块或流程并不一定是实施本申请所必须的。
本领域技术人员可以理解实施场景中的装置中的模块可以按照实施场景描述进行分布于实施场景的装置中,也可以进行相应变化位于不同于本实施场景的一个或多个装置中。上述实施场景的模块可以合并为一个模块,也可以进一步拆分成多个子模块。
上述本申请序号仅仅为了描述,不代表实施场景的优劣。
以上公开的仅为本申请的几个具体实施场景,但是,本申请并非局限于此,任何本领域的技术人员能思之的变化都应落入本申请的保护范围。

Claims (18)

  1. 一种图像主体提取方法,其特征在于,包括:
    判断待处理图像中是否存在与指定特征对应的待识别区域;
    若存在所述待识别区域,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域;
    将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
  2. 如权利要求1所述的方法,其特征在于,在判断待处理图像中是否存在与指定特征对应的待识别区域之前,还包括:
    将所述待处理图像进行超像素分割处理,并确定处理后各超像素块的颜色的平均值;
    根据所述各超像素块的颜色的平均值确定所述待处理图像的背景色;
    确定所述待处理图像的纯色背景分数,并根据所述纯色背景分数判断所述待处理图像是否为纯色背景,所述纯色背景分数为符合所述背景色的颜色阈值的超像素块中所包含的像素数与所述待处理图像的总像素数的比值。
  3. 如权利要求1所述的方法,其特征在于,所述判断待处理图像中是否存在与指定特征对应的待识别区域,具体为:
    检测所述待处理图像中是否存在与所述指定特征对应的区域;
    若存在,对所述待处理图像中与所述指定特征对应的区域进行轮廓检测,并根据与所述指定特征对应的参数化表观模型对已检测的区域进行验证,确认所述区域是否为与所述指定特征对应的待识别区域。
  4. 如权利要求3所述的方法,其特征在于,所述指定特征具体为人体头肩区域,所述对所述待处理图像中与所述指定特征对应的区域进行轮廓检测,并根据与所述指定特征对应的参数化表观模型对已检测的区域进行验证,具体为:
    根据上半身检测算法以及头肩轮廓检测器对所述待处理图像中与人体头肩区域对应的区域进行检测,并基于头肩表观模型对已检测的区域进行验证。
  5. 如权利要求4所述的方法,其特征在于,
    所述头肩轮廓检测器是通过使用AdaBoost算法结合HoG特征来训练得到的,所述头肩表观模型基于深度网络模型建立。
  6. 如权利要求1所述的方法,其特征在于,还包括:
    若不存在所述待识别区域,对所述待处理图像进行显著性检测,并根据检测后所述待处理图像中各像素点的显著性值确定所述待处理图像的主体区域。
  7. 如权利要求6所述的方法,其特征在于,所述根据检测后所述待处理图像中各像素点的显著性值确定所述待处理图像的主体区域,具体为:
    对由所述显著性值所构成的显著性图进行二值化处理;
    对二值化处理后的图像中的较小的色块进行清除处理;
    查找清除处理后的图像中色块的轮廓线,并使用多边形近似各所述色块的轮廓线;
    利用矩形包围各所述多边形,将包围各所述矩形的外围矩形框作为所述主体区域。
  8. 如权利要求1所述的方法,其特征在于,在将所述主体区域中的图像作为前景图像进行前景目标提取处理之前,还包括:
    对所述待处理图像进行超像素分割处理,并确定处理后各超像素块中所有像素的平均显著性值;
    依次判断各所述超像素块的平均显著性值是否高于预设的阈值;
    若所述超像素块的平均显著性值高于设定阈值,则设置所述超像素块中各像素图像掩码值为非0;
    利用包围掩码图像非0像素的外围矩形框更新所述主体区域。
  9. 如权利要求1所述的方法,其特征在于,将所述主体区域中的图像作为前景图像进行前景目标提取处理,具体为:
    将所述主体区域中的图像设置为前景图像,并初始化GrabCut算法;
    对所述前景图像执行所述GrabCut算法。
  10. 一种图像主体提取系统,其特征在于,包括:
    判断模块,用于判断待处理图像中是否存在与指定特征对应的待识别区域;
    矩形框获取模块,用于在存在所述待识别区域时,根据预设的主体区域特征参数、所述待识别区域的坐标及大小,确定包含所述待处理图像的主体部分的主体区域;
    提取模块,用于将所述主体区域中的图像作为前景图像进行前景目标提取处理,并将提取出的图像作为所述待处理图像的主体部分。
  11. 如权利要求10所述的系统,其特征在于,还包括:
    分割模块,用于将所述待处理图像进行超像素分割处理,并确定处理后各超像素块的颜色的平均值;
    背景色模块,用于根据所述各超像素块的颜色的平均值确定所述待处理图像的背景色;
    纯色判断模块,用于确定所述待处理图像的纯色背景分数,并根据所述纯色背景分数判断所述待处理图像是否为纯色背景,所述纯色背景分数为符合所述背景色的颜色阈值的超像素块中所包含的像素数与所述待处理图像的总像素数的比值。
  12. 如权利要求10所述的系统,其特征在于,所述判断模块,具体包括:
    检测子模块,用于检测所述待处理图像中是否存在与所述指定特征对应的区域;
    确认子模块,用于所述检测子模块确认存在与所述指定特征对应的区域时,对所述待处理图像中与所述指定特征对应的区域进行轮廓检测,并根据与所述指定特征对应的参数化表观模型对已检测的区域进行验证,确认所述区域是否为与所述指定特征对应的待识别区域。
  13. 如权利要求12所述的系统,其特征在于,所述指定特征具体为人体头肩区域,所述确认子模块,具体用于根据上半身检测算法以及头肩轮廓检测器对所述待处理图像中与人体头肩区域对应的区域进行检测,并基于头肩表观模型对已检测的区域进行验证。
  14. 如权利要求13所述的系统,其特征在于,
    所述头肩轮廓检测器是通过使用AdaBoost算法结合HoG特征来训练得到的,所述头肩表观模型基于深度网络模型建立。
  15. 如权利要求10所述的系统,其特征在于,还包括:
    显著性检测模块,用于在不存在所述待识别区域时,对所述待处理图像进行显著性检测,并根据检测后所述待处理图像中各像素点的显著性值确定所述待处理图像的主体区域。
  16. 如权利要求15所述的系统,其特征在于,
    所述显著性检测模块,具体用于对由所述显著性值所构成的显著性图进行二值化处理,对二值化处理后的图像中的较小的色块进行清除处理,查找清除处理后的图像中色块的轮廓线,并使用多边形近似各所述色块的轮廓线,利用矩形包围各所述多边形,将包围各所述矩形的外围矩形框作为所述主体区域。
  17. 如权利要求10所述的系统,其特征在于,还包括:
    矩形框改进模块,用于对所述待处理图像进行超像素分割处理,并确定处理后各超像素块中所有像素的平均显著性值,依次判断各所述超像素块的平均显著性值是否高于预设的阈值,并在所述超像素块的平均显著性值高于 设定阈值时设置所述超像素块中各像素图像掩码值为非0,利用包围掩码图像非0像素的外围矩形框更新所述主体区域。
  18. 如权利要求10所述的系统,其特征在于,
    所述提取模块,具体用于将所述主体区域中的图像设置为前景图像,并初始化GrabCut算法,对所述前景图像执行所述GrabCut算法。
PCT/CN2015/092505 2014-10-27 2015-10-22 一种图像主体提取方法及系统 WO2016066038A1 (zh)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2017521554A JP6719457B2 (ja) 2014-10-27 2015-10-22 画像の主要被写体を抽出する方法とシステム
US15/499,840 US10497121B2 (en) 2014-10-27 2017-04-27 Method and system for extracting a main subject of an image

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201410584982.3 2014-10-27
CN201410584982.3A CN105631455B (zh) 2014-10-27 2014-10-27 一种图像主体提取方法及系统

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/499,840 Continuation US10497121B2 (en) 2014-10-27 2017-04-27 Method and system for extracting a main subject of an image

Publications (1)

Publication Number Publication Date
WO2016066038A1 true WO2016066038A1 (zh) 2016-05-06

Family

ID=55856591

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2015/092505 WO2016066038A1 (zh) 2014-10-27 2015-10-22 一种图像主体提取方法及系统

Country Status (4)

Country Link
US (1) US10497121B2 (zh)
JP (1) JP6719457B2 (zh)
CN (1) CN105631455B (zh)
WO (1) WO2016066038A1 (zh)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107452013A (zh) * 2017-05-27 2017-12-08 深圳市美好幸福生活安全系统有限公司 基于Harris角点检测和Sugeno模糊积分的显著性检测方法
US10497121B2 (en) 2014-10-27 2019-12-03 Alibaba Group Holding Limited Method and system for extracting a main subject of an image
US20200184697A1 (en) * 2016-11-08 2020-06-11 Adobe Inc. Image Modification Using Detected Symmetry

Families Citing this family (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017203705A1 (ja) * 2016-05-27 2017-11-30 楽天株式会社 画像処理装置、画像処理方法及び画像処理プログラム
CN106780306B (zh) * 2016-12-09 2020-07-24 腾讯音乐娱乐(深圳)有限公司 一种重构稿生成方法及装置
CN108269260B (zh) * 2016-12-30 2021-08-27 粉迷科技股份有限公司 动态影像去背方法、系统与计算机可读取存储装置
CN106815848A (zh) * 2017-01-17 2017-06-09 厦门可睿特信息科技有限公司 基于grubcut和人工智能的人像背景分离及轮廓提取方法
CN108694719B (zh) * 2017-04-05 2020-11-03 北京京东尚科信息技术有限公司 图像输出方法和装置
CN109003289B (zh) * 2017-12-11 2021-04-30 罗普特科技集团股份有限公司 一种基于颜色标签的目标跟踪快速初始化方法
CN108198172B (zh) * 2017-12-28 2022-01-28 北京大学深圳研究生院 图像显著性检测方法和装置
CN108968991B (zh) * 2018-05-08 2022-10-11 平安科技(深圳)有限公司 手骨x光片骨龄评估方法、装置、计算机设备和存储介质
CN109410169B (zh) * 2018-09-11 2020-06-05 广东智媒云图科技股份有限公司 一种图像背景干扰度的识别方法及装置
CN109389582B (zh) * 2018-09-11 2020-06-26 广东智媒云图科技股份有限公司 一种图像主体亮度的识别方法及装置
CN109584251A (zh) * 2018-12-06 2019-04-05 湘潭大学 一种基于单目标区域分割的舌体图像分割方法
CN111476253B (zh) * 2019-01-23 2024-04-02 阿里巴巴集团控股有限公司 服装图像分类、图像分类方法、装置及设备
CN110084801A (zh) * 2019-04-28 2019-08-02 深圳回收宝科技有限公司 一种终端屏幕的检测方法、装置、便携式终端和存储介质
CN110717865B (zh) * 2019-09-02 2022-07-29 苏宁云计算有限公司 图片检测方法及装置
CN110569380B (zh) * 2019-09-16 2021-06-04 腾讯科技(深圳)有限公司 一种图像标签获取方法、装置及存储介质和服务器
JP7439431B2 (ja) * 2019-09-25 2024-02-28 株式会社Jvcケンウッド 情報配信装置、情報生成方法、及び情報生成プログラム
US11039037B2 (en) * 2019-10-31 2021-06-15 Kyocera Document Solutions Inc. Image processing apparatus, image forming apparatus and image processing method for improving efficiency of clipping process
CN113256361A (zh) * 2020-02-10 2021-08-13 阿里巴巴集团控股有限公司 商品发布方法及图像处理方法、装置、设备和存储介质
CN111383244B (zh) * 2020-02-28 2023-09-01 浙江大华技术股份有限公司 一种目标检测跟踪方法
CN113627453A (zh) * 2020-05-08 2021-11-09 珠海金山办公软件有限公司 一种纯色背景图像抠图方法、装置及电子设备
CN111724396B (zh) * 2020-06-17 2023-07-14 泰康保险集团股份有限公司 图像分割方法及装置、计算机可读存储介质、电子设备
US11776184B2 (en) * 2020-10-18 2023-10-03 Adobe, Inc. Sky replacement
CN113194282A (zh) * 2021-04-13 2021-07-30 杭州三信网络技术有限公司 一种基于后端人工智能的视播融合应急广播系统
CN115439334A (zh) * 2021-06-02 2022-12-06 中国科学院分子植物科学卓越创新中心 整穗图像处理方法和装置
CN114108757B (zh) * 2021-12-13 2024-04-05 广东威力电器有限公司 一种厕用虹吸式人体红外自动感应冲水系统
CN114494730A (zh) * 2022-04-15 2022-05-13 深圳安牌信息技术有限公司 基于图像识别的商标自动分类处理系统

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102598113A (zh) * 2009-06-30 2012-07-18 安芯美特控股有限公司 匹配出现在两个或多个图像内的对象或人的方法、电路和系统
CN102779270A (zh) * 2012-06-21 2012-11-14 西南交通大学 一种针对购物图像搜索的目标衣物图像提取方法
CN103164858A (zh) * 2013-03-20 2013-06-19 浙江大学 基于超像素和图模型的粘连人群分割与跟踪方法

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7139411B2 (en) * 2002-06-14 2006-11-21 Honda Giken Kogyo Kabushiki Kaisha Pedestrian detection and tracking with night vision
US7200266B2 (en) * 2002-08-27 2007-04-03 Princeton University Method and apparatus for automated video activity analysis
JP4318465B2 (ja) * 2002-11-08 2009-08-26 コニカミノルタホールディングス株式会社 人物検出装置および人物検出方法
US7864989B2 (en) * 2006-03-31 2011-01-04 Fujifilm Corporation Method and apparatus for adaptive context-aided human classification
JP4626692B2 (ja) * 2008-09-12 2011-02-09 ソニー株式会社 物体検出装置、撮像装置、物体検出方法およびプログラム
JP2011035636A (ja) 2009-07-31 2011-02-17 Casio Computer Co Ltd 画像処理装置及び方法
US8638993B2 (en) * 2010-04-05 2014-01-28 Flashfoto, Inc. Segmenting human hairs and faces
WO2011152844A1 (en) 2010-06-01 2011-12-08 Hewlett-Packard Development Company, L.P. Image clustering using a personal clothing model
EP2395452A1 (en) * 2010-06-11 2011-12-14 Toyota Motor Europe NV/SA Detection of objects in an image using self similarities
CN102446347B (zh) * 2010-10-09 2014-10-01 株式会社理光 图像白平衡方法和装置
CN102467740A (zh) * 2010-11-08 2012-05-23 北京大学 一种大尺寸彩色图像的前景和背景交互式分割方法及系统
US9031286B2 (en) * 2010-12-09 2015-05-12 Panasonic Corporation Object detection device and object detection method
US9153031B2 (en) * 2011-06-22 2015-10-06 Microsoft Technology Licensing, Llc Modifying video regions using mobile device input
JP6032921B2 (ja) 2012-03-30 2016-11-30 キヤノン株式会社 物体検出装置及びその方法、プログラム
JP6192271B2 (ja) * 2012-08-22 2017-09-06 キヤノン株式会社 画像処理装置、画像処理方法及びプログラム
CN105631455B (zh) 2014-10-27 2019-07-05 阿里巴巴集团控股有限公司 一种图像主体提取方法及系统

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102598113A (zh) * 2009-06-30 2012-07-18 安芯美特控股有限公司 匹配出现在两个或多个图像内的对象或人的方法、电路和系统
CN102779270A (zh) * 2012-06-21 2012-11-14 西南交通大学 一种针对购物图像搜索的目标衣物图像提取方法
CN103164858A (zh) * 2013-03-20 2013-06-19 浙江大学 基于超像素和图模型的粘连人群分割与跟踪方法

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10497121B2 (en) 2014-10-27 2019-12-03 Alibaba Group Holding Limited Method and system for extracting a main subject of an image
US20200184697A1 (en) * 2016-11-08 2020-06-11 Adobe Inc. Image Modification Using Detected Symmetry
US11551388B2 (en) * 2016-11-08 2023-01-10 Adobe Inc. Image modification using detected symmetry
CN107452013A (zh) * 2017-05-27 2017-12-08 深圳市美好幸福生活安全系统有限公司 基于Harris角点检测和Sugeno模糊积分的显著性检测方法

Also Published As

Publication number Publication date
JP6719457B2 (ja) 2020-07-08
CN105631455B (zh) 2019-07-05
US20170228872A1 (en) 2017-08-10
US10497121B2 (en) 2019-12-03
CN105631455A (zh) 2016-06-01
JP2017531883A (ja) 2017-10-26

Similar Documents

Publication Publication Date Title
WO2016066038A1 (zh) 一种图像主体提取方法及系统
WO2017190656A1 (zh) 行人再识别方法和装置
CN107862667B (zh) 一种基于高分辨率遥感影像的城市阴影检测与去除方法
WO2020107717A1 (zh) 视觉显著性区域检测方法及装置
WO2018024030A1 (zh) 一种夜视红外图像中的基于显著性的道路目标提取方法
WO2018072233A1 (zh) 一种基于选择性搜索算法的车标检测识别方法及系统
CN103443804B (zh) 面部标志检测方法
CN103473564B (zh) 一种基于敏感区域的正面人脸检测方法
US9443137B2 (en) Apparatus and method for detecting body parts
JP4905931B2 (ja) 人体領域抽出方法および装置並びにプログラム
US20100128927A1 (en) Image processing apparatus and image processing method
CN102521579B (zh) 一种基于二维平面摄像头推的动作识别方法及系统
WO2019071976A1 (zh) 基于区域增长和眼动模型的全景图像显著性检测方法
CN105740945A (zh) 一种基于视频分析的人群计数方法
CN107944403B (zh) 一种图像中的行人属性检测方法及装置
CN103310194A (zh) 视频中基于头顶像素点梯度方向的行人头肩部检测方法
Thalji et al. Iris Recognition using robust algorithm for eyelid, eyelash and shadow avoiding
WO2023000253A1 (zh) 攀爬行为预警方法和装置、电子设备、存储介质
CN109934112B (zh) 一种人脸对齐方法和相机
JP6338429B2 (ja) 被写体検出装置、被写体検出方法及びプログラム
KR101296734B1 (ko) 가우시안 혼합 모델 및 알지비 클러스터링을 이용한 오브젝트 분할방법
Chen et al. Fresh tea sprouts detection via image enhancement and fusion SSD
CN113177439B (zh) 一种行人翻越马路护栏检测方法
Shen et al. Image based hair segmentation algorithm for the application of automatic facial caricature synthesis
Jacques et al. Head-shoulder human contour estimation in still images

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15855939

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2017521554

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 15855939

Country of ref document: EP

Kind code of ref document: A1