WO2013075295A1 - Clothing identification method and system for low-resolution video - Google Patents

Clothing identification method and system for low-resolution video Download PDF

Info

Publication number
WO2013075295A1
WO2013075295A1 PCT/CN2011/082705 CN2011082705W WO2013075295A1 WO 2013075295 A1 WO2013075295 A1 WO 2013075295A1 CN 2011082705 W CN2011082705 W CN 2011082705W WO 2013075295 A1 WO2013075295 A1 WO 2013075295A1
Authority
WO
WIPO (PCT)
Prior art keywords
clothing
human body
target
foreground image
frame
Prior art date
Application number
PCT/CN2011/082705
Other languages
French (fr)
Chinese (zh)
Inventor
李响
李俐
张超
陈晓娟
Original Assignee
浙江晨鹰科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 浙江晨鹰科技有限公司 filed Critical 浙江晨鹰科技有限公司
Priority to PCT/CN2011/082705 priority Critical patent/WO2013075295A1/en
Publication of WO2013075295A1 publication Critical patent/WO2013075295A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features

Definitions

  • the present invention relates to the field of image information processing technologies, and more particularly to a garment recognition method and system for low resolution video.
  • Face recognition based on face recognition mainly adopts multi-level detection system, which is roughly divided into four levels: face detection, uniform area detection, accessory detection and collar recognition, and filtering out a large amount of irrelevant data in the process of executing each level. To improve detection accuracy and efficiency.
  • the specific identification process is shown in Figure 1.
  • the face data obtained after the face detection is the next level of execution of the uniform area inspection service; the uniform area mask obtained after the inspection of the uniform area is the next level of accessories and the collar detection service. .
  • the present invention provides a clothing recognition method and system for low-resolution video, which overcomes the method of face recognition based on the prior art, and cannot realize character clothing and identity recognition in low-resolution video. The problem.
  • the present invention provides the following technical solutions:
  • a clothing recognition method for low resolution video comprising:
  • Determining a current time series in the received video stream extracting a foreground image in the video stream, determining a human body target from the foreground image, and extracting contour information of the human body target;
  • Decomposing the contour information of the human body target extracting a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category; and a clothing category of each of the blocks in the frame;
  • step of determining a current time sequence in the video stream acquiring a clothing category of the same human target in each frame in different time series in the video stream, and performing a voting decision according to the pre-stored clothing category, determining The clothing category of the sports target.
  • a garment recognition system for low resolution video comprising:
  • Extracting means configured to determine a current time sequence in the received video stream, and extract a foreground image of the video stream time series, determine a human body target from the foreground image, and extract contour information of the human body target;
  • Decomposing means configured to decompose the contour information of the human body target, and extract a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category;
  • a comparison identifying device configured to compare the obtained clothing feature values of each segment with a preset clothing feature threshold, and identify a clothing category of each segment in the current frame
  • a merging device configured to fuse a clothing category of each of the blocks in the same time sequence or different time series in the video stream
  • a determining device configured to perform a voting decision according to the pre-stored clothing category, determine a clothing category of the human target in the current time series; and a clothing category of the same human target in each time frame of different time series After the fusion, the judgment of the clothing category of the human target.
  • the present invention discloses a clothing recognition method and system for low resolution video. Based on the spatio-temporal classifier fusion technique, first, extracting the foreground image in the acquired video stream, and extracting the contour information of the moving human body; and then, identifying the moving human body object according to the extracted contour information; and passing the same frame in the video frame Different blocks of the same human target in the image are processed by multi-point feature recognition, and the recognition result is voted. Finally, the voting result is determined according to the judgment result of the same human target in the multi-frame image in the video stream, and finally the human body is determined.
  • the clothing category of the target is a clothing recognition method and system for low resolution video.
  • the method for performing moving human target recognition according to the background model preprocesses the algorithm, and can eliminate objects in the video background that are similar to the target color, and reduce interference.
  • the result of the clothing feature judgment in multiple video frames of the same human target finally determining the clothing category of the moving target, thereby achieving high efficiency and high quality. , high-accuracy identity and clothing identification purposes.
  • FIG. 1 is a flow chart of a method for identifying a face recognition based on the prior art
  • FIG. 2 is a flowchart of a method for recognizing a low resolution video according to an embodiment of the present invention
  • a flowchart of extracting a foreground image disclosed in the embodiment
  • FIG. 4a-4c are diagrams showing an effect of the process of identifying a human body object according to an embodiment of the present invention.
  • FIG. 5 is a flowchart of performing various feature information extraction according to an embodiment of the present invention.
  • FIG. 6 is a flowchart of performing multi-feature weak classifier fusion according to an embodiment of the present invention
  • FIG. 7 is an effect diagram of finalizing garment recognition processing in low-resolution video according to an embodiment of the present invention.
  • FIG. 8 is a schematic diagram of a clothing recognition system for low resolution video disclosed in an embodiment of the present invention.
  • the technical solutions in the embodiments of the present invention are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. example. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present invention without departing from the inventive scope are the scope of the present invention.
  • Embodiment 1 discloses a clothing recognition method and system for low-resolution video. Based on the spatio-temporal classifier fusion technology, various uniforms, general clothing, camouflage clothes, etc. can be identified and classified, and high efficiency and high quality are adopted. And realize the identification of the final identity information in a reliable and accurate manner. The specific process is described in detail by the following examples.
  • Embodiment 1 is described in detail by the following examples.
  • FIG. 2 it is a flow chart of a method for recognizing a low-resolution video according to the present invention, which mainly includes the following steps:
  • Step S101 Extract a foreground image in the received video stream.
  • step S101 is:
  • Step S1011 The video stream is read into a computer or a related device that can be analyzed, and the obtained video stream is decomposed, and a plurality of single-frame video sequences are obtained according to a time series.
  • Step S1012 Acquire a foreground image corresponding to the plurality of single-frame video sequences.
  • the process of acquiring the foreground image corresponding to a single frame video sequence in the step S1012 is: first, performing background modeling on the video according to the content of the video sequence; secondly, determining the current single frame video sequence and the current background frame; Determining a foreground image corresponding to the current single-frame video sequence by using a difference between the current frame video sequence and the background frame; and finally, updating the background frame according to the current single-frame video sequence to ensure the accuracy of the background frame in the next frame,
  • the update process is updated in real time.
  • the process of determining the current background frame determined by the above is: background modeling implemented by a single Gaussian or mixed Gaussian method. And further adopting the frame difference principle to obtain a corresponding foreground image according to the difference between the current single frame video sequence and the background frame.
  • the above background modeling of video can adopt single Gaussian, mixed Gaussian, Kernel-based, Eigen-Background and the like.
  • the background modeling is performed by using a mixed Gaussian method, that is, the background frame is obtained, and the mixed Gaussian model is defined as:
  • each pixel has a probability of 3 ⁇ 4 PC ⁇ K Gaussian mixture to facilitate background modeling in the video stream.
  • the foreground image of the moving object is extracted by the above process, that is, the foreground image is extracted by subtracting the video image from the background image, that is, the frame difference can be used to improve the foreground image extraction effect, thereby obtaining a more accurate foreground of the moving target.
  • the image that is, the contour foreground image.
  • Fig. 4a ⁇ ®4c The effect diagram after performing the above process is shown in Fig. 4a ⁇ ®4c, wherein Fig. 4a is a video image; Fig. 4b is a background image (background frame); Fig. 4c is a foreground image corresponding to the current moving target.
  • Step S102 Determine a current time series in the video stream, determine a moving target from the foreground image, and extract contour information of the moving human target.
  • the human body target generally refers to the moving human body, that is, the human body appearing in the foreground image in the current time series.
  • the contour information acquired during the execution of step S102 is determined based on the contour width and the contour height of the human body target.
  • the process of specifically extracting the contour information of the human body target is: first, extracting the feature of the moving object from the foreground image, analyzing according to the ratio of the width and height of the moving object, and identifying the moving human body; and then analyzing and acquiring the human body Outline information.
  • a more specific description is: Extracting the contour features of the moving object based on the plane geometry knowledge.
  • the distance between the leftmost point and the rightmost point of the contour of each animal is taken as the width of the object; the distance between the uppermost point and the lowermost point is taken as the height of the moving object.
  • Calculating the aspect ratio of the moving object and comparing the length-to-width ratio of each obtained moving object with the threshold value of the shoulder width and the height ratio of the conventional human body, excluding other moving objects such as vehicles, and determining the moving target, that is, the moving human body, and Extract the contour information of the moving human body from it.
  • the threshold of height ratio can effectively overcome the influence of objects such as trees and buildings on the recognition result, and reduce the interference of non-moving human body in the moving object to recognize the moving human body.
  • Step S103 Decompose the contour information of the human body target, and extract a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category. In contrast, the clothing category of each block in the current frame is identified.
  • step S105 the clothing category of each of the segments is merged, and a voting decision is performed according to the pre-stored clothing category, and the clothing category of the human target in the current time series is determined.
  • Step S1031 Decompose the contour information of the human body, and divide the human body according to the biological characteristics of the human body.
  • Step S1032 Perform eigenvalue training, and perform calculation of the corresponding clothing feature value according to the preset clothing category.
  • Step S1033 Extract a clothing feature value corresponding to each block in the contour information of the human body.
  • Identifying the clothing category of each of the segments in the current frame for step S104, identifying and determining the clothing category of the human target in the current time series is then for step S105, and the fusion of the plurality of characteristic information for each of the segments is also directed to step S105.
  • the process can also be specifically as follows: First, according to the human target obtained in step S102, one of the personal goals is determined. Decomposing the contour information of the determined human target, that is, dividing the same individual body, for example, dividing the human body into parts such as arms, tops, pants, etc. according to human body characteristics (step S1031); and then performing each block The calculation and extraction of the clothing feature values; the different segment clothing categories are identified, and the clothing categories of the respective blocks are merged. That is, the fusion is implemented based on the spatiotemporal classifier fusion technique; finally, the voting decision is made by the large number decision, and the clothing category of the human target in the current time series is determined.
  • the above process can be recorded as: for the clothing category of the same individual body target in the same frame image in the video stream, using the spatial correlation of the image, voting judgment is performed on the multiple recognition results, and the same human body is judged by using the large number judgment.
  • the plurality of recognition results are combined and aggregated to obtain a clothing recognition result.
  • any feature is not stable, therefore,
  • a plurality of weak classifiers are fused to form a strong classifier that is stable for a single block image to implement clothing recognition for the block image.
  • the process of specifically performing multi-feature weak classifier fusion can be seen in FIG. 6.
  • the color and texture features are selected as the classification features for the calculation of the feature values, and the calculation process corresponds to the feature value calculation portion in FIG.
  • training for different categories of clothing color and texture features can be performed offline. Select hundreds of samples to calculate clothing color and texture features separately. That is, the eigenvalue calculations such as color and gray level symmetry are performed for each identified block (human body region).
  • the color feature calculation in the mega area of the segmentation sample, converts the RGB color space into the HSV color space, and the HSV color system is closer to human visual perception.
  • the specific conversion method is as follows:
  • R, G, and B represent the color values of the RGB color space, respectively
  • H, S, and V represent the color values in the HSV color space, respectively.
  • the three channels of H, S, and V are separated for the converted picture.
  • the range of values of the preset clothing color feature threshold can be determined.
  • the gray level co-occurrence matrix is used for texture feature calculation.
  • the gray level co-occurrence matrix has a total of 15 eigenvalues.
  • the four second-order moments, the contrast, the correlation, and the entropy are selected for calculation.
  • the second moment of the angle, also called energy can be expressed by the formula:
  • ASM ⁇ P(i ⁇ ( 5 )
  • the second-order moment of the angle is a measure of the uniformity of the gray scale of the image texture, which is used to reflect the uniformity of the image gray distribution and the texture thickness.
  • the contrast CON is used to measure how the values of the matrix are distributed and how much of the image changes locally, reflecting the sharpness of the image and the depth of the texture.
  • Correlated CORRLN is used to measure the similarity of spatial gray level co-occurrence matrix elements in the row or column direction. Therefore, the magnitude of the correlation value can reflect the local grayscale correlation in the image. Specifically expressed as:
  • the second-order moment, contrast, correlation and entropy information of different categories of clothing angles are calculated for the sample RGB images converted into grayscale images, and the preset clothing texture feature threshold range is finally determined.
  • the color and texture features of the human body region in the video image of the recognized human body are subjected to block calculation according to the above calculation method; and a plurality of different features for the selected color feature and texture;
  • the weak classifier 126 is constructed, and the value of each weak classifier of each image clothing is determined by the threshold range; since each classifier has the same feature weight, for each of the N features, each feature weight is 1/1 N.
  • the weak classifier fusion is cascaded into a strong classifier output that is stable for a single block image. That is, the clothing category of each of the sub-blocks is fused by the clothing category weak classifier corresponding to each sub-block, and the result of the fusion is output through the strong classifier.
  • the clothing identification is performed using the large number judgment, and the final recognition result is obtained.
  • the embodiment of the invention is formed by multiple weak classifiers by comprehensively considering multiple features such as color and texture.
  • the strong classifier after a large number of judgments, uses the recognition results of different blocks of the same human body to optimize the clothing recognition results. Ensure the recognition effect with good effect and high reliability.
  • the preset clothing texture feature threshold and the preset clothing color feature threshold are both part of the preset clothing feature threshold. That is to say, in the process of comparing, the obtained clothing feature values of each block of the same type are compared with the feature thresholds of the same type in the preset clothing feature threshold.
  • Step S106 returning to step S102, the tracking of the human body target is realized, that is, the recognition of the clothing category of the human body target in the next time sequence or the adjacent time series is performed.
  • Step S107 Acquire a clothing category of the same human target in each frame in the different time series in the video stream, and perform a voting decision according to the pre-stored clothing category to determine a clothing category of the moving target.
  • Step S106 is performed by using the most single moving target tracking method to perform the same motion on the moving human body target in the next time series or adjacent time series in the video sequence according to the human body morphological characteristics (aspect ratio) and the position correlation of the target. Target Tracking. And repeating the above steps S102 to S105 to obtain the clothing category of the human target in the current time series of the same moving target. Finally, step S107 is performed, according to the identification result of the clothing category of the plurality of adjacent video sequences, the voting is performed, and the result of the large number is used to smooth out the erroneous recognition result. Thereby completing the garment recognition processing in the low resolution video. See Figure 7 for the specific recognition effect.
  • the adjacent multi-frame recognition result is summarized by using the time series correlation in the embodiment of the present invention, and the multi-frame clothing recognition result of the same human body is voted based on the tracking result.
  • the embodiments disclosed in the present invention can improve the real-time performance of the present invention for clothing recognition by reducing the number of detection targets by motion detection, performing feature training by offline, and using a method of setting a weighted weight calculation method.
  • the above-mentioned disclosed embodiments of the present invention describe in detail a clothing recognition method for low-resolution video.
  • the method of the present invention can be implemented by various forms of systems, and thus the present invention also discloses a low-resolution video clothing.
  • the identification system is described in detail below with reference to specific embodiments.
  • a garment recognition system for low resolution video disclosed in the embodiment of the present invention mainly includes: an extracting device 11, a disassembling device 12, a comparison identifying device 13, and a fusing device 14. And decision device 15.
  • the extracting device 11 is configured to determine a current time sequence in the received video stream, and extract a foreground image in the video stream, determine a moving human target from the foreground image, and extract contour information of the human target.
  • the decomposing device 12 is configured to decompose the contour information of the human body target, and extract the clothing feature value corresponding to each segment in the contour information of the human body target according to the preset clothing category.
  • the comparison identifying means 13 is configured to compare the obtained clothing feature values of the respective blocks with the preset clothing feature thresholds, and identify the clothing categories of the respective blocks in the current frame.
  • the merging device 14 is configured to fuse the clothing categories of each of the blocks in the same time series or different time series in the video stream.
  • the determining device 15 is configured to perform a voting decision according to the pre-stored clothing category, determine a clothing category of the human target in the current time series, and a clothing category of the same human target in each frame in different time series, the human body The judgment of the target clothing category.
  • the method further includes:
  • the removing device 16 is configured to perform noise and cavity removal operations on the acquired foreground image.
  • the above-disclosed system of the present invention corresponds to the method disclosed in the above-mentioned first embodiment, and the principle or the process of execution of each part can be referred to the above-disclosed method and its related parts.
  • the method and system disclosed by the present invention are based on a spatiotemporal classifier fusion technique, based on motion detection, human body recognition and clothing recognition, determine clothing characteristics in multiple video streams of the same human target, and finally determine the clothing category and identity of the moving target. In order to achieve high efficiency, high quality, high accuracy identity and clothing recognition purposes.

Abstract

Disclosed are a clothing identification method and system for low-resolution video. The method is based on the time and space classifier convergence technology. The method concerns particularly: extracting a foreground image in a video stream, and extracting contour information about a moving object in the foreground image; identifying a moving human body target according to the extracted contour information; performing multi-point feature identification processing on different blocks of the same human body target in the same frame image in the video stream and voting to decide the identification result; and performing the voting decision according to the decision result of the same human body target in a plurality of frame images in the video stream and finally determining the clothing type of the moving human body target. By way of the method disclosed on the basis of the time and space classifier convergence technology in the present invention, the clothing features of a plurality of video frames of the same moving target are decided on the basis of motion detection, human body identification and clothing identification, and the clothing type and identity of the human body target are finally determined, thus realizing the purpose of identity and clothing identification with high efficiency, high quality, and high accuracy.

Description

低分辨率视频的服装识别方法及系统  Garment recognition method and system for low resolution video
技术领域 本发明涉及图像信息处理技术领域, 更具体的说,是涉及一种低分辨率视 频的服装识别方法及系统。 TECHNICAL FIELD The present invention relates to the field of image information processing technologies, and more particularly to a garment recognition method and system for low resolution video.
背景技术 随着科技的不断进步, 传统的生物识别技术已经难以满足安全敏感场合 (如部队大院、 武警大院等军事管辖区)的安全防护的要求。 因此, 提出了对 能实现自动实时识别人物身份的智能视觉监控系统的需求,近年来非接触式远 距离的人物身份识别技术备受研究人员的广泛关注, 也得到了相应的发展。 BACKGROUND OF THE INVENTION With the continuous advancement of technology, traditional biometrics technology has been difficult to meet the requirements for security protection in security-sensitive situations (such as military complexes such as military complexes and armed police complexes). Therefore, the demand for intelligent visual monitoring system that can realize the automatic identification of characters in real time is put forward. In recent years, the non-contact long-distance character identification technology has received extensive attention from researchers and has also been developed accordingly.
当前, 主要在多模式、 大范围的视觉监控技术实现远距离的人员检测、 分 类与识别技术方面进行相关研究。 具体包括 "基于步态的身份识别" 和 "基于 人脸识别基础的身份识别" 的两种方式。  At present, mainly in multi-mode, large-scale visual monitoring technology to achieve long-distance personnel detection, classification and identification technology related research. These include "gait-based identification" and "face recognition based identification".
其中, 由于针对基于步态的远距离人员身份识别, 需要建立丰富的步态知 识库。 因此, 无法满足在人员数量大又有制服区分的场合, 实现非接触式远距 离的人物身份识别的要求。  Among them, due to the gait-based long-distance personnel identification, it is necessary to establish a rich gait knowledge base. Therefore, it is impossible to satisfy the requirement of non-contact remote character identification in the case where the number of people is large and there is a uniform distinction.
基于人脸识别基础的身份识别, 主要采用多级检测体系, 大致分为四级: 人脸检测、 军装区域检测、饰物检测及领花识别, 并在执行每一级的过程中过 滤掉大量无关数据, 以提高检测精度和效率。 其具体识别过程如图 1所示, 在 人脸检测后得到的人脸数据为下一级执行军装区域检测服务;在军装区域检测 后得到的军装区域掩码为下一级饰物及领花检测服务。  Face recognition based on face recognition, mainly adopts multi-level detection system, which is roughly divided into four levels: face detection, uniform area detection, accessory detection and collar recognition, and filtering out a large amount of irrelevant data in the process of executing each level. To improve detection accuracy and efficiency. The specific identification process is shown in Figure 1. The face data obtained after the face detection is the next level of execution of the uniform area inspection service; the uniform area mask obtained after the inspection of the uniform area is the next level of accessories and the collar detection service. .
在现有技术上述基于人脸识别基础的身份识别过程中所进行的服装识别, 需要先进行人脸识别, 然后再根据饰物领花特征进行军装识别。 但是, 在进行 人脸识别和军装的识别, 要求高分辨率、 背景效果好的图像, 如针对低分辨率 图像, 识别准确率将降低且漏检率高。 而且该方法适用于图像, 而非视频进行 服装识别, 因此无法满足或无法适应视频乃至低分辨率视频中的服装识别。 发明内容 有鉴于此, 本发明提供了一种低分辨率视频的服装识别方法及系统, 以克 服采用现有技术中基于人脸识别的方法,无法实现低分辨率视频中的人物服装 及身份识别的问题。 In the prior art clothing recognition based on the face recognition based on the face recognition process, it is necessary to perform face recognition first, and then perform military uniform recognition according to the feature of the accessory collar flower. However, in the recognition of face recognition and uniforms, images with high resolution and good background effects are required. For low-resolution images, the recognition accuracy is lowered and the missed detection rate is high. Moreover, the method is suitable for image recognition, not video, and therefore cannot meet or adapt to clothing recognition in video or even low resolution video. SUMMARY OF THE INVENTION In view of this, the present invention provides a clothing recognition method and system for low-resolution video, which overcomes the method of face recognition based on the prior art, and cannot realize character clothing and identity recognition in low-resolution video. The problem.
为实现上述目的, 本发明提供如下技术方案:  To achieve the above object, the present invention provides the following technical solutions:
一种低分辨率视频的服装识别方法, 包括:  A clothing recognition method for low resolution video, comprising:
确定接收到的视频流中的当前时间序列,提取所述视频流中的前景图像,从所 述前景图像中确定人体目标, 并提取所述人体目标的轮廓信息; Determining a current time series in the received video stream, extracting a foreground image in the video stream, determining a human body target from the foreground image, and extracting contour information of the human body target;
分解所述人体目标的轮廓信息,依据预设服装类别提取所述人体目标的轮 廓信息中各分块对应的服装特征值; 帧中各分块的服装类别;  Decomposing the contour information of the human body target, extracting a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category; and a clothing category of each of the blocks in the frame;
融合所述各分块的服装类别, 并依据预存储的服装类别进行投票判决,确 定当前时间序列中所述人体目标的服装类别;  Converging the clothing categories of the respective segments, and performing a voting decision according to the pre-stored clothing categories to determine a clothing category of the human target in the current time series;
返回执行确定所述视频流中的当前时间序列这一步骤,获取所述视频流中 不同时间序列中各帧中同一人体目标的服装类别进行融合,并依据预存储的服 装类别进行投票判决, 确定所述运动目标的服装类别。  Returning to the step of determining a current time sequence in the video stream, acquiring a clothing category of the same human target in each frame in different time series in the video stream, and performing a voting decision according to the pre-stored clothing category, determining The clothing category of the sports target.
一种低分辨率视频的服装识别系统, 包括:  A garment recognition system for low resolution video, comprising:
提取装置, 用于确定接收到的视频流中的当前时间序列, 以及提取所述视 频流时间序列的前景图像,从所述前景图像中确定人体目标, 并提取所述人体 目标的轮廓信息;  Extracting means, configured to determine a current time sequence in the received video stream, and extract a foreground image of the video stream time series, determine a human body target from the foreground image, and extract contour information of the human body target;
分解装置, 用于分解所述人体目标的轮廓信息,依据预设服装类别提取所 述人体目标的轮廓信息中各分块对应的服装特征值;  Decomposing means, configured to decompose the contour information of the human body target, and extract a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category;
比较识别装置,用于将获取到的各分块的服装特征值与预设服装特征阈值 进行比较, 识别当前帧中各分块的服装类别;  a comparison identifying device, configured to compare the obtained clothing feature values of each segment with a preset clothing feature threshold, and identify a clothing category of each segment in the current frame;
融合装置,用于融合所述视频流中同一时间序列或不同时间序列中各帧所 述各分块的服装类别;  a merging device, configured to fuse a clothing category of each of the blocks in the same time sequence or different time series in the video stream;
判决装置, 用于依据预存储的服装类别进行投票判决,确定当前时间序列 中所述人体目标的服装类别;以及不同时间序列各帧中同一人体目标的服装类 别进行融合后, 所述人体目标的服装类别的判决。 a determining device, configured to perform a voting decision according to the pre-stored clothing category, determine a clothing category of the human target in the current time series; and a clothing category of the same human target in each time frame of different time series After the fusion, the judgment of the clothing category of the human target.
经由上述的技术方案可知, 与现有技术相比, 本发明公开了一种低分辨率 视频的服装识别方法及系统。 基于时空分类器融合技术, 首先, 提取获取到的 视频流中的前景图像, 以及提取运动的人体的轮廓信息; 然后, 依据提取的轮 廓信息进行运动人体目标的识别;通过对视频帧中同一帧图像内同一人体目标 的不同分块, 分别进行多点特征识别进行处理, 并对识别结果进行投票判决; 最后根据视频流内多帧图像中同一人体目标的判决结果进行投票判决,最终确 定该人体目标的服装类别。通过上述本发明基于时空分类器融合技术,根据背 景模型进行运动人体目标识别的方法对算法进行预处理,能够排除视频背景中 与识别目标颜色相近的物体, 降低干扰。 同时综合考虑多个服装特征, 基于运 动检测、人体识别和服装识别,对同一人体目标多个视频帧中的服装特征判决 的结果, 最终确定该运动目标的服装类别, 从而实现高效率、 高质量, 高准确 度的身份及服装识别目的。  According to the above technical solution, compared with the prior art, the present invention discloses a clothing recognition method and system for low resolution video. Based on the spatio-temporal classifier fusion technique, first, extracting the foreground image in the acquired video stream, and extracting the contour information of the moving human body; and then, identifying the moving human body object according to the extracted contour information; and passing the same frame in the video frame Different blocks of the same human target in the image are processed by multi-point feature recognition, and the recognition result is voted. Finally, the voting result is determined according to the judgment result of the same human target in the multi-frame image in the video stream, and finally the human body is determined. The clothing category of the target. According to the above-described method based on spatio-temporal classifier fusion technology, the method for performing moving human target recognition according to the background model preprocesses the algorithm, and can eliminate objects in the video background that are similar to the target color, and reduce interference. At the same time, considering a plurality of clothing features, based on motion detection, human body recognition and clothing recognition, the result of the clothing feature judgment in multiple video frames of the same human target, finally determining the clothing category of the moving target, thereby achieving high efficiency and high quality. , high-accuracy identity and clothing identification purposes.
附图说明 为了更清楚地说明本发明实施例或现有技术中的技术方案,下面将对实施 例或现有技术描述中所需要使用的附图作筒单地介绍,显而易见地, 下面描述 中的附图仅仅是本发明的实施例,对于本领域普通技术人员来讲,在不付出创 造性劳动的前提下, 还可以根据提供的附图获得其他的附图。 BRIEF DESCRIPTION OF THE DRAWINGS In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described below. Obviously, in the following description The drawings are merely examples of the invention, and those skilled in the art can obtain other drawings based on the drawings provided without any creative work.
图 1为现有技术公开的一种基于人脸识别的身份识别的方法流程图; 图 2为本发明实施例公开的一种低分辨率视频的服装识别方法的流程图; 图 3为本发明实施例公开的提取前景图像的流程图;  1 is a flow chart of a method for identifying a face recognition based on the prior art; FIG. 2 is a flowchart of a method for recognizing a low resolution video according to an embodiment of the present invention; A flowchart of extracting a foreground image disclosed in the embodiment;
图 4a~ 4c为本发明实施例公开人体目标的识别流程中的效果图; 图 5为本发明实施例公开的进行多种特征信息提取的流程图;  4a-4c are diagrams showing an effect of the process of identifying a human body object according to an embodiment of the present invention; FIG. 5 is a flowchart of performing various feature information extraction according to an embodiment of the present invention;
图 6为本发明实施例公开的进行多特征弱分类器融合的流程图; 图 7 为本发明实施例公开的最终完成低分辨率视频中服装识别处理的效 果图;  6 is a flowchart of performing multi-feature weak classifier fusion according to an embodiment of the present invention; FIG. 7 is an effect diagram of finalizing garment recognition processing in low-resolution video according to an embodiment of the present invention;
图 8为本发明实施例公开的一种低分辨率视频的服装识别系统的框架图。 具体实施方式 下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清 楚、 完整地描述, 显然, 所描述的实施例仅仅是本发明一部分实施例, 而不是 全部的实施例。基于本发明中的实施例, 本领域普通技术人员在没有做出创造 性劳动前提下所获得的所有其他实施例, 都属于本发明保护的范围。 FIG. 8 is a schematic diagram of a clothing recognition system for low resolution video disclosed in an embodiment of the present invention. The technical solutions in the embodiments of the present invention are clearly and completely described in the following with reference to the accompanying drawings in the embodiments of the present invention. It is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. example. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present invention without departing from the inventive scope are the scope of the present invention.
本发明以下实施例公开了一种低分辨率视频的服装识别方法和系统,基于 时空分类器融合技术, 能够对各种制服、 普通服装、 迷彩服等进行识别分类, 并采用高效率、 高质量、 且可靠准确的方式实现最终的人物身份信息的识别。 具体过程通过以下实施例进行详细说明。 实施例一  The following embodiments of the present invention disclose a clothing recognition method and system for low-resolution video. Based on the spatio-temporal classifier fusion technology, various uniforms, general clothing, camouflage clothes, etc. can be identified and classified, and high efficiency and high quality are adopted. And realize the identification of the final identity information in a reliable and accurate manner. The specific process is described in detail by the following examples. Embodiment 1
请参阅附图 2, 为本发明公开的一种低分辨率视频的服装识别方法的流程 图, 主要包括以下步骤:  Referring to FIG. 2, it is a flow chart of a method for recognizing a low-resolution video according to the present invention, which mainly includes the following steps:
步骤 S101 , 提取接收到的视频流中的前景图像。  Step S101: Extract a foreground image in the received video stream.
请参见附图 3 , 执行步骤 S101的具体过程为:  Referring to FIG. 3, the specific process of performing step S101 is:
步骤 S1011 , 将视频流读取至计算机或相关可进行分析的设备中, 分解获 取到的视频流, 按照时间序列得到多个单帧视频序列。  Step S1011: The video stream is read into a computer or a related device that can be analyzed, and the obtained video stream is decomposed, and a plurality of single-frame video sequences are obtained according to a time series.
步骤 S1012, 获取多个所述单帧视频序列对应的前景图像。 针对该步骤 S1012中获取一个单帧视频序列对应的前景图像的过程为: 首先, 根据视频序 列的内容,对视频进行背景建模;其次,确定当前单帧视频序列和当前背景帧; 其次,依据当前帧视频序列与背景帧之间的差值,确定所述当前单帧视频序列 对应的前景图像; 最后, 为保证下一帧中的背景帧的准确, 根据当前单帧视频 序列更新背景帧, 该更新过程为实时更新。  Step S1012: Acquire a foreground image corresponding to the plurality of single-frame video sequences. The process of acquiring the foreground image corresponding to a single frame video sequence in the step S1012 is: first, performing background modeling on the video according to the content of the video sequence; secondly, determining the current single frame video sequence and the current background frame; Determining a foreground image corresponding to the current single-frame video sequence by using a difference between the current frame video sequence and the background frame; and finally, updating the background frame according to the current single-frame video sequence to ensure the accuracy of the background frame in the next frame, The update process is updated in real time.
需要说明的是, 上述所确定的当前背景帧, 其确定的过程为: 采用单高斯 或混合高斯方法实现的背景建模。并进一步的采用帧差原理依据当前单帧视频 序列与背景帧的差值获取对应的前景图像。  It should be noted that the process of determining the current background frame determined by the above is: background modeling implemented by a single Gaussian or mixed Gaussian method. And further adopting the frame difference principle to obtain a corresponding foreground image according to the difference between the current single frame video sequence and the background frame.
上述对视频进行背景建模可以采用单高斯、 混合高斯、 Kernel-based、 Eigen-Background等方法。在本发明所公开的该实施例中采用混合高斯的方法 进行背景建模, 即获取背景帧, 该混合高斯模型的定义为:
Figure imgf000007_0001
The above background modeling of video can adopt single Gaussian, mixed Gaussian, Kernel-based, Eigen-Background and the like. In the embodiment disclosed by the present invention, the background modeling is performed by using a mixed Gaussian method, that is, the background frame is obtained, and the mixed Gaussian model is defined as:
Figure imgf000007_0001
其中, 是第 j个高斯核权重; K为高斯核个数; ;;(; ^;∑ 是中值为 /、 方差为∑ 的第 j个高斯分布。 上述(1 ) 式表示为: 在 N时刻, 每个像素拥有 ¾的概率 PC^^ K个高 斯混合所描述, 以便于用于视频流中的背景建模。 Where is the jth Gaussian kernel weight; K is the Gaussian kernel number; ;; (; ^; ∑ is the jth Gaussian distribution with a median of /, and the variance is 。. The above (1) is expressed as: At the moment, each pixel has a probability of 3⁄4 PC^^ K Gaussian mixture to facilitate background modeling in the video stream.
采用上述过程对运动物体的前景图像进行提取,即通过视频图像与背景图 像相减实现前景图像的提取, 即采用帧差的方式可以提高前景图像提取的效 果, 进而获得较为准确的运动目标的前景图像, 即轮廓前景图像。  The foreground image of the moving object is extracted by the above process, that is, the foreground image is extracted by subtracting the video image from the background image, that is, the frame difference can be used to improve the foreground image extraction effect, thereby obtaining a more accurate foreground of the moving target. The image, that is, the contour foreground image.
此外, 为了消除在提取前景图像的过程中噪声和空洞等因素的影响, 采用 滤波方法去除噪声; 采用数学形态学方法去除空洞,从而得到效果更好的运动 目标的前景图像。 执行上述过程后的效果图如图 4a~ ® 4c所示, 其中, 图 4a 为视频图像; 图 4b为背景图像(背景帧); 图 4c为当前运动目标对应的前景 图像。  In addition, in order to eliminate the influence of noise and cavity in the process of extracting the foreground image, the filtering method is used to remove the noise; the mathematical morphology is used to remove the cavity, thereby obtaining a foreground image of the moving target with better effect. The effect diagram after performing the above process is shown in Fig. 4a~®4c, wherein Fig. 4a is a video image; Fig. 4b is a background image (background frame); Fig. 4c is a foreground image corresponding to the current moving target.
步骤 S102, 确定所述视频流中的当前时间序列, 从所述前景图像中确定 运动目标, 并提取运动人体目标的轮廓信息。其中人体目标一般情况下指运动 的人体, 即当前时间序列中出现在前景图像中的人体。  Step S102: Determine a current time series in the video stream, determine a moving target from the foreground image, and extract contour information of the moving human target. The human body target generally refers to the moving human body, that is, the human body appearing in the foreground image in the current time series.
在执行步骤 S102的过程中获取的轮廓信息基于人体目标的轮廓宽度和轮 廓高度确定。 且具体提取人体目标的轮廓信息的过程为: 首先, 从所述前景图 像中提取运动物体的特征,按照运动物体的宽度和高度的比例进行分析,识别 运动的人体; 然后, 分析获取所述人体的轮廓信息。  The contour information acquired during the execution of step S102 is determined based on the contour width and the contour height of the human body target. And the process of specifically extracting the contour information of the human body target is: first, extracting the feature of the moving object from the foreground image, analyzing according to the ratio of the width and height of the moving object, and identifying the moving human body; and then analyzing and acquiring the human body Outline information.
更具体的说明为: 根据平面几何知识提取运动物体的轮廓特征。将每个运 动物体的轮廓最左边的点和最右边的点之间的距离作为物体的宽度;将最上面 的点与最下面的点之间的距离作为运动物体的高度。 计算运动物体的长宽比, 并将获取到的各个运动物体的长宽比与常规人体的肩宽、身高比例的阈值进行 对比, 排除车辆等其他运动物体, 确定运动目标, 即运动人体, 并从中提取运 动人体的轮廓信息。  A more specific description is: Extracting the contour features of the moving object based on the plane geometry knowledge. The distance between the leftmost point and the rightmost point of the contour of each animal is taken as the width of the object; the distance between the uppermost point and the lowermost point is taken as the height of the moving object. Calculating the aspect ratio of the moving object, and comparing the length-to-width ratio of each obtained moving object with the threshold value of the shoulder width and the height ratio of the conventional human body, excluding other moving objects such as vehicles, and determining the moving target, that is, the moving human body, and Extract the contour information of the moving human body from it.
采用上述步骤 S101和步骤 S102中前景图像的提取方法,以及识别前景图 像中的运动目标的方式, 以及针对运动目标的预处理(设置常规人体的肩宽、 身高比例的阈值 ), 能够有效的克服树木、 建筑物等物体对识别结果的影响, 以及降低运动物体中非运动人体对识别运动人体的干扰。 Adopting the extraction method of the foreground image in the above steps S101 and S102, and the manner of recognizing the moving object in the foreground image, and the preprocessing for the moving target (setting the shoulder width of the conventional human body, The threshold of height ratio can effectively overcome the influence of objects such as trees and buildings on the recognition result, and reduce the interference of non-moving human body in the moving object to recognize the moving human body.
步骤 S103 , 分解所述人体目标的轮廓信息, 依据预设服装类别提取所述 人体目标的轮廓信息中各分块对应的服装特征值。 较, 识别当前帧中各分块的服装类别。  Step S103: Decompose the contour information of the human body target, and extract a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category. In contrast, the clothing category of each block in the current frame is identified.
步骤 S105 , 融合所述各分块的服装类别, 并依据预存储的服装类别进行 投票判决, 确定当前时间序列中所述人体目标的服装类别。  In step S105, the clothing category of each of the segments is merged, and a voting decision is performed according to the pre-stored clothing category, and the clothing category of the human target in the current time series is determined.
其中, 针对上述各个分块的多种特征信息提取、 融合以及识别的过程。 多 种特征信息提取具体针对步骤 S103 , 请参见附图 5 , 主要包括以下步骤: 步骤 S1031 , 分解所述人体的轮廓信息, 按照人体生物特征对人体进行分 块。  The process of extracting, merging, and identifying a plurality of feature information for each of the above blocks. The plurality of feature information extractions are specifically directed to step S103. Referring to FIG. 5, the method further includes the following steps: Step S1031: Decompose the contour information of the human body, and divide the human body according to the biological characteristics of the human body.
步骤 S1032, 进行特征值训练, 依据预设服装类别进行对应的服装特征值 的计算。  Step S1032: Perform eigenvalue training, and perform calculation of the corresponding clothing feature value according to the preset clothing category.
步骤 S1033 , 提取所述人体的轮廓信息中各分块对应的服装特征值。  Step S1033: Extract a clothing feature value corresponding to each block in the contour information of the human body.
识别当前帧中各分块的服装类别针对步骤 S104, 识别和确定当前时间序 列中所述人体目标的服装类别则针对步骤 S105 , 且对各个分块的多种特征信 息的融合也针对步骤 S105。 程, 也可以具体为: 首先, 根据执行步骤 S102获取的人体目标, 确定其中一 个人体目标。 分解确定的人体目标的轮廓信息, 即对同一个人体进行分块, 如 将该人体按照人体特征分为胳膊、上衣、裤子等部分进行分块(如步骤 S1031 ); 然后, 进行各个分块中的服装特征值的计算和提取; 再对不同分块服装类别进 行识别, 并将各个分块的服装类别进行融合。 即基于时空分类器融合技术实现 该融合; 最后通过大数判决进行投票判决,确定当前时间序列中所述人体目标 的服装类别。  Identifying the clothing category of each of the segments in the current frame for step S104, identifying and determining the clothing category of the human target in the current time series is then for step S105, and the fusion of the plurality of characteristic information for each of the segments is also directed to step S105. The process can also be specifically as follows: First, according to the human target obtained in step S102, one of the personal goals is determined. Decomposing the contour information of the determined human target, that is, dividing the same individual body, for example, dividing the human body into parts such as arms, tops, pants, etc. according to human body characteristics (step S1031); and then performing each block The calculation and extraction of the clothing feature values; the different segment clothing categories are identified, and the clothing categories of the respective blocks are merged. That is, the fusion is implemented based on the spatiotemporal classifier fusion technique; finally, the voting decision is made by the large number decision, and the clothing category of the human target in the current time series is determined.
上述过程可筒单记载为: 针对视频流中同一帧图像中, 同一个人体目标不 同分块的服装类别, 利用图像空间相关性, 对多块识别结果进行投票判决, 运 用大数判决对同一人体的多个识别结果进行融合汇总, 得到服装识别结果。  The above process can be recorded as: for the clothing category of the same individual body target in the same frame image in the video stream, using the spatial correlation of the image, voting judgment is performed on the multiple recognition results, and the same human body is judged by using the large number judgment. The plurality of recognition results are combined and aggregated to obtain a clothing recognition result.
另外, 对于低分辨率视频来说, 任何一种特征都稳定性都不高, 因此, 为 保证对低分辨率视频中的运动目标的服装识别,在本发明中利用多个弱分类器 融合级联成对于单块图像稳定的强分类器, 以实现对分块图像进行服装识别。 具体进行多特征弱分类器融合的过程可参见附图 6。 In addition, for low-resolution video, any feature is not stable, therefore, To ensure garment recognition for moving objects in low-resolution video, in the present invention, a plurality of weak classifiers are fused to form a strong classifier that is stable for a single block image to implement clothing recognition for the block image. The process of specifically performing multi-feature weak classifier fusion can be seen in FIG. 6.
针对各分块中对应的服装特征值,在本实施例中选取颜色和纹理特征作为 分类特征进行特征值的计算, 该计算过程对应图 5中的特征值计算部分。  For the corresponding clothing feature values in each block, in the present embodiment, the color and texture features are selected as the classification features for the calculation of the feature values, and the calculation process corresponds to the feature value calculation portion in FIG.
首先,针对不同类别服装颜色和纹理特征及进行训练, 该训练可以离线进 行。 选取数百个样本, 分别计算服装颜色和纹理特征。 即对每个识别出的分块 (人体区域)进行色彩、 灰度共生阵等特征值计算。  First, training for different categories of clothing color and texture features can be performed offline. Select hundreds of samples to calculate clothing color and texture features separately. That is, the eigenvalue calculations such as color and gray level symmetry are performed for each identified block (human body region).
其中, 颜色特征计算, 在分割样本的巨型区域, 将 RGB颜色空间转化为 HSV颜色空间, HSV颜色系统更接近人类视觉感知。 具体转化方法如下:  Among them, the color feature calculation, in the mega area of the segmentation sample, converts the RGB color space into the HSV color space, and the HSV color system is closer to human visual perception. The specific conversion method is as follows:
Figure imgf000009_0001
Figure imgf000009_0001
( 2 )  ( 2 )
max(R, G, B) - min(R, G, B)  Max(R, G, B) - min(R, G, B)
max(R, G, B)  Max(R, G, B)
( 3 ) (3)
max(R, G, B)  Max(R, G, B)
255  255
( 4 )  (4)
其中, R、 G、 B分别代表 RGB颜色空间的颜色值, H、 S、 V分别代表 HSV颜色空间中的颜色值。  Where R, G, and B represent the color values of the RGB color space, respectively, and H, S, and V represent the color values in the HSV color space, respectively.
利用上述公式(2 )至公式(4 )对于转化后的图片, 拆分出 H、 S、 V三 通道。 此外, 通过统计不同类别服装11、 S、 V值, 可以确定预设服装颜色特 征阈值的取值范围。  Using the above formula (2) to formula (4), the three channels of H, S, and V are separated for the converted picture. In addition, by counting the values of the different categories of clothing 11, S, V, the range of values of the preset clothing color feature threshold can be determined.
纹理特征计算, 对不同服装样本, 选用灰度共生矩阵进行纹理特征计算。 灰度共生矩阵共有 15个特征值,在本发明所公开的该实施例中选用角二阶矩、 对比度、 相关和熵四个统计效果较好的特征进行计算。 其中, 角二阶矩, 又称能量, 具体用公式可以表示为: Texture feature calculation, for different clothing samples, the gray level co-occurrence matrix is used for texture feature calculation. The gray level co-occurrence matrix has a total of 15 eigenvalues. In the embodiment disclosed in the present invention, the four second-order moments, the contrast, the correlation, and the entropy are selected for calculation. Among them, the second moment of the angle, also called energy, can be expressed by the formula:
ASM =∑∑P(i † ( 5 ) 该角二阶矩是影像纹理灰度变化均一的度量,用于反映影像灰度分布均匀 程度和纹理粗细度。
Figure imgf000010_0001
该对比度 CON用于度量矩阵的值是如何分布和影像中局部变化的多少, 反映了影像的清晰度和纹理的沟纹深浅。
ASM = ∑∑P(i † ( 5 ) The second-order moment of the angle is a measure of the uniformity of the gray scale of the image texture, which is used to reflect the uniformity of the image gray distribution and the texture thickness.
Figure imgf000010_0001
The contrast CON is used to measure how the values of the matrix are distributed and how much of the image changes locally, reflecting the sharpness of the image and the depth of the texture.
相关 CORRLN, 则用于度量空间灰度共生矩阵元素在行或列方向上的相 似程度。 因此, 相关值的大小可以反映影像中局部灰度相关性。 具体表示为:  Correlated CORRLN is used to measure the similarity of spatial gray level co-occurrence matrix elements in the row or column direction. Therefore, the magnitude of the correlation value can reflect the local grayscale correlation in the image. Specifically expressed as:
CORRLN = |∑∑ (W)P(i, j)) - μχμγ } / ^ ( 7 ) 熵 ΕΝΤ用于度量影像纹理的随机性。当空间共生矩阵中所有值均相等时, 它取得最大值; 相反, 如果共生矩阵中的值非常不均匀时, 其值较小。 具体表 示为:CORRLN = |∑∑(W)P(i, j)) - μ χ μ γ } / ^ ( 7 ) Entropy ΕΝΤ is used to measure the randomness of image texture. When all the values in the spatial co-occurrence matrix are equal, it takes the maximum value; conversely, if the values in the co-occurrence matrix are very uneven, the value is small. Specifically expressed as:
Figure imgf000010_0002
利用上述公式(5 )至(8 ), 针对转化为灰度图像的样本 RGB图像, 计算 统计不同类别服装角二阶矩、 对比度、 相关度和熵信息, 最终确定预设服装纹 理特征阈值范围。
Figure imgf000010_0002
Using the above formulas (5) to (8), the second-order moment, contrast, correlation and entropy information of different categories of clothing angles are calculated for the sample RGB images converted into grayscale images, and the preset clothing texture feature threshold range is finally determined.
在进行不同类别服装颜色和纹理特征的训练之后,根据上述计算方法对识 别出来人体的视频图像中人体区域进行颜色和纹理特征进行分块计算;对于所 选的颜色特征和纹理多个不同特征; 对于第 j个特征, 构造弱分类器 1¾,由阈值 范围确定每块图像服装各弱分类器的值; 由于, 各分类器特征权重相同, 对于 N个特征, 每个特征权值分别为 1/N。 最后将弱分类器融合级联成对于单块图 像稳定的强分类器输出。即利用对应各分块的服装类别弱分类器融合所述各分 块的服装类别, 并将融合的结果通过强分类器输出。  After training the color and texture features of different categories of clothing, the color and texture features of the human body region in the video image of the recognized human body are subjected to block calculation according to the above calculation method; and a plurality of different features for the selected color feature and texture; For the j-th feature, the weak classifier 126 is constructed, and the value of each weak classifier of each image clothing is determined by the threshold range; since each classifier has the same feature weight, for each of the N features, each feature weight is 1/1 N. Finally, the weak classifier fusion is cascaded into a strong classifier output that is stable for a single block image. That is, the clothing category of each of the sub-blocks is fused by the clothing category weak classifier corresponding to each sub-block, and the result of the fusion is output through the strong classifier.
最后, 根据属于同一人体目标, 即同一人体的不同分块的识别结果, 运用 大数判决进行服装识别, 获取最终的识别结果。  Finally, according to the recognition results of the same human body target, that is, the different blocks of the same human body, the clothing identification is performed using the large number judgment, and the final recognition result is obtained.
本发明实施例通过综合考虑颜色、 纹理等多个特征, 由多个弱分类器形成 强分类器, 经过大数判决, 利用同一人体不同分块的识别结果进行服装识别结 果优化。 保证获取效果较好、 可靠性高的识别效果。 The embodiment of the invention is formed by multiple weak classifiers by comprehensively considering multiple features such as color and texture. The strong classifier, after a large number of judgments, uses the recognition results of different blocks of the same human body to optimize the clothing recognition results. Ensure the recognition effect with good effect and high reliability.
需要说明的是,预设服装纹理特征阈值和预设服装颜色特征阈值都属于预 设服装特征阈值的一部分。 也就是说, 在进行对比的过程中, 获取的同类型的 各分块的服装特征值对应与预设服装特征阈值中同类型的特征阈值进行比较。  It should be noted that the preset clothing texture feature threshold and the preset clothing color feature threshold are both part of the preset clothing feature threshold. That is to say, in the process of comparing, the obtained clothing feature values of each block of the same type are compared with the feature thresholds of the same type in the preset clothing feature threshold.
步骤 S106, 返回步骤 S102, 实现对所述人体目标的跟踪, 即执行下一时 间序列或相邻时间序列中人体目标的服装类别的识别。  Step S106, returning to step S102, the tracking of the human body target is realized, that is, the recognition of the clothing category of the human body target in the next time sequence or the adjacent time series is performed.
步骤 S107 , 获取所述视频流中不同时间序列中各帧同一人体目标的服装 类别进行融合, 并依据预存储的服装类别进行投票判决,确定所述运动目标的 服装类别。  Step S107: Acquire a clothing category of the same human target in each frame in the different time series in the video stream, and perform a voting decision according to the pre-stored clothing category to determine a clothing category of the moving target.
执行步骤 S106采用最筒单的运动目标跟踪方法, 对视频序列中下一时间 序列或相邻时间序列中运动人体目标, 根据该目标的人体形态特性(长宽比) 和位置相关性进行同一运动目标跟踪。并重复执行上述步骤 S102至步骤 S105 , 获取同一运动目标在当前时间序列中所述人体目标的服装类别。最后执行步骤 S107, 根据多个相邻视频序列服装类别识别结果, 进行投票表决, 利用大数判 决结果平滑掉有误的识别结果。从而完成低分辨率视频中服装识别处理。具体 识别后的效果参见图示 7。  Step S106 is performed by using the most single moving target tracking method to perform the same motion on the moving human body target in the next time series or adjacent time series in the video sequence according to the human body morphological characteristics (aspect ratio) and the position correlation of the target. Target Tracking. And repeating the above steps S102 to S105 to obtain the clothing category of the human target in the current time series of the same moving target. Finally, step S107 is performed, according to the identification result of the clothing category of the plurality of adjacent video sequences, the voting is performed, and the result of the large number is used to smooth out the erroneous recognition result. Thereby completing the garment recognition processing in the low resolution video. See Figure 7 for the specific recognition effect.
在执行步骤 S106和步骤 S107的过程中,通过上述本发明实施例中利用时 间序列相关性, 对相邻多帧识别结果进行汇总, 根据跟踪结果, 对同一人体的 多帧服装识别结果进行投票判决, 运用相邻多帧的识别结果平滑有误的结果, 最终借助时空融合结果实现或完成快速、 高精度的服装识别。  In the process of performing step S106 and step S107, the adjacent multi-frame recognition result is summarized by using the time series correlation in the embodiment of the present invention, and the multi-frame clothing recognition result of the same human body is voted based on the tracking result. The use of adjacent multi-frame recognition results to smooth the results of errors, and finally achieve or complete fast, high-precision clothing recognition by means of spatiotemporal fusion results.
此外, 在发明公开的实施例基于通过运动检测, 减少检测目标数量; 利用 离线进行特征训练; 采用筒化权值计算设定方法等措施, 能够提高本发明进行 服装识别的实时性。 上述本发明公开的实施例中详细描述了一种低分辨率视频的服装识别方 法,对于本发明的方法可采用多种形式的系统实现, 因此本发明还公开了一种 低分辨率视频的服装识别系统, 下面给出具体的实施例进行详细说明。  Further, the embodiments disclosed in the present invention can improve the real-time performance of the present invention for clothing recognition by reducing the number of detection targets by motion detection, performing feature training by offline, and using a method of setting a weighted weight calculation method. The above-mentioned disclosed embodiments of the present invention describe in detail a clothing recognition method for low-resolution video. The method of the present invention can be implemented by various forms of systems, and thus the present invention also discloses a low-resolution video clothing. The identification system is described in detail below with reference to specific embodiments.
请参见附图 8 , 为本发明实施例公开的一种低分辨率视频的服装识别系 统, 主要包括: 提取装置 11、 分解装置 12、 比较识别装置 13、 融合装置 14 和判决装置 15。 Referring to FIG. 8, a garment recognition system for low resolution video disclosed in the embodiment of the present invention mainly includes: an extracting device 11, a disassembling device 12, a comparison identifying device 13, and a fusing device 14. And decision device 15.
提取装置 11 , 用于确定接收到的视频流中的当前时间序列, 以及提取所 述视频流中的前景图像,从所述前景图像中确定运动人体目标, 并提取所述人 体目标的轮廓信息。  The extracting device 11 is configured to determine a current time sequence in the received video stream, and extract a foreground image in the video stream, determine a moving human target from the foreground image, and extract contour information of the human target.
分解装置 12, 用于分解所述人体目标的轮廓信息, 依据预设服装类别提 取所述人体目标的轮廓信息中各分块对应的服装特征值。  The decomposing device 12 is configured to decompose the contour information of the human body target, and extract the clothing feature value corresponding to each segment in the contour information of the human body target according to the preset clothing category.
比较识别装置 13 , 用于将获取到的各分块的服装特征值与预设服装特征 阈值进行比较, 识别当前帧中各分块的服装类别。  The comparison identifying means 13 is configured to compare the obtained clothing feature values of the respective blocks with the preset clothing feature thresholds, and identify the clothing categories of the respective blocks in the current frame.
融合装置 14 , 用于融合所述视频流中同一时间序列或不同时间序列中各 帧所述各分块的服装类别。  The merging device 14 is configured to fuse the clothing categories of each of the blocks in the same time series or different time series in the video stream.
判决装置 15 , 用于依据预存储的服装类别进行投票判决, 确定当前时间 序列中所述人体目标的服装类别;以及不同时间序列中各帧中同一人体目标的 服装类别进行融合后, 所述人体目标的服装类别的判决。  The determining device 15 is configured to perform a voting decision according to the pre-stored clothing category, determine a clothing category of the human target in the current time series, and a clothing category of the same human target in each frame in different time series, the human body The judgment of the target clothing category.
在上述本发明实施例公开的低分辨率视频的服装识别系统的基础上,还包 括:  Based on the clothing recognition system for low resolution video disclosed in the above embodiments of the present invention, the method further includes:
去除装置 16, 用于对获取到的所述前景图像进行噪声和空洞的去除操作。 上述本发明公开的该系统对应上述实施例一公开的方法,各部分具体执行 的原理或执行的过程可参见上述公开的方法与其相关的部分。 综上所述:  The removing device 16 is configured to perform noise and cavity removal operations on the acquired foreground image. The above-disclosed system of the present invention corresponds to the method disclosed in the above-mentioned first embodiment, and the principle or the process of execution of each part can be referred to the above-disclosed method and its related parts. In summary:
本发明所公开的方法和系统基于时空分类器融合技术,基于运动检测、人 体识别和服装识别,对同一人体目标多个视频流中的服装特征进行判决, 最终 确定该运动目标的服装类别和身份, 从而实现高效率、 高质量, 高准确度的身 份及服装识别目的。  The method and system disclosed by the present invention are based on a spatiotemporal classifier fusion technique, based on motion detection, human body recognition and clothing recognition, determine clothing characteristics in multiple video streams of the same human target, and finally determine the clothing category and identity of the moving target. In order to achieve high efficiency, high quality, high accuracy identity and clothing recognition purposes.
本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是 与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于 实施例公开的装置而言, 由于其与实施例公开的方法相对应, 所以描述的比较 筒单, 相关之处参见方法部分说明即可。  The various embodiments in the present specification are described in a progressive manner, and each embodiment focuses on differences from other embodiments, and the same similar parts between the various embodiments may be referred to each other. For the device disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the comparison is described, and the relevant part can be referred to the method part.
结合本文中所公开的实施例描述的方法或算法的步骤可以直接用硬件、处 理器执行的软件模块, 或者二者的结合来实施。软件模块可以置于随机存储器 ( RAM )、内存、只读存储器( ROM )、电可编程 ROM、电可擦除可编程 ROM、 寄存器、 硬盘、 可移动磁盘、 CD-ROM, 或技术领域内所公知的任意其它形式 的存储介质中。 The steps of a method or algorithm described in connection with the embodiments disclosed herein may be implemented directly in hardware, a software module executed by a processor, or a combination of both. Software modules can be placed in random access memory (RAM), memory, read only memory (ROM), electrically programmable ROM, electrically erasable programmable ROM, registers, hard disk, removable disk, CD-ROM, or any other form of storage known in the art. In the medium.
对所公开的实施例的上述说明,使本领域专业技术人员能够实现或使用本 发明。 对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见 的, 本文中所定义的一般原理可以在不脱离本发明的精神或范围的情况下, 在 其它实施例中实现。 因此, 本发明将不会被限制于本文所示的这些实施例, 而 是要符合与本文所公开的原理和新颖特点相一致的最宽的范围。  The above description of the disclosed embodiments enables those skilled in the art to make or use the invention. Various modifications to these embodiments are obvious to those skilled in the art, and the general principles defined herein may be implemented in other embodiments without departing from the spirit or scope of the invention. Therefore, the present invention is not to be limited to the embodiments shown herein, but is to be accorded to the broadest scope of the principles and novel features disclosed herein.

Claims

1、 一种低分辨率视频的服装识别方法, 其特征在于, 包括: A garment recognition method for low resolution video, comprising:
提取接收到的视频流中的前景图像;  Extracting a foreground image in the received video stream;
确定所述视频流中的当前时间序列,从所述前景图像中确定运动目标,识 别人体目标, 并提取所述人体目标的轮廓信息;  Determining a current time series in the video stream, determining a moving target from the foreground image, identifying a body object, and extracting contour information of the human body target;
分解所述人体目标的轮廓信息,依据预设服装类别提取所述人体目标的轮 廓信息中各分块对应的服装特征值; 帧中各分块的服装类别;  Decomposing the contour information of the human body target, extracting a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category; and a clothing category of each of the blocks in the frame;
融合所述各分块的服装类别, 并依据预存储的服装类别进行投票判决,确 定当前时间序列中所述人体目标的服装类别;  Converging the clothing categories of the respective segments, and performing a voting decision according to the pre-stored clothing categories to determine a clothing category of the human target in the current time series;
返回执行确定所述视频流中的当前时间序列这一步骤,获取所述视频流中 不同时间序列中各帧中同一人体目标的服装类别进行融合,并依据预存储的服 装类别进行投票判决, 确定所述运动目标的服装类别。  Returning to the step of determining a current time sequence in the video stream, acquiring a clothing category of the same human target in each frame in different time series in the video stream, and performing a voting decision according to the pre-stored clothing category, determining The clothing category of the sports target.
2、根据权利要求 1所述的方法, 其特征在于, 在接收所述前景图像之后, 从所述前景图像中确定运动目标之前, 还包括:  The method according to claim 1, wherein, after the foreground image is received, before determining the motion target from the foreground image, the method further includes:
对获取到的所述前景图像进行噪声和空洞的去除操作。  A noise and hole removal operation is performed on the acquired foreground image.
3、 根据权利要求 1或 2所述的方法, 其特征在于, 提取接收到的视频流 中的前景图像, 具体过程包括:  The method according to claim 1 or 2, wherein the foreground image in the received video stream is extracted, and the specific process includes:
分解获取到的视频流, 按照时间序列得到多个单帧视频序列;  Decomposing the acquired video stream, and obtaining a plurality of single-frame video sequences according to a time series;
获取多个所述单帧视频序列对应的前景图像。  Obtaining foreground images corresponding to the plurality of single-frame video sequences.
4、 根据权利要求 3所述的方法, 其特征在于, 获取多个所述单帧视频序 列对应的前景图像, 具体过程包括:  The method according to claim 3, wherein the foreground image corresponding to the plurality of single-frame video sequences is obtained, and the specific process includes:
根据前面若干帧视频序列的内容, 对视频进行背景建模;  Background modeling of the video according to the content of the previous several video sequences;
确定当前单帧视频序列和当前背景帧;  Determining a current single frame video sequence and a current background frame;
依据当前单帧视频序列与背景帧之间的差值,确定所述当前单帧视频序列 对应的前景图像, 并根据当前帧视频序列实时更新背景帧。  And determining, according to a difference between the current single-frame video sequence and the background frame, a foreground image corresponding to the current single-frame video sequence, and updating the background frame in real time according to the current frame video sequence.
5、 根据权利要求 1或 2所述的方法, 其特征在于, 从所述前景图像中确 定人体目标的具体过程包括:  5. The method according to claim 1 or 2, wherein the specific process of determining a human body target from the foreground image comprises:
从所述前景图像中提取运动物体的特征,分析获取所述运动物体的轮廓信 息; Extracting features of the moving object from the foreground image, and analyzing the contour information of the moving object Interest rate
求解运动物体长宽比,根据常规人体肩宽和身高比例设定阈值,识别人体 目标。  The aspect ratio of the moving object is solved, and the threshold is set according to the conventional shoulder width and the height ratio to identify the human target.
6、 根据权利要求 1或 2所述的方法, 其特征在于, 分解所述人体的轮廓 信息, 依据预设服装类别提取所述人体的轮廓信息中各分块对应的服装特征 值, 具体过程包括:  The method according to claim 1 or 2, wherein the contour information of the human body is decomposed, and the clothing feature value corresponding to each segment in the contour information of the human body is extracted according to the preset clothing category, and the specific process includes :
分解所述人体的轮廓信息, 按照人体生物特征对人体进行分块; 进行特征值训练, 依据预设服装类别进行对应的服装特征值的计算; 提取所述人体的轮廓信息中各分块对应的服装特征值。  Decomposing the contour information of the human body, and dividing the human body according to the biological characteristics of the human body; performing eigenvalue training, performing calculation of the corresponding clothing feature value according to the preset clothing category; extracting corresponding blocks of the contour information of the human body Clothing feature value.
7、 根据权利要求 4所述的方法, 其特征在于, 包括: 采用单高斯、 混合 高斯、 Kernel-based或 Eigen-Background的方法建立当前背景帧。  7. The method according to claim 4, comprising: establishing a current background frame by using a single Gaussian, mixed Gaussian, Kernel-based or Eigen-Background method.
8、 根据权利要求 1或 2所述的方法, 其特征在于, 包括:  8. The method according to claim 1 or 2, comprising:
利用对应各分块的服装类别弱分类器融合所述各分块的服装类别,并将融 合的结果形成强分类器。  The clothing category of each of the blocks is fused by the clothing category weak classifier corresponding to each of the blocks, and the result of the fusion is formed into a strong classifier.
9、 一种低分辨率视频的服装识别系统, 其特征在于, 包括:  9. A garment recognition system for low resolution video, comprising:
提取装置, 用于提取接收到的视频流中的前景图像, 以及在确定所述视频 流中的当前时间序列后,从所述前景图像中确定人体目标, 并提取所述人体目 标的轮廓信息;  An extracting device, configured to extract a foreground image in the received video stream, and after determining a current time sequence in the video stream, determine a human body target from the foreground image, and extract contour information of the human body target;
分解装置, 用于分解所述人体目标的轮廓信息,依据预设服装类别提取所 述人体目标的轮廓信息中各分块对应的服装特征值;  Decomposing means, configured to decompose the contour information of the human body target, and extract a clothing feature value corresponding to each block in the contour information of the human body target according to the preset clothing category;
比较识别装置,用于将获取到的各分块的服装特征值与预设服装特征阈值 进行比较, 识别当前帧中各分块的服装类别;  a comparison identifying device, configured to compare the obtained clothing feature values of each segment with a preset clothing feature threshold, and identify a clothing category of each segment in the current frame;
融合装置,用于融合所述视频流中同一时间序列或不同时间序列中各帧所 述各分块的服装类别;  a merging device, configured to fuse a clothing category of each of the blocks in the same time sequence or different time series in the video stream;
判决装置, 用于依据预存储的服装类别进行投票判决,确定当前时间序列 中所述人体目标的服装类别;以及不同时间序列各帧中同一人体目标的服装类 别进行融合后, 所述人体目标的服装类别的判决。  a determining device, configured to perform a voting decision according to the pre-stored clothing category, determine a clothing category of the human target in the current time series; and, after the clothing category of the same human target in each frame of different time series, the human target The judgment of the clothing category.
10、 根据权利要求 9所述的系统, 其特征在于, 还包括:  10. The system according to claim 9, further comprising:
去除装置, 用于对获取到的所述前景图像进行噪声和空洞的去除操作。  And a removing device, configured to perform noise and cavity removal operations on the acquired foreground image.
PCT/CN2011/082705 2011-11-23 2011-11-23 Clothing identification method and system for low-resolution video WO2013075295A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2011/082705 WO2013075295A1 (en) 2011-11-23 2011-11-23 Clothing identification method and system for low-resolution video

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2011/082705 WO2013075295A1 (en) 2011-11-23 2011-11-23 Clothing identification method and system for low-resolution video

Publications (1)

Publication Number Publication Date
WO2013075295A1 true WO2013075295A1 (en) 2013-05-30

Family

ID=48468998

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2011/082705 WO2013075295A1 (en) 2011-11-23 2011-11-23 Clothing identification method and system for low-resolution video

Country Status (1)

Country Link
WO (1) WO2013075295A1 (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103886305A (en) * 2014-04-08 2014-06-25 中国人民解放军国防科学技术大学 Specific face searching method for grassroots policing, safeguard stability and counter-terrorism
CN111476336A (en) * 2019-01-23 2020-07-31 阿里巴巴集团控股有限公司 Piece counting method, device and equipment for clothes
CN112613376A (en) * 2020-12-17 2021-04-06 深圳集智数字科技有限公司 Re-recognition method and device and electronic equipment
CN112711966A (en) * 2019-10-24 2021-04-27 阿里巴巴集团控股有限公司 Video file processing method and device and electronic equipment
CN113221928A (en) * 2020-01-21 2021-08-06 海信集团有限公司 Clothing classification information display device, method and storage medium
CN113283369A (en) * 2021-06-08 2021-08-20 苏州市伏泰信息科技股份有限公司 Port wharf operating personnel safety protection measure monitoring system and method
CN113409076A (en) * 2021-06-11 2021-09-17 广州天辰信息科技有限公司 Method and system for constructing user portrait based on big data and cloud platform
CN114040140A (en) * 2021-11-15 2022-02-11 北京医百科技有限公司 Video matting method, device and system and storage medium
US20220180551A1 (en) * 2020-12-04 2022-06-09 Shopify Inc. System and method for generating recommendations during image capture of a product
CN114040140B (en) * 2021-11-15 2024-04-12 北京医百科技有限公司 Video matting method, device, system and storage medium

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007164690A (en) * 2005-12-16 2007-06-28 Matsushita Electric Ind Co Ltd Image processor and image processing method
US20070239764A1 (en) * 2006-03-31 2007-10-11 Fuji Photo Film Co., Ltd. Method and apparatus for performing constrained spectral clustering of digital image data
JP2007293912A (en) * 2004-06-09 2007-11-08 Matsushita Electric Ind Co Ltd Image processing method, and image processing apparatus
US20090116698A1 (en) * 2007-11-07 2009-05-07 Palo Alto Research Center Incorporated Intelligent fashion exploration based on clothes recognition
WO2009091259A1 (en) * 2008-01-18 2009-07-23 Nederlandse Organisatie Voor Toegepast-Natuurwetenschappelijk Onderzoek Tno Method of improving the resolution of a moving object in a digital image sequence
CN101527838A (en) * 2008-03-04 2009-09-09 华为技术有限公司 Method and system for feedback-type object detection and tracing of video object
CN101763634A (en) * 2009-08-03 2010-06-30 北京智安邦科技有限公司 simple objective classification method and device
CN101833653A (en) * 2010-04-02 2010-09-15 上海交通大学 Figure identification method in low-resolution video
CN101882217A (en) * 2010-02-26 2010-11-10 杭州海康威视软件有限公司 Target classification method of video image and device
CN102521565A (en) * 2011-11-23 2012-06-27 浙江晨鹰科技有限公司 Garment identification method and system for low-resolution video

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007293912A (en) * 2004-06-09 2007-11-08 Matsushita Electric Ind Co Ltd Image processing method, and image processing apparatus
JP2007164690A (en) * 2005-12-16 2007-06-28 Matsushita Electric Ind Co Ltd Image processor and image processing method
US20070239764A1 (en) * 2006-03-31 2007-10-11 Fuji Photo Film Co., Ltd. Method and apparatus for performing constrained spectral clustering of digital image data
US20090116698A1 (en) * 2007-11-07 2009-05-07 Palo Alto Research Center Incorporated Intelligent fashion exploration based on clothes recognition
WO2009091259A1 (en) * 2008-01-18 2009-07-23 Nederlandse Organisatie Voor Toegepast-Natuurwetenschappelijk Onderzoek Tno Method of improving the resolution of a moving object in a digital image sequence
CN101527838A (en) * 2008-03-04 2009-09-09 华为技术有限公司 Method and system for feedback-type object detection and tracing of video object
CN101763634A (en) * 2009-08-03 2010-06-30 北京智安邦科技有限公司 simple objective classification method and device
CN101882217A (en) * 2010-02-26 2010-11-10 杭州海康威视软件有限公司 Target classification method of video image and device
CN101833653A (en) * 2010-04-02 2010-09-15 上海交通大学 Figure identification method in low-resolution video
CN102521565A (en) * 2011-11-23 2012-06-27 浙江晨鹰科技有限公司 Garment identification method and system for low-resolution video

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103886305B (en) * 2014-04-08 2017-01-25 中国人民解放军国防科学技术大学 Specific face searching method for grassroots policing, safeguard stability and counter-terrorism
CN103886305A (en) * 2014-04-08 2014-06-25 中国人民解放军国防科学技术大学 Specific face searching method for grassroots policing, safeguard stability and counter-terrorism
CN111476336B (en) * 2019-01-23 2023-06-20 阿里巴巴集团控股有限公司 Method, device and equipment for counting clothes
CN111476336A (en) * 2019-01-23 2020-07-31 阿里巴巴集团控股有限公司 Piece counting method, device and equipment for clothes
CN112711966B (en) * 2019-10-24 2024-03-01 阿里巴巴集团控股有限公司 Video file processing method and device and electronic equipment
CN112711966A (en) * 2019-10-24 2021-04-27 阿里巴巴集团控股有限公司 Video file processing method and device and electronic equipment
CN113221928B (en) * 2020-01-21 2023-07-18 海信集团有限公司 Clothing classification information display device, method and storage medium
CN113221928A (en) * 2020-01-21 2021-08-06 海信集团有限公司 Clothing classification information display device, method and storage medium
US20220180551A1 (en) * 2020-12-04 2022-06-09 Shopify Inc. System and method for generating recommendations during image capture of a product
US11645776B2 (en) * 2020-12-04 2023-05-09 Shopify Inc. System and method for generating recommendations during image capture of a product
CN112613376A (en) * 2020-12-17 2021-04-06 深圳集智数字科技有限公司 Re-recognition method and device and electronic equipment
CN112613376B (en) * 2020-12-17 2024-04-02 深圳集智数字科技有限公司 Re-identification method and device and electronic equipment
CN113283369A (en) * 2021-06-08 2021-08-20 苏州市伏泰信息科技股份有限公司 Port wharf operating personnel safety protection measure monitoring system and method
CN113409076A (en) * 2021-06-11 2021-09-17 广州天辰信息科技有限公司 Method and system for constructing user portrait based on big data and cloud platform
CN114040140A (en) * 2021-11-15 2022-02-11 北京医百科技有限公司 Video matting method, device and system and storage medium
CN114040140B (en) * 2021-11-15 2024-04-12 北京医百科技有限公司 Video matting method, device, system and storage medium

Similar Documents

Publication Publication Date Title
CN110837784B (en) Examination room peeping and cheating detection system based on human head characteristics
WO2013075295A1 (en) Clothing identification method and system for low-resolution video
US8395676B2 (en) Information processing device and method estimating a posture of a subject in an image
JP4273359B2 (en) Age estimation system and age estimation method
JP5675229B2 (en) Image processing apparatus and image processing method
Avgerinakis et al. Recognition of activities of daily living for smart home environments
Seow et al. Neural network based skin color model for face detection
JP5959093B2 (en) People search system
Lu et al. A novel approach for video text detection and recognition based on a corner response feature map and transferred deep convolutional neural network
CN102521565A (en) Garment identification method and system for low-resolution video
JP2000003452A (en) Method for detecting face surface in digital picture, its detecting device, picture judging method, picture judging device and computer readable record medium
Zheng et al. Attention-based spatial-temporal multi-scale network for face anti-spoofing
CN111126240B (en) Three-channel feature fusion face recognition method
Wang et al. Head pose estimation with combined 2D SIFT and 3D HOG features
CN110991315A (en) Method for detecting wearing state of safety helmet in real time based on deep learning
WO2015131468A1 (en) Method and system for estimating fingerprint pose
WO2018019149A1 (en) Human body gender automatic recognition method and apparatus
CN106529441B (en) Depth motion figure Human bodys' response method based on smeared out boundary fragment
CN107392105B (en) Expression recognition method based on reverse collaborative salient region features
CN103577804A (en) Abnormal human behavior identification method based on SIFT flow and hidden conditional random fields
Gürel et al. Design of a face recognition system
Hu et al. Fast face detection based on skin color segmentation using single chrominance Cr
CN111639562A (en) Intelligent positioning method for palm region of interest
Duan et al. Local feature learning for face recognition under varying poses
Gul et al. A machine learning approach to detect occluded faces in unconstrained crowd scene

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11876088

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 11876088

Country of ref document: EP

Kind code of ref document: A1