WO2020173024A1 - Multi-gesture precise segmentation method for smart home scenario - Google Patents

Multi-gesture precise segmentation method for smart home scenario Download PDF

Info

Publication number
WO2020173024A1
WO2020173024A1 PCT/CN2019/092970 CN2019092970W WO2020173024A1 WO 2020173024 A1 WO2020173024 A1 WO 2020173024A1 CN 2019092970 W CN2019092970 W CN 2019092970W WO 2020173024 A1 WO2020173024 A1 WO 2020173024A1
Authority
WO
WIPO (PCT)
Prior art keywords
gesture
image
smart home
area
segmentation method
Prior art date
Application number
PCT/CN2019/092970
Other languages
French (fr)
Chinese (zh)
Inventor
张晖
张迪
赵海涛
孙雁飞
朱洪波
Original Assignee
南京邮电大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 南京邮电大学 filed Critical 南京邮电大学
Priority to JP2020515730A priority Critical patent/JP6932402B2/en
Publication of WO2020173024A1 publication Critical patent/WO2020173024A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Disclosed in the present invention is a multi-gesture precise segmentation method for a smart home scenario, comprising the following steps: S1: pre-processing a gesture image Image0 to obtain an image Image1; S2: performing skin tone segmentation on the pre-processed image Image1 to obtain a processed image Image4; S3: constructing an image minimum bounding rectangle (MBR) in the image Image4; S4: excluding the non-gesture areas in the image Image4 by means of a non-gesture area exclusion criterion to acquire a gesture image Image5; and S5: processing the image Image5 by means of an arm redundancy removal algorithm based on hand shape features to implement removal of arm redundancy. The present invention can locally and smartly segment gestures, and the entire process is quick and accurate, significantly increasing the comfort of use of gesture-based human-machine interaction systems.

Description

面向智能家居场景的多手势精准分割方法Multi-gesture accurate segmentation method for smart home scene 技术领域Technical field
本发明涉及一种智能识别方法,具体涉及一种面向智能家居场景的多手势精准分割方法,属于智能家居领域。The invention relates to an intelligent recognition method, in particular to a multi-gesture accurate segmentation method oriented to a smart home scene, and belongs to the field of smart home.
背景技术Background technique
手势分割是指将手势信息从复杂的图像背景中分割出来的技术,手势分割质量(准确性、完整性、冗余性)的好坏对基于手势的人机交互系统的识别与检测准确率有着重要的影响。Gesture segmentation refers to the technology of segmenting gesture information from the complex image background. The quality of gesture segmentation (accuracy, completeness, redundancy) is good or bad for the recognition and detection accuracy of gesture-based human-computer interaction systems. important influence.
面向家居场景下的实时手势分割更为复杂,用户手势不仅更为复杂多变,而且易受背景、光照与拍摄角度等因素的影响。在目前的计算机视觉领域内,尚未存在一种能够自适应的手势分割算法。现行的一些具有代表性的手势分割方法主要依赖于外部设备或者需要对用户的手部进行特殊处理,但由于其限制了人们活动范围,同时需要配套的硬件设备,造价及其昂贵等原因,这些技术也很难在实际应用中得到大规模的推广。Real-time gesture segmentation in home-oriented scenarios is more complex. User gestures are not only more complex and changeable, but also vulnerable to factors such as background, lighting, and shooting angle. In the current computer vision field, there is no adaptive gesture segmentation algorithm. Some of the current representative gesture segmentation methods mainly rely on external devices or require special processing of the user’s hands. However, they limit the range of people's activities and require supporting hardware equipment, cost and expensive. The technology is also difficult to be widely promoted in practical applications.
与技术发展相对应的是,当今市场上,具有手势分割功能的智能家居设备还较为少见,大多数的手势分割产品仅仅停留在对皮肤的分割上,并不能完全且准确地针对手势进行分割,分割效果并不理想。并且这些设备大多依靠云端服务器、过于依赖网络,在没有网络的情况下将无法进行工作。Corresponding to technological development, smart home devices with gesture segmentation are still relatively rare in the market today. Most gesture segmentation products only focus on segmentation of the skin, and cannot completely and accurately segment the gestures. The segmentation effect is not ideal. And most of these devices rely on cloud servers and rely too much on the network, and will not be able to work without a network.
综上所述,如何在现有技术的基础上提出一种全新的面向智能家居场景的多手势精准分割方法,以实现手势分割技术在智能家居设备上的大规 模推广应用,也就成为了目前业内人士所亟待解决的问题。In summary, how to propose a new multi-gesture precision segmentation method for smart home scenes based on the existing technology to realize the large-scale promotion and application of gesture segmentation technology on smart home devices has become the current Problems that need to be solved urgently by the insiders.
发明内容Summary of the invention
鉴于现有技术存在上述缺陷,本发明的目的是提出一种面向智能家居场景的多手势精准分割方法,包括如下步骤:In view of the above-mentioned defects in the prior art, the purpose of the present invention is to propose a multi-gesture accurate segmentation method for smart home scenes, including the following steps:
S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;
S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;
S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
优选地,S1中所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。Preferably, the preprocessing in S1 at least includes: gesture image denoising, gesture image binarization, and morphological processing.
优选地,S2具体包括如下步骤:Preferably, S2 specifically includes the following steps:
S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3;S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then compare each pixel with the threshold through the global fixed threshold binarization method to obtain the binarized image Image3;
S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
优选地,S3具体包括如下步骤:将S2中所获得到的二值化手势图像的轮廓信息存放至列表contours中,并根据坐标信息获得外接矩形的四个顶点坐标,分别为top_left,top_right,bottom_left与bottom_right。Preferably, S3 specifically includes the following steps: store the contour information of the binarized gesture image obtained in S2 into the list contours, and obtain the coordinates of the four vertices of the circumscribed rectangle according to the coordinate information, which are top_left, top_right, and bottom_left. With bottom_right.
优选地,S4中所述非手势区域排除准则,具体包括:Preferably, the non-gesture area exclusion criterion in S4 specifically includes:
1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;
2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;
3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
优选地,S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。Preferably, the arm redundancy removal algorithm based on hand shape features described in S5 specifically includes: counting the hand width distribution histogram and gradient distribution histogram of the image Image6, wherein the width of the gesture width distribution histogram is the largest The value and its corresponding coordinate are the thumb carpal joint, and the coordinates of the wrist dividing line are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point.
优选地,S5步骤中所述手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定,确定标准为:当前点的梯度为0,且下一点的梯度大于等于0。Preferably, the coordinates of the wrist dividing line in step S5 are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point, and the determination criterion is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
与现有技术相比,本发明的优点主要体现在以下几个方面:Compared with the prior art, the advantages of the present invention are mainly embodied in the following aspects:
本发明所提出的面向智能家居场景的多手势精准分割方法,可以在本地智能化地对手势进行分割,克服了现有技术过于依赖网络的弊端,使得应用本方法的设备在没有网络连接的情况下仍然可以正常工作。The multi-gesture precise segmentation method for smart home scenes proposed by the present invention can segment the gestures intelligently locally, overcomes the drawbacks of the prior art that is too dependent on the network, and makes the device applying this method not connected to the network. It can still work normally.
本发明通过将手势图片从RGB颜色空间转化到YCbCr颜色空间、再通过全局固定阈值二值化的方法,完成了对肤色的分割。随后,对非手势区域进行排除,构建出手势轮廓的MBR与MABR,并对手势图像进行旋转以统计手部宽度、构建出宽度分布直方图与基于宽度的梯度分布直方图,确定手腕分割线。最后,完成对手臂冗余的去除,得到完整的手势图像。本发明能够快速且准确地对家居环境图像中的手势进行分割,显著地提高了基于手势的人机交互系统的使用舒适度,提升了用户满意度。The present invention completes the segmentation of the skin color by converting the gesture picture from the RGB color space to the YCbCr color space, and then by the method of global fixed threshold binarization. Subsequently, the non-gesture area is excluded, the MBR and MABR of the gesture contour are constructed, the gesture image is rotated to count the hand width, the width distribution histogram and the width-based gradient distribution histogram are constructed, and the wrist division line is determined. Finally, the removal of arm redundancy is completed, and a complete gesture image is obtained. The invention can quickly and accurately segment the gestures in the home environment image, significantly improves the use comfort of the gesture-based human-computer interaction system, and improves user satisfaction.
此外,本发明也为同领域内的其他相关问题提供了参考,可以以此为依据进行拓展延伸,运用于其他关于手势分割的技术方案中,具有十分广阔的应用前景。In addition, the present invention also provides a reference for other related issues in the same field, which can be used as a basis for expansion and application in other technical solutions related to gesture segmentation, and has very broad application prospects.
以下便结合实施例附图,对本发明的具体实施方式作进一步的详述,以使本发明技术方案更易于理解、掌握。The specific implementation of the present invention will be described in further detail below in conjunction with the accompanying drawings of the embodiments, so as to make the technical solution of the present invention easier to understand and grasp.
附图说明Description of the drawings
图1为本发明所提供的对手势图像进行肤色分割的步骤流程示意图;FIG. 1 is a schematic diagram of the steps of performing skin color segmentation on gesture images provided by the present invention;
图2为本发明所提供的对手势图像进行手臂冗余去除的步骤流程示意图;FIG. 2 is a schematic flow chart of steps for removing arm redundancy from gesture images provided by the present invention;
图3为本发明所提供的面向智能家居场景的多手势精准分割方法的总体步骤流程示意图。FIG. 3 is a schematic flow diagram of the overall steps of the multi-gesture precise segmentation method for smart home scenes provided by the present invention.
具体实施方式detailed description
本发明揭示了一种面向智能家居场景的多手势精准分割方法,所述方法基于YCbCr颜色空间的肤色分割算法,非手势区域排除准则,基于手部形状特征的手臂冗余去除算法。本发明的方法包括如下步骤:The present invention discloses a multi-gesture precise segmentation method for smart home scenes. The method is based on a skin color segmentation algorithm of YCbCr color space, non-gesture region exclusion criteria, and an arm redundancy removal algorithm based on hand shape features. The method of the present invention includes the following steps:
S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;
S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;
S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
由上述步骤表述可以看出,本发明的方法主要包括两大方面,即肤色分割及手臂冗余去除。It can be seen from the above steps that the method of the present invention mainly includes two major aspects, namely skin color segmentation and arm redundancy removal.
以下结合附图对本发明的方法进行具体说明,图1显示了本发明实施例提供的一种对手势图像进行肤色分割的方法,该方法的步骤主要包含:Hereinafter, the method of the present invention will be described in detail with reference to the accompanying drawings. FIG. 1 shows a method for segmenting a gesture image according to an embodiment of the present invention. The steps of the method mainly include:
S1、对手势图像Image0进行预处理,得到图像Image1。S1. Preprocess the gesture image Image0 to obtain the image Image1.
由于手势图像在获取时不可避免的会存在噪声的干扰,会对手势图像的分割与识别造成严重的影响,因此在对手势分割前对图像进行预处理显得尤为重要。所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。As the gesture image is inevitably interfered by noise when it is acquired, it will have a serious impact on the segmentation and recognition of the gesture image, so it is particularly important to preprocess the image before the gesture segmentation. The preprocessing at least includes: gesture image denoising, gesture image binarization and morphological processing.
其中,手势图像去噪主要使用高斯滤波器,该滤波器属于线性滤波器,其滤波器窗口的像素取值服从高斯分布,随着距离模板中心的距离增大而减小,其二维高斯函数为:Among them, the gesture image denoising mainly uses a Gaussian filter, which is a linear filter. The pixel value of the filter window obeys the Gaussian distribution, and decreases with the increase of the distance from the center of the template. Its two-dimensional Gaussian function for:
Figure PCTCN2019092970-appb-000001
Figure PCTCN2019092970-appb-000001
其中,h(x,y)表示高斯滤波器中(x,y)坐标上的取值,σ表示标准差。Among them, h(x,y) represents the value on the (x,y) coordinate in the Gaussian filter, and σ represents the standard deviation.
S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3。S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then use the global fixed threshold binarization method to compare each pixel with the threshold value to obtain the binarized image Image3.
YCbCr颜色空间是视频图像和数字图像中常用的色彩空间。包含三个分量:Y(luma,亮度),表示的是图像的亮暗程度,取值范围为0~255;Cb分量表示的是RGB颜色空间中蓝色分量与RGB颜色空间中亮度值之间差异,取值范围为0~255;Cr分量表示的是RGB颜色空间中红色分量的值与RGB颜色空间中亮度之间的差异,取值范围为0~255。其中Cb分量和Cr分量是相互独立的,并且与Y分量能有效地分离。YCbCr color space is a commonly used color space in video images and digital images. Contains three components: Y (luma, brightness), which represents the brightness and darkness of the image, ranging from 0 to 255; Cb component represents the blue component in the RGB color space and the brightness value in the RGB color space The value range of the difference is 0-255; the Cr component represents the difference between the value of the red component in the RGB color space and the brightness in the RGB color space, and the value range is 0-255. The Cb component and the Cr component are independent of each other and can be effectively separated from the Y component.
RGB颜色空间到YCbCr颜色空间的转换公式如下:The conversion formula from RGB color space to YCbCr color space is as follows:
Figure PCTCN2019092970-appb-000002
Figure PCTCN2019092970-appb-000002
转化为矩阵形式为:Converted to matrix form:
Figure PCTCN2019092970-appb-000003
Figure PCTCN2019092970-appb-000003
上述步骤中所述的将每个像素与阈值进行比较,具体操作为,人的肤色的Y,Cb,Cr值大约为[0:256,130:174,77:128],如果图像中像素的YCbCr值属于这个区间,则该像素值置为255,否则置为0,则可得到二值化图像Image3。In the above steps, each pixel is compared with the threshold. The specific operation is that the Y, Cb, and Cr values of human skin are approximately [0:256,130:174,77:128], if the YCbCr value of the pixel in the image If it belongs to this interval, the pixel value is set to 255, otherwise it is set to 0, and the binary image Image3 can be obtained.
S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
手势图像经过二值化处理后,得到的图像会存在空隙、残缺等现象。形态学的作用就是去除孤立的小点、毛刺、填充小孔、弥合小缝隙等,形态学操作主要有以下4种:After the gesture image is binarized, the resulting image will have gaps and defects. The role of morphology is to remove isolated dots, burrs, fill small holes, bridge small gaps, etc. There are four main types of morphological operations:
1、膨胀。形态学操作中的膨胀运算过程是:将物体接触到的背景点合并到该物体。其结果使目标物体的面积变大;其意义在于对目标区域中存在的空洞与缝隙进行填充。1. Expansion. The process of dilation operation in morphological operation is to merge the background points touched by the object into the object. As a result, the area of the target object becomes larger; its meaning is to fill the holes and gaps in the target area.
2、腐蚀。形态学操作中的腐蚀运算过程是:将物体的所有边界点进行消除。其结果使目标物体的面积变小;其意义在于对目标区域中存在的一些较小的没有意义的孤立点进行消除。2. Corrosion. The corrosion calculation process in the morphological operation is to eliminate all boundary points of the object. As a result, the area of the target object becomes smaller; its significance is to eliminate some small meaningless isolated points in the target area.
3、开运算。形态学操作中的开运算过程先对二值化图像进行腐蚀运算,然后在对其进行膨胀运算。其意义在于对目标区域中存在的孤立小点、毛 刺等没有意义的点进行消除(腐蚀运算),空洞与缝隙进行填充(膨胀运算)。3. Open operation. The opening operation process in the morphological operation first performs the erosion operation on the binary image, and then performs the expansion operation on it. Its significance is to eliminate the isolated small dots, burrs and other meaningless points in the target area (corrosion operation), and fill cavities and gaps (expansion operation).
4、闭运算。形态学操作中的闭运算过程先对二值化图像进行膨胀运算,然后在对其进行腐蚀运算。其意义在于对目标区域中存在的空洞与缝隙进行填充(膨胀运算),孤立小点、毛刺等没有意义的点进行消除(腐蚀运算)。4. Close operation. The closing operation process in the morphological operation first performs dilation operation on the binary image, and then performs erosion operation on it. Its significance lies in filling the voids and gaps in the target area (expansion operation), and eliminating isolated small dots, burrs and other meaningless points (corrosion operation).
所述的中值滤波器属于非线性滤波器,主要是对当前点的周围像素点进行统计与排序,选取其中值作为当前点的像素值,从而消除孤立的噪声点。主要用中值滤波器平滑手势二值化图像边缘的毛刺,使其边缘变得平滑,降低对手腕分割线的查找带来的影响。The median filter is a non-linear filter, which mainly counts and sorts the surrounding pixels of the current point, and selects the median value as the pixel value of the current point, thereby eliminating isolated noise points. The median filter is mainly used to smooth the burrs on the edges of the gesture binarization image, so that the edges are smoothed, and the influence of the search of the wrist division line is reduced.
图2为本发明实施例提供的对手势图像进行手臂冗余去除方法,该方法的步骤主要包含:2 is a method for removing arm redundancy from a gesture image provided by an embodiment of the present invention. The steps of the method mainly include:
S3、构建手势图像的最小面积绑定矩形MABR;S3. Construct the minimum area binding rectangle MABR of the gesture image;
在图像Image4中构建手势图像的最小绑定矩形MBR,其顶点坐标信息为,Construct the smallest bound rectangle MBR of the gesture image in Image4, and its vertex coordinate information is,
Figure PCTCN2019092970-appb-000004
Figure PCTCN2019092970-appb-000004
在MBR的基础上构建图像的MABR,在已知轮廓的前提下可以根据Graham扫描法求手势轮廓的凸包,将图形以其MBR的中心为原点,在其90度范围内以β为尺度等间隔的进行旋转。同时记录相应旋转角度下图形的MBR面积,则在记录中最小的MBR面积所对应的MBR就是所求的MABR。The MABR of the image is constructed on the basis of the MBR. Under the premise of the known contour, the convex hull of the gesture contour can be obtained according to the Graham scanning method. The center of the MBR of the graphic is the origin, and the β is the scale within the 90 degree range, etc. Rotate at intervals. At the same time, the MBR area of the graphics under the corresponding rotation angle is recorded, and the MBR corresponding to the smallest MBR area in the record is the required MABR.
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排 除,获取手势图像Image5。S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5.
所述非手势区域排除准则,具体包括:The non-gesture area exclusion criteria specifically include:
1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;
2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;
3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
随后逆时针旋转二值化手势图像,在上述步骤中可以得到MABR对应的旋转角度,逆时针旋转手势图像,即可使手势方向变为垂直。Then rotate the binarized gesture image counterclockwise, the rotation angle corresponding to the MABR can be obtained in the above steps, and rotate the gesture image counterclockwise to make the gesture direction vertical.
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。The arm redundancy removal algorithm based on hand shape features described in S5 specifically includes: Counting image Image6's hand width distribution histogram and gradient distribution histogram, where the maximum width of the gesture width distribution histogram and its The corresponding coordinate is the thumb carpal joint, and the coordinates of the wrist dividing line are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point.
所述宽度直方图的计算程序如下:The calculation procedure of the width histogram is as follows:
Figure PCTCN2019092970-appb-000005
Figure PCTCN2019092970-appb-000005
所述梯度直方图的计算程序如下:The calculation procedure of the gradient histogram is as follows:
gradient=[0]gradient=[0]
for index in range(1,len(width)):for index in range(1,len(width)):
  gradient.append(width[index]-width[index-1])。Gradient.append(width[index]-width[index-1]).
随后,确定手腕分割线,由于手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标可以通过查找拇指腕掌关节点之后的梯度分布直方图中的值确定,其确定方法为:当前点的梯度为0,且下一点的梯度大于等于0。Then, determine the wrist segmentation line. Since the maximum width of the gesture width distribution histogram and its corresponding coordinates are the thumb carpal joint, the coordinates of the wrist segmentation line can be found by looking for the value in the gradient distribution histogram after the thumb carpal joint point To determine, the method for determining is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
最后,完成手臂冗余去除。在上述步骤中获取得到手腕分割线的坐标信息,令在手腕分割线下面的像素值为0,即只保留上部手势图像,手臂部分去除掉。Finally, complete the arm redundancy removal. In the above steps, the coordinate information of the wrist dividing line is obtained, and the pixel value under the wrist dividing line is set to 0, that is, only the upper gesture image is retained, and the arm part is removed.
以下结合图3,提出一面向智能家居场景的多手势精准分割方法的具体实施例,该实施例主要包括如下步骤:In the following, in conjunction with FIG. 3, a specific embodiment of a multi-gesture precise segmentation method for smart home scenes is proposed. This embodiment mainly includes the following steps:
步骤S301,图像采集;Step S301, image acquisition;
主要通过2D摄像头采集家居图像。The home image is collected mainly through a 2D camera.
步骤S302,对采集到的图像进行预处理;Step S302, preprocessing the collected image;
对图像进行滤波处理,形态学处理,二值化处理等。Perform filtering, morphological, and binarization processing on the image.
步骤S303,对图像进行肤色分割;Step S303, perform skin tone segmentation on the image;
通过把在YCbCr颜色空间中使用全局固定阈值方法,对其进行二值化处理,并通过八邻域法获取每个区域的轮廓信息。By using the global fixed threshold method in the YCbCr color space, it is binarized, and the contour information of each area is obtained by the eight-neighbor method.
步骤S304,对非手势区域进行过滤;Step S304, filtering the non-gesture area;
通过对步骤S303中分割出的手势图像进行非手势区域过滤,首先构建手势图像的MBR,对不符合条件的区域进行过滤,符合条件的则进行手势分割处理。By performing non-gesture region filtering on the gesture image segmented in step S303, the MBR of the gesture image is first constructed, regions that do not meet the conditions are filtered, and those that meet the conditions are subjected to gesture segmentation processing.
步骤S305,对图像进行手势分割;Step S305, perform gesture segmentation on the image;
在手势图像MBR的基础上构建MABR,并获取得到手势图像的偏转角度,通过分析手部宽度分布直方图与手部梯度分布直方图,获取得到手势的手腕分割线,并对手臂区域进行过滤。The MABR is constructed on the basis of the gesture image MBR, and the deflection angle of the gesture image is obtained. By analyzing the histogram of the hand width distribution and the hand gradient distribution histogram, the division line of the wrist of the gesture is obtained, and the arm area is filtered.
步骤S306,获取得到完整的手势图像;Step S306, obtaining a complete gesture image;
经过手势分割后,会产生0~多个手势,可以把图像中的手势都提取出来,用于后续需要,主要用于基于手势的人机交互系统,实现人们通过手势对家居设备的控制。After the gesture segmentation, 0 to multiple gestures will be generated, and all the gestures in the image can be extracted for subsequent needs. It is mainly used for gesture-based human-computer interaction systems to realize people's control of home equipment through gestures.
本发明所提出的面向智能家居场景的多手势精准分割方法,可以在本地智能化地对手势进行分割,克服了现有技术过于依赖网络的弊端,使得应用本方法的设备在没有网络连接的情况下仍然可以正常工作。The multi-gesture precise segmentation method for smart home scenes proposed by the present invention can segment the gestures intelligently locally, overcomes the drawbacks of the prior art that is too dependent on the network, and makes the device applying this method not connected to the network. It can still work normally.
本发明通过将手势图片从RGB颜色空间转化到YCbCr颜色空间、再通过全局固定阈值二值化的方法,完成了对肤色的分割。随后,对非手势区域进行排除,构建出手势轮廓的MBR与MABR,并对手势图像进行旋转以统计手部宽度、构建出宽度分布直方图与基于宽度的梯度分布直方图,确定手腕分割线。最后,完成对手臂冗余的去除,得到完整的手势图像。本发明能够快速且准确地对家居环境图像中的手势进行分割,显著地提高了基于手势的人机交互系统的使用舒适度,提升了用户满意度。The present invention completes the segmentation of the skin color by converting the gesture picture from the RGB color space to the YCbCr color space, and then by the method of global fixed threshold binarization. Subsequently, the non-gesture area is excluded, the MBR and MABR of the gesture contour are constructed, the gesture image is rotated to count the hand width, the width distribution histogram and the width-based gradient distribution histogram are constructed, and the wrist division line is determined. Finally, the removal of arm redundancy is completed, and a complete gesture image is obtained. The invention can quickly and accurately segment the gestures in the home environment image, significantly improves the use comfort of the gesture-based human-computer interaction system, and improves user satisfaction.
此外,本发明也为同领域内的其他相关问题提供了参考,可以以此为依据进行拓展延伸,运用于其他关于手势分割的技术方案中,具有十分广阔的应用前景。In addition, the present invention also provides a reference for other related issues in the same field, which can be used as a basis for expansion and application in other technical solutions related to gesture segmentation, and has very broad application prospects.
对于本领域技术人员而言,显然本发明不限于上述示范性实施例的细节,而且在不背离本发明的精神和基本特征的情况下,能够以其他的具体形式实现本发明。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本发明的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化囊括在本发明内,不应将权利要求中的任何附图标记视为限制所涉及的权利要求。For those skilled in the art, it is obvious that the present invention is not limited to the details of the above exemplary embodiments, and the present invention can be implemented in other specific forms without departing from the spirit and basic characteristics of the present invention. Therefore, from any point of view, the embodiments should be regarded as exemplary and non-restrictive. The scope of the present invention is defined by the appended claims rather than the above description, and therefore it is intended to fall within the claims. All changes within the meaning and scope of equivalent elements of are included in the present invention, and any reference signs in the claims should not be regarded as limiting the involved claims.
此外,应当理解,虽然本说明书按照实施方式加以描述,但并非每个实施方式仅包含一个独立的技术方案,说明书的这种叙述方式仅仅是为清楚起见,本领域技术人员应当将说明书作为一个整体,各实施例中的技术方案也可以经适当组合,形成本领域技术人员可以理解的其他实施方式。In addition, it should be understood that although this specification is described in accordance with the implementation manners, not each implementation manner only contains an independent technical solution. This narration in the specification is only for clarity, and those skilled in the art should regard the specification as a whole The technical solutions in each embodiment can also be appropriately combined to form other implementations that can be understood by those skilled in the art.

Claims (7)

  1. 一种面向智能家居场景的多手势精准分割方法,其特征在于,包括如下步骤:A multi-gesture accurate segmentation method for smart home scenes is characterized in that it includes the following steps:
    S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;
    S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;
    S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;
    S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;
    S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
  2. 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S1中所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein the preprocessing in S1 at least includes: gesture image denoising, gesture image binarization, and morphological processing.
  3. 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S2具体包括如下步骤:The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein S2 specifically includes the following steps:
    S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3;S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then compare each pixel with the threshold through the global fixed threshold binarization method to obtain the binarized image Image3;
    S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
  4. 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S3具体包括如下步骤:将S2中所获得到的二值化手势图像的 轮廓信息存放至列表contours中,并根据坐标信息获得外接矩形的四个顶点坐标,分别为top_left,top_right,bottom_left与bottom_right。The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein S3 specifically includes the following steps: storing the contour information of the binarized gesture image obtained in S2 in the list contours, and According to the coordinate information, the coordinates of the four vertices of the bounding rectangle are obtained, which are top_left, top_right, bottom_left and bottom_right.
  5. 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S4中所述非手势区域排除准则,具体包括:The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein the non-gesture area exclusion criterion in S4 specifically includes:
    1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;
    2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;
    3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
  6. 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。The multi-gesture accurate segmentation method for smart home scenes according to claim 1, wherein the arm redundancy removal algorithm based on hand shape features in S5 specifically includes: counting the hand width distribution of the image Image6 Histogram and gradient distribution histogram, where the maximum width of the gesture width distribution histogram and its corresponding coordinates are the thumb carpal joint, and the coordinates of the wrist division line are found in the gradient distribution histogram after the thumb carpal joint point The value is determined.
  7. 根据权利要求6所述的面向智能家居场景的多手势精准分割方法,其特征在于,S5步骤中所述手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定,确定标准为:当前点的梯度为0,且下一点的梯度大于等于0。The multi-gesture precise segmentation method for smart home scenes according to claim 6, wherein the coordinates of the wrist segmentation line in step S5 are determined by finding the value in the histogram of the gradient distribution after the thumb palmar joint point , The determination criterion is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
PCT/CN2019/092970 2019-02-26 2019-06-26 Multi-gesture precise segmentation method for smart home scenario WO2020173024A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2020515730A JP6932402B2 (en) 2019-02-26 2019-06-26 Multi-gesture fine division method for smart home scenes

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910140430.6A CN109961016B (en) 2019-02-26 2019-02-26 Multi-gesture accurate segmentation method for smart home scene
CN201910140430.6 2019-02-26

Publications (1)

Publication Number Publication Date
WO2020173024A1 true WO2020173024A1 (en) 2020-09-03

Family

ID=67023818

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/092970 WO2020173024A1 (en) 2019-02-26 2019-06-26 Multi-gesture precise segmentation method for smart home scenario

Country Status (3)

Country Link
JP (1) JP6932402B2 (en)
CN (1) CN109961016B (en)
WO (1) WO2020173024A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112613355A (en) * 2020-12-07 2021-04-06 北京理工大学 Gesture segmentation method based on island searching algorithm

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112070859B (en) * 2020-09-16 2021-05-04 山东晨熙智能科技有限公司 Photo image automatic filling method and system for photo book
CN112949542A (en) * 2021-03-17 2021-06-11 哈尔滨理工大学 Wrist division line determining method based on convex hull detection
CN113204991B (en) 2021-03-25 2022-07-15 南京邮电大学 Rapid face detection method based on multilayer preprocessing

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110282140A1 (en) * 2010-05-14 2011-11-17 Intuitive Surgical Operations, Inc. Method and system of hand segmentation and overlay using depth data
CN103426000A (en) * 2013-08-28 2013-12-04 天津大学 Method for detecting static gesture fingertip
CN106557173A (en) * 2016-11-29 2017-04-05 重庆重智机器人研究院有限公司 Dynamic gesture identification method and device
CN109190496A (en) * 2018-08-09 2019-01-11 华南理工大学 A kind of monocular static gesture identification method based on multi-feature fusion

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH11167455A (en) * 1997-12-05 1999-06-22 Fujitsu Ltd Hand form recognition device and monochromatic object form recognition device
JP4332649B2 (en) * 1999-06-08 2009-09-16 独立行政法人情報通信研究機構 Hand shape and posture recognition device, hand shape and posture recognition method, and recording medium storing a program for executing the method
CN106325485B (en) * 2015-06-30 2019-09-10 芋头科技(杭州)有限公司 A kind of gestures detection recognition methods and system
CN108345867A (en) * 2018-03-09 2018-07-31 南京邮电大学 Gesture identification method towards Intelligent household scene
CN109214297A (en) * 2018-08-09 2019-01-15 华南理工大学 A kind of static gesture identification method of combination depth information and Skin Color Information

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110282140A1 (en) * 2010-05-14 2011-11-17 Intuitive Surgical Operations, Inc. Method and system of hand segmentation and overlay using depth data
CN103426000A (en) * 2013-08-28 2013-12-04 天津大学 Method for detecting static gesture fingertip
CN106557173A (en) * 2016-11-29 2017-04-05 重庆重智机器人研究院有限公司 Dynamic gesture identification method and device
CN109190496A (en) * 2018-08-09 2019-01-11 华南理工大学 A kind of monocular static gesture identification method based on multi-feature fusion

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CAO, XINYAN : "Monocular Vision Gesture Segmentation Based on Skin Color and Motion Detection", JOURNAL OF HUNAN UNIVERSITY (NATURAL SCIENCE), vol. 38, no. 01, 31 January 2011 (2011-01-31), pages 1 - 83, XP009522952 *
GONG, TAOBO: "Computer Vision-Based Static Hand Gesture Recognition System", ELECTRONIC TECHNOLOGY & INFORMATION SCIENCE, CHINA MASTER’S THESES FULL-TEXT DATABASE, 15 September 2008 (2008-09-15), pages 1 - 56, XP009522953, ISSN: 1674-0246 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112613355A (en) * 2020-12-07 2021-04-06 北京理工大学 Gesture segmentation method based on island searching algorithm
CN112613355B (en) * 2020-12-07 2022-07-26 北京理工大学 Gesture segmentation method based on island searching algorithm

Also Published As

Publication number Publication date
JP6932402B2 (en) 2021-09-08
CN109961016A (en) 2019-07-02
JP2021517281A (en) 2021-07-15
CN109961016B (en) 2022-10-14

Similar Documents

Publication Publication Date Title
WO2020173024A1 (en) Multi-gesture precise segmentation method for smart home scenario
EP1969559B1 (en) Contour finding in segmentation of video sequences
EP1969560B1 (en) Edge-controlled morphological closing in segmentation of video sequences
JP5538909B2 (en) Detection apparatus and method
EP1969562B1 (en) Edge-guided morphological closing in segmentation of video sequences
EP1969561A1 (en) Segmentation of video sequences
EP2863362B1 (en) Method and apparatus for scene segmentation from focal stack images
US20150319373A1 (en) Method and device to compose an image by eliminating one or more moving objects
JPH0877334A (en) Automatic feature point extracting method for face image
US10885321B2 (en) Hand detection method and system, image detection method and system, hand segmentation method, storage medium, and device
CN105894464A (en) Median filtering image processing method and apparatus
JP2018045693A (en) Method and system for removing background of video
JP2010057105A (en) Three-dimensional object tracking method and system
CN111539980A (en) Multi-target tracking method based on visible light
Ansari et al. An approach for human machine interaction using dynamic hand gesture recognition
CN109934152B (en) Improved small-bent-arm image segmentation method for sign language image
Avinash et al. Color hand gesture segmentation for images with complex background
CN108717699B (en) Ultrasonic image segmentation method based on continuous minimum segmentation
CN108491820B (en) Method, device and equipment for identifying limb representation information in image and storage medium
Huang et al. M2-Net: multi-stages specular highlight detection and removal in multi-scenes
CN111831123B (en) Gesture interaction method and system suitable for desktop mixed reality environment
Chuang et al. Moving object segmentation and tracking using active contour and color classification models
CN114283157A (en) Ellipse fitting-based ellipse object segmentation method
CN111932470A (en) Image restoration method, device, equipment and medium based on visual selection fusion
Malavika et al. Moving object detection and velocity estimation using MATLAB

Legal Events

Date Code Title Description
ENP Entry into the national phase

Ref document number: 2020515730

Country of ref document: JP

Kind code of ref document: A

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19805113

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 19805113

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE