WO2020173024A1 - Multi-gesture precise segmentation method for smart home scenario - Google Patents
Multi-gesture precise segmentation method for smart home scenario Download PDFInfo
- Publication number
- WO2020173024A1 WO2020173024A1 PCT/CN2019/092970 CN2019092970W WO2020173024A1 WO 2020173024 A1 WO2020173024 A1 WO 2020173024A1 CN 2019092970 W CN2019092970 W CN 2019092970W WO 2020173024 A1 WO2020173024 A1 WO 2020173024A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- gesture
- image
- smart home
- area
- segmentation method
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/50—Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/56—Extraction of image or video features relating to colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Disclosed in the present invention is a multi-gesture precise segmentation method for a smart home scenario, comprising the following steps: S1: pre-processing a gesture image Image0 to obtain an image Image1; S2: performing skin tone segmentation on the pre-processed image Image1 to obtain a processed image Image4; S3: constructing an image minimum bounding rectangle (MBR) in the image Image4; S4: excluding the non-gesture areas in the image Image4 by means of a non-gesture area exclusion criterion to acquire a gesture image Image5; and S5: processing the image Image5 by means of an arm redundancy removal algorithm based on hand shape features to implement removal of arm redundancy. The present invention can locally and smartly segment gestures, and the entire process is quick and accurate, significantly increasing the comfort of use of gesture-based human-machine interaction systems.
Description
本发明涉及一种智能识别方法,具体涉及一种面向智能家居场景的多手势精准分割方法,属于智能家居领域。The invention relates to an intelligent recognition method, in particular to a multi-gesture accurate segmentation method oriented to a smart home scene, and belongs to the field of smart home.
手势分割是指将手势信息从复杂的图像背景中分割出来的技术,手势分割质量(准确性、完整性、冗余性)的好坏对基于手势的人机交互系统的识别与检测准确率有着重要的影响。Gesture segmentation refers to the technology of segmenting gesture information from the complex image background. The quality of gesture segmentation (accuracy, completeness, redundancy) is good or bad for the recognition and detection accuracy of gesture-based human-computer interaction systems. important influence.
面向家居场景下的实时手势分割更为复杂,用户手势不仅更为复杂多变,而且易受背景、光照与拍摄角度等因素的影响。在目前的计算机视觉领域内,尚未存在一种能够自适应的手势分割算法。现行的一些具有代表性的手势分割方法主要依赖于外部设备或者需要对用户的手部进行特殊处理,但由于其限制了人们活动范围,同时需要配套的硬件设备,造价及其昂贵等原因,这些技术也很难在实际应用中得到大规模的推广。Real-time gesture segmentation in home-oriented scenarios is more complex. User gestures are not only more complex and changeable, but also vulnerable to factors such as background, lighting, and shooting angle. In the current computer vision field, there is no adaptive gesture segmentation algorithm. Some of the current representative gesture segmentation methods mainly rely on external devices or require special processing of the user’s hands. However, they limit the range of people's activities and require supporting hardware equipment, cost and expensive. The technology is also difficult to be widely promoted in practical applications.
与技术发展相对应的是,当今市场上,具有手势分割功能的智能家居设备还较为少见,大多数的手势分割产品仅仅停留在对皮肤的分割上,并不能完全且准确地针对手势进行分割,分割效果并不理想。并且这些设备大多依靠云端服务器、过于依赖网络,在没有网络的情况下将无法进行工作。Corresponding to technological development, smart home devices with gesture segmentation are still relatively rare in the market today. Most gesture segmentation products only focus on segmentation of the skin, and cannot completely and accurately segment the gestures. The segmentation effect is not ideal. And most of these devices rely on cloud servers and rely too much on the network, and will not be able to work without a network.
综上所述,如何在现有技术的基础上提出一种全新的面向智能家居场景的多手势精准分割方法,以实现手势分割技术在智能家居设备上的大规 模推广应用,也就成为了目前业内人士所亟待解决的问题。In summary, how to propose a new multi-gesture precision segmentation method for smart home scenes based on the existing technology to realize the large-scale promotion and application of gesture segmentation technology on smart home devices has become the current Problems that need to be solved urgently by the insiders.
发明内容Summary of the invention
鉴于现有技术存在上述缺陷,本发明的目的是提出一种面向智能家居场景的多手势精准分割方法,包括如下步骤:In view of the above-mentioned defects in the prior art, the purpose of the present invention is to propose a multi-gesture accurate segmentation method for smart home scenes, including the following steps:
S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;
S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;
S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
优选地,S1中所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。Preferably, the preprocessing in S1 at least includes: gesture image denoising, gesture image binarization, and morphological processing.
优选地,S2具体包括如下步骤:Preferably, S2 specifically includes the following steps:
S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3;S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then compare each pixel with the threshold through the global fixed threshold binarization method to obtain the binarized image Image3;
S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
优选地,S3具体包括如下步骤:将S2中所获得到的二值化手势图像的轮廓信息存放至列表contours中,并根据坐标信息获得外接矩形的四个顶点坐标,分别为top_left,top_right,bottom_left与bottom_right。Preferably, S3 specifically includes the following steps: store the contour information of the binarized gesture image obtained in S2 into the list contours, and obtain the coordinates of the four vertices of the circumscribed rectangle according to the coordinate information, which are top_left, top_right, and bottom_left. With bottom_right.
优选地,S4中所述非手势区域排除准则,具体包括:Preferably, the non-gesture area exclusion criterion in S4 specifically includes:
1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;
2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;
3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
优选地,S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。Preferably, the arm redundancy removal algorithm based on hand shape features described in S5 specifically includes: counting the hand width distribution histogram and gradient distribution histogram of the image Image6, wherein the width of the gesture width distribution histogram is the largest The value and its corresponding coordinate are the thumb carpal joint, and the coordinates of the wrist dividing line are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point.
优选地,S5步骤中所述手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定,确定标准为:当前点的梯度为0,且下一点的梯度大于等于0。Preferably, the coordinates of the wrist dividing line in step S5 are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point, and the determination criterion is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
与现有技术相比,本发明的优点主要体现在以下几个方面:Compared with the prior art, the advantages of the present invention are mainly embodied in the following aspects:
本发明所提出的面向智能家居场景的多手势精准分割方法,可以在本地智能化地对手势进行分割,克服了现有技术过于依赖网络的弊端,使得应用本方法的设备在没有网络连接的情况下仍然可以正常工作。The multi-gesture precise segmentation method for smart home scenes proposed by the present invention can segment the gestures intelligently locally, overcomes the drawbacks of the prior art that is too dependent on the network, and makes the device applying this method not connected to the network. It can still work normally.
本发明通过将手势图片从RGB颜色空间转化到YCbCr颜色空间、再通过全局固定阈值二值化的方法,完成了对肤色的分割。随后,对非手势区域进行排除,构建出手势轮廓的MBR与MABR,并对手势图像进行旋转以统计手部宽度、构建出宽度分布直方图与基于宽度的梯度分布直方图,确定手腕分割线。最后,完成对手臂冗余的去除,得到完整的手势图像。本发明能够快速且准确地对家居环境图像中的手势进行分割,显著地提高了基于手势的人机交互系统的使用舒适度,提升了用户满意度。The present invention completes the segmentation of the skin color by converting the gesture picture from the RGB color space to the YCbCr color space, and then by the method of global fixed threshold binarization. Subsequently, the non-gesture area is excluded, the MBR and MABR of the gesture contour are constructed, the gesture image is rotated to count the hand width, the width distribution histogram and the width-based gradient distribution histogram are constructed, and the wrist division line is determined. Finally, the removal of arm redundancy is completed, and a complete gesture image is obtained. The invention can quickly and accurately segment the gestures in the home environment image, significantly improves the use comfort of the gesture-based human-computer interaction system, and improves user satisfaction.
此外,本发明也为同领域内的其他相关问题提供了参考,可以以此为依据进行拓展延伸,运用于其他关于手势分割的技术方案中,具有十分广阔的应用前景。In addition, the present invention also provides a reference for other related issues in the same field, which can be used as a basis for expansion and application in other technical solutions related to gesture segmentation, and has very broad application prospects.
以下便结合实施例附图,对本发明的具体实施方式作进一步的详述,以使本发明技术方案更易于理解、掌握。The specific implementation of the present invention will be described in further detail below in conjunction with the accompanying drawings of the embodiments, so as to make the technical solution of the present invention easier to understand and grasp.
图1为本发明所提供的对手势图像进行肤色分割的步骤流程示意图;FIG. 1 is a schematic diagram of the steps of performing skin color segmentation on gesture images provided by the present invention;
图2为本发明所提供的对手势图像进行手臂冗余去除的步骤流程示意图;FIG. 2 is a schematic flow chart of steps for removing arm redundancy from gesture images provided by the present invention;
图3为本发明所提供的面向智能家居场景的多手势精准分割方法的总体步骤流程示意图。FIG. 3 is a schematic flow diagram of the overall steps of the multi-gesture precise segmentation method for smart home scenes provided by the present invention.
本发明揭示了一种面向智能家居场景的多手势精准分割方法,所述方法基于YCbCr颜色空间的肤色分割算法,非手势区域排除准则,基于手部形状特征的手臂冗余去除算法。本发明的方法包括如下步骤:The present invention discloses a multi-gesture precise segmentation method for smart home scenes. The method is based on a skin color segmentation algorithm of YCbCr color space, non-gesture region exclusion criteria, and an arm redundancy removal algorithm based on hand shape features. The method of the present invention includes the following steps:
S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;
S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;
S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
由上述步骤表述可以看出,本发明的方法主要包括两大方面,即肤色分割及手臂冗余去除。It can be seen from the above steps that the method of the present invention mainly includes two major aspects, namely skin color segmentation and arm redundancy removal.
以下结合附图对本发明的方法进行具体说明,图1显示了本发明实施例提供的一种对手势图像进行肤色分割的方法,该方法的步骤主要包含:Hereinafter, the method of the present invention will be described in detail with reference to the accompanying drawings. FIG. 1 shows a method for segmenting a gesture image according to an embodiment of the present invention. The steps of the method mainly include:
S1、对手势图像Image0进行预处理,得到图像Image1。S1. Preprocess the gesture image Image0 to obtain the image Image1.
由于手势图像在获取时不可避免的会存在噪声的干扰,会对手势图像的分割与识别造成严重的影响,因此在对手势分割前对图像进行预处理显得尤为重要。所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。As the gesture image is inevitably interfered by noise when it is acquired, it will have a serious impact on the segmentation and recognition of the gesture image, so it is particularly important to preprocess the image before the gesture segmentation. The preprocessing at least includes: gesture image denoising, gesture image binarization and morphological processing.
其中,手势图像去噪主要使用高斯滤波器,该滤波器属于线性滤波器,其滤波器窗口的像素取值服从高斯分布,随着距离模板中心的距离增大而减小,其二维高斯函数为:Among them, the gesture image denoising mainly uses a Gaussian filter, which is a linear filter. The pixel value of the filter window obeys the Gaussian distribution, and decreases with the increase of the distance from the center of the template. Its two-dimensional Gaussian function for:
其中,h(x,y)表示高斯滤波器中(x,y)坐标上的取值,σ表示标准差。Among them, h(x,y) represents the value on the (x,y) coordinate in the Gaussian filter, and σ represents the standard deviation.
S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3。S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then use the global fixed threshold binarization method to compare each pixel with the threshold value to obtain the binarized image Image3.
YCbCr颜色空间是视频图像和数字图像中常用的色彩空间。包含三个分量:Y(luma,亮度),表示的是图像的亮暗程度,取值范围为0~255;Cb分量表示的是RGB颜色空间中蓝色分量与RGB颜色空间中亮度值之间差异,取值范围为0~255;Cr分量表示的是RGB颜色空间中红色分量的值与RGB颜色空间中亮度之间的差异,取值范围为0~255。其中Cb分量和Cr分量是相互独立的,并且与Y分量能有效地分离。YCbCr color space is a commonly used color space in video images and digital images. Contains three components: Y (luma, brightness), which represents the brightness and darkness of the image, ranging from 0 to 255; Cb component represents the blue component in the RGB color space and the brightness value in the RGB color space The value range of the difference is 0-255; the Cr component represents the difference between the value of the red component in the RGB color space and the brightness in the RGB color space, and the value range is 0-255. The Cb component and the Cr component are independent of each other and can be effectively separated from the Y component.
RGB颜色空间到YCbCr颜色空间的转换公式如下:The conversion formula from RGB color space to YCbCr color space is as follows:
转化为矩阵形式为:Converted to matrix form:
上述步骤中所述的将每个像素与阈值进行比较,具体操作为,人的肤色的Y,Cb,Cr值大约为[0:256,130:174,77:128],如果图像中像素的YCbCr值属于这个区间,则该像素值置为255,否则置为0,则可得到二值化图像Image3。In the above steps, each pixel is compared with the threshold. The specific operation is that the Y, Cb, and Cr values of human skin are approximately [0:256,130:174,77:128], if the YCbCr value of the pixel in the image If it belongs to this interval, the pixel value is set to 255, otherwise it is set to 0, and the binary image Image3 can be obtained.
S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
手势图像经过二值化处理后,得到的图像会存在空隙、残缺等现象。形态学的作用就是去除孤立的小点、毛刺、填充小孔、弥合小缝隙等,形态学操作主要有以下4种:After the gesture image is binarized, the resulting image will have gaps and defects. The role of morphology is to remove isolated dots, burrs, fill small holes, bridge small gaps, etc. There are four main types of morphological operations:
1、膨胀。形态学操作中的膨胀运算过程是:将物体接触到的背景点合并到该物体。其结果使目标物体的面积变大;其意义在于对目标区域中存在的空洞与缝隙进行填充。1. Expansion. The process of dilation operation in morphological operation is to merge the background points touched by the object into the object. As a result, the area of the target object becomes larger; its meaning is to fill the holes and gaps in the target area.
2、腐蚀。形态学操作中的腐蚀运算过程是:将物体的所有边界点进行消除。其结果使目标物体的面积变小;其意义在于对目标区域中存在的一些较小的没有意义的孤立点进行消除。2. Corrosion. The corrosion calculation process in the morphological operation is to eliminate all boundary points of the object. As a result, the area of the target object becomes smaller; its significance is to eliminate some small meaningless isolated points in the target area.
3、开运算。形态学操作中的开运算过程先对二值化图像进行腐蚀运算,然后在对其进行膨胀运算。其意义在于对目标区域中存在的孤立小点、毛 刺等没有意义的点进行消除(腐蚀运算),空洞与缝隙进行填充(膨胀运算)。3. Open operation. The opening operation process in the morphological operation first performs the erosion operation on the binary image, and then performs the expansion operation on it. Its significance is to eliminate the isolated small dots, burrs and other meaningless points in the target area (corrosion operation), and fill cavities and gaps (expansion operation).
4、闭运算。形态学操作中的闭运算过程先对二值化图像进行膨胀运算,然后在对其进行腐蚀运算。其意义在于对目标区域中存在的空洞与缝隙进行填充(膨胀运算),孤立小点、毛刺等没有意义的点进行消除(腐蚀运算)。4. Close operation. The closing operation process in the morphological operation first performs dilation operation on the binary image, and then performs erosion operation on it. Its significance lies in filling the voids and gaps in the target area (expansion operation), and eliminating isolated small dots, burrs and other meaningless points (corrosion operation).
所述的中值滤波器属于非线性滤波器,主要是对当前点的周围像素点进行统计与排序,选取其中值作为当前点的像素值,从而消除孤立的噪声点。主要用中值滤波器平滑手势二值化图像边缘的毛刺,使其边缘变得平滑,降低对手腕分割线的查找带来的影响。The median filter is a non-linear filter, which mainly counts and sorts the surrounding pixels of the current point, and selects the median value as the pixel value of the current point, thereby eliminating isolated noise points. The median filter is mainly used to smooth the burrs on the edges of the gesture binarization image, so that the edges are smoothed, and the influence of the search of the wrist division line is reduced.
图2为本发明实施例提供的对手势图像进行手臂冗余去除方法,该方法的步骤主要包含:2 is a method for removing arm redundancy from a gesture image provided by an embodiment of the present invention. The steps of the method mainly include:
S3、构建手势图像的最小面积绑定矩形MABR;S3. Construct the minimum area binding rectangle MABR of the gesture image;
在图像Image4中构建手势图像的最小绑定矩形MBR,其顶点坐标信息为,Construct the smallest bound rectangle MBR of the gesture image in Image4, and its vertex coordinate information is,
在MBR的基础上构建图像的MABR,在已知轮廓的前提下可以根据Graham扫描法求手势轮廓的凸包,将图形以其MBR的中心为原点,在其90度范围内以β为尺度等间隔的进行旋转。同时记录相应旋转角度下图形的MBR面积,则在记录中最小的MBR面积所对应的MBR就是所求的MABR。The MABR of the image is constructed on the basis of the MBR. Under the premise of the known contour, the convex hull of the gesture contour can be obtained according to the Graham scanning method. The center of the MBR of the graphic is the origin, and the β is the scale within the 90 degree range, etc. Rotate at intervals. At the same time, the MBR area of the graphics under the corresponding rotation angle is recorded, and the MBR corresponding to the smallest MBR area in the record is the required MABR.
S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排 除,获取手势图像Image5。S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5.
所述非手势区域排除准则,具体包括:The non-gesture area exclusion criteria specifically include:
1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;
2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;
3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
随后逆时针旋转二值化手势图像,在上述步骤中可以得到MABR对应的旋转角度,逆时针旋转手势图像,即可使手势方向变为垂直。Then rotate the binarized gesture image counterclockwise, the rotation angle corresponding to the MABR can be obtained in the above steps, and rotate the gesture image counterclockwise to make the gesture direction vertical.
S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。The arm redundancy removal algorithm based on hand shape features described in S5 specifically includes: Counting image Image6's hand width distribution histogram and gradient distribution histogram, where the maximum width of the gesture width distribution histogram and its The corresponding coordinate is the thumb carpal joint, and the coordinates of the wrist dividing line are determined by finding the value in the histogram of the gradient distribution after the thumb carpal joint point.
所述宽度直方图的计算程序如下:The calculation procedure of the width histogram is as follows:
所述梯度直方图的计算程序如下:The calculation procedure of the gradient histogram is as follows:
gradient=[0]gradient=[0]
for index in range(1,len(width)):for index in range(1,len(width)):
gradient.append(width[index]-width[index-1])。Gradient.append(width[index]-width[index-1]).
随后,确定手腕分割线,由于手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标可以通过查找拇指腕掌关节点之后的梯度分布直方图中的值确定,其确定方法为:当前点的梯度为0,且下一点的梯度大于等于0。Then, determine the wrist segmentation line. Since the maximum width of the gesture width distribution histogram and its corresponding coordinates are the thumb carpal joint, the coordinates of the wrist segmentation line can be found by looking for the value in the gradient distribution histogram after the thumb carpal joint point To determine, the method for determining is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
最后,完成手臂冗余去除。在上述步骤中获取得到手腕分割线的坐标信息,令在手腕分割线下面的像素值为0,即只保留上部手势图像,手臂部分去除掉。Finally, complete the arm redundancy removal. In the above steps, the coordinate information of the wrist dividing line is obtained, and the pixel value under the wrist dividing line is set to 0, that is, only the upper gesture image is retained, and the arm part is removed.
以下结合图3,提出一面向智能家居场景的多手势精准分割方法的具体实施例,该实施例主要包括如下步骤:In the following, in conjunction with FIG. 3, a specific embodiment of a multi-gesture precise segmentation method for smart home scenes is proposed. This embodiment mainly includes the following steps:
步骤S301,图像采集;Step S301, image acquisition;
主要通过2D摄像头采集家居图像。The home image is collected mainly through a 2D camera.
步骤S302,对采集到的图像进行预处理;Step S302, preprocessing the collected image;
对图像进行滤波处理,形态学处理,二值化处理等。Perform filtering, morphological, and binarization processing on the image.
步骤S303,对图像进行肤色分割;Step S303, perform skin tone segmentation on the image;
通过把在YCbCr颜色空间中使用全局固定阈值方法,对其进行二值化处理,并通过八邻域法获取每个区域的轮廓信息。By using the global fixed threshold method in the YCbCr color space, it is binarized, and the contour information of each area is obtained by the eight-neighbor method.
步骤S304,对非手势区域进行过滤;Step S304, filtering the non-gesture area;
通过对步骤S303中分割出的手势图像进行非手势区域过滤,首先构建手势图像的MBR,对不符合条件的区域进行过滤,符合条件的则进行手势分割处理。By performing non-gesture region filtering on the gesture image segmented in step S303, the MBR of the gesture image is first constructed, regions that do not meet the conditions are filtered, and those that meet the conditions are subjected to gesture segmentation processing.
步骤S305,对图像进行手势分割;Step S305, perform gesture segmentation on the image;
在手势图像MBR的基础上构建MABR,并获取得到手势图像的偏转角度,通过分析手部宽度分布直方图与手部梯度分布直方图,获取得到手势的手腕分割线,并对手臂区域进行过滤。The MABR is constructed on the basis of the gesture image MBR, and the deflection angle of the gesture image is obtained. By analyzing the histogram of the hand width distribution and the hand gradient distribution histogram, the division line of the wrist of the gesture is obtained, and the arm area is filtered.
步骤S306,获取得到完整的手势图像;Step S306, obtaining a complete gesture image;
经过手势分割后,会产生0~多个手势,可以把图像中的手势都提取出来,用于后续需要,主要用于基于手势的人机交互系统,实现人们通过手势对家居设备的控制。After the gesture segmentation, 0 to multiple gestures will be generated, and all the gestures in the image can be extracted for subsequent needs. It is mainly used for gesture-based human-computer interaction systems to realize people's control of home equipment through gestures.
本发明所提出的面向智能家居场景的多手势精准分割方法,可以在本地智能化地对手势进行分割,克服了现有技术过于依赖网络的弊端,使得应用本方法的设备在没有网络连接的情况下仍然可以正常工作。The multi-gesture precise segmentation method for smart home scenes proposed by the present invention can segment the gestures intelligently locally, overcomes the drawbacks of the prior art that is too dependent on the network, and makes the device applying this method not connected to the network. It can still work normally.
本发明通过将手势图片从RGB颜色空间转化到YCbCr颜色空间、再通过全局固定阈值二值化的方法,完成了对肤色的分割。随后,对非手势区域进行排除,构建出手势轮廓的MBR与MABR,并对手势图像进行旋转以统计手部宽度、构建出宽度分布直方图与基于宽度的梯度分布直方图,确定手腕分割线。最后,完成对手臂冗余的去除,得到完整的手势图像。本发明能够快速且准确地对家居环境图像中的手势进行分割,显著地提高了基于手势的人机交互系统的使用舒适度,提升了用户满意度。The present invention completes the segmentation of the skin color by converting the gesture picture from the RGB color space to the YCbCr color space, and then by the method of global fixed threshold binarization. Subsequently, the non-gesture area is excluded, the MBR and MABR of the gesture contour are constructed, the gesture image is rotated to count the hand width, the width distribution histogram and the width-based gradient distribution histogram are constructed, and the wrist division line is determined. Finally, the removal of arm redundancy is completed, and a complete gesture image is obtained. The invention can quickly and accurately segment the gestures in the home environment image, significantly improves the use comfort of the gesture-based human-computer interaction system, and improves user satisfaction.
此外,本发明也为同领域内的其他相关问题提供了参考,可以以此为依据进行拓展延伸,运用于其他关于手势分割的技术方案中,具有十分广阔的应用前景。In addition, the present invention also provides a reference for other related issues in the same field, which can be used as a basis for expansion and application in other technical solutions related to gesture segmentation, and has very broad application prospects.
对于本领域技术人员而言,显然本发明不限于上述示范性实施例的细节,而且在不背离本发明的精神和基本特征的情况下,能够以其他的具体形式实现本发明。因此,无论从哪一点来看,均应将实施例看作是示范性的,而且是非限制性的,本发明的范围由所附权利要求而不是上述说明限定,因此旨在将落在权利要求的等同要件的含义和范围内的所有变化囊括在本发明内,不应将权利要求中的任何附图标记视为限制所涉及的权利要求。For those skilled in the art, it is obvious that the present invention is not limited to the details of the above exemplary embodiments, and the present invention can be implemented in other specific forms without departing from the spirit and basic characteristics of the present invention. Therefore, from any point of view, the embodiments should be regarded as exemplary and non-restrictive. The scope of the present invention is defined by the appended claims rather than the above description, and therefore it is intended to fall within the claims. All changes within the meaning and scope of equivalent elements of are included in the present invention, and any reference signs in the claims should not be regarded as limiting the involved claims.
此外,应当理解,虽然本说明书按照实施方式加以描述,但并非每个实施方式仅包含一个独立的技术方案,说明书的这种叙述方式仅仅是为清楚起见,本领域技术人员应当将说明书作为一个整体,各实施例中的技术方案也可以经适当组合,形成本领域技术人员可以理解的其他实施方式。In addition, it should be understood that although this specification is described in accordance with the implementation manners, not each implementation manner only contains an independent technical solution. This narration in the specification is only for clarity, and those skilled in the art should regard the specification as a whole The technical solutions in each embodiment can also be appropriately combined to form other implementations that can be understood by those skilled in the art.
Claims (7)
- 一种面向智能家居场景的多手势精准分割方法,其特征在于,包括如下步骤:A multi-gesture accurate segmentation method for smart home scenes is characterized in that it includes the following steps:S1、对手势图像Image0进行预处理,得到图像Image1;S1. Preprocess the gesture image Image0 to obtain the image Image1;S2、对预处理后的图像Image1进行肤色分割,得到经过处理后的图像Image4;S2. Perform skin color segmentation on the preprocessed image Image1 to obtain a processed image Image4;S3、在图像Image4中构建图像的最小绑定矩形MBR;S3. Construct the smallest bound rectangle MBR of the image in the image Image4;S4、通过非手势区域排除准则对图像Image4中的非手势区域进行排除,获取手势图像Image5;S4. Exclude the non-gesture area in the image Image4 according to the non-gesture area exclusion criterion to obtain the gesture image Image5;S5、通过基于手部形状特征的手臂冗余去除算法对图像Image5进行处理,完成对手臂冗余的去除。S5. The image Image5 is processed by the arm redundancy removal algorithm based on the hand shape feature to complete the removal of the arm redundancy.
- 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S1中所述预处理至少包括:手势图像去噪、手势图像二值化及形态学处理。The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein the preprocessing in S1 at least includes: gesture image denoising, gesture image binarization, and morphological processing.
- 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S2具体包括如下步骤:The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein S2 specifically includes the following steps:S21、将Image1图像从RGB颜色空间转换到YCbCr颜色空间,得到图像Image2,再通过全局固定阈值二值化法对每个像素与阈值进行比较,得到二值化图像Image3;S21. Convert the Image1 image from the RGB color space to the YCbCr color space to obtain the image Image2, and then compare each pixel with the threshold through the global fixed threshold binarization method to obtain the binarized image Image3;S22、使用形态学中的膨胀腐蚀运算对二值化图像Image3中的孔洞与缝隙进行消除,并使用中值滤波器处理二值化图像,得到图像Image4。S22. Use the dilation and corrosion operation in the morphology to eliminate the holes and gaps in the binarized image Image3, and use the median filter to process the binarized image to obtain the image Image4.
- 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S3具体包括如下步骤:将S2中所获得到的二值化手势图像的 轮廓信息存放至列表contours中,并根据坐标信息获得外接矩形的四个顶点坐标,分别为top_left,top_right,bottom_left与bottom_right。The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein S3 specifically includes the following steps: storing the contour information of the binarized gesture image obtained in S2 in the list contours, and According to the coordinate information, the coordinates of the four vertices of the bounding rectangle are obtained, which are top_left, top_right, bottom_left and bottom_right.
- 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S4中所述非手势区域排除准则,具体包括:The multi-gesture precise segmentation method for smart home scenes according to claim 1, wherein the non-gesture area exclusion criterion in S4 specifically includes:1)外接矩形的面积小于2500时,则认定为非手势区域,其采集到的图像尺寸为640*480;1) When the area of the circumscribed rectangle is less than 2500, it is regarded as a non-gesture area, and the collected image size is 640*480;2)外接矩形的长度与宽度之比大于5时,则认定为非手势区域;2) When the ratio of the length to the width of the circumscribed rectangle is greater than 5, it is regarded as a non-gesture area;3)矩形内像素值为255的点与矩形面积的比大于0.8或小于0.4时,则认定为非手势区域。3) When the ratio of the point with the pixel value of 255 to the area of the rectangle is greater than 0.8 or less than 0.4, it is regarded as a non-gesture area.
- 根据权利要求1所述的面向智能家居场景的多手势精准分割方法,其特征在于,S5中所述基于手部形状特征的手臂冗余去除算法,具体包括:对图像Image6统计其手部宽度分布直方图与梯度分布直方图,其中,手势宽度分布直方图中的宽度最大值及其对应坐标为拇指腕掌关节,手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定。The multi-gesture accurate segmentation method for smart home scenes according to claim 1, wherein the arm redundancy removal algorithm based on hand shape features in S5 specifically includes: counting the hand width distribution of the image Image6 Histogram and gradient distribution histogram, where the maximum width of the gesture width distribution histogram and its corresponding coordinates are the thumb carpal joint, and the coordinates of the wrist division line are found in the gradient distribution histogram after the thumb carpal joint point The value is determined.
- 根据权利要求6所述的面向智能家居场景的多手势精准分割方法,其特征在于,S5步骤中所述手腕分割线的坐标通过查找拇指腕掌关节点之后的梯度分布直方图中的值进行确定,确定标准为:当前点的梯度为0,且下一点的梯度大于等于0。The multi-gesture precise segmentation method for smart home scenes according to claim 6, wherein the coordinates of the wrist segmentation line in step S5 are determined by finding the value in the histogram of the gradient distribution after the thumb palmar joint point , The determination criterion is: the gradient of the current point is 0, and the gradient of the next point is greater than or equal to 0.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2020515730A JP6932402B2 (en) | 2019-02-26 | 2019-06-26 | Multi-gesture fine division method for smart home scenes |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910140430.6A CN109961016B (en) | 2019-02-26 | 2019-02-26 | Multi-gesture accurate segmentation method for smart home scene |
CN201910140430.6 | 2019-02-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020173024A1 true WO2020173024A1 (en) | 2020-09-03 |
Family
ID=67023818
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/092970 WO2020173024A1 (en) | 2019-02-26 | 2019-06-26 | Multi-gesture precise segmentation method for smart home scenario |
Country Status (3)
Country | Link |
---|---|
JP (1) | JP6932402B2 (en) |
CN (1) | CN109961016B (en) |
WO (1) | WO2020173024A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112613355A (en) * | 2020-12-07 | 2021-04-06 | 北京理工大学 | Gesture segmentation method based on island searching algorithm |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112070859B (en) * | 2020-09-16 | 2021-05-04 | 山东晨熙智能科技有限公司 | Photo image automatic filling method and system for photo book |
CN112949542A (en) * | 2021-03-17 | 2021-06-11 | 哈尔滨理工大学 | Wrist division line determining method based on convex hull detection |
CN113204991B (en) | 2021-03-25 | 2022-07-15 | 南京邮电大学 | Rapid face detection method based on multilayer preprocessing |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110282140A1 (en) * | 2010-05-14 | 2011-11-17 | Intuitive Surgical Operations, Inc. | Method and system of hand segmentation and overlay using depth data |
CN103426000A (en) * | 2013-08-28 | 2013-12-04 | 天津大学 | Method for detecting static gesture fingertip |
CN106557173A (en) * | 2016-11-29 | 2017-04-05 | 重庆重智机器人研究院有限公司 | Dynamic gesture identification method and device |
CN109190496A (en) * | 2018-08-09 | 2019-01-11 | 华南理工大学 | A kind of monocular static gesture identification method based on multi-feature fusion |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH11167455A (en) * | 1997-12-05 | 1999-06-22 | Fujitsu Ltd | Hand form recognition device and monochromatic object form recognition device |
JP4332649B2 (en) * | 1999-06-08 | 2009-09-16 | 独立行政法人情報通信研究機構 | Hand shape and posture recognition device, hand shape and posture recognition method, and recording medium storing a program for executing the method |
CN106325485B (en) * | 2015-06-30 | 2019-09-10 | 芋头科技(杭州)有限公司 | A kind of gestures detection recognition methods and system |
CN108345867A (en) * | 2018-03-09 | 2018-07-31 | 南京邮电大学 | Gesture identification method towards Intelligent household scene |
CN109214297A (en) * | 2018-08-09 | 2019-01-15 | 华南理工大学 | A kind of static gesture identification method of combination depth information and Skin Color Information |
-
2019
- 2019-02-26 CN CN201910140430.6A patent/CN109961016B/en active Active
- 2019-06-26 JP JP2020515730A patent/JP6932402B2/en active Active
- 2019-06-26 WO PCT/CN2019/092970 patent/WO2020173024A1/en active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20110282140A1 (en) * | 2010-05-14 | 2011-11-17 | Intuitive Surgical Operations, Inc. | Method and system of hand segmentation and overlay using depth data |
CN103426000A (en) * | 2013-08-28 | 2013-12-04 | 天津大学 | Method for detecting static gesture fingertip |
CN106557173A (en) * | 2016-11-29 | 2017-04-05 | 重庆重智机器人研究院有限公司 | Dynamic gesture identification method and device |
CN109190496A (en) * | 2018-08-09 | 2019-01-11 | 华南理工大学 | A kind of monocular static gesture identification method based on multi-feature fusion |
Non-Patent Citations (2)
Title |
---|
CAO, XINYAN : "Monocular Vision Gesture Segmentation Based on Skin Color and Motion Detection", JOURNAL OF HUNAN UNIVERSITY (NATURAL SCIENCE), vol. 38, no. 01, 31 January 2011 (2011-01-31), pages 1 - 83, XP009522952 * |
GONG, TAOBO: "Computer Vision-Based Static Hand Gesture Recognition System", ELECTRONIC TECHNOLOGY & INFORMATION SCIENCE, CHINA MASTER’S THESES FULL-TEXT DATABASE, 15 September 2008 (2008-09-15), pages 1 - 56, XP009522953, ISSN: 1674-0246 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112613355A (en) * | 2020-12-07 | 2021-04-06 | 北京理工大学 | Gesture segmentation method based on island searching algorithm |
CN112613355B (en) * | 2020-12-07 | 2022-07-26 | 北京理工大学 | Gesture segmentation method based on island searching algorithm |
Also Published As
Publication number | Publication date |
---|---|
JP6932402B2 (en) | 2021-09-08 |
CN109961016A (en) | 2019-07-02 |
JP2021517281A (en) | 2021-07-15 |
CN109961016B (en) | 2022-10-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020173024A1 (en) | Multi-gesture precise segmentation method for smart home scenario | |
EP1969559B1 (en) | Contour finding in segmentation of video sequences | |
EP1969560B1 (en) | Edge-controlled morphological closing in segmentation of video sequences | |
JP5538909B2 (en) | Detection apparatus and method | |
EP1969562B1 (en) | Edge-guided morphological closing in segmentation of video sequences | |
EP1969561A1 (en) | Segmentation of video sequences | |
EP2863362B1 (en) | Method and apparatus for scene segmentation from focal stack images | |
US20150319373A1 (en) | Method and device to compose an image by eliminating one or more moving objects | |
JPH0877334A (en) | Automatic feature point extracting method for face image | |
US10885321B2 (en) | Hand detection method and system, image detection method and system, hand segmentation method, storage medium, and device | |
CN105894464A (en) | Median filtering image processing method and apparatus | |
JP2018045693A (en) | Method and system for removing background of video | |
JP2010057105A (en) | Three-dimensional object tracking method and system | |
CN111539980A (en) | Multi-target tracking method based on visible light | |
Ansari et al. | An approach for human machine interaction using dynamic hand gesture recognition | |
CN109934152B (en) | Improved small-bent-arm image segmentation method for sign language image | |
Avinash et al. | Color hand gesture segmentation for images with complex background | |
CN108717699B (en) | Ultrasonic image segmentation method based on continuous minimum segmentation | |
CN108491820B (en) | Method, device and equipment for identifying limb representation information in image and storage medium | |
Huang et al. | M2-Net: multi-stages specular highlight detection and removal in multi-scenes | |
CN111831123B (en) | Gesture interaction method and system suitable for desktop mixed reality environment | |
Chuang et al. | Moving object segmentation and tracking using active contour and color classification models | |
CN114283157A (en) | Ellipse fitting-based ellipse object segmentation method | |
CN111932470A (en) | Image restoration method, device, equipment and medium based on visual selection fusion | |
Malavika et al. | Moving object detection and velocity estimation using MATLAB |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 2020515730 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19805113 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19805113 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |