CN106778452A - Service robot is based on human testing and the tracking of binocular vision - Google Patents

Service robot is based on human testing and the tracking of binocular vision Download PDF

Info

Publication number
CN106778452A
CN106778452A CN201510830932.3A CN201510830932A CN106778452A CN 106778452 A CN106778452 A CN 106778452A CN 201510830932 A CN201510830932 A CN 201510830932A CN 106778452 A CN106778452 A CN 106778452A
Authority
CN
China
Prior art keywords
tracking
human
service robot
binocular vision
human body
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510830932.3A
Other languages
Chinese (zh)
Inventor
姜楠
李邦宇
邹风山
徐方
张强
张涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenyang Siasun Robot and Automation Co Ltd
Original Assignee
Shenyang Siasun Robot and Automation Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenyang Siasun Robot and Automation Co Ltd filed Critical Shenyang Siasun Robot and Automation Co Ltd
Priority to CN201510830932.3A priority Critical patent/CN106778452A/en
Publication of CN106778452A publication Critical patent/CN106778452A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/285Selection of pattern recognition techniques, e.g. of classifiers in a multi-classifier system

Abstract

The invention provides human testing and tracking that a kind of service robot is based on binocular vision, including human testing step and human body tracking step;The human testing step, being connected by five-category device carries out human body image detection;The human body tracking step, by being obtained per frame human body information after human testing, performs tracking, confirms target body.The function that service robot intelligently leads, follows is realized by above-mentioned interaction.

Description

Service robot is based on human testing and the tracking of binocular vision
【Technical field】
The people of binocular vision is based on the present invention relates to robotics, more particularly to a kind of service robot Physical examination survey and tracking.
【Background technology】
Human testing and the study hotspot that tracking technique is in computer vision field.The technology is mainly used In the human-computer interaction function of indoor video monitoring system, auxiliary driving and service robot.At present, this side The matured product in face is relatively fewer, especially in service robot field.
The scheme conventional with tracking for human testing, the difference according to sensor is divided three classes:1st, based on list Mesh camera;2nd, based on laser sensor;3rd, based on RGBD sensors.
During using monocular camera, full detail only has image, for common human testing algorithm, easily goes out The situation of existing error detection, certain region that will be in scene is as being human body.Because scene information is two dimension, Three-dimensional information will be lost, this kind of error detection is difficult to avoid;Simultaneously for tracking process, due to lacking three-dimensional letter Breath, it is impossible to pose estimation is carried out to target body, to obtain the movement locus of target.
The detection and tracking of human leg are realized using 2D laser sensors, the benefit of the method is processing frequency Comparatively fast, required system resources in computation is relatively fewer.But, it also has obvious defect, due to information content mistake It is few, cause the detection of human leg to be easy to error detection, need by a relatively simple scene just can correctly recognize People's leg;Simultaneously for tracking process, easily with wrong object if there is nearer people's leg apart.
Using the algorithm of RGBD sensors, the algorithm sharpest edges are the letter that sensor can provide comprehensive and abundant Breath, including image two-dimensional signal and corresponding depth information.Due to there is depth information, therefore can obtain The track of human motion is obtained to help robot to be tracked;Simultaneously because informative, can be by depth Information effectively reduces false recognition rate.
【The content of the invention】
Based on this, it is an object of the invention to provide the human testing that a kind of service robot is based on binocular vision With tracking.
In order to realize the purpose of the present invention, there is provided a kind of service robot be based on binocular vision human testing with Tracking, including human testing step and human body tracking step;The human testing step, by Pyatyi Grader series connection carries out human body image detection;The human body tracking step, by being obtained per frame after human testing Human body information, performs tracking, confirms target body.
Preferably, the first order grader in the five-category device, detects figure in the first order grader As the size and human body of detection block are fixed with the distance of camera, and the abscissa positions of detection image detection block and Human body is associated with the distance of camera.
Preferably, the second level grader in the five-category device, the second level grader passes through Haar-like features carry out human body image detection with Adaboost graders.
Preferably, the Haar features include edge feature, linear character, central feature and diagonal feature, And it is combined into feature templates.
Preferably, the feature masterplate includes two kinds of white and black, sets the characteristic value of the feature masterplate It is white rectangle pixel and the value for subtracting black rectangle pixel sum.
Preferably, the training process of the Adaboost graders is:S11, given training sample (x1, Y1) ..., (xi, yi) ..., (xn, yn), wherein xi represents i-th sample, and yi=0 is expressed as negative sample This, yi=1 is expressed as positive sample, and n is training sample sum;The weight of S12, initialization training sample;S13、 First time iteration, trains Weak Classifier, and calculate error rates of weak classifiers;Selected threshold so that error It is minimum;Update sample weights;S14, T circulation, obtain T Weak Classifier, evaluate each weak typing The weight of the importance of device is weighted superposition, finally gives strong classifier.
Preferably, the third level grader in the five-category device, the third level grader passes through Haar-like features are detected that the third level grader is enterprising in gray-scale map with Adaboost graders Row training, classification.
Preferably, the level V grader in the five-category device, the level V grader passes through Hog Feature is combined with SVM classifier.
Preferably, the HOG feature extractions are, S21, carry out gray processing to image;S22, image is entered The standardization of row color space;S23, the gradient for calculating each pixel of image;S24, divide an image into it is small Unit;S25, the histogram of gradients for counting each unit, form the describer of each unit;S26, will not Fixed number unit constitutes a block, and the profiler series connection of all units, obtains described piece in described piece HOG profilers;S27, the HOG profilers series connection by all pieces in image, obtain the figure The HOG profilers of picture.
Preferably, the human body tracking using color histogram match confirm in currently detected human body whether There is target body.
Prior art is different from, above-mentioned service robot is based on human testing and the tracking of binocular vision, Including human testing step and human body tracking step.Wherein, the human testing step, by five-category Device series connection carries out human body image detection;The human body tracking step, by being obtained per frame human body after human testing Information, performs tracking, confirms target body.Service robot is set intelligently to lead, follow by above-mentioned interaction Function.
【Brief description of the drawings】
Fig. 1 is human testing and track side of the service robot based on binocular vision in one embodiment of the invention The classifier training procedure chart of method.
Fig. 2 is human testing and track side of the service robot based on binocular vision in one embodiment of the invention The HOG feature extraction figures of method.
【Specific embodiment】
To describe technology contents of the invention, structural feature, the objects and the effects in detail, below in conjunction with Implementation method simultaneously coordinates accompanying drawing to be explained in detail.It should be appreciated that specific embodiment described herein is only used to solve The present invention is released, is not used to limit the present invention.
A kind of service robot is based on human testing and the tracking of binocular vision, including human testing step With human body tracking step.Wherein, the human testing step, being connected by five-category device carries out human figure As detection;The human body tracking step, by being obtained per frame human body information after human testing, performs tracking, Confirm target body.
In one particular embodiment of the present invention, it is a kind of based on binocular vision service robot detection with Track method, including human testing step and human body tracking step.
Wherein, the human testing step, being connected by five-category device carries out human body image detection;It is described Human body tracking step, by being obtained per frame human body information after human testing, performs tracking, confirms target body
First order grader in the five-category device, can be greatly reduced the number of detection block, lift human body The time of detection process, and for the correct classification of subsequent classifier provides basis.The first order grader Size and human body and the distance of camera of middle detection image detection block are fixed, and detection image detection block horizontal seat Cursor position and human body are associated with the distance of camera.
Based on this, the excursion of relevant parameter is obtained using logistic regression algorithm, the number of detection block is relative There is the reduction of matter in the method for general sliding window.
It may be preferred that the second level grader in the five-category device, the second level grader passes through Haar-like features carry out human body image detection with Adaboost graders.
Wherein, Haar features are used to characterize characteristics of human body, and Adaboost graders are used for the human figure for recognizing Picture.Haar features include edge feature, linear character, central feature and diagonal feature, and are combined into spy Levy template.
Feature masterplate includes two kinds of white and black, and it is white rectangle pixel to set the characteristic value of feature masterplate With the value for subtracting black rectangle pixel sum.The Haar characteristic values reflect the grey scale change situation of image.
, only to some simple graphic structures, such as edge, line segment is more sensitive, therefore can only describe for rectangular characteristic Particular orientation, level, vertical, diagonal structure.
As shown in figure 1, Adaboost enters the strong classifier that row set is obtained by by multiple Weak Classifiers, tool Body ground the Adaboost graders training process be:
S11, given training sample (x1, y1) ..., (xi, yi) ..., (xn, yn), wherein xi is represented I-th sample, yi=0 is expressed as negative sample, and yi=1 is expressed as positive sample, and n is training sample sum;
The weight of S12, initialization training sample;
S13, first time iteration, train Weak Classifier, and calculate error rates of weak classifiers;Selected threshold, So that error is minimum;Update sample weights;
S14, by T times circulate after, obtain T Weak Classifier, evaluate the importance of each Weak Classifier Weight be weighted superposition, finally give strong classifier.
It may be preferred that the third level grader in the five-category device, the third level grader passes through Haar-like features are detected with Adaboost graders, wherein, the third level grader is in gray scale It is trained on figure, is classified.The third level grader is acted on disparity map, color by disparity map relative two dimensional Color image, possesses and becomes apparent and terse feature, it is easier to carry out the confirmation of human body target.
Fourth stage grader in the five-category device is made using Haar-like features combination supporting vector machine It is grader.The SVMs (SVM) is a classifying quality preferably grader, by feature space It is middle to construct a hyperplane for optimum segmentation to realize two-value classification.
In certain embodiments, the level V grader in the five-category device, the level V grader Combined with SVM classifier by Hog features.
As shown in Fig. 2 the HOG feature extractions are,
S21, gray processing is carried out to image;Will image regard a 3-D view for x, y, z gray scale as;
S22, the standardization that color space is carried out to image;
Specifically, the standardization (normalization) of color space is carried out to input picture using Gamma correction methods; To adjust the contrast of image, the influence caused by the shade and illumination variation of image local is reduced, while can To suppress the interference of noise.
S23, the gradient for calculating each pixel of image;The gradient information includes size and Orientation, is taken turns for capturing Wide information, while the interference that further weakened light shines;
S24, junior unit cells is divided an image into, in the embodiment of the present invention, for example, 6*6 pixels/cell;
S25, the histogram of gradients for counting each unit cell, the number of different gradients form each unit Describer descriptor;
S26, indefinite several unit cell are constituted into a block block, all units in described piece of block The profiler series connection of cell, obtains described piece of HOG profilers descriptor;
S27, the HOG profilers series connection by all pieces of block in image, obtain the HOG of described image Profiler descriptor.
The human body tracking step, by being obtained per frame human body information after human testing, performs tracking, confirms Target body.Produced tracking target is floated during can effectively avoiding tracking for a long time using which The problem of shifting.
The human body tracking confirms to whether there is target in currently detected human body using color histogram match Human body.Specifically, color histogram calculating is carried out using two passages of the H in hsv color space and V, And build hsv color histogram model.
The embodiment of the present invention, is built in embedded board in concrete application, is applied to indoor service robot On.To reduce the cost of system, biocular systems are voluntarily built using two network USB cameras, use OpenCV The stereo calibration of binocular camera is carried out, and the image of 320*240 resolution sizes is used using its SGBM algorithm Stereo matching is carried out, so as to obtain the RGBD data of scene.
To complete human testing process, using multistage human body grader, human sample is made first.This hair Bright first order grader has a direct relation with the parameter of camera, the height etc. installed, therefore by binocular camera After on service robot, human body image and corresponding anaglyph are gathered.
In the embodiment of the present invention, gather positive negative sample each 600 multiple, the mode for manually marking, by people Body region is drawn as sample, then is read out training.Because feature is in actual implementation process, carry out many The simplification of the degree of kind, therefore the feature calculation time is very short, so as to ensure the training speed of algorithm quickly, is easy to The adjustment of model.
Service robot based on the binocular vision detection of the embodiment of the present invention and tracking, can be very good Realize that the intelligence of robot leads, follows function, so that the intelligent of service robot product is effectively lifted, Lift the practical value of its product.Function is led by adding service robot, makes robot application in exhibition Shop, dining room and other needs show and lead the place of function.
It should be noted that in the present invention, such as first and second or the like relational terms are used merely to One entity or operation are made a distinction with another entity or operation, and is not necessarily required or is implied this There is any this actual relation or order between a little entities or operation.And, term " including ", " bag Containing " or any other variant thereof is intended to cover non-exclusive inclusion, so that including a series of key elements Process, method, article or terminal device not only include those key elements, but also including not arranging clearly Other key elements for going out, or also include intrinsic for this process, method, article or terminal device Key element.In the absence of more restrictions, limited by sentence " including ... " or " including ... " Key element, it is not excluded that also exist in the process including above-mentioned key element, method, article or terminal device another Outer key element.Additionally, herein, " being more than ", " being less than ", " exceeding " etc. are interpreted as not including this number; " more than ", " below ", " within " etc. be interpreted as including this number.
Although being described to the various embodiments described above, those skilled in the art once know Basic creative concept, then can make other change and modification to these embodiments, so the above is above-mentioned only It is the embodiment of type of the present invention, not thereby limits scope of patent protection of the invention, it is every using the present invention Equivalent structure or equivalent flow conversion that specification and accompanying drawing content are made, or directly or indirectly it is used in other Related technical field, is similarly included within scope of patent protection of the invention.

Claims (10)

1. service robot is based on human testing and the tracking of binocular vision, it is characterised in that including people Body detecting step and human body tracking step;The human testing step, is connected into pedestrian by five-category device Body image detection;The human body tracking step, by being obtained after human testing per frame human body information, perform with Track, confirms target body.
2. service robot according to claim 1 is based on human testing and the tracking of binocular vision, Characterized in that, the first order grader in the five-category device, detects figure in the first order grader As the size and human body of detection block are fixed with the distance of camera, and the abscissa positions of detection image detection block and Human body is associated with the distance of camera.
3. service robot according to claim 1 is based on human testing and the tracking of binocular vision, Characterized in that, the second level grader in the five-category device, the second level grader passes through Haar-like features carry out human body image detection with Adaboost graders.
4. service robot according to claim 3 is based on human testing and the tracking of binocular vision, Characterized in that, the Haar features include edge feature, linear character, central feature and diagonal feature, And it is combined into feature templates.
5. service robot according to claim 4 is based on human testing and the tracking of binocular vision, Characterized in that, the feature masterplate includes two kinds of white and black, the characteristic value of the feature masterplate is set It is white rectangle pixel and the value for subtracting black rectangle pixel sum.
6. service robot according to claim 3 is based on human testing and the tracking of binocular vision, Characterized in that, the training process of the Adaboost graders is:
S11, given training sample (x1, y1) ..., (xi, yi) ..., (xn, yn), wherein xi is represented I-th sample, yi=0 is expressed as negative sample, and yi=1 is expressed as positive sample, and n is training sample sum;
The weight of S12, initialization training sample;
S13, first time iteration, train Weak Classifier, and calculate error rates of weak classifiers;Selected threshold, So that error is minimum;Update sample weights;
S14, T circulation, obtain T Weak Classifier, evaluate the weight of the importance of each Weak Classifier Superposition is weighted, strong classifier is finally given.
7. service robot according to claim 1 is based on human testing and the tracking of binocular vision, Characterized in that, the third level grader in the five-category device, the third level grader passes through Haar-like features are detected that the third level grader is enterprising in gray-scale map with Adaboost graders Row training, classification.
8. service robot according to claim 1 is based on human testing and the tracking of binocular vision, Characterized in that, the level V grader in the five-category device, the level V grader is by Hog Feature is combined with SVM classifier.
9. service robot according to claim 8 is based on human testing and the tracking of binocular vision, Characterized in that, the HOG feature extractions are,
S21, gray processing is carried out to image;
S22, the standardization that color space is carried out to image;
S23, the gradient for calculating each pixel of image;
S24, divide an image into junior unit;
S25, the histogram of gradients for counting each unit, form the describer of each unit;
S26, indefinite several units are constituted into a block, the profiler series connection of all units in described piece, Obtain described piece of HOG profilers;
S27, the HOG profilers series connection by all pieces in image, obtain the HOG features of described image Describer.
10. service robot according to claim 1 is based on human testing and the track side of binocular vision Method, it is characterised in that the human body tracking is confirmed in currently detected human body using color histogram match With the presence or absence of target body.
CN201510830932.3A 2015-11-24 2015-11-24 Service robot is based on human testing and the tracking of binocular vision Pending CN106778452A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510830932.3A CN106778452A (en) 2015-11-24 2015-11-24 Service robot is based on human testing and the tracking of binocular vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510830932.3A CN106778452A (en) 2015-11-24 2015-11-24 Service robot is based on human testing and the tracking of binocular vision

Publications (1)

Publication Number Publication Date
CN106778452A true CN106778452A (en) 2017-05-31

Family

ID=58964083

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510830932.3A Pending CN106778452A (en) 2015-11-24 2015-11-24 Service robot is based on human testing and the tracking of binocular vision

Country Status (1)

Country Link
CN (1) CN106778452A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108737362A (en) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 Register method, device, equipment and storage medium
CN110163033A (en) * 2018-02-13 2019-08-23 京东方科技集团股份有限公司 Positive sample acquisition methods, pedestrian detection model generating method and pedestrian detection method

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101196991A (en) * 2007-12-14 2008-06-11 同济大学 Close passenger traffic counting and passenger walking velocity automatic detection method and system thereof
CN101980242A (en) * 2010-09-30 2011-02-23 徐勇 Human face discrimination method and system and public safety system
CN102411368A (en) * 2011-07-22 2012-04-11 北京大学 Active vision human face tracking method and tracking system of robot
CN102496001A (en) * 2011-11-15 2012-06-13 无锡港湾网络科技有限公司 Method of video monitor object automatic detection and system thereof
CN104063689A (en) * 2014-06-20 2014-09-24 南京航空航天大学 Face image identification method based on binocular stereoscopic vision
CN104240264A (en) * 2014-08-29 2014-12-24 青岛海信电器股份有限公司 Height detection method and device for moving object
CN104821010A (en) * 2015-05-04 2015-08-05 清华大学深圳研究生院 Binocular-vision-based real-time extraction method and system for three-dimensional hand information

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101196991A (en) * 2007-12-14 2008-06-11 同济大学 Close passenger traffic counting and passenger walking velocity automatic detection method and system thereof
CN101980242A (en) * 2010-09-30 2011-02-23 徐勇 Human face discrimination method and system and public safety system
CN102411368A (en) * 2011-07-22 2012-04-11 北京大学 Active vision human face tracking method and tracking system of robot
CN102496001A (en) * 2011-11-15 2012-06-13 无锡港湾网络科技有限公司 Method of video monitor object automatic detection and system thereof
CN104063689A (en) * 2014-06-20 2014-09-24 南京航空航天大学 Face image identification method based on binocular stereoscopic vision
CN104240264A (en) * 2014-08-29 2014-12-24 青岛海信电器股份有限公司 Height detection method and device for moving object
CN104821010A (en) * 2015-05-04 2015-08-05 清华大学深圳研究生院 Binocular-vision-based real-time extraction method and system for three-dimensional hand information

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
张淑平: "基于视觉的并联机器人位姿检测方法研究", 《中国博士学位论文全文数据库 信息科技辑》 *
耿盛涛: "基于双目视觉的机器人目标检测与跟踪研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110163033A (en) * 2018-02-13 2019-08-23 京东方科技集团股份有限公司 Positive sample acquisition methods, pedestrian detection model generating method and pedestrian detection method
US11238296B2 (en) 2018-02-13 2022-02-01 Boe Technology Group Co., Ltd. Sample acquisition method, target detection model generation method, target detection method, computing device and computer readable medium
CN108737362A (en) * 2018-03-21 2018-11-02 北京猎户星空科技有限公司 Register method, device, equipment and storage medium

Similar Documents

Publication Publication Date Title
CN103530599B (en) The detection method and system of a kind of real human face and picture face
CN101406390B (en) Method and apparatus for detecting part of human body and human, and method and apparatus for detecting objects
CN106682603B (en) Real-time driver fatigue early warning system based on multi-source information fusion
CN103824070B (en) A kind of rapid pedestrian detection method based on computer vision
CN106127137A (en) A kind of target detection recognizer based on 3D trajectory analysis
CN107301378B (en) Pedestrian detection method and system based on multi-classifier integration in image
CN106446779B (en) Personal identification method and device
CN102609724B (en) Method for prompting ambient environment information by using two cameras
CN103870843B (en) Head posture estimation method based on multi-feature-point set active shape model (ASM)
CN107305635A (en) Object identifying method, object recognition equipment and classifier training method
CN106650619A (en) Human action recognition method
CN103530640B (en) Unlicensed vehicle checking method based on AdaBoost Yu SVM
CN109255375A (en) Panoramic picture method for checking object based on deep learning
CN104077577A (en) Trademark detection method based on convolutional neural network
CN104715238A (en) Pedestrian detection method based on multi-feature fusion
CN107784291A (en) target detection tracking method and device based on infrared video
CN110032932B (en) Human body posture identification method based on video processing and decision tree set threshold
CN108388871B (en) Vehicle detection method based on vehicle body regression
CN103390164A (en) Object detection method based on depth image and implementing device thereof
KR20170006355A (en) Method of motion vector and feature vector based fake face detection and apparatus for the same
CN104881662A (en) Single-image pedestrian detection method
CN105335725A (en) Gait identification identity authentication method based on feature fusion
CN106682641A (en) Pedestrian identification method based on image with FHOG- LBPH feature
CN103218605A (en) Quick eye locating method based on integral projection and edge detection
CN105096307A (en) Method for detecting objects in paired stereo images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20170531