CN102436301B - Human-machine interaction method and system based on reference region and time domain information - Google Patents

Human-machine interaction method and system based on reference region and time domain information Download PDF

Info

Publication number
CN102436301B
CN102436301B CN201110239672.4A CN201110239672A CN102436301B CN 102436301 B CN102436301 B CN 102436301B CN 201110239672 A CN201110239672 A CN 201110239672A CN 102436301 B CN102436301 B CN 102436301B
Authority
CN
China
Prior art keywords
reference zone
motion
image
human
rgb color
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201110239672.4A
Other languages
Chinese (zh)
Other versions
CN102436301A (en
Inventor
谭文伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCL Corp
Original Assignee
TCL Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by TCL Corp filed Critical TCL Corp
Priority to CN201110239672.4A priority Critical patent/CN102436301B/en
Publication of CN102436301A publication Critical patent/CN102436301A/en
Application granted granted Critical
Publication of CN102436301B publication Critical patent/CN102436301B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses a human-machine interaction method and system based on reference region and time domain information. The human-machine interaction method comprises the following steps of: acquiring a colored RGB (Red, Green and Blue) image of the reference region; acquiring a consecutive frame subtraction image; counting the motion characteristics of a two-dimensional space and the motion characteristics of the time domain; carrying out characteristic comparison on the calculated motion characteristic vector and an action type template in a preset action type template characteristic library; calculating the similarity of the motion characteristic vector and the motion characteristic template; judging the human body motion in the reference region according to the result obtained by the calculating similarity; and finally, outputting standard data of an action event for calling an outer-layer application program. According to the human-machine interaction method and system disclosed by the invention, the processing flow is simplified and the processing speed is increased. According to a self-adaptive region selection algorithm contained in the invention, the constraint for the human-machine interaction action design is reduced, the flexibility of the human body action is increased and more and more convenient human-machine interaction action types are contained.

Description

A kind of man-machine interaction method based on reference zone and time-domain information and system
Technical field
The present invention relates to Multi-media image processing technical field, in particular a kind of man-machine interaction method based on reference zone and time-domain information and system.
Background technology
Man-machine interaction (Human-Computer Interaction, writes a Chinese character in simplified form HCI): be realize people and computing system to carry out friendship mutual.Along with the fast development of computer technology, the interacting activity of people and computing machine becomes an important component part of people's daily life gradually.The traditional human-computer interaction device such as mouse, keyboard has some limitations in the naturality used and friendly etc., and the human-computer interaction technology that therefore research meets interpersonal communication custom becomes current development trend.
Man-machine interaction mode in the past needs additional specific sensor, such as handle, telepilot, data glove.This mode brings many inconveniences to user operation.The man-machine interaction mode also had has been come by the detection and tracking of staff hand shape, and because the freedom of movement of staff is very large, change of shape is large, and therefore under the natural scene of complexity, verification and measurement ratio is not high.Such method makes the more time burden of mutual needs, is difficult to the demand meeting the application of many real time human-machine interactions.
Therefore, prior art has yet to be improved and developed.
Summary of the invention
The technical problem to be solved in the present invention is, for the above-mentioned defect of prior art, a kind of man-machine interaction method based on reference zone and time-domain information and system are provided, provide and a kind ofly contain more human action dirigibilities and real-time man-machine interaction method based on computer vision, by just having skipped the detection and tracking of staff hand shape based on the system of selection of reference zone and time-domain information process, thus simplify processing procedure, improve man-machine interaction efficiency.
The technical scheme that technical solution problem of the present invention adopts is as follows:
Based on a man-machine interaction method for reference zone and time-domain information, wherein, comprise step:
A, by camera collection human body image, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select in human face region side certain region with reference to region adaptivity, and obtain reference zone RGB color image;
The neighbor frame difference image of reference zone RGB color image in B, obtaining step A;
C, the neighbor frame difference image obtained according to step B calculate the motion feature of human body in two-dimensional space reference zone, by the motion feature of computing reference region RGB color image in certain time domain, and calculate temporal motion proper vector;
D, the motion feature calculated in step C vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculate motion feature vector and the similarity of type of action masterplate;
E, according to step D calculate similarity obtain result the human motion in reference zone is judged, last output action event criteria data, for outer application call;
Described steps A specifically comprises the steps:
A1, calculate the integrogram of the human body image arrived by camera collection, extract the class rectangular characteristic of this human body image, according to predetermined sorter feature database, the method running cascade cascade searches for human face region in this human body image;
A2, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select in human face region side certain region with reference to region adaptivity, the position of computing reference region RGB color image and size: , wherein, P be comprise reference zone RGB color image central point horizontal ordinate, ordinate, width and height vector, T is mapping function, independent variable represent the center horizontal ordinate of input human face region, ordinate, width and height respectively.
The described man-machine interaction method based on reference zone and time-domain information, wherein, the predetermined sorter feature database in described steps A 1 comprises the steps:
A11, calculate the integrogram of described human body image, extract the class rectangular characteristic of described human body image;
A12, screen effective feature according to Adaboost algorithm, form Weak Classifier;
A13, by combination multiple Weak Classifier, form strong classifier;
A14, the multiple strong classifier of cascade, form the sorter feature database of Face datection.
The described man-machine interaction method based on reference zone and time-domain information, wherein, described step B specifically comprises the steps:
B1, by described reference zone RGB color image by following formula:
Carry out gray processing process; Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB color image, and f (x, y) represents gray level image, and its value is between 0 ~ 255;
B2, pass through formula: , wherein , be respectively the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this motion change value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.
The described man-machine interaction method based on reference zone and time-domain information, wherein, the statistics two-dimensional space motion feature in described step C specifically comprises:
C11, calculate motor point number according to motion change value described in neighbor frame difference image in step B2, computing formula is , wherein W, H represent the wide and high of neighbor frame difference image respectively;
C12, calculate motion centroid position according to the motor point number of neighbor frame difference image in step C11, computing formula is , ;
C13, setting time domain window N, the information of record N continuous frame reference zone RGB color image;
C14, according to step C12, the motion characteristic value of computing reference region RGB color image, computing formula is , this v value to illustrate in reference zone RGB color image human action in motion state sometime; Then temporal motion proper vector is calculated: , definition i=1 ~ N, i are natural number, then represent the i-th frame motion characteristic value .
The described man-machine interaction method based on reference zone and time-domain information, wherein, described step D comprises:
D1, predefine K type of action masterplate T i, wherein i=1 ~ K, i are natural number;
D2, utilize absolute value distance motion feature vector V that calculation procedure C14 obtains and type of action masterplate similarity ;
If D3 , and <TH, TH are predetermined threshold value, then this type of action is attributed to kth class; If do not met, do not belong to any class.
Based on a man-machine interactive system for reference zone and time-domain information, wherein, comprising:
Reference zone image collection module, for passing through camera collection human body image, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select around face certain region, to obtain reference zone RGB color image with reference to region adaptivity;
Neighbor frame difference acquisition module, for obtaining the neighbor frame difference image of described reference zone RGB color image;
Statistics and computing module, for calculating the motion feature of human body in two-dimensional space reference zone according to the neighbor frame difference image obtained, by the motion feature of computing reference region RGB color image in certain time domain, and calculate temporal motion proper vector;
Characteristic Contrast module, for calculated motion feature vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculates the similarity of motion feature vector and type of action masterplate;
Output module, for judging the human motion in reference zone according to the result calculating similarity acquisition, last output action event criteria data, for outer application call;
Described reference zone image collection module comprises:
Face search unit, for calculating the integrogram of the human body image arrived by camera collection, extracts the class rectangular characteristic of this human body image, and according to predetermined sorter feature database, the method running cascade cascade searches for human face region in this human body image;
Reference zone computing unit, for being with reference to object with face, adopts the adaptively selected algorithm of reference zone to select around face certain region with reference to region adaptivity, the position of computing reference region RGB color image and size: , wherein, P comprises reference zone RGB color image central point horizontal ordinate, ordinate, width, and the vector of height, T is mapping function, independent variable represent the center horizontal ordinate of input human face region respectively, ordinate, width and height.
The described man-machine interactive system based on reference zone and time-domain information, wherein, described neighbor frame difference acquisition module comprises:
Gray proces unit, for described reference zone RGB color image is passed through following formula:
Carry out gray processing process; Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB color image, and f (x, y) represents gray level image, and its value is between 0 ~ 255;
Neighbor frame difference image acquisition unit, for passing through formula: , wherein , for the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this motion change value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.
Man-machine interaction method based on reference zone and time-domain information provided by the present invention and system, a kind of method for testing motion made for common RGB image or depth image and motion recognition method, by just having skipped the detection and tracking of staff hand shape based on the system of selection of reference zone and time-domain information process, thus simplify processing procedure, improve man-machine interaction efficiency.
Man-machine interaction method based on reference zone and time-domain information provided by the invention and system.The method, has abandoned required staff hand-type detection and positioning in invention in the past and has followed the tracks of, simplified processing procedure, improve processing speed.The adaptive region selection algorithm that the present invention comprises, decreases the constraint to interactive action design, improves human action dirigibility, contain the type of more interactive actions more easily.
The man-machine interaction method based on reference zone and time-domain information of the present embodiment, owing to having abandoned staff hand shape detection and tracking, having which saved the processing time simultaneously, having decreased the erroneous judgement because causing when staff can't detect.Reduce time loss, the introducing of adaptive reference regional selection method, only need process reference zone and do not need process view picture two field picture, so also saved the time of many image procossing.Automatically set because this reference zone is relative position according to face, so after position of human body changes, reference zone is followed automatically.When the limb motion of people, very naturally drop in this reference zone.Such man-machine interaction is carried out very naturally easily, does not need human body to constrain in specific position.The more human action freedom and flexibility of such containing, promote the convenience of user interactions, improve man-machine interaction efficiency.
Accompanying drawing explanation
Fig. 1 is the man-machine interaction method process flow diagram that the present invention is based on reference zone and time-domain information.
Fig. 2 is the structural representation of step S100 camera collection human body image in the man-machine interaction method that the present invention is based on reference zone and time-domain information.
Fig. 3 is step S100 reference zone selection course schematic diagram in the man-machine interaction method that the present invention is based on reference zone and time-domain information.
Fig. 4 is that in the man-machine interaction method that the present invention is based on reference zone and time-domain information, reference zone confirms procedure chart.
Fig. 5 is the temporal motion feature calculation example to right translation in the man-machine interaction method step S300 that the present invention is based on reference zone and time-domain information.
Fig. 6 is the multiple reference zone selection course schematic diagram that the present invention is based in the man-machine interaction method step S100 of reference zone and time-domain information.
Fig. 7 is the theory diagram of the man-machine interactive system that the present invention is based on reference zone and time-domain information.
Embodiment
The present invention is based on the man-machine interaction method of reference zone and time-domain information and system to provide and a kind ofly contain more human action dirigibilities and real-time man-machine interaction method and system based on computer vision, for making the object of the invention, technical scheme clearly, clearly, also the present invention is elaborated referring to accompanying drawing in conjunction with the embodiments.Should be appreciated that specific embodiment described herein only in order to explain the present invention, be not intended to limit the present invention.
The invention provides a kind of man-machine interaction method based on reference zone and time-domain information, as shown in Figure 1, mainly comprise step:
Step S100: by camera collection human body image is with reference to object with face, adopts the adaptively selected algorithm of reference zone to select in human face region side certain region with reference to region adaptivity, and obtains reference zone RGB color image.As shown in Figure 2, be the collection being carried out human body image by common camera 10 by camera collection human body image, common camera 10 comprises the USB camera on PC, the camera of handheld device, the camera on intelligent television.Common camera is commonly used, and can arbitrarily gather the natural scene comprising human motion, use cost is lower.
Reach a conclusion according to human cinology's principle and a large amount of field experiment, man-machine interaction 80% upper limks movements be very naturally complete at the shoulder near zone of human body, these actions comprise waves, to left, to right translation, upwards translation, downward translation, pressings etc., therefore, the selection of reference zone mainly can be placed on this region.Step S100 concrete methods of realizing is as follows:
S110, detection face, adopt Adaboost Face datection algorithm, by judging whether have face to exist in the image that common camera is absorbed based on the Adaboost Face datection algorithm of class rectangular characteristic.First calculate the human body image integrogram arrived by camera collection, extract the class rectangular characteristic of this human body image, according to the sorter feature database trained, the method running cascade cascade searches for human face region in this human body image.
Wherein sorter feature database training method comprises: A11, calculate the integrogram of the human body image arrived by camera collection, extracts the class rectangular characteristic of this human body image; A12, screen effective feature according to Adaboost algorithm, form Weak Classifier; A13, by combination multiple Weak Classifier, form strong classifier; The multiple strong classifier of A14 cascade, forms the sorter feature database of Face datection.As shown in Figure 3, region 50 shown in solid box is exactly the human face region automatically detected.
S120, be with reference to object with face, reference zone adaptively selected algorithm is adopted to select in human face region side certain region with reference to region adaptivity, as shown in Figure 3, reference zone is shown in dotted line frame 30, after determining reference zone 30, get final product position and the size of computing reference region RGB color image: , wherein, P is the vector comprising the RGB color image central point horizontal ordinate of reference zone 30, ordinate, width and height, and T is mapping function, independent variable represent the center horizontal ordinate of human face region 50, ordinate, width and height respectively.
" reference zone " in the embodiment of the present invention take human face region as reference, the region obtained by above reference zone adaptive algorithm.Reference zone is also within the effective angular field of view of camera simultaneously.Such as mapping function T is set as linear transformation function, then , .Wherein a, b, l, m are preset value, a represents the transversal displacement of reference zone relative to human face region center, b represents the vertical misalignment amount of reference zone relative to human face region center, m is the multiplication factor of reference zone relative to human face region width, and l is the multiplication factor of reference zone relative to human face region height.
As shown in Figure 4, solid box 50 is human face region, order represents the transversal displacement a=30cm of reference zone 30 relative to human face region 50 center, represent the vertical misalignment amount b=20cm of reference zone 30 relative to human face region 50 center, represent the multiplication factor m=2 of reference zone 30 relative to human face region 50 width, represent the multiplication factor l=2 of reference zone relative to human face region 50 height, shown in the 30(dotted rectangle of region) select certain region near human face region 50 (near such as left shoulder adaptively, preaxial, region above the waist, top area etc.).The introducing of reference zone is the process being different from the entire image frame adopted when in the past doing image procossing, and reference zone is much smaller compared with entire image, and such data processing amount reduces much relatively.
There is reference zone, as long as locking reference zone is observed, just can complete the identification of most common human action.So, the data volume of image procossing can significantly reduce, thus has saved the processing time.
Fig. 6 shows more reference zone example, and solid box 50 is human face region, and dotted line frame 31 is the first reference zone, and dotted line frame 32 is the second reference zone, and dotted line frame 33 is the 3rd reference zone.
First reference zone 31 is relative to the transversal displacement a=30cm of human face region 50 center, relative to the vertical misalignment amount b=30cm of human face region 50 center, relative to the multiplication factor m=2 of human face region 50 width, relative to the multiplication factor l=6 of human face region 50 height.
Second reference zone 32 is relative to the transversal displacement a=10cm of human face region 50 center, relative to the vertical misalignment amount b=20cm of human face region 50 center, relative to the multiplication factor m=6 of human face region 50 width, relative to the multiplication factor l=1.5 of human face region 50 height.
3rd reference zone 33 is relative to the transversal displacement a=30cm of human face region 50 center, relative to the vertical misalignment amount b=30cm of human face region 50 center, relative to the multiplication factor m=2 of human face region 50 width, relative to the multiplication factor l=6 of human face region 50 height.The visual concrete condition of example of above-mentioned each reference zone setting is oppositely arranged according to face location, is not limited to above-mentioned parameter.
After obtaining reference zone RGB color image, namely enter step S200.
Step S200: the neighbor frame difference image of reference zone RGB color image in obtaining step S100;
Step S200 specifically comprises: S210, to step S100 obtain reference zone RGB color image carry out gray processing process, gray processing process expression formula is:
Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB image.F (x, y) represents that gray level image is also luminance graph, and its value is between 0 ~ 255.
S220: pass through formula , wherein , be respectively the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.
After obtaining the neighbor frame difference image of reference zone RGB color image, enter step S300.
Step S300: the motion feature calculating human body in two-dimensional space reference zone according to the neighbor frame difference image obtained in step S200, by the motion feature of computing reference region RGB color image in certain time domain, calculates temporal motion proper vector.
Specifically following steps are adopted during statistics two-dimensional space motion feature:
S311, calculate motor point number according to motion change value described in neighbor frame difference image in step S220: computing formula is , wherein W, H represent the wide and high of neighbor frame difference image respectively.
S312, calculate motion centroid position according to the motor point number of neighbor frame difference image in step S311, computing formula is , .
S313, setting time domain window N, the information of record N continuous frame reference zone RGB color image.
S314, according to step S312, the motion characteristic value of computing reference region RGB color image, computing formula is , this v value to illustrate in reference zone RGB color image human action in motion state sometime; Then temporal motion proper vector is calculated: , definition i=1 ~ N, i are natural number, then represent the i-th frame motion characteristic value .
The time domain reference figure (moving right 7 frame pictures continuously for staff) of to be a width shown in Fig. 5 be reference zone image, in this figure, continuous 7 frames are respectively the first frame reference zone t1, the second frame reference zone t2, the 3rd frame reference zone t3, the 4th frame reference zone t4, the 5th frame reference zone t5, the 6th frame reference zone t6, the 7th frame reference zone t7, calculate the motion feature that this N continuous=7 frame is corresponding, i.e. V=(0,0,1,1,1,0,0), this proper vector indicates the continuous state of human action.
After calculating temporal motion proper vector, enter step S400.
Step S400: the motion feature calculated in step S300 vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculates the similarity of motion feature vector and type of action masterplate.
Step S400 specifically adopts following steps:
S410, predefine K type of action masterplate , i is natural number (such as upwards translation, downward translation, to left, to right translation, waves, pressing), such as a large amount of the results shows, reasonable set N=7;
So and ( represent the motion barycenter of the i-th frame, for setting threshold value) represent in image coordinate to right translation.
and represent to left, represent waving motion.
S420, utilize absolute value distance motion feature vector V that calculation procedure S314 obtains and type of action masterplate similarity , the absolute value distance of similarity selection here calculates.
If , and <TH, TH are predetermined threshold value, then this type of action is attributed to kth class; If do not met, do not belong to any class; The motion feature vector being about to calculate carries out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, judges whether to satisfy condition, if otherwise return step S100 acquisition reference zone RGB color image.If satisfied condition, then enter step S500 and input out corresponding type of action.
It is to the right translation that arrow shown in Fig. 3 just illustrates this type of action.
Step S500: the result according to calculating similarity acquisition in step S400 judges the human motion in reference zone, and last output action event criteria data, for calling of outer application program, with output action type.
The present invention is based on the man-machine interaction method of reference zone and time-domain information, may be used for robot and automatically control, internet browsing and guidance operation, the application demands such as game interactive.Such as Internet user is surfed the web by internet television, first human body be in be equipped with camera televisor before, in effective angular field of view of camera, will automatic search to human face region, then a by presetting, b, m and l value self-adaptation finds reference zone and locks, the movable information of record reference zone successive frame, determine the front and back change of certain time domain internal reference area image, when in reference zone, the motion of human body is identified, namely compare according to the type of action in the motion feature calculated and predetermined action template types feature database, determine the type of action exported, as shown in Figure 5, when action is to right translation, trigger the event of page turning to the right.
The man-machine interaction method that the present invention is based on reference zone and time-domain information has abandoned staff hand shape detection and tracking, while saving the processing time, decreases the erroneous judgement because causing when staff can't detect.Reduce time loss, the introducing of adaptive reference regional selection method, only need process reference zone and do not need to process whole two field picture, so also saved the time of many image procossing.And human face detection tech relative maturity is stablized, real-time, be widely used in industrial products.Automatically reference zone sets because this reference zone is relative position according to face, so after position of human body changes, can be followed automatically.When the limb motion of people, very naturally drop in this reference zone.Such man-machine interaction is carried out very naturally easily, does not need human body to constrain in specific position.The more human action freedom and flexibility of such containing, promote the convenience of user interactions, improve man-machine interaction efficiency.
Based on the embodiment of above-mentioned man-machine interaction method, present invention also offers a kind of man-machine interactive system based on reference zone and time-domain information, as shown in Figure 7, it comprises:
Reference zone image collection module 710, for passing through camera collection human body image, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select around face certain region, to obtain reference zone RGB color image with reference to region adaptivity; Specifically as described in above-mentioned steps S100.
Neighbor frame difference acquisition module 720, for obtaining the neighbor frame difference image of described reference zone RGB color image; As detailed above described in step S200.
Statistics and computing module 730, for calculating the motion feature of human body in two-dimensional space reference zone according to the neighbor frame difference image obtained, by the motion feature of computing reference region RGB color image in certain time domain, and calculate temporal motion proper vector; Specifically as shown in above-mentioned steps S300.
Characteristic Contrast module 740, for calculated motion feature vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculates the similarity of motion feature vector and type of action masterplate; Specifically as shown in above-mentioned steps S400.
Output module 750, for judging the human motion in reference zone according to the result calculating similarity acquisition, last output action event criteria data, for outer application call.Specifically as shown in above-mentioned steps S500.
Wherein, described reference zone image collection module comprises:
Face search unit, for calculating the integrogram of the human body image arrived by camera collection, extracts the class rectangular characteristic of this human body image, and according to predetermined sorter feature database, the method running cascade cascade searches for human face region in this human body image; Specifically as shown in above-mentioned S110.
Reference zone computing unit, for being with reference to object with face, adopts the adaptively selected algorithm of reference zone to select around face certain region with reference to region adaptivity, the position of computing reference region RGB color image and size: , wherein, P comprises reference zone RGB color image central point horizontal ordinate, ordinate, width, and the vector of height, T is mapping function, independent variable represent the center horizontal ordinate of input human face region respectively, ordinate, width and height, specifically as shown in above-mentioned S120.
And described neighbor frame difference acquisition module comprises:
Gray proces unit, for described reference zone RGB color image is passed through following formula:
Carry out gray processing process; Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB color image, and f (x, y) represents gray level image, and its value is between 0 ~ 255; Specifically as shown in above-mentioned S210.
Neighbor frame difference image acquisition unit, for passing through formula: , wherein , for the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.Specifically as shown in above-mentioned S220.
A kind of man-machine interaction method based on reference zone and time-domain information provided by the present invention and system, make motion for common RGB image or depth image to detect and motion identification, by having skipped the detection and tracking of staff hand shape based on the system of selection of reference zone and time-domain information process, thus simplify processing procedure, improve man-machine interaction efficiency.The staff hand-type detection and positioning that the method has been abandoned to be needed in prior art is followed the tracks of, and simplifies processing procedure, improves processing speed.The adaptive region selection algorithm that the present invention comprises, decreases the constraint to interactive action design, improves human action dirigibility, contain the type of more interactive actions more easily.
Should be understood that, application of the present invention is not limited to above-mentioned citing, for those of ordinary skills, can be improved according to the above description or convert, and all these improve and convert the protection domain that all should belong to claims of the present invention.

Claims (7)

1., based on a man-machine interaction method for reference zone and time-domain information, it is characterized in that, comprise step:
A, by camera collection human body image, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select in human face region side certain region with reference to region adaptivity, and obtain reference zone RGB color image;
The neighbor frame difference image of reference zone RGB color image in B, obtaining step A;
C, the neighbor frame difference image obtained according to step B calculate the motion feature of human body in two-dimensional space reference zone, by the motion feature of computing reference region RGB color image in certain time domain, calculate temporal motion proper vector;
D, the motion feature calculated in step C vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculate motion feature vector and the similarity of type of action masterplate;
E, according to step D calculate similarity obtain result the human motion in reference zone is judged, last output action event criteria data, for outer application call;
Described steps A specifically comprises the steps:
A1, calculate the integrogram of the human body image arrived by camera collection, extract the class rectangular characteristic of this human body image, according to predetermined sorter feature database, the method running cascade cascade searches for human face region in this human body image;
A2, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select in human face region side certain region with reference to region adaptivity, the position of computing reference region RGB color image and size: , wherein, P be comprise reference zone RGB color image central point horizontal ordinate, ordinate, width and height vector, T is mapping function, independent variable represent the center horizontal ordinate of input human face region, ordinate, width and height respectively.
2., according to claim 1 based on the man-machine interaction method of reference zone and time-domain information, it is characterized in that, the predetermined sorter feature database in described steps A 1 comprises the steps:
A11, calculate the integrogram of described human body image, extract the class rectangular characteristic of described human body image;
A12, screen effective feature according to Adaboost algorithm, form Weak Classifier;
A13, by combination multiple Weak Classifier, form strong classifier;
A14, the multiple strong classifier of cascade, form the sorter feature database of Face datection.
3., according to claim 1 based on the man-machine interaction method of reference zone and time-domain information, it is characterized in that, described step B specifically comprises the steps:
B1, by described reference zone RGB color image by following formula:
Carry out gray processing process; Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB color image, and f (x, y) represents gray level image, and its value is between 0 ~ 255;
B2, pass through formula: , wherein , be respectively the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this motion change value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.
4., according to claim 3 based on the man-machine interaction method of reference zone and time-domain information, it is characterized in that, the statistics two-dimensional space motion feature in described step C specifically comprises:
C11, calculate motor point number according to motion change value described in neighbor frame difference image in step B2, computing formula is , wherein W, H represent the wide and high of neighbor frame difference image respectively;
C12, calculate motion centroid position according to the motor point number of neighbor frame difference image in step C11, computing formula is , ;
C13, setting time domain window N, the information of record N continuous frame reference zone RGB color image;
C14, according to step C12, the motion characteristic value of computing reference region RGB color image, computing formula is , this v value to illustrate in reference zone RGB color image human action in motion state sometime; Then temporal motion proper vector is calculated: , definition i=1 ~ N, i are natural number, then represent the i-th frame motion characteristic value .
5., according to claim 4 based on the man-machine interaction method of reference zone and time-domain information, it is characterized in that, described step D comprises:
D1, predefine K type of action masterplate T i, wherein i=1 ~ K, i are natural number;
D2, utilize absolute value distance motion feature vector V that calculation procedure C14 obtains and type of action masterplate similarity ;
If D3 , and <TH, TH are predetermined threshold value, then this type of action is attributed to kth class; If do not met, do not belong to any class.
6., based on a man-machine interactive system for reference zone and time-domain information, it is characterized in that, comprising:
Reference zone image collection module, for passing through camera collection human body image, be with reference to object with face, adopt the adaptively selected algorithm of reference zone to select around face certain region, to obtain reference zone RGB color image with reference to region adaptivity;
Neighbor frame difference acquisition module, for obtaining the neighbor frame difference image of described reference zone RGB color image;
Statistics and computing module, for calculating the motion feature of human body in two-dimensional space reference zone according to the neighbor frame difference image obtained, by the motion feature of computing reference region RGB color image in certain time domain, and calculate temporal motion proper vector;
Characteristic Contrast module, for calculated motion feature vector is carried out aspect ratio pair with the type of action masterplate in predetermined action type exemplary feature storehouse, calculates the similarity of motion feature vector and type of action masterplate;
Output module, for judging the human motion in reference zone according to the result calculating similarity acquisition, last output action event criteria data, for outer application call;
Described reference zone image collection module comprises:
Face search unit, for calculating the integrogram of the human body image arrived by camera collection, extracts the class rectangular characteristic of this human body image, and according to predetermined sorter feature database, the method running cascade cascade searches for human face region in this human body image;
Reference zone computing unit, for being with reference to object with face, adopts the adaptively selected algorithm of reference zone to select around face certain region with reference to region adaptivity, the position of computing reference region RGB color image and size: , wherein, P comprises reference zone RGB color image central point horizontal ordinate, ordinate, width, and the vector of height, T is mapping function, independent variable represent the center horizontal ordinate of input human face region respectively, ordinate, width and height.
7., according to claim 6 based on the man-machine interactive system of reference zone and time-domain information, it is characterized in that, described neighbor frame difference acquisition module comprises:
Gray proces unit, for described reference zone RGB color image is passed through following formula:
Carry out gray processing process; Wherein x, y are horizontal ordinate and the ordinate of any pixel in described reference zone RGB color image, and f (x, y) represents gray level image, and its value is between 0 ~ 255;
Neighbor frame difference image acquisition unit, for passing through formula: , wherein , for the gray-scale value of adjacent front and back frame pixel, for the change threshold of setting, calculate the motion change value of any pixel in neighbor frame difference image , this motion change value be 1 pixel indicate motor point, be that the pixel of 0 represents do not have motor point.
CN201110239672.4A 2011-08-20 2011-08-20 Human-machine interaction method and system based on reference region and time domain information Expired - Fee Related CN102436301B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201110239672.4A CN102436301B (en) 2011-08-20 2011-08-20 Human-machine interaction method and system based on reference region and time domain information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201110239672.4A CN102436301B (en) 2011-08-20 2011-08-20 Human-machine interaction method and system based on reference region and time domain information

Publications (2)

Publication Number Publication Date
CN102436301A CN102436301A (en) 2012-05-02
CN102436301B true CN102436301B (en) 2015-04-15

Family

ID=45984397

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201110239672.4A Expired - Fee Related CN102436301B (en) 2011-08-20 2011-08-20 Human-machine interaction method and system based on reference region and time domain information

Country Status (1)

Country Link
CN (1) CN102436301B (en)

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150002419A1 (en) * 2013-06-26 2015-01-01 Microsoft Corporation Recognizing interactions with hot zones
CN103353935B (en) * 2013-07-19 2016-06-08 电子科技大学 A kind of 3D dynamic gesture identification method for intelligent domestic system
DE102014224898A1 (en) * 2014-12-04 2016-06-09 Robert Bosch Gmbh Method for operating an input device, input device
CN104991684A (en) 2015-07-23 2015-10-21 京东方科技集团股份有限公司 Touch control device and working method therefor
CN107330424B (en) * 2017-08-03 2020-10-16 重庆邮电大学 Interaction area and interaction time period identification method, storage device and mobile terminal

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010019024A2 (en) * 2008-08-13 2010-02-18 Mimos Berhad Method and system for tracking and tagging objects
CN101777114A (en) * 2009-01-08 2010-07-14 北京中星微电子有限公司 Intelligent analysis system and intelligent analysis method for video monitoring, and system and method for detecting and tracking head and shoulder
CN101989326A (en) * 2009-07-31 2011-03-23 三星电子株式会社 Human posture recognition method and device
CN102081918A (en) * 2010-09-28 2011-06-01 北京大学深圳研究生院 Video image display control method and video image display device

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010019024A2 (en) * 2008-08-13 2010-02-18 Mimos Berhad Method and system for tracking and tagging objects
CN101777114A (en) * 2009-01-08 2010-07-14 北京中星微电子有限公司 Intelligent analysis system and intelligent analysis method for video monitoring, and system and method for detecting and tracking head and shoulder
CN101989326A (en) * 2009-07-31 2011-03-23 三星电子株式会社 Human posture recognition method and device
CN102081918A (en) * 2010-09-28 2011-06-01 北京大学深圳研究生院 Video image display control method and video image display device

Also Published As

Publication number Publication date
CN102436301A (en) 2012-05-02

Similar Documents

Publication Publication Date Title
CN106598227B (en) Gesture identification method based on Leap Motion and Kinect
Ke et al. Computer vision for human–machine interaction
Chen et al. Repetitive assembly action recognition based on object detection and pose estimation
Zhao et al. Combing rgb and depth map features for human activity recognition
US7308112B2 (en) Sign based human-machine interaction
CN103530613B (en) Target person hand gesture interaction method based on monocular video sequence
CN103376890B (en) The gesture remote control system of view-based access control model
CN103353935A (en) 3D dynamic gesture identification method for intelligent home system
Voronin et al. Action recognition for the robotics and manufacturing automation using 3-D binary micro-block difference
Wu et al. Robust fingertip detection in a complex environment
CN102436301B (en) Human-machine interaction method and system based on reference region and time domain information
US8698092B2 (en) Method and apparatus for motion recognition
CN105536205A (en) Upper limb training system based on monocular video human body action sensing
CN105107200B (en) Face Changing system and method based on real-time deep body feeling interaction and augmented reality
CN103135753A (en) Gesture input method and system
CN109800676A (en) Gesture identification method and system based on depth information
CN105589553A (en) Gesture control method and system for intelligent equipment
TW201331891A (en) Activity recognition method
CN108197534A (en) A kind of head part&#39;s attitude detecting method, electronic equipment and storage medium
CN106502390B (en) A kind of visual human&#39;s interactive system and method based on dynamic 3D Handwritten Digit Recognition
CN108983980A (en) A kind of mobile robot basic exercise gestural control method
CN105929947B (en) Man-machine interaction method based on scene situation perception
Wang et al. Gesture recognition by using kinect skeleton tracking system
Niranjani et al. System application control based on Hand gesture using Deep learning
Osipov et al. Real-time static custom gestures recognition based on skeleton hand

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20150415

Termination date: 20200820