CN103176607A - Eye-controlled mouse realization method and system - Google Patents

Eye-controlled mouse realization method and system Download PDF

Info

Publication number
CN103176607A
CN103176607A CN2013101303929A CN201310130392A CN103176607A CN 103176607 A CN103176607 A CN 103176607A CN 2013101303929 A CN2013101303929 A CN 2013101303929A CN 201310130392 A CN201310130392 A CN 201310130392A CN 103176607 A CN103176607 A CN 103176607A
Authority
CN
China
Prior art keywords
eyes image
eye
screen
pupil
module
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2013101303929A
Other languages
Chinese (zh)
Other versions
CN103176607B (en
Inventor
王东强
唐云建
黄丽丽
韩鹏
孙怀义
莫斌
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Academy of Science and Technology
Original Assignee
Chongqing Academy of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Academy of Science and Technology filed Critical Chongqing Academy of Science and Technology
Priority to CN201310130392.9A priority Critical patent/CN103176607B/en
Publication of CN103176607A publication Critical patent/CN103176607A/en
Application granted granted Critical
Publication of CN103176607B publication Critical patent/CN103176607B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Processing (AREA)

Abstract

The invention provides an eye-controlled mouse realization method and system. The method comprises the following steps: obtaining an eye image; detecting and positioning a pupil position in the eye image; establishing a mapping relation between an eye gazing direction and a screen; monitoring the eye gazing direction in real time, determining a fixation point of the eye sight on the screen according to the mapping relation between the eye gazing direction and the screen, determining the working mode of the mouse, displaying the mouse at a fixation point of the eye sight on the screen and working according to the working mode. The movement of the mouse can be accurately controlled by using the eye gazing direction, and the function of controlling the mouse by eye is realized.

Description

A kind of controlled mouse implementation method and system
Technical field
The present invention relates to the Visual Tracking field, be specifically related to a kind of controlled mouse implementation method and system.
Background technology
The man-computer relation of exploring natural harmony has become a key areas of computer research, and natural, efficient, intelligentized human-computer interaction interface is the important trend of computer nowadays development.But, for physical disabilities, if the trick campaign is inconvenient, utilize conventional mouse to realize that man-machine interaction is just very difficult.
In field of human-computer interaction, eyes are as important information interaction passage, and sight line reaction people's attention direction, thereby line-of-sight applications had characteristics such as its naturality, substantivity and interactivity in field of human-computer interaction, enjoy people's concern, therefore how to realize utilizing the eye sight line direction to control and realize that the eye controlled mouse of mouse function is the technical matters of needing solution badly.
Summary of the invention
In order to overcome the defective that exists in above-mentioned prior art, the purpose of this invention is to provide a kind of controlled mouse implementation method and system, can utilize the eye sight line direction to control and realize mouse function.
In order to realize above-mentioned purpose of the present invention, according to an aspect of the present invention, the invention provides a kind of controlled mouse implementation method, comprise the steps:
S1: obtain eyes image;
S2: the pupil position in described eyes image is detected and locates;
S3: the mapping relations of setting up described eye direction of visual lines and screen;
S4: Real Time Monitoring eye direction of visual lines, according to the mapping relations of described eye direction of visual lines and screen, determine the blinkpunkt of described eye sight line on described screen;
S5: determine the mode of operation of mouse, described mode of operation comprises the mouse Move Mode, clicks pattern and double-click pattern;
S6: the blinkpunkt place of eye sight line shows mouse and carries out work according to described mode of operation on described screen.
Eye controlled mouse implementation method of the present invention is by setting up the mapping relations of eye direction of visual lines and screen, then determine pupil center, mapping relations according to eye direction of visual lines and screen, determine the blinkpunkt of eye sight line on described screen, the blinkpunkt place of eye sight line shows mouse and carries out work according to default mode of operation on screen, realizes rapidly and accurately the function of a controlled mouse.
In a preferred embodiment of the present invention, in described step S2, the localization method of pupil is:
S21: described eyes image is carried out pre-service, increase the overall brightness of described eyes image, and the noise of the described eyes image of filtering;
S22: the maximum between-cluster variance of all pixels of the described eyes image of statistics;
S23: according to described maximum between-cluster variance, calculate the optimal threshold of described eyes image binaryzation, realize the eyes image binary conversion treatment;
S24: extract pupil edge information;
S25: determine pupil center and radius.
Pupil positioning method of the present invention can position pupil rapidly and accurately, even in the situation that eye image is fuzzy, also the center of pupil can be located out fast.
In another kind of preferred embodiment of the present invention, in described eyes image was carried out pretreated step, the method that increases the overall brightness of described eyes image was:
S211: described eyes image is converted into corresponding histogram;
S212: the gray-scale value that goes out all pixels in described eyes image according to described statistics with histogram;
S213: determine the benchmark of grey scale pixel value, the gray-scale value of the pixel of described eyes image is carried out light compensation according to described benchmark.
The present invention is by increasing the overall brightness of eyes image, also can carry out rapidly and accurately the through hole location for some vague image vegetarian refreshments or blurred picture, improved accuracy.
In another preferred embodiment of the present invention, the method for the benchmark of described definite grey scale pixel value is: select the mean value of the gray-scale value of the pixel of 5%-10% in all pixels as benchmark.
The present invention selects the mean value of gray-scale value of the partial pixel point in all pixels as benchmark, and the ratio of choosing can be adjusted according to actual needs, has improved the dirigibility of method.
In a preferred embodiment of the present invention, in described eyes image is carried out pretreated step, adopt the noise of the described eyes image of median filter filtering of two dimension.
The present invention adopts the preprocess method of medium filtering and light compensation, has ability fuzzy, partially dark, promoted brightness and filtering noise by the eyes image of noise, has improved the accuracy of pupil location.
In another kind of preferred embodiment of the present invention, in described step S22, the method for the maximum between-cluster variance of all pixels of the described eyes image of statistics is:
S221: the histogram that obtains described eyes image;
S222: determine that in described histogram in described eyes image, number of pixels is zero grey scale pixel value;
S223: determine the non-vanishing grey scale pixel value of number of pixels in described eyes image, calculate the maximum between-cluster variance of the non-vanishing grey scale pixel value of described number of pixels.
Maximum between-cluster variance computing method of the present invention only need the maximum between-cluster variance of the non-vanishing grey scale pixel value of calculating pixel number, have improved computing velocity.
In a preferred embodiment of the present invention, choose the segmentation threshold of prospect and the background of described eyes image, when described maximum between-cluster variance was maximum, described segmentation threshold was the optimal threshold of cutting apart.
The present invention utilizes follow-on maximum variance between clusters to realize that adaptive threshold extracts, and has that threshold value is suitable, a self-adaptation, fireballing advantage.
In a preferred embodiment of the present invention, the method for extracting pupil edge information is:
If the pixel value of central pixel point is 255, no matter why the pixel value of all the other 8 adjacent pixels is worth, the pixel value that keeps without exception central pixel point is 255;
If the pixel value of central pixel point is 0, and the pixel value of 8 adjacent pixels is 0, the pixel value of central pixel point changed into 255;
During all the other situations, all the pixel value with central pixel point changes 0 into.
In another kind of preferred embodiment of the present invention, utilize least square method to determine pupil center and radius.
The extracting method of pupil edge information of the present invention and based on the accurate location that the ellipse fitting method of definite pupil center of least square method and radius realizes pupil has characteristics quickly and accurately.
In a preferred embodiment of the present invention, the step of setting up the mapping relations of described eye direction of visual lines and screen is:
S31: show successively M calibration point on screen, described M is positive integer;
S32: when eyes when watching each calibration point attentively, video camera detect light that infrared light supply sends through pupil center's reflection after N corresponding point on screen, described N is positive integer;
S33: N the corresponding point that each calibration point detects are analyzed, obtained an equivalent correspondence point;
S34: the corresponding relation of setting up pupil center and calibration point.
Thereby determine the blinkpunkt of eyes on screen by detecting the eye sight line direction, realize the function of eyelet mouse.
In order to realize above-mentioned purpose of the present invention, according to two aspects of the present invention, the invention provides a kind of controlled mouse and realize system, comprising: infrared light supply, video camera, screen and control module, described control module comprises eyes image acquisition module and core processing module; Described infrared light supply and video camera are set up the mapping relations of eye direction of visual lines and screen; Described eyes image acquisition module is used for gathering eyes image; Described core processing module is connected with described eyes image acquisition module, is used for receiving described eyes image and judges the center of pupil, and realize the work of eye controlled mouse and show by screen.
Eye controlled mouse of the present invention realizes that system can utilize the motion of mouse beacon exactly of eye sight line direction, realizes the function of eye controlled mouse.
in a preferred embodiment of the present invention, described core processing module comprises pupil center's judge module, demarcating module and eye tracking module, described pupil center judge module is connected with described eyes image acquisition module, the center that is used for receiving eyes image and judges pupil, described demarcating module is used for setting up the mapping relations of described eye direction of visual lines and screen, described eye tracking module is connected with described demarcating module with the pupil center judge module respectively, described eye tracking module is determined blinkpunkt and the analog mouse realization tracking operation of eyes on screen according to the pupil center of described pupil center judge module judgement and the mapping relations of described eye direction of visual lines and screen.Thereby utilize the motion of mouse beacon exactly of eye sight line direction, realize the function of eye controlled mouse.
In a preferred embodiment of the present invention, also comprise data conversion module, described data conversion module is connected with described core processing module with described eyes image acquisition module respectively, and the eyes image after also changing for the format conversion that realizes described eyes image data transfers to described core processing module.
Data conversion module of the present invention carries out format conversion, when the eyes image data of the data acquisition of eyes image acquisition module collection and data type that core processing module can be processed not simultaneously, utilize this data conversion module that the data layout of eyes image acquisition module collection is changed, improved the compatibility of pupil positioning system.
Additional aspect of the present invention and advantage part in the following description provide, and part will become obviously from the following description, or recognize by practice of the present invention.
Description of drawings
Above-mentioned and/or additional aspect of the present invention and advantage are from obviously and easily understanding becoming the description of embodiment in conjunction with following accompanying drawing, wherein:
Fig. 1 is the process flow diagram of eye controlled mouse implementation method of the present invention;
Fig. 2 is the process flow diagram of pupil positioning method of the present invention;
Fig. 3 is the corresponding histogram of eyes image in a kind of preferred implementation of the present invention;
Fig. 4 is the eyes image figure that gathers in a kind of preferred implementation of the present invention;
Fig. 5 is the pretreating effect figure to eyes image shown in Figure 4;
Fig. 6 is the binaryzation design sketch of eyes image shown in Figure 4;
Fig. 7 is the locating effect figure of pupil center of eyes image shown in Figure 4.
Embodiment
The below describes embodiments of the invention in detail, and the example of described embodiment is shown in the drawings, and wherein same or similar label represents same or similar element or the element with identical or similar functions from start to finish.Be exemplary below by the embodiment that is described with reference to the drawings, only be used for explaining the present invention, and can not be interpreted as limitation of the present invention.
In description of the invention, unless otherwise prescribed and limit, need to prove, term " installation ", " being connected ", " connection " should be done broad understanding, for example, can be mechanical connection or electrical connection, can be also the connection of two element internals, can be directly to be connected, and also can indirectly be connected by intermediary, for the ordinary skill in the art, can understand as the case may be the concrete meaning of above-mentioned term.
The invention provides a kind of controlled mouse implementation method, as shown in Figure 1, comprise the steps:
S1: obtain eyes image;
S2: the pupil position in eyes image is detected and locates;
S3: the mapping relations of setting up eye direction of visual lines and screen;
S4: Real Time Monitoring eye direction of visual lines, according to the mapping relations of eye direction of visual lines and screen, determine the blinkpunkt of eye sight line on screen;
S5: determine the mode of operation of mouse, wherein, mode of operation comprises the mouse Move Mode, clicks pattern and double-click pattern;
S6: the blinkpunkt place of eye sight line shows mouse and carries out work according to described mode of operation on screen.
In the present embodiment, the method to the pupil position in eyes image detects and locates as shown in Figure 2, comprises the steps:
S21: described eyes image is carried out pre-service, increase the overall brightness of described eyes image, and the noise of the described eyes image of filtering;
S22: the maximum between-cluster variance of all pixels of the described eyes image of statistics;
S23: according to described maximum between-cluster variance, calculate the optimal threshold of described eyes image binaryzation, realize the eyes image binary conversion treatment;
S24: extract pupil edge information;
S25: determine pupil center and radius.
In a kind of preferred implementation of the present invention, obtain eyes image, gather the eye video image of eye rotation by the eye movement instrument, and be stored in computing machine; Then utilize step S21 to gather each frame eye video image to step S23, realize that the pre-service of image and self-adaption binaryzation process; Utilize at last step S24 pupil to be positioned to step S25, specifically extract pupil edge information and determine pupil center and radius.Pupil positioning method of the present invention can position pupil rapidly and accurately, even in the situation that eye image is fuzzy, also the center of pupil can be located out fast.
In a kind of preferred implementation of the present invention, the concrete steps of this pupil positioning method are:
The first step: obtain eyes image, in the present embodiment, utilize the video camera of eye movement instrument obtain the eyes image video and be stored in the storer of computing machine, Fig. 4 shows the eyes image of the three width different angles that gathered by the eye movement instrument.
Eyes image is carried out pre-service, increase the overall brightness of eyes image, and the noise of filtering eyes image, image after processing as shown in Figure 5, in the present embodiment, gather each frame eyes image and do the image pre-service, increase integral image brightness and filtering noise by light compensation.In eyes image was carried out pretreated step, the method that increases the overall brightness of eyes image by light compensation was:
S211: eyes image is converted into corresponding histogram, in the present embodiment, as shown in Figure 3, in conjunction with software Visual studio2010, eyes image is converted into corresponding histogram by the image vision storehouse, its horizontal ordinate is gray-scale value, and its ordinate is the corresponding number of pixels of gray-scale value.
S212: the gray-scale value that goes out all pixels in eyes image according to statistics with histogram.
S213: determine the benchmark of grey scale pixel value, the gray-scale value benchmark of the pixel of eyes image is carried out light compensation.In the present embodiment, the method of determining the benchmark of grey scale pixel value is: select the mean value of the gray-scale value of the pixel of some in all pixels as benchmark, in an embodiment that is more preferably of the present invention, choose the mean value of gray-scale value of pixel of 5%-10% as benchmark.The mean value of the gray-scale value of the pixel of preferred employing 8% is as benchmark.The present invention selects the mean value of gray-scale value of the partial pixel point in all pixels as benchmark, and the ratio of choosing can be adjusted according to actual needs, has improved the dirigibility of method.
After determining benchmark, the gray-scale value benchmark of the pixel of eyes image is carried out light compensation.In the present embodiment, can carry out light compensation to the gray-scale value of all pixels, also can carry out light compensation to the gray-scale value of other all pixels except determining the selected pixel of benchmark.
In the present embodiment, the method of concrete light compensation can for but be not limited to the method that coefficient in proportion compensates, the certain proportion coefficient of getting the benchmark gray-scale value is added on the gray-scale value of all pixels, for example getting 50% of benchmark gray-scale value is added on the gray-scale value of all pixels, thereby realize the light compensation of the gray-scale value of all pixels has been strengthened the brightness of eyes image.The present invention is by increasing the overall brightness of eyes image, also can carry out rapidly and accurately the through hole location for some vague image vegetarian refreshments or blurred picture, improved accuracy.
After eyes image is carried out light compensation, the noise of filtering eyes image, in the present embodiment, wave filter adopts nonlinear filter, can adopt but be not limited to mean filter or and median filter, preferably adopt median filter, in a kind of embodiment that is more preferably of the present invention, adopt the noise of the median filter filtering eyes image of two dimension, concrete implementation procedure is as follows:
Slide on image with a moving window that contains odd point, the gray-scale value of window center being put the correspondence image pixel replaces with the intermediate value in window.
Because eyes image is two-dimentional, adopt two dimensional filter, this wave filter can be expressed from the next:
y i , j = Med s { f i , j } - - - ( 1 )
Wherein, S represents filter window, { f i,jBe the sequence of image pixel, y i,jOutput for wave filter.
The present invention adopts the preprocess method of medium filtering and light compensation, has ability fuzzy, partially dark, promoted brightness and filtering noise by the eyes image of noise, has improved the accuracy of pupil location.
Second step: the maximum between-cluster variance of all pixels of statistics eyes image, in another kind of preferred implementation of the present invention, also can adopt improved maximum between-cluster variance computing method, it is not namely the maximum between-cluster variance that calculates all pixels, but calculate the maximum between-cluster variance of certain one part of pixel point, concrete method can for but be not limited to according to the eyes image histogram, count the grey scale pixel value that does not occur in eyes image, calculate the maximum between-cluster variance of remaining pixel, in the present embodiment, the method for calculating maximum between-cluster variance is:
S231: the histogram that obtains eyes image;
S232: determine that in histogram in eyes image, number of pixels is zero grey scale pixel value;
S233: determine the non-vanishing grey scale pixel value of number of pixels in eyes image, the maximum between-cluster variance of the grey scale pixel value that the calculating pixel number is non-vanishing.
Maximum between-cluster variance computing method of the present invention only need the maximum between-cluster variance of the non-vanishing grey scale pixel value of calculating pixel number, have improved computing velocity.
In a preferred embodiment of the present invention, choose the segmentation threshold of prospect and the background of eyes image, when maximum between-cluster variance was maximum, segmentation threshold was the optimal threshold of cutting apart.The present invention utilizes follow-on maximum variance between clusters to realize that adaptive threshold extracts, and has that threshold value is suitable, a self-adaptation, fireballing advantage.
In the present embodiment, after in the first step, pretreated eyes image being done the histogram processing, then adopt improved maximum variance between clusters to ask for and count the grey scale pixel value that does not occur in image, only calculate the maximum between-cluster variance of remaining pixel.As shown in Figure 3, this histogram is that the corresponding number of pixels of this two part gray-scale value that reaches between 0-35 between 210-255 is all zero in the gradation of image value, therefore when asking for maximum between-cluster variance, first this two parts grey scale pixel value in image is come out, do not calculate its corresponding inter-class variance value, only the pixel of remainder is calculated.
The principle of calculating maximum between-cluster variance is as follows: meter t is the segmentation threshold of picture prospect and background, wherein to count and account for whole image scaled be w0 to prospect, its image averaging gray-scale value is u0, background is counted and accounted for image scaled is w1, its image averaging gray-scale value is u1, is u so can calculate image overall average gray scale.Traversal is t from the minimum gradation value to the maximum gradation value, and when t made inter-class variance value g maximum, this moment, t was the optimal threshold of cutting apart.
If the image size is the M*N pixel, L is the rank of gradation of image value, and in the present embodiment, the value of L is 1-255, and the gray-scale value of establishing pixel in image is denoted as N0 less than the number of pixels of t, and pixel grey scale is denoted as N1 greater than the number of pixels of t, can calculate:
w 0 = N 0 M × N - - - ( 2 )
N 0+N 1=M×N (3)
w 0+w 1=1 (4)
u=w 0u 0+w 1u 1 (5)
g=w 0(u 0-u) 2+w 1(u 1-u) 2 (6)
Formula (5) is brought in formula (6), obtains formula (7):
g=w 0w 1(u 0-u 1) 2 (7)
T will be from 0 to L-1 value successively, and the t value when g gets maximum is the optimal threshold of image segmentation.
The 3rd step: according to maximum between-cluster variance, after the optimal threshold that the calculating eyes image is cut apart, carry out binary conversion treatment, gray-scale value is taken as p1 greater than the gray-scale value of the pixel of t value, and gray-scale value is taken as p2 less than the gray-scale value of the pixel of t value, and p1〉p2.Thereby realize the eyes image binary conversion treatment, result as shown in Figure 5.In the present embodiment, the optimal threshold of cutting apart according to the motion conditions Dynamic Acquisition eyes image of eye, and carry out dynamic binary conversion treatment according to optimal threshold.In the present embodiment, can also filter the eyes image after binary conversion treatment and make an uproar and repair, concrete method is:
extract through adaptive threshold the hole that can there be the pupil target that causes because cornea is reflective in the binaryzation eyes image that obtains, eyelash blocks the shade that brings, the noise that burr etc. causes and the situation of defective, therefore need to carry out image to the binaryzation eyes image and filter the repairing of making an uproar, in the present embodiment, adopt the corrosion operation, expansive working, opening operation and closed operation realize filtering the repairing of making an uproar, in a kind of embodiment that is more preferably of the present invention, the concrete employing corroded operation 5 times, 5 expansive workings, 1 opening operation, 1 closed operation realizes filtering preferably the repair efficiency of making an uproar, wherein, concrete function and the parameter that adopts is set to:
The corrosion handling function is: cvErode (threshold, threshold, NULL, 5);
The expansive working function is: cvDilate (threshold, threshold, NULL, 5);
The opening operation function is:
cvMorphologyEx(threshold,threshold,0,0,CV_MOP_OPEN,1);
The closed operation function is:
cvMorphologyEx(threshold,threshold,0,0,CV_MOP_CLOSE,1)。
The 4th step: extract pupil edge information on the eyes image after binary conversion treatment.In the present embodiment, the method for extraction pupil edge information is:
If the pixel value of central pixel point is 255, no matter why the pixel value of all the other 8 adjacent pixels is worth, the pixel value that keeps without exception central pixel point is 255;
If the pixel value of central pixel point is 0, and the pixel value of 8 adjacent pixels is 0, the pixel value of central pixel point changed into 255;
During all the other situations, all the pixel value with central pixel point changes 0 into.
The 5th step: determine pupil center and radius.In the present embodiment, as shown in Figure 7, utilize least square method to determine pupil center and radius.Employing realizes that based on least square method the basic thought of the match link of edge point is: ask each candidate point minimum to the quadratic sum of the upper distance of circle.If the elliptic equation of pupil is:
Ax 2+Bxy+Cy 2+Dx+Ey+F=0 (8)
Its constraint condition is:
B 2-4AC<0 (9)
Can be got by principle of least square method:
min f ( A , B , C , D , E , F ) = min Σ i n ( Ax i 2 + Bx i y i + Cy i 2 + Dx i + Ey i + F ) 2 - - - ( 10 )
By extreme value theorem, the value of wanting to make f (A, B, C, D, E) is minimum, must have:
∂ f ∂ A = ∂ f ∂ B = ∂ f ∂ C = ∂ f ∂ D = ∂ f ∂ E = ∂ f ∂ F = 0 - - - ( 11 )
Can obtain thus a system of linear equations, then in conjunction with constraint condition, can solve equation coefficient A, B, C, D, E, the value of F.So oval centre coordinate (xp, yp), major axis and minor axis (a, b) all can be asked by following formula, thereby obtained radius and the centre coordinate value of pupil:
| a , b | = 2 { - 2 F A + C ± [ B 2 + ( A - C F ) 2 ] 1 / 2 } 1 / 2 x p = ( BE - 2 CD ) / ( 4 AC - B 2 ) , y p = ( BD - 2 AE ) / ( 4 AC - B 2 ) , - - - ( 12 )
The extracting method of pupil edge information of the present invention and based on the accurate location that the ellipse fitting method of definite pupil center of least square method and radius realizes pupil has characteristics quickly and accurately.
Pupil positioning method based on the modified maximum variance between clusters of the present invention is by image pretreatment operation and the grey scale pixel value that exists in image is calculated the method for maximum between-cluster variance, fast and accurately asked for the optimal threshold of image segmentation, for the intact extraction of pupil edge information lays the foundation; Last adopt on this basis contour extraction method and realize the accurate location of pupil and improved arithmetic speed based on the ellipse fitting method of least square method.
S3: the mapping relations of setting up eye direction of visual lines and screen.In the present embodiment, step S3 can carry out before step S1, and the step of specifically setting up the mapping relations of eye direction of visual lines and screen is:
S31: show successively M calibration point on screen, described M is positive integer, in the present embodiment, preferably adopts 5 calibration points, and namely four angles and the central authorities at screen arrange respectively a calibration point; In the other preferred implementation of the present invention, adopt 9 calibration points, being about to screen divider is the sphere of movements for the elephants shape, the intersection point of every two lines is as a calibration point.
S32: when eyes when watching each calibration point attentively, N corresponding point after the light that video camera detection infrared light supply sends reflects through pupil center on screen, described N is positive integer, in the present embodiment, 30-50 corresponding point of preferred employing, in a kind of embodiment that is more preferably of the present invention, adopt 50 corresponding point, the raising corresponding point are got accuracy a little.
S33: N the corresponding point that each calibration point detects are analyzed, obtain an equivalent correspondence point, in the present embodiment, coordinate to N corresponding point is averaged, obtain the equivalent correspondence point, simultaneously N pupil center location corresponding to corresponding point also averaged, obtain equivalent pupil center, thereby the movement of pupil is reflected as the equivalent correspondence point movement that screen coordinate is fastened.
S34: the corresponding relation of setting up equivalent pupil center and calibration point.
After the mapping relations of setting up eye direction of visual lines and screen, be kept in calibration matrix, for the tracking that realizes follow-up sight line provides real time data.
S4: Real Time Monitoring eye direction of visual lines, according to the mapping relations of eye direction of visual lines and screen, determine the blinkpunkt of eye sight line on screen.
S5: determine the mode of operation of mouse, wherein, mode of operation comprises the mouse Move Mode, clicks pattern and double-click pattern.In the present embodiment, can define the movement of the corresponding mouse of pupil uniform motion, twice nictation of interval M1 time corresponding single-click operation, twice nictation of interval M2 time corresponding single-click operation, in an embodiment that is more preferably of the present invention, M1 is 1 second, and M2 is 2 seconds.
S6: the blinkpunkt place of eye sight line shows mouse and carries out work according to the mode of operation of determining on screen.
The present invention also provides a kind of controlled mouse to realize system, and it comprises infrared light supply, video camera, screen and control module, and control module comprises eyes image acquisition module and core processing module.The light that infrared light supply sends is used for shining pupil center, the light that video camera detects pupil center's reflection on screen reflection spot and reflective information is transferred to control module, control module shows reflection spot information on screen, and sets up the mapping relations of eye direction of visual lines and screen.The eyes image acquisition module is used for gathering eyes image; Core processing module is connected with described eyes image acquisition module, is used for receiving described eyes image and judges the center of pupil, realizes that the work of eye controlled mouse also shows by screen.
in the present embodiment, core processing module comprises pupil center's judge module, demarcating module and eye tracking module, described pupil center judge module is connected with described eyes image acquisition module, the center that is used for receiving eyes image and judges pupil, described demarcating module is used for setting up the mapping relations of described eye direction of visual lines and screen, described eye tracking module is connected with described demarcating module with the pupil center judge module respectively, described eye tracking module is determined blinkpunkt and the analog mouse realization tracking operation of eyes on screen according to the pupil center of described pupil center judge module judgement and the mapping relations of described eye direction of visual lines and screen.
In the present embodiment, core processing module is positioned at computing machine, also has storer in computing machine, be used for the eyes image that storage eyes image acquisition module gathers, the user can call the eyes image in storer and check at any time by image display by core processing module.
In a kind of preferred implementation of the present invention, also comprise data conversion module, data conversion module is connected with core processing module with the eyes image acquisition module respectively, and the eyes image after also changing for the format conversion that realizes the eyes image data transfers to described core processing module.Data conversion module of the present invention carries out format conversion, when the eyes image data of the data acquisition of eyes image acquisition module collection and data type that core processing module can be processed not simultaneously, utilize this data conversion module that the data layout of eyes image acquisition module collection is changed, improved the compatibility of pupil positioning system, concrete conversion method can be carried out according to prior art.
In the present embodiment, the eyes image acquisition module is eye movement instrument or optical sensor, preferably adopts the eye movement instrument, and the present invention adopts the eye movement instrument to gather eyes image, guarantees that eyes image is accurate, has improved accuracy and has been convenient to timely processing.
In the other preferred implementation of the present invention, this system also comprises power supply and peripheral circuit, is used to infrared light supply, video camera, screen and control module that electric energy is provided, and guarantees the normal operation of infrared light supply, video camera, screen and control module.
In the present embodiment, the eyes image acquisition module selects the MT9V011 chip of MICRON company as optical sensor SENSOR, image resolution ratio is 640*480, the frame per second of the eyes image of taking was 15 frame/seconds, was that fineness and the filming frequency of image all satisfies the usual oculomotor requirement of the mankind.Data conversion module is selected CP2102, has the full speed interface of USB2.0, satisfies the large transmission requirement of data volume, and compatible with the process chip of back.Core processing module is selected the ARM11 series of processes chip S3C6410 of Samsung, and speed and performance can satisfy the requirement of image and Algorithm Analysis.Power supply and peripheral circuit can be selected existing power supply chip, as long as the operating voltage that provides meets the demands.
In the description of this instructions, the description of reference term " embodiment ", " some embodiment ", " example ", " concrete example " or " some examples " etc. means to be contained at least one embodiment of the present invention or example in conjunction with specific features, structure, material or the characteristics of this embodiment or example description.In this manual, the schematic statement of above-mentioned term not necessarily referred to identical embodiment or example.And the specific features of description, structure, material or characteristics can be with suitable mode combinations in any one or more embodiment or example.
Although illustrated and described embodiments of the invention, those having ordinary skill in the art will appreciate that: in the situation that do not break away from principle of the present invention and aim can be carried out multiple variation, modification, replacement and modification to these embodiment, scope of the present invention is limited by claim and equivalent thereof.

Claims (10)

1. an eye controlled mouse implementation method, is characterized in that, comprises the steps:
S1: obtain eyes image;
S2: the pupil position in described eyes image is detected and locates;
S3: the mapping relations of setting up described eye direction of visual lines and screen;
S4: Real Time Monitoring eye direction of visual lines, according to the mapping relations of described eye direction of visual lines and screen, determine the blinkpunkt of described eye sight line on described screen;
S5: determine the mode of operation of mouse, described mode of operation comprises the mouse Move Mode, clicks pattern and double-click pattern;
S6: the blinkpunkt place of eye sight line shows mouse and carries out work according to described mode of operation on described screen.
2. eye controlled mouse implementation method as claimed in claim 1, is characterized in that, in described step S2, the localization method of pupil is:
S21: described eyes image is carried out pre-service, increase the overall brightness of described eyes image, and the noise of the described eyes image of filtering;
S22: the maximum between-cluster variance of all pixels of the described eyes image of statistics;
S23: according to described maximum between-cluster variance, calculate the optimal threshold of described eyes image binaryzation, realize the eyes image binary conversion treatment;
S24: extract pupil edge information;
S25: determine pupil center and radius.
3. eye controlled mouse implementation method as claimed in claim 2, is characterized in that, in described eyes image was carried out pretreated step, the method that increases the overall brightness of described eyes image was:
S211: described eyes image is converted into corresponding histogram;
S212: the gray-scale value that goes out all pixels in described eyes image according to described statistics with histogram;
S213: determine the benchmark of grey scale pixel value, the gray-scale value of the pixel of described eyes image is carried out light compensation according to described benchmark.
4. eye controlled mouse implementation method as claimed in claim 2, is characterized in that, the method for the benchmark of described definite grey scale pixel value is: select the mean value of the gray-scale value of the pixel of 5%-10% in all pixels as benchmark.
5. eye controlled mouse implementation method as claimed in claim 2, is characterized in that, in described step S22, the method for the maximum between-cluster variance of all pixels of the described eyes image of statistics is:
S221: the histogram that obtains described eyes image;
S222: determine that in described histogram in described eyes image, number of pixels is zero grey scale pixel value;
S223: determine the non-vanishing grey scale pixel value of number of pixels in described eyes image, calculate the maximum between-cluster variance of the non-vanishing grey scale pixel value of described number of pixels.
6. as 2 or 5 described controlled mouse implementation methods of claim, it is characterized in that, choose the segmentation threshold of prospect and the background of described eyes image, when described maximum between-cluster variance was maximum, described segmentation threshold was the optimal threshold of cutting apart.
7. eye controlled mouse implementation method as claimed in claim 2, is characterized in that, the method for extracting pupil edge information is:
If the pixel value of central pixel point is 255, no matter why the pixel value of all the other 8 adjacent pixels is worth, the pixel value that keeps without exception central pixel point is 255;
If the pixel value of central pixel point is 0, and the pixel value of 8 adjacent pixels is 0, the pixel value of central pixel point changed into 255;
During all the other situations, all the pixel value with central pixel point changes 0 into.
8. eye controlled mouse implementation method as claimed in claim 1, is characterized in that, the step of setting up the mapping relations of described eye direction of visual lines and screen is:
S31: show successively M calibration point on screen, described M is positive integer;
S32: when eyes when watching each calibration point attentively, video camera detect light that infrared light supply sends through pupil center's reflection after N corresponding point on screen, described N is positive integer;
S33: N the corresponding point that each calibration point detects are analyzed, obtained an equivalent correspondence point;
S34: the corresponding relation of setting up pupil center and calibration point.
9. an eye controlled mouse is realized system, it is characterized in that, comprising: infrared light supply, video camera, screen and control module, and described control module comprises eyes image acquisition module and core processing module;
Described infrared light supply and video camera are set up the mapping relations of eye direction of visual lines and screen according to the method shown in claim 8;
Described eyes image acquisition module is used for gathering eyes image;
Described core processing module is connected with described eyes image acquisition module, is used for receiving described eyes image and judges the center of pupil, and realize the work of eye controlled mouse and show by screen according to the described method of one of claim 1-7.
10. eye controlled mouse as claimed in claim 9 is realized system, it is characterized in that: described core processing module comprises pupil center's judge module, demarcating module and eye tracking module, described pupil center judge module is connected with described eyes image acquisition module, the center that is used for receiving eyes image and judges pupil, described demarcating module is used for setting up the mapping relations of described eye direction of visual lines and screen, described eye tracking module is connected with described demarcating module with the pupil center judge module respectively, described eye tracking module is determined blinkpunkt and the analog mouse realization tracking operation of eyes on screen according to the pupil center of described pupil center judge module judgement and the mapping relations of described eye direction of visual lines and screen.
CN201310130392.9A 2013-04-16 2013-04-16 A kind of eye-controlled mouse realization method and system Expired - Fee Related CN103176607B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310130392.9A CN103176607B (en) 2013-04-16 2013-04-16 A kind of eye-controlled mouse realization method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310130392.9A CN103176607B (en) 2013-04-16 2013-04-16 A kind of eye-controlled mouse realization method and system

Publications (2)

Publication Number Publication Date
CN103176607A true CN103176607A (en) 2013-06-26
CN103176607B CN103176607B (en) 2016-12-28

Family

ID=48636542

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310130392.9A Expired - Fee Related CN103176607B (en) 2013-04-16 2013-04-16 A kind of eye-controlled mouse realization method and system

Country Status (1)

Country Link
CN (1) CN103176607B (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103336581A (en) * 2013-07-30 2013-10-02 黄通兵 Human eye movement characteristic design-based human-computer interaction method and system
CN103425970A (en) * 2013-08-29 2013-12-04 大连理工大学 Human-computer interaction method based on head postures
CN104238120A (en) * 2013-12-04 2014-12-24 全蕊 Smart glasses and control method
CN104680122A (en) * 2013-11-29 2015-06-03 展讯通信(天津)有限公司 Tracking method and device based on skin color detection
CN104680552A (en) * 2013-11-29 2015-06-03 展讯通信(天津)有限公司 Tracking method and device based on skin color detection
CN104898823A (en) * 2014-03-04 2015-09-09 中国电信股份有限公司 Method and device for controlling sighting mark motion
CN105094604A (en) * 2015-06-30 2015-11-25 联想(北京)有限公司 Information processing method and electronic equipment
CN105138965A (en) * 2015-07-31 2015-12-09 东南大学 Near-to-eye sight tracking method and system thereof
CN105137601A (en) * 2015-10-16 2015-12-09 上海斐讯数据通信技术有限公司 Intelligent glasses
CN105425967A (en) * 2015-12-16 2016-03-23 中国科学院西安光学精密机械研究所 Sight tracking and human eye region-of-interest positioning system
CN105676458A (en) * 2016-04-12 2016-06-15 王鹏 Wearable calculation device and control method thereof, and wearable equipment with wearable calculation device
CN106020461A (en) * 2016-05-13 2016-10-12 陈盛胜 Video interaction method based on eyeball tracking technology
CN106502423A (en) * 2016-11-21 2017-03-15 武汉理工大学 Based on the automation microoperation method that human eye vision is positioned
CN106527705A (en) * 2016-10-28 2017-03-22 努比亚技术有限公司 Operation realization method and apparatus
CN106774862A (en) * 2016-12-03 2017-05-31 西安科锐盛创新科技有限公司 VR display methods and VR equipment based on sight line
CN106774863A (en) * 2016-12-03 2017-05-31 西安中科创星科技孵化器有限公司 A kind of method that Eye-controlling focus are realized based on pupil feature
CN103412643B (en) * 2013-07-22 2017-07-25 深圳Tcl新技术有限公司 Terminal and its method for remote control
CN107067441A (en) * 2017-04-01 2017-08-18 海信集团有限公司 Camera marking method and device
CN104007820B (en) * 2014-05-26 2017-09-29 联想(北京)有限公司 A kind of information processing method and electronic equipment
CN107436675A (en) * 2016-05-25 2017-12-05 深圳纬目信息技术有限公司 A kind of visual interactive method, system and equipment
CN107609516A (en) * 2017-09-13 2018-01-19 重庆爱威视科技有限公司 Adaptive eye moves method for tracing
CN107831900A (en) * 2017-11-22 2018-03-23 中国地质大学(武汉) The man-machine interaction method and system of a kind of eye-controlled mouse
CN108282699A (en) * 2018-01-24 2018-07-13 北京搜狐新媒体信息技术有限公司 A kind of display interface processing method and processing device
CN108427503A (en) * 2018-03-26 2018-08-21 京东方科技集团股份有限公司 Human eye method for tracing and human eye follow-up mechanism
CN109656373A (en) * 2019-01-02 2019-04-19 京东方科技集团股份有限公司 One kind watching independent positioning method and positioning device, display equipment and storage medium attentively
CN109960405A (en) * 2019-02-22 2019-07-02 百度在线网络技术(北京)有限公司 Mouse operation method, device and storage medium
CN110780739A (en) * 2019-10-18 2020-02-11 天津理工大学 Eye control auxiliary input method based on fixation point estimation
CN112183200A (en) * 2020-08-25 2021-01-05 中电海康集团有限公司 Eye movement tracking method and system based on video image
CN113891002A (en) * 2021-11-12 2022-01-04 维沃移动通信有限公司 Shooting method and device
CN114020155A (en) * 2021-11-05 2022-02-08 沈阳飞机设计研究所扬州协同创新研究院有限公司 High-precision sight line positioning method based on eye tracker

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101359365A (en) * 2008-08-07 2009-02-04 电子科技大学中山学院 Iris positioning method based on Maximum between-Cluster Variance and gray scale information
CN102012742A (en) * 2010-11-24 2011-04-13 广东威创视讯科技股份有限公司 Method and device for correcting eye mouse
CN102509095A (en) * 2011-11-02 2012-06-20 青岛海信网络科技股份有限公司 Number plate image preprocessing method
CN102830797A (en) * 2012-07-26 2012-12-19 深圳先进技术研究院 Man-machine interaction method and system based on sight judgment
CN102938058A (en) * 2012-11-14 2013-02-20 南京航空航天大学 Method and system for video driving intelligent perception and facing safe city

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101359365A (en) * 2008-08-07 2009-02-04 电子科技大学中山学院 Iris positioning method based on Maximum between-Cluster Variance and gray scale information
CN102012742A (en) * 2010-11-24 2011-04-13 广东威创视讯科技股份有限公司 Method and device for correcting eye mouse
CN102509095A (en) * 2011-11-02 2012-06-20 青岛海信网络科技股份有限公司 Number plate image preprocessing method
CN102830797A (en) * 2012-07-26 2012-12-19 深圳先进技术研究院 Man-machine interaction method and system based on sight judgment
CN102938058A (en) * 2012-11-14 2013-02-20 南京航空航天大学 Method and system for video driving intelligent perception and facing safe city

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
黄丽丽等: "基于改进型最大类间方差法的瞳孔定位方法", 《计算机工程与应用》 *

Cited By (45)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103412643B (en) * 2013-07-22 2017-07-25 深圳Tcl新技术有限公司 Terminal and its method for remote control
CN103336581A (en) * 2013-07-30 2013-10-02 黄通兵 Human eye movement characteristic design-based human-computer interaction method and system
CN103425970A (en) * 2013-08-29 2013-12-04 大连理工大学 Human-computer interaction method based on head postures
CN104680122B (en) * 2013-11-29 2019-03-19 展讯通信(天津)有限公司 A kind of tracking and device based on Face Detection
CN104680122A (en) * 2013-11-29 2015-06-03 展讯通信(天津)有限公司 Tracking method and device based on skin color detection
CN104680552A (en) * 2013-11-29 2015-06-03 展讯通信(天津)有限公司 Tracking method and device based on skin color detection
CN104680552B (en) * 2013-11-29 2017-11-21 展讯通信(天津)有限公司 A kind of tracking and device based on Face Detection
CN104238120A (en) * 2013-12-04 2014-12-24 全蕊 Smart glasses and control method
CN104898823A (en) * 2014-03-04 2015-09-09 中国电信股份有限公司 Method and device for controlling sighting mark motion
CN104898823B (en) * 2014-03-04 2018-01-23 中国电信股份有限公司 The method and apparatus for controlling sighting target motion
CN104007820B (en) * 2014-05-26 2017-09-29 联想(北京)有限公司 A kind of information processing method and electronic equipment
CN105094604A (en) * 2015-06-30 2015-11-25 联想(北京)有限公司 Information processing method and electronic equipment
CN105138965B (en) * 2015-07-31 2018-06-19 东南大学 A kind of near-to-eye sight tracing and its system
CN105138965A (en) * 2015-07-31 2015-12-09 东南大学 Near-to-eye sight tracking method and system thereof
CN105137601A (en) * 2015-10-16 2015-12-09 上海斐讯数据通信技术有限公司 Intelligent glasses
CN105137601B (en) * 2015-10-16 2017-11-14 上海斐讯数据通信技术有限公司 A kind of intelligent glasses
CN105425967A (en) * 2015-12-16 2016-03-23 中国科学院西安光学精密机械研究所 Sight tracking and human eye region-of-interest positioning system
CN105425967B (en) * 2015-12-16 2018-08-28 中国科学院西安光学精密机械研究所 Sight tracking and human eye region-of-interest positioning system
CN105676458A (en) * 2016-04-12 2016-06-15 王鹏 Wearable calculation device and control method thereof, and wearable equipment with wearable calculation device
CN106020461A (en) * 2016-05-13 2016-10-12 陈盛胜 Video interaction method based on eyeball tracking technology
CN107436675A (en) * 2016-05-25 2017-12-05 深圳纬目信息技术有限公司 A kind of visual interactive method, system and equipment
CN106527705A (en) * 2016-10-28 2017-03-22 努比亚技术有限公司 Operation realization method and apparatus
CN106502423B (en) * 2016-11-21 2019-04-30 武汉理工大学 Automation microoperation method based on human eye vision positioning
CN106502423A (en) * 2016-11-21 2017-03-15 武汉理工大学 Based on the automation microoperation method that human eye vision is positioned
CN106774863A (en) * 2016-12-03 2017-05-31 西安中科创星科技孵化器有限公司 A kind of method that Eye-controlling focus are realized based on pupil feature
CN106774862A (en) * 2016-12-03 2017-05-31 西安科锐盛创新科技有限公司 VR display methods and VR equipment based on sight line
CN107067441A (en) * 2017-04-01 2017-08-18 海信集团有限公司 Camera marking method and device
CN107067441B (en) * 2017-04-01 2020-02-11 海信集团有限公司 Camera calibration method and device
CN107609516A (en) * 2017-09-13 2018-01-19 重庆爱威视科技有限公司 Adaptive eye moves method for tracing
CN107609516B (en) * 2017-09-13 2019-10-08 重庆爱威视科技有限公司 Adaptive eye movement method for tracing
CN107831900B (en) * 2017-11-22 2019-12-10 中国地质大学(武汉) human-computer interaction method and system of eye-controlled mouse
CN107831900A (en) * 2017-11-22 2018-03-23 中国地质大学(武汉) The man-machine interaction method and system of a kind of eye-controlled mouse
CN108282699A (en) * 2018-01-24 2018-07-13 北京搜狐新媒体信息技术有限公司 A kind of display interface processing method and processing device
CN108427503A (en) * 2018-03-26 2018-08-21 京东方科技集团股份有限公司 Human eye method for tracing and human eye follow-up mechanism
CN108427503B (en) * 2018-03-26 2021-03-16 京东方科技集团股份有限公司 Human eye tracking method and human eye tracking device
CN109656373A (en) * 2019-01-02 2019-04-19 京东方科技集团股份有限公司 One kind watching independent positioning method and positioning device, display equipment and storage medium attentively
US11205070B2 (en) 2019-01-02 2021-12-21 Beijing Boe Optoelectronics Technology Co., Ltd. Method, an apparatus, a display device and a storage medium for positioning a gaze point
CN109960405A (en) * 2019-02-22 2019-07-02 百度在线网络技术(北京)有限公司 Mouse operation method, device and storage medium
CN110780739B (en) * 2019-10-18 2023-11-03 天津理工大学 Eye control auxiliary input method based on gaze point estimation
CN110780739A (en) * 2019-10-18 2020-02-11 天津理工大学 Eye control auxiliary input method based on fixation point estimation
CN112183200A (en) * 2020-08-25 2021-01-05 中电海康集团有限公司 Eye movement tracking method and system based on video image
CN112183200B (en) * 2020-08-25 2023-10-17 中电海康集团有限公司 Eye movement tracking method and system based on video image
CN114020155A (en) * 2021-11-05 2022-02-08 沈阳飞机设计研究所扬州协同创新研究院有限公司 High-precision sight line positioning method based on eye tracker
WO2023083279A1 (en) * 2021-11-12 2023-05-19 维沃移动通信有限公司 Photographing method and apparatus
CN113891002A (en) * 2021-11-12 2022-01-04 维沃移动通信有限公司 Shooting method and device

Also Published As

Publication number Publication date
CN103176607B (en) 2016-12-28

Similar Documents

Publication Publication Date Title
CN103176607A (en) Eye-controlled mouse realization method and system
CN103136512A (en) Pupil positioning method and system
CN103340637B (en) Move and driver's Alertness intelligent monitor system of brain electro' asion and method based on eye
CN102508110B (en) Texture-based insulator fault diagnostic method
CN102692216B (en) Real-time optical fiber winding defect detection method based on machine vision technology
CN105184246A (en) Living body detection method and living body detection system
CN101241548A (en) Eyelid detection apparatus and programs thereof
CN104615978A (en) Sight direction tracking method and device
CN104766059A (en) Rapid and accurate human eye positioning method and sight estimation method based on human eye positioning
CN103105924B (en) Man-machine interaction method and device
CN104224204A (en) Driver fatigue detection system on basis of infrared detection technology
CN104503472A (en) Intelligent real-time human body position tracking energy saving heating system based on infrared radiation
CN102980535A (en) Angle measurement method and device
CN103164022A (en) Multi-finger touch method, device and portable type terminal device
US20230057878A1 (en) Industrial internet of things, control methods and storage medium based on machine visual detection
CN103761011A (en) Method, system and computing device of virtual touch screen
CN104123728A (en) Hough transform based round pipe detection method
US11195490B1 (en) Smart contact lens with adjustable light transmittance
CN108634925A (en) A kind of vision testing system based on gesture identification
CN113160260A (en) Head-eye double-channel intelligent man-machine interaction system and operation method
CN105068643A (en) Method and device for adjusting brightness of intelligent mirror
CN107770359A (en) A kind of obstacle based reminding method and mobile device
CN106094969A (en) The image aided diagnosis technique that a kind of maximum power point of photovoltaic power generation system is followed the trail of
CN114661152B (en) AR display control system and method for reducing visual fatigue
CN103926922A (en) Control and monitor system of smart vehicle

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20161228

Termination date: 20180416

CF01 Termination of patent right due to non-payment of annual fee