CN106599994B - A kind of gaze estimation method based on depth Recurrent networks - Google Patents

A kind of gaze estimation method based on depth Recurrent networks Download PDF

Info

Publication number
CN106599994B
CN106599994B CN201611036387.1A CN201611036387A CN106599994B CN 106599994 B CN106599994 B CN 106599994B CN 201611036387 A CN201611036387 A CN 201611036387A CN 106599994 B CN106599994 B CN 106599994B
Authority
CN
China
Prior art keywords
depth
indicate
layers
parameter
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201611036387.1A
Other languages
Chinese (zh)
Other versions
CN106599994A (en
Inventor
潘力立
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201611036387.1A priority Critical patent/CN106599994B/en
Publication of CN106599994A publication Critical patent/CN106599994A/en
Application granted granted Critical
Publication of CN106599994B publication Critical patent/CN106599994B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/50Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis

Abstract

This patent proposes a kind of gaze estimation method based on depth Recurrent networks, belongs to computer vision and machine learning field.The main thought of this method is the mapping relations established between input picture feature and sight by depth Recurrent networks.Firstly, extracting the gradient orientation histogram feature in eyes image region;Then, 5 layers of depth regression model is established, input picture feature is fitted and exports the mapping relations between direction of visual lines;Later, the parameter of gradient descent method optimization depth regression model is utilized;Finally, for eyes image to be estimated, direction of visual lines is estimated using the depth model succeeded in school.

Description

A kind of gaze estimation method based on depth Recurrent networks
Technical field
The invention belongs to technical field of computer vision, are related to deep learning method, mainly solve sight estimation and sight The visions mapping problems such as tracking can be applied to the fields such as vehicle security drive and area-of-interest detection.
Background technique
In computer vision, sight estimation refers to the face-image according to input, positions human eye area and according to eyeball Automatically estimate direction of visual lines in position.Existing gaze estimation method includes two classes: (1) method based on geometry and (2) base In the method for recurrence.It is detailed in bibliography: Takahiro Ishikawa, Simon Baker, Iain Matthews, and Takeo Kanade, Passive Driver Gaze Tracking with Active Appearance Models, tech.report CMU-RI-TR-04-08,2004。
Gaze estimation method based on geometry mainly passes through positioning pupil center, upper palpebra inferior characteristic point and canthus position Etc. calculating direction of visual lines.The characteristic point for positioning above-mentioned ocular mainly utilizes movable appearance model (Active Appearance Model), this is a kind of side by face overall situation appearance positioning facial feature points (canthus point, corners of the mouth point etc.) Method.The advantages of gaze estimation method based on geometry, is that sight accuracy of estimation is high when eye feature point location is accurate, and Disadvantage be active appearance models vulnerable to illumination, block influence with posture, cause eye feature point location inaccurate.It is detailed in ginseng Examine document: Takahiro Ishikawa, Simon Baker, Iain Matthews, and Takeo Kanade, Passive Driver Gaze Tracking with Active Appearance Models,tech.report CMU-RI-TR-04- 08,2004 and Iain Matthews, and Simon Baker, Active Appearance Models Revisited, International Journal of Computer Vision,Vol.60,No.2,pp.135-164,2004.
Gaze estimation method based on recurrence mainly passes through detection human eye area, and establish eyes image feature and sight it Between mapping relations.Existing method mainly passes through support vector regression, Gaussian process recurrence etc. and establishes mapping relations.Based on recurrence Gaze estimation method major advantage be it is simple under the premise of the human eye area accurate positioning easily realize, and the disadvantage is that existing Homing method is difficult accurately to describe very much the mapping relations between human eye feature and sight.It is detailed in bibliography: Zhiwei Zhu,Qiang Ji,and Kristin P.Bennett,Nonlinear Eye Gaze Mapping Function Estimation via Support Vector Regression,The 18th International Conference on Pattern Recognition, Vol.1, pp.1132-1135,2006. and Oliver Williams, Andrew Blake, and Roberto Cipolla.Sparse and Semi-supervised Visual Mapping with the S3GP.The 2006IEEE Computer Society Conference on Computer Vision and Pattern Recognition,Vol.1,pp.230-237,2006.
In the gaze estimation method based on homing method, sixty-four dollar question be establish it is non-from eye feature to sight Linear regression model (LRM).In nonlinear regression model (NLRM), present depth regression model has been found mould best to solve this problem One of type.Due to its high accuracy, flexibility and strong versatility, depth method is currently widely used.In recent years, to encode certainly Device is the research work of core, is more and more applied in practical problem.It is detailed in bibliography: Yan LeCun, Bengio Yoshua Bengio,and Geoffrey Hinton,Deep Learning,Nature,Vol.521,pp 436-444。
Summary of the invention
The present invention provides a kind of gaze estimation methods based on depth network.Collected eyes image is carried out first Size normalization;Later, depth is established between input eyes image gradient orientation histogram feature and corresponding sight return mould Type;Then, the parameter of depth regression model is initialized, and solves depth model parameter using gradient descent method;Estimate finally, treating It counts eyes image and extracts gradient orientation histogram feature, and utilize the depth regression model estimation direction of visual lines acquired.Algorithm Schematic diagram is referring to fig. 2.
In order to easily describe the content of present invention, some terms are defined first.
Definition 1: direction of visual lines.The angle in eyeball fixes direction is usually indicated by vector in two-dimensional space, this to Amount is made of two elements, and first element is level angle, and second element is vertical angle.
Definition 2: gradient orientation histogram feature.Piece image is described using the directional spreding of image pixel intensities gradient or edge In object presentation and shape Visual Feature Retrieval Process method.Its implementation, which first divides the image into, small is called pane location Connected region;Then the gradient direction or edge orientation histogram of each pixel in pane location are acquired;It is finally that these are straight Square figure, which combines, can be formed by Feature Descriptor.It, can also be these local histograms in image in order to improve accuracy Bigger section (block) in degree of comparing normalization (contrast-normalized), the method is each by first calculating Density of the histogram in this section (block), then does each pane location in section according to this density value and returns One changes.There can be stronger robustness to illumination variation and shade by the normalization.
Definition 3: depth Recurrent networks.Refer to the neural network successively returned, each layer of output is next layer of input.
Define 4:S shape function.Sigmoid function (sigmoid function) can generally be expressed as σ (), and expression formula is
Definition 5: back-propagation algorithm.It is a kind of supervised learning algorithm, is often used to train multilayer neural network.General packet Containing two stages: (1) the propagated forward stage, which inputs training, is sent into network to obtain exciter response;(2) back-propagation phase will Exciter response inputs corresponding target output with training and asks poor, to obtain the response error of hidden layer and output layer.
Definition 6: gradient descent method.It is a kind of unconfined optimization method, when solving objective function minimum value, finds ladder Direction is spent, and is searched for along gradient opposite direction, the method until reaching local minimum.
A kind of gaze estimation method based on depth Recurrent networks according to the invention, it is comprised the steps of:
Step 1: acquisition N width includes the eyes image (see Fig. 1) of different sight, and records corresponding when acquisition each image Direction of visual linesynOne-dimensional representation horizontal direction, two-dimensional representation vertical direction, subscript n indicate the n-th width image Corresponding direction of visual lines;
Step 2: the eyes image normalized that will be acquired in step 1, and extract gradient orientation histogram feature (Histogramof Oriented Gradient,HOG)。
Step 3: the range of the corresponding direction of visual lines of N width image is normalized into [0,1] section, specific practice are as follows:
WhereinIndicate the component of the calibration sight jth dimension of n-th of sample, ynjNumerical value after indicating dimension normalization.
Step 4: the corresponding mapping function of projected depth regression model is to input feature vectorWherein s1Indicate special The dimension of sign successively maps input feature vector,
Indicate the input of l+1 layers of i-th of unit,Indicate connection deep neural network L layers of all slParameter between a unit and l+1 layers of i-th of unit.Specifically,It indicates to connect l layers Parameter between j-th of unit and l+1 i-th of unit of layer,For bias term relevant to l+1 layers of hidden unit i, sl+1 For the number of l+1 layers of hidden unit.Whether l+1 layers of i-th of unit is activated, and is determined by the output of sigmoid function, it may be assumed that
Above formula can also indicate are as follows:
The output layer of the depth regression model of this patent design shares 2 units, uses symbolIt indicates, to estimate The level angle and vertical angle of direction of visual lines are counted, subscript (L) indicates the label of output layer.Entire depth regression model function hW, b(xn) indicate to input as xnWhen sight estimated value, it may be assumed that
Step 5: by gradient orientation histogram feature [x normalized in step 31..., xN] as depth regression model Input, corresponding calibration direction of visual lines are [y1..., yN], establish the objective function of depth regression model:
WhereinL is the layer of depth Recurrent networks Number,
Step 6: inputting to indicate to work as xnWhen, any unit i of any one layer of l is to error sum of squares The size of contribution defines error termI=1 ..., sl, l=2 ..., L.For output Layer (L layers), the corresponding error term of each unit i are as follows:
It indicatesDerivative calculate l=2 using Back Propagation Algorithm, each node at 3, L-1 layers The corresponding error term of j
Following objective function J (w, b) is finally obtained about parameterWithPartial derivative
WhereinWithIt indicates to work as and input as xnWhen corresponding l layers of j-th of unit output and l+1 layer i-th The corresponding error term of a unit.Objective function J (w, b) is finally obtained about parameter vector w, the gradient of bWith
Step 7: parameter w and b in order to acquire optimal deep neural network, our first initiation parameters, parameter are initial Change value is the reconstructed error minimum so that input signal, obtains initial value w[0]And b[0]And then it is carried out using gradient descent method Optimization;That is:
Wherein subscript [t] and [t+1] indicate the t times and t+1 iteration.Stop iteration when w and b meet the condition of convergence.
Step 8: for new eyes image, detecting ocular and extract gradient orientation histogram feature, numerical value is returned One change after be sent into trained depth network, obtain corresponding direction of visual lines estimated value, and numberical range is reverted to- 90 ° Dao+90 °.
It should be understood that
Further, the specific method of affiliated step 2 is that the eyes image that will be acquired in step 1 is normalized to size and is 100 × 60 pixels, during gradient orientation histogram feature calculation, the parameter of areal is set as 2 × 2, Mei Yiqu The number parameter of elementary area is set as 4 × 4 in domain, and the number of steering column is set as 9, and it is corresponding to finally obtain any piece image Gradient orientation histogram feature dimension be 1152, and remember the corresponding gradient orientation histogram feature of any n-th width image to AmountAnd then to every one-dimensional progress numerical value normalization, data area is compressed to [0,1] section, for n-th of sample This, the data x of i-th dimensionniNormalize formula
For the minimum value in all sample i-th dimensions,It is similar fixed Justice is maximum value in all sample i-th dimensions.
The stack self-encoding encoder mentioned in step 4, each layer of number of unit are respectively s1=1440, s2=300, s3= 250 and s4=200, output layer only has 2 units, s5=2.
When seeking depth network parameter using gradient descent method in step 7, the condition of convergence be front and back twice iteration parameter not Change again, that is, reaches local best points;Gradient descent method optimizes method are as follows:
Wherein subscript [t] and [t+1] indicate the t times and t+1 iteration.Stop iteration when w and b meet the condition of convergence.
Innovation of the invention is:
It proposes to utilize depth Recurrent networks, establishes the Nonlinear Mapping relationship between eyes image and direction of visual lines.This hair The bright N eyes image of acquisition first is normalized to the image that size is 100*60 as training sample, and depth image, mentions simultaneously 1152 dimension gradient orientation histogram features are taken, corresponding direction of visual lines is then recorded.Later, projected depth Recurrent networks, the depth It spends network and removes input layer and output layer, totally 3 layers of middle layer.Then, on training sample and calibration direction of visual lines data, ladder is utilized Each layer parameter of degree decline calligraphy learning depth Recurrent networks.Finally, for the eyes image of sight to be estimated, gradient direction is extracted Histogram feature estimates direction of visual lines according to the above-mentioned depth Recurrent networks model succeeded in school.With traditional gaze estimation method Compare, this method can simulation input feature to direction of visual lines complex mapping relation, effectively overcome shallow Model estimation The not high problem of accuracy.
Detailed description of the invention
Fig. 1 is that sight estimates schematic diagram;
Fig. 2 is depth Recurrent networks schematic diagram.
Specific embodiment
According to the method for the present invention, the training pattern of depth Recurrent networks is write first with Matlab or C language;It connects The collected training sample of input and training depth Recurrent networks parameter;Then it is straight gradient direction to be extracted to acquired image Square figure feature is input in trained depth Recurrent networks as source data and is handled;The direction of visual lines estimated.This The method of invention can be used in the sight estimation problem of eye in natural scene.
A kind of gaze estimation method based on depth Recurrent networks, it is comprised the steps of:
Step 1: acquisition N width includes the eyes image (see Fig. 1) of different sight, and records corresponding when acquisition each image Direction of visual linesynOne-dimensional representation horizontal direction, two-dimensional representation vertical direction, subscript n indicate the n-th width image Corresponding direction of visual lines;
Step 2: it is 100 × 60 pixels that the eyes image acquired in step 1, which is normalized to size, and extracts gradient direction Histogram feature (Histogram of Oriented Gradient, HOG).In the process of gradient orientation histogram feature calculation In, the parameter of areal is set as 2 × 2, and the number parameter of elementary area is set as 4 × 4 in each region, of steering column Number is set as 9, and the dimension for finally obtaining the corresponding gradient orientation histogram feature of any piece image is 1152, and remembers any the The corresponding gradient orientation histogram feature vector of n width imageAnd then it to every one-dimensional progress numerical value normalization, will count According to Ratage Coutpressioit to [0,1] section, for n-th of sample, the data x of i-th dimensionniNormalize formula
For the minimum value in all sample i-th dimensions,It is similar fixed Justice is maximum value in all sample i-th dimensions.
Step 3: the range of the corresponding direction of visual lines of N width image is normalized into [0,1] section, specific practice are as follows:
WhereinIndicate the component of the calibration sight jth dimension of n-th of sample, ynjNumerical value after indicating dimension normalization.
Step 4: the corresponding mapping function of projected depth regression model is to input feature vectorWherein s1Indicate special The dimension of sign successively maps input feature vector,
Indicate the input of l+1 layers of i-th of unit,Indicate connection deep neural network L layers of all slParameter between a unit and l+1 layers of i-th of unit.Specifically,It indicates to connect l layers Parameter between j-th of unit and l+1 i-th of unit of layer,For bias term relevant to l+1 layers of hidden unit i, Sl+1 For the number of l+1 layers of hidden unit.Whether l+1 layers of i-th of unit is activated, and is determined by the output of sigmoid function, it may be assumed that
Above formula can also indicate are as follows:
The output layer of the depth regression model of this patent design shares 2 units, uses symbolIt indicates, to estimate The level angle and vertical angle of direction of visual lines are counted, subscript (L) indicates the label of output layer.Entire depth regression model function hW, b(xn) indicate to input as xnWhen sight estimated value, it may be assumed that
Step 5: by gradient orientation histogram feature [x normalized in step 31..., xN] as depth regression model Input, corresponding calibration direction of visual lines are [y1..., yN], establish the objective function of depth regression model:
WhereinL is the layer of depth Recurrent networks Number,
Step 6: inputting to indicate to work as xnWhen, any unit i of any one layer of l is to error sum of squares The size of contribution defines error termI=1 ..., sl, l=2 ..., L.For output Layer (L layers), the corresponding error term of each unit i are as follows:
It indicatesDerivative calculate l=2 using Back Propagation Algorithm, each node at 3, L-1 layers The corresponding error term of j
Following objective function J (w, b) is finally obtained about parameterWithPartial derivative
WhereinWithIt indicates to work as and input as xnWhen corresponding l layers of j-th of unit output and l+1 layer i-th The corresponding error term of a unit.Objective function J (w, b) is finally obtained about parameter vector w, the gradient of bWith
Step 7: parameter w and b in order to acquire optimal deep neural network, our first initiation parameters, parameter are initial Change value is the reconstructed error minimum so that input signal, obtains initial value w[0]And b[0]And then it is carried out using gradient descent method Optimization, that is:
Wherein subscript [t] and [t+1] indicate the t times and t+1 iteration.Stop iteration when w and b meet the condition of convergence.
Step 8: for new eyes image, detecting ocular and extract gradient orientation histogram feature, numerical value is returned One change after be sent into trained depth network, obtain corresponding direction of visual lines estimated value, and numberical range is reverted to- 90 ° Dao+90 °.
It should be understood that
The stack self-encoding encoder mentioned in step 3, each layer of number of unit are respectively s1=1440, s2=300, s3= 250 and s4=200, output layer only has 2 units, it may be assumed that s5=2.
When seeking depth network parameter using gradient descent method in step 7, the condition of convergence be front and back twice iteration parameter not Change again, that is, reaches local best points.

Claims (3)

1. a kind of gaze estimation method based on depth Recurrent networks, comprising the following steps:
Step 1: acquisition N width includes the eyes image of different sight, and records corresponding direction of visual lines when acquisition each imageynOne-dimensional representation horizontal direction, two-dimensional representation vertical direction, subscript n indicates the corresponding view of the n-th width image Line direction;
Step 2: the eyes image normalized that will be acquired in step 1, and gradient orientation histogram feature is extracted, obtain N width Gradient orientation histogram feature [the x of image1..., xN];
Step 3: the range that N width image corresponds to direction of visual lines being normalized into [0,1] section, obtaining calibration direction of visual lines is [y1..., yN], specific practice are as follows:
WhereinIndicate the component of the calibration sight jth dimension of the n-th width image, ynjNumerical value after indicating dimension normalization, wherein for Expression facilitates yniIt is expressed as yn
Step 4: the corresponding mapping function of projected depth regression model is to input feature vectorWherein s1Indicate feature Dimension successively maps input feature vector,
Indicate the input of l+1 layers of i-th of unit,Indicate l layers of deep neural network of connection All slParameter between a unit and l+1 layers of i-th of unit;Specifically,Indicate j-th of l layers of connection Parameter between unit and l+1 i-th of unit of layer,Indicate the output of the sigmoid function of l+1 layers of i-th of unit, For bias term relevant to l+1 layers of hidden unit i, Sl+1For the number of l+1 layers of hidden unit;L+1 layers of i-th of unit Whether it is activated, is determined by the output of sigmoid function, it may be assumed that
The output layer of the depth regression model of design shares 2 units, uses symbolIt indicates, to estimate direction of visual lines Level angle and vertical angle, subscript (L) indicate output layer label;Entire depth regression model function hW, b(xn) table Show to work as and input as xnWhen sight estimated value, it may be assumed thatσ () indicates sigmoid function;
Step 5: by the normalized gradient orientation histogram feature [x of step 21..., xN] input as depth regression model, Corresponding calibration direction of visual lines is [y1..., yN], establish the objective function of depth regression model:
WhereinL is the number of plies of depth Recurrent networks,The intensity of λ Section 2 bound term;
Step 6: inputting to indicate to work as xnWhen, any unit i of any one layer of l is to error sum of squares The size of contribution defines an error termOutput layer is used as L layers When, the corresponding error term of each unit i are as follows:
It indicatesDerivative,It indicates to work as and input as xnWhen L i-th of node of layer input, to biography after utilization It broadcasts algorithm, calculates l=2, each node j corresponding error term at 3, L-1 layers
Following objective function J (w, b) is finally obtained about parameterWithPartial derivative
WhereinWithIt indicates to work as and input as xnWhen corresponding l layers of j-th of unit output and i-th of l+1 layer it is single The corresponding error term of member;Objective function J (w, b) is finally obtained about parameter vector w, the gradient of bWith
Step 7: parameter w and b in order to acquire optimal deep neural network, our first initiation parameters, parameter initialization value It is the reconstructed error minimum so that input signal, obtains initial value w[0]And b[0]And then it is optimized using gradient descent method; That is:
Wherein subscript [t] and [t+1] indicate the t times and t+1 iteration;Stop iteration when w and b meet the condition of convergence, α is indicated The step-length of gradient decline;
Step 8: for new eyes image, detects ocular and extract gradient orientation histogram feature, numerical value normalization Be sent into trained depth network later, obtain corresponding direction of visual lines estimated value, and by numberical range revert to -90 ° to+ 90°。
2. a kind of gaze estimation method based on depth Recurrent networks as described in claim 1, it is characterised in that the step 2 Method particularly includes: it is 100 × 60 pixels that the eyes image acquired in step 1, which is normalized to size, in gradient direction histogram During figure feature calculation, the parameter of areal is set as 2 × 2, the number parameter setting of elementary area in each region It is 4 × 4, the number of steering column is set as 9, finally obtains the dimension of the corresponding gradient orientation histogram feature of any piece image It is 1152, and remembers the corresponding gradient orientation histogram feature vector of any n-th width imageAnd then to it is every it is one-dimensional into The normalization of line number value, is compressed to [0,1] section for data area, for the n-th width image, the data x of i-th dimensionniNormalization is public Formula
For the minimum value in all sample i-th dimensions,It is similar to be defined as Maximum value in all sample i-th dimensions,Indicate the n-th width image, the gradient orientation histogram feature vector of the data of i-th dimension, To indicate to facilitate xniIt is expressed as xn
3. a kind of gaze estimation method based on depth Recurrent networks as described in claim 1, it is characterised in that sharp in step 7 When seeking depth network parameter with gradient descent method, the condition of convergence is that the parameter of front and back iteration twice no longer changes, that is, reaches part Optimum point;Gradient descent method optimizes method are as follows:
Wherein subscript [t] and [t+1] indicate the t times and t+1 iteration;Stop iteration when w and b meet the condition of convergence.
CN201611036387.1A 2016-11-23 2016-11-23 A kind of gaze estimation method based on depth Recurrent networks Expired - Fee Related CN106599994B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201611036387.1A CN106599994B (en) 2016-11-23 2016-11-23 A kind of gaze estimation method based on depth Recurrent networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201611036387.1A CN106599994B (en) 2016-11-23 2016-11-23 A kind of gaze estimation method based on depth Recurrent networks

Publications (2)

Publication Number Publication Date
CN106599994A CN106599994A (en) 2017-04-26
CN106599994B true CN106599994B (en) 2019-02-15

Family

ID=58591880

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201611036387.1A Expired - Fee Related CN106599994B (en) 2016-11-23 2016-11-23 A kind of gaze estimation method based on depth Recurrent networks

Country Status (1)

Country Link
CN (1) CN106599994B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109409173B (en) * 2017-08-18 2021-06-04 安徽三联交通应用技术股份有限公司 Driver state monitoring method, system, medium and equipment based on deep learning
CN109409172B (en) * 2017-08-18 2021-08-13 安徽三联交通应用技术股份有限公司 Driver sight line detection method, system, medium, and apparatus
CN108345843B (en) * 2018-01-25 2020-12-29 电子科技大学 Head posture estimation method based on mixed depth regression network
EP3750029A1 (en) 2018-02-09 2020-12-16 Pupil Labs GmbH Devices, systems and methods for predicting gaze-related parameters using a neural network
WO2019154509A1 (en) 2018-02-09 2019-08-15 Pupil Labs Gmbh Devices, systems and methods for predicting gaze-related parameters
US11393251B2 (en) 2018-02-09 2022-07-19 Pupil Labs Gmbh Devices, systems and methods for predicting gaze-related parameters
CN108681699A (en) * 2018-05-04 2018-10-19 上海像我信息科技有限公司 A kind of gaze estimation method and line-of-sight estimation device based on deep learning
CN111277857B (en) * 2018-12-04 2021-04-13 清华大学 Streaming media scheduling method and device
EP3912013A1 (en) 2019-01-16 2021-11-24 Pupil Labs GmbH Methods for generating calibration data for head-wearable devices and eye tracking system
US11676422B2 (en) 2019-06-05 2023-06-13 Pupil Labs Gmbh Devices, systems and methods for predicting gaze-related parameters
CN110347837B (en) * 2019-07-17 2022-02-18 电子科技大学 Cardiovascular disease unplanned hospitalization risk prediction method
CN110533065A (en) * 2019-07-18 2019-12-03 西安电子科技大学 Based on the shield attitude prediction technique from coding characteristic and deep learning regression model
CN114241179A (en) * 2021-12-06 2022-03-25 电子科技大学 Sight estimation method based on self-learning

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120166083A1 (en) * 2010-12-28 2012-06-28 Aisin Aw Co., Ltd. Navigation device, navigation method, and navigation program
CN102830793A (en) * 2011-06-16 2012-12-19 北京三星通信技术研究有限公司 Sight tracking method and sight tracking device
CN104809443A (en) * 2015-05-05 2015-07-29 上海交通大学 Convolutional neural network-based license plate detection method and system
CN105354548A (en) * 2015-10-30 2016-02-24 武汉大学 Surveillance video pedestrian re-recognition method based on ImageNet retrieval
CN106056039A (en) * 2016-05-18 2016-10-26 电子科技大学 Robust mixed regression method for line of sight estimation

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20120166083A1 (en) * 2010-12-28 2012-06-28 Aisin Aw Co., Ltd. Navigation device, navigation method, and navigation program
CN102830793A (en) * 2011-06-16 2012-12-19 北京三星通信技术研究有限公司 Sight tracking method and sight tracking device
CN104809443A (en) * 2015-05-05 2015-07-29 上海交通大学 Convolutional neural network-based license plate detection method and system
CN105354548A (en) * 2015-10-30 2016-02-24 武汉大学 Surveillance video pedestrian re-recognition method based on ImageNet retrieval
CN106056039A (en) * 2016-05-18 2016-10-26 电子科技大学 Robust mixed regression method for line of sight estimation

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
3D Road Environment Modeling Applied to Visibility Mapping: an Experimental;Jean-Philippe Tarel等;《ACM》;20121231;全文 *
Eye Gaze Calculation Based on nonlinear polynomial and Generalized Regression Neural Network;Chi Jian-nan等;《IEEE》;20091231;全文 *
基于单目摄像头的驾驶员注视区域估计方法;冮俊冶;《中国优秀硕士学位论文全文数据库 信息科技辑》;20160715;全文 *

Also Published As

Publication number Publication date
CN106599994A (en) 2017-04-26

Similar Documents

Publication Publication Date Title
CN106599994B (en) A kind of gaze estimation method based on depth Recurrent networks
CN106897675B (en) Face living body detection method combining binocular vision depth characteristic and apparent characteristic
CN103942577B (en) Based on the personal identification method for establishing sample database and composite character certainly in video monitoring
CN110210551A (en) A kind of visual target tracking method based on adaptive main body sensitivity
CN108182397B (en) Multi-pose multi-scale human face verification method
CN109102547A (en) Robot based on object identification deep learning model grabs position and orientation estimation method
CN107145842A (en) With reference to LBP characteristic patterns and the face identification method of convolutional neural networks
CN108665481A (en) Multilayer depth characteristic fusion it is adaptive resist block infrared object tracking method
CN104268539A (en) High-performance human face recognition method and system
CN103971106B (en) Various visual angles facial image gender identification method and device
CN106599810B (en) A kind of head pose estimation method encoded certainly based on stack
CN109389045A (en) Micro- expression recognition method and device based on mixing space-time convolution model
Sang et al. Pose-invariant face recognition via RGB-D images
CN109255319A (en) For the recognition of face payment information method for anti-counterfeit of still photo
CN106599785A (en) Method and device for building human body 3D feature identity information database
CN108921064B (en) Pedestrian re-identification method based on multi-feature fusion
CN104036299B (en) A kind of human eye contour tracing method based on local grain AAM
Ravi et al. Sign language recognition with multi feature fusion and ANN classifier
He et al. Human segmentation of infrared image for mobile robot search
Ejbali et al. Face recognition based on beta 2D elastic bunch graph matching
CN106056039A (en) Robust mixed regression method for line of sight estimation
CN105404883B (en) A kind of heterogeneous three-dimensional face identification method
Pathak et al. Entropy based CNN for segmentation of noisy color eye images using color, texture and brightness contour features
Baqar et al. Efficient iris recognition system based on dual boundary detection using robust variable learning rate multilayer feed forward neural network
Lin et al. An object tracking method based on CNN and optical flow

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190215

Termination date: 20211123

CF01 Termination of patent right due to non-payment of annual fee