CN113052064B - Attention detection method based on face orientation, facial expression and pupil tracking - Google Patents
Attention detection method based on face orientation, facial expression and pupil tracking Download PDFInfo
- Publication number
- CN113052064B CN113052064B CN202110310469.5A CN202110310469A CN113052064B CN 113052064 B CN113052064 B CN 113052064B CN 202110310469 A CN202110310469 A CN 202110310469A CN 113052064 B CN113052064 B CN 113052064B
- Authority
- CN
- China
- Prior art keywords
- face
- image
- pupil
- parameter
- person
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 210000001747 pupil Anatomy 0.000 title claims abstract description 116
- 230000008921 facial expression Effects 0.000 title claims abstract description 19
- 238000001514 detection method Methods 0.000 title claims description 41
- 239000011159 matrix material Substances 0.000 claims abstract description 56
- 238000000034 method Methods 0.000 claims abstract description 32
- 230000001815 facial effect Effects 0.000 claims abstract description 25
- 206010062519 Poor quality sleep Diseases 0.000 claims abstract description 11
- 206010041349 Somnolence Diseases 0.000 claims abstract description 11
- 238000012545 processing Methods 0.000 claims description 35
- 238000010606 normalization Methods 0.000 claims description 30
- 210000003128 head Anatomy 0.000 claims description 20
- 230000008909 emotion recognition Effects 0.000 claims description 9
- 230000009466 transformation Effects 0.000 claims description 8
- 230000002146 bilateral effect Effects 0.000 claims description 3
- 238000005260 corrosion Methods 0.000 claims description 3
- 230000007797 corrosion Effects 0.000 claims description 3
- 238000001914 filtration Methods 0.000 claims description 3
- 238000012502 risk assessment Methods 0.000 abstract description 2
- 238000011156 evaluation Methods 0.000 description 6
- 230000008569 process Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000006870 function Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000002093 peripheral effect Effects 0.000 description 2
- 238000003491 array Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000011840 criminal investigation Methods 0.000 description 1
- 210000005069 ears Anatomy 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000011022 operating instruction Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000000750 progressive effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/60—Rotation of whole images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
- G06T5/30—Erosion or dilatation, e.g. thinning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/20—Analysis of motion
- G06T7/246—Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/174—Facial expression recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20024—Filtering details
- G06T2207/20028—Bilateral filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- General Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Ophthalmology & Optometry (AREA)
- Geometry (AREA)
- Image Analysis (AREA)
Abstract
The embodiment of the application provides a method for detecting attention based on facial orientation, facial expression and pupil tracking, which relates to the technical field of risk assessment and comprises the steps of extracting a face image of an assessed person; determining key point coordinates in the face image; determining the face orientation angle of the person to be evaluated according to the key point coordinates, the camera internal parameter matrix and the camera distortion parameters; extracting an eye region image of the subject; determining the eye pupil position of the person to be evaluated; determining a pupil deflection parameter based on the eye region image and the eye pupil position; determining a first expression parameter representing the positive/negative degree and a second expression parameter representing the wakefulness/drowsiness degree according to the pixel values of each pixel point of the face image; and determining an attention parameter corresponding to the person to be evaluated based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter. The method provided by the embodiment of the invention can accurately evaluate the attention of the evaluated person.
Description
Technical Field
The present document relates to the field of risk assessment technologies, and in particular, to a method for detecting attention based on facial orientation, facial expression, and pupil tracking.
Background
Attention detection is often widely used in various fields, such as attention detection of drivers during traffic driving, attention detection of students during teaching, attention detection for fraud recognition during criminal investigation, and the like.
Currently, attention detection is mostly performed based on the face orientation of the subject. However, the face orientation can reflect only the possibility that the person to be detected is concerned with something, and often cannot accurately reflect the attention of the person to be monitored.
Therefore, how to provide an effective solution to improve the accuracy of attention detection has become a urgent problem in the prior art.
Disclosure of Invention
The embodiment of the application provides an attention detection method based on facial orientation, facial expression and pupil tracking, which is used for solving the problem of low accuracy of attention detection in the prior art.
In order to solve the technical problems, the embodiment of the application is realized as follows:
the embodiment of the application provides a method for detecting attention based on facial orientation, facial expression and pupil tracking, which comprises the following steps:
Acquiring a target image frame containing a head portrait of an evaluated person;
extracting a face image of the person to be evaluated from the target image frame;
carrying out normalization processing on the pixel values of the face image;
calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image, wherein the plurality of key point coordinates comprise eye contour coordinates;
determining the face orientation angle of the person to be evaluated according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates;
determining the eye pupil position of the person to be evaluated according to the eye region image;
determining a pupil deflection parameter of the subject based on the eye region image and the eye pupil position;
calculating pixel values of all pixel points of the face image after normalization processing as input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated, wherein the first expression parameter represents the positive/negative degree of the person to be evaluated, and the second expression parameter represents the wakefulness/drowsiness degree of the person to be evaluated;
And determining an attention parameter corresponding to the evaluative person based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, wherein the attention parameter characterizes the attention concentration degree of the evaluative person.
Optionally, the format of the target image frame is an RGB format, and the extracting the face image of the person under evaluation from the target image frame includes:
sequentially converting the image channels of the target image frames into RGB;
scaling the target image frames after the image channels are sequentially converted into RGB to a first specified size;
normalizing the target image frames scaled to the first specified size;
calculating a matrix formed by pixel values of all pixel points of the target image frame subjected to normalization processing as input of a pre-trained face detection model to obtain face boundary point coordinates;
and extracting the face image from the target image frame based on the face boundary point coordinates.
Optionally, the face boundary point coordinates include first boundary point coordinates and second boundary point coordinates, and the extracting the face image from the target image frame based on the face boundary point coordinates includes:
And taking the first boundary point coordinates and the second boundary point coordinates as diagonal angles, and extracting the rectangular face image from the target image frame.
Optionally, the method further comprises:
scaling the face image to a second specified size;
the normalizing the pixel value of the face image includes:
and carrying out normalization processing on the pixel values of the face image scaled to the second designated size.
Optionally, the determining the face orientation angle of the person under evaluation according to the coordinates of the plurality of key points, the preset camera internal parameter matrix and the preset camera distortion parameter includes:
determining a plurality of head three-dimensional key point reference coordinates corresponding to the plurality of key point coordinates one by one according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
determining a rotation vector and a transformation vector of a camera according to the preset camera internal parameter matrix, the preset camera distortion parameter, the target key point coordinates and the target head three-dimensional key point reference coordinates;
converting the rotation vector into a rotation matrix;
splicing the rotation matrix with the transformation vector to obtain a posture matrix;
Decomposing the gesture matrix to obtain the face orientation angle of the person to be evaluated;
the target key point coordinate is one of the plurality of key point coordinates, and the target head three-dimensional key point reference coordinate is a head three-dimensional key point reference coordinate corresponding to the target key point coordinate.
Optionally, the determining the position of the pupil of the eye of the person under evaluation according to the eye region image includes:
determining a first face image with pupil ratio in a preset ratio range in the face image corresponding to the target image frame, wherein the pupil ratio is the ratio of the pupil area to the eye area;
when the number of the first face images reaches a preset number, calculating the average value of the pupil ratio of each first face image;
finding out the target pupil duty ratio closest to the average value in the pupil duty ratios of the first face images;
selecting a first face image corresponding to the target pupil ratio as a target face image;
and taking the center of the pupil area of the eyes in the target face image as the pupil position.
Optionally, the determining the first face image with the pupil ratio in the range of the preset ratio in the face image corresponding to the target image frame includes:
Based on eye contour coordinates in a face image corresponding to the target image frame, an externally connected rectangular area is obtained;
expanding the circumscribed matrix area outwards to designate pixels;
performing bilateral filtering on the expanded circumscribed matrix region, and then performing corrosion operation to obtain a corroded image;
performing binarization processing on the corroded image to obtain a binarized image;
inwards shrinking the specified pixels on the binarized image to obtain a shrunk image;
calculating the duty ratio of non-zero pixel values in the contracted image to obtain the pupil duty ratio of the face image;
and taking the face image with the pupil ratio in the range of the preset ratio as the first face image.
The above technical scheme that this application embodiment adopted can reach following beneficial effect:
the attention parameter corresponding to the subject is then determined based on the face orientation angle, the pupil deflection parameter, the first expression parameter, and the second expression parameter by determining the face orientation angle of the subject, the pupil deflection parameter of the subject, the first expression parameter that characterizes the degree of aggressiveness/negativity of the subject, and the second expression parameter that characterizes the degree of wakefulness/drowsiness of the subject. Thus, attention detection can be performed by comprehensively considering parameters related to attention from different dimensions, so that the attention concentration degree of an evaluative person can be accurately evaluated, and the robustness is high.
Drawings
The accompanying drawings, which are included to provide a further understanding of the present document, illustrate and explain the present document, and are not to be construed as limiting the document. In the drawings:
fig. 1 is a flowchart of an attention detection method based on facial orientation, facial expression and pupil tracking according to an embodiment of the present application.
Fig. 2 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Fig. 3 is a schematic structural diagram of an attention detection device based on facial orientation, facial expression and pupil tracking according to an embodiment of the present application.
Detailed Description
For the purposes, technical solutions and advantages of this document, the technical solutions of this document will be clearly and completely described below with reference to specific embodiments of this document and corresponding drawings. It will be apparent that the embodiments described are only some, but not all, of the embodiments of this document. All other embodiments, based on the embodiments in this document, which would be within the purview of one of ordinary skill in the art without the creative effort, are contemplated within the scope of protection of this document.
In order to ensure the accuracy of attention assessment, the embodiment of the application provides an attention detection method based on facial orientation, facial expression and pupil tracking, which can accurately assess the attention concentration degree and has high robustness.
The following describes in detail the attention detection method based on the face orientation, the facial expression and the pupil tracking provided in the embodiment of the present application.
The attention detection method based on facial orientation, facial expression and pupil tracking provided by the embodiment of the application can be applied to a user terminal, wherein the user terminal can be, but is not limited to, a personal computer, a smart phone, a tablet computer, a laptop portable computer, a personal digital assistant and the like.
It is understood that the execution bodies do not constitute limitations on the embodiments of the present application.
Optionally, the flow of the method for detecting the attention based on the face orientation, the facial expression and the pupil tracking is shown in fig. 1, and may include the following steps:
step S101, a target image frame including a head portrait of the subject to be evaluated is acquired.
The target image frame may be a video acquired by a camera, or may be a picture shot by a camera, and the target image frame may include one or more frames of images, which is not specifically limited in the embodiment of the present application.
Step S102, extracting face images of the evaluated person from the target image frames.
If the target image frame is a frame of image, the face image of the person to be evaluated can be directly extracted according to the target image frame. If the target image frame includes multiple frames of images, when the face image of the person to be evaluated is extracted from the target image frame, each frame in the multiple frames of images may be extracted to obtain multiple face images corresponding to the multiple frames of images one by one, or one frame in the multiple frames of images may be extracted to obtain one face image, or one frame may be selected from the multiple frames of images at intervals of a certain frame number to be extracted to obtain multiple face images.
The target image frame is not limited in format, and taking the target image frame as an image in an RGB format as an example, the step of extracting the face image of the subject to be evaluated from the target image frame may include the steps of:
step S1021, sequentially converts the image channels of the target image frame into RGB.
In step S1022, the target image frame after the image channel is sequentially converted into RGB is scaled to the first specified size.
The first specified size may be determined according to a face detection model that is used to determine coordinates of a boundary point of a face, for example, an input of the face detection model is a 300×300 matrix, and the first specified size may be 300×300 pixels.
Step S1023, the target image frame scaled to the first specified size is normalized.
Specifically, the pixel value of each pixel in the target image frame scaled to the first specified size may be subtracted by 127.5 and then divided by 127.5, such that the pixel value is distributed within the interval of [ -1,1 ]. For use in subsequent operations.
Step S1024, the matrix formed by the pixel values of each pixel point of the normalized target image frame is used as the input of the pre-trained face detection model to calculate, and the face boundary point coordinates are obtained.
In the embodiment of the application, a face detection model for face detection is pre-established. After normalization processing, a matrix can be constructed by taking the pixel value of each pixel point in the target image frame of the normalization processing as one value in the matrix. And then, the matrix is used as the input of a pre-trained face detection model to operate, so as to obtain the face boundary point coordinates.
For example, if the target image frame of the normalization process is an image frame with a size of 300×300 pixels, the constructed matrix is a 300×300 matrix.
In step S1025, a face image is extracted from the target image frame based on the face boundary point coordinates.
In this embodiment of the present invention, the face boundary point coordinates include a first boundary point coordinate and a second boundary point coordinate, and when the face image is extracted, the first boundary point coordinate and the second boundary point coordinate may be diagonal, and a rectangular face image may be extracted from the target image frame.
For example, the first boundary point coordinates are (x_min, y_min), the second boundary point coordinates are (x_max, y_max), and rectangular area images with the horizontal coordinate ranges of x_min to x_max and the vertical coordinate ranges of y_min to y_max in the target image frame can be used as face images.
Step S103, carrying out normalization processing on the pixel values of the face image.
Specifically, the pixel value of each pixel point in the face image may be divided by 256, so that the pixel value of each pixel point in the normalized face image is distributed in the interval of [0,1], so as to be convenient for calculating the coordinates of the key point in the face image.
Furthermore, before the normalization processing is performed on the pixel values of the face image, the face image may be scaled to a second specified size, where the second specified size may be determined according to a face key point detection model that is used to calculate key point coordinates in the face image. For example, the input requirement of the face key point detection model is a 112×112 matrix, and the second specified size may be 112×112 pixel size.
Step S104, calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image.
In the embodiment of the application, a face key point detection model for calculating the coordinates of the face key points is trained in advance, wherein the face key points can be, but are not limited to, eyes, ears, nose and other areas on the face.
When the coordinates of the key points of the face are calculated, the pixel value of each pixel point of the face image which is processed by normalization can be used as one value in the matrix to construct a matrix. And then, the matrix is used as the input of a pre-trained face key point detection model to operate, so that a plurality of key point coordinates in the face image are obtained.
The plurality of key point coordinates comprise eye contour coordinates.
Step S105, determining the face orientation angle of the person to be evaluated according to the coordinates of the key points, the parameter matrix in the preset camera and the distortion parameters of the preset camera.
The preset camera internal parameter matrix refers to an internal parameter matrix of the camera for acquiring the target image frame, and the preset camera distortion parameter refers to a distortion parameter of the camera for acquiring the target image frame. The parameter matrix in the preset camera and the distortion parameters of the preset camera are preset, and the cameras of different manufacturers can be set differently.
In the embodiment of the present application, determining the face orientation angle of the evaluated person may include the following steps:
step S1051, determining a plurality of head three-dimensional key point reference coordinates corresponding to the plurality of key point coordinates one by one according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters.
In the embodiment of the present application, the reference coordinates of the three-dimensional key points of the head can be determined through opencv, which is the prior art, and specific description is not made in the embodiment of the present application.
Step S1052, determining a rotation vector and a transformation vector of the camera according to the preset camera internal parameter matrix, the preset camera distortion parameter, the target key point coordinates and the target head three-dimensional key point reference coordinates.
The target key point coordinates are one of a plurality of key point coordinates, and the target head three-dimensional key point reference coordinates are head three-dimensional key point reference coordinates corresponding to the target key point coordinates.
Specifically, the function solvepnp can be used for receiving the target key point coordinates and the target head three-dimensional key point reference coordinates, and the preset camera internal parameter matrix and the preset camera distortion parameters for back-pushing to obtain the rotation vector and the transformation vector of the camera.
In step S1053, the rotation vector is converted into a rotation matrix.
The rotation vector can be converted into a rotation matrix by using a function Rodrigues during conversion, and the embodiment of the application is not specifically described.
Step S1054, splice the rotation matrix and the transformation vector to obtain an attitude matrix.
For example, the rotation matrix is a 3*3 matrix, the transformation vector is a 3-dimensional vector, and the spliced posture matrix is a 3×4 matrix.
In step S1055, the pose matrix is decomposed to obtain the face orientation angle of the subject.
Wherein the face orientation angle includes a pitch angle, a yaw angle, and a roll angle.
Step S106, extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates.
In the embodiment of the application, the coordinate index of the coordinate corresponding to each pixel point in the face image can be set, and after the eye contour coordinate is obtained, rough clipping can be performed according to the coordinate index of the eye contour coordinate, so that an eye region image of the person to be evaluated is obtained. The coordinate index of the eye contour coordinate comprises a coordinate index of a left eye contour coordinate and a coordinate index of a right eye contour coordinate, and the obtained eye area image comprises a left eye area image and a right eye area image.
Step S107, determining the pupil position of the eyes of the evaluated person according to the eye area image.
In an embodiment of the present application, determining the pupil position of the eye of the subject may include the following steps:
step S1071, determining a first face image with a pupil ratio in a preset ratio range in the face images corresponding to the target image frames.
Wherein, the pupil ratio is the ratio of the pupil area to the eye area.
Specifically, a face image corresponding to one frame of image can be selected from the target image frames at a certain frame number interval, and a corresponding circumscribed rectangular area is obtained according to the eye contour coordinates.
For example, if the fixed frame number is 5 frames, the 5 th frame image in the target image frame may be selected first, and the corresponding circumscribed rectangular region may be obtained according to the eye contour coordinates in the face image corresponding to the 5 th frame image.
The determined circumscribed matrix area is then expanded outwardly by a specified number of pixels, for example 5 pixels. And performing corrosion operation after bilateral filtering on the expanded circumscribed matrix area to obtain a corroded image. And then carrying out binarization processing on the corroded image to obtain a binarized image. And inwardly shrinking the binarized image by specified pixels to obtain a shrunk image, wherein the number of the shrunk pixels is the same as the number of the outwardly expanded pixels. And calculating the duty ratio of the non-zero pixel value in the contracted image to obtain the pupil duty ratio of the face image. If the pupil ratio of the face image is in the range of the preset ratio, the face image is taken as the first face image, otherwise, a frame of image is selected from the target image frame at intervals of a certain frame number (such as 5 frames) to calculate the pupil ratio of the face image.
The preset ratio range can be determined according to the ratio of the pupil to the eye area of an ordinary person. For example, the ratio of the pupils of the average person to the eye area is 0.46 to 0.50, and the preset ratio may be in the range of 0.46 to 0.50.
Step S1072, when the number of the first face images reaches the preset number, the average value of the pupil duty ratio of each first face image is obtained.
After each first face image is determined, whether the number of the first face images reaches the preset number or not can be judged, if the number of the first face images does not reach the preset number, one frame of image is continuously selected from the target image frames at intervals to calculate the pupil ratio of the face images, and the number of the first face images reaches the preset number. The preset number can be set according to actual conditions.
When the number of the first face images reaches the preset number, the average value of the pupil ratio of each first face image is calculated.
Step S1073, finding out the target pupil ratio closest to the mean value in the pupil ratios of the first face images.
After the average value of the pupil duty ratios of the first face images is obtained, one pupil duty ratio closest to the average value can be found out from the pupil duty ratios corresponding to the first face images to be used as the target pupil duty ratio.
Step S1074, selecting a first face image corresponding to the target pupil ratio as a target face image.
Step S1075, the center of the pupil area of the eye in the target face image is set as the pupil position.
And step S108, determining pupil deflection parameters of the evaluated person based on the eye region image and the eye pupil position.
The pupil deflection parameter may be a distance of left/right deflection of the pupil, or may be a proportion of left/right deflection of the pupil, which is not specifically limited in the embodiment of the present application.
Specifically, the distance between the two side corners of the eye and the pupil can be calculated according to the coordinates of the key points of the eye area and the coordinates corresponding to the pupil positions, and then the pupil deflection parameter of the person to be evaluated is determined according to the distances between the two side corners and the pupil.
Step S109, calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated.
In this embodiment of the present application, a facial emotion recognition model for facial emotion recognition is trained in advance, after normalization processing is performed on pixel values of a face image, pixel values of each pixel point of the face image after normalization processing may be used as one value in a matrix to construct a matrix, and then the matrix is used as input of the facial emotion recognition model trained in advance to perform operation, so as to obtain a first expression parameter and a second expression parameter corresponding to an evaluated person. Wherein the first expression parameter characterizes the positive/negative degree of the subject and the second expression parameter characterizes the wakefulness/drowsiness degree of the subject.
Step S110, determining the attention parameter corresponding to the person to be evaluated based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter.
Wherein the attention parameter characterizes a degree of concentration of the subject.
The face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter can reflect whether the evaluative person is concentrated or not to a certain extent. Therefore, when determining the attention parameter corresponding to the person to be evaluated, a score may be respectively assigned to the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, and then a final score corresponding to the person to be evaluated is determined according to the sum of the scores corresponding to the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, so that the attention concentration degree of the person to be evaluated is obtained according to the final score.
Specifically, the total score of the face orientation angle may be set to 100, and the score corresponding to the face orientation angle may be expressed as ss=100-pitch a-yaw b-roll c, where pitch represents the degree of pitch angle in the face orientation angle, yaw represents the degree of yaw angle in the face orientation angle, and roll represents the degree of roll angle in the face orientation angle. When the absolute value of pitch is less than or equal to 15, the value of a is 0.8, and when the absolute value of pitch is greater than 15, the value of a is 1.5. When the absolute value of the yaw is less than or equal to 15, the value of b is 0.8, and when the absolute value of the yaw is more than 15, the value of b is 1.5. When the absolute value of the roll is less than or equal to 15, the value of c is 0.8, and when the absolute value of the roll is greater than 15, the value of c is 1.5
The score corresponding to the pupil deflection parameter may be the minimum value of the score corresponding to the pupil deflection parameter of the left eye and the score corresponding to the pupil deflection parameter of the right eye. The score corresponding to the pupil deflection parameter of the left eye may be expressed as els=100-abs ((1-LR)). 50, LR represents the pupil deflection parameter of the left eye. The score corresponding to the pupil deflection parameter for the right eye may be expressed as ers=100-abs ((1-RR)). 50, RR representing the pupil deflection parameter for the right eye. The score corresponding to the pupil deflection parameter is marked as EMS, and EMS is the minimum value in ELS and ERS.
The first expression parameter characterizes the positive/negative degree of the person being evaluated, wherein a first expression parameter greater than 0 indicates a positive state and less than 0 indicates a negative state, denoted V. The second expression parameter characterizes the wakefulness/drowsiness of the subject, wherein the second expression parameter is greater than 0 for the wakefulness state and less than 0 for the drowsiness state, denoted as A. The score corresponding to the first expression parameter may be denoted as sv=10 (1+V), and the score corresponding to the second expression parameter may be denoted as sa=10 (1+a).
The final score fs= (ss+ems+sa+sv)/200 for the evaluators. The corresponding final score FS of the evaluators characterizes the degree of concentration of the evaluators, wherein a larger final score FS indicates a more concentrated concentration of the evaluators.
According to the attention detection method based on facial orientation, facial expression and pupil tracking, the facial orientation angle of the person to be evaluated, the pupil deflection parameter of the person to be evaluated, the first expression parameter representing the positive/negative degree of the person to be evaluated and the second expression parameter representing the wakefulness/drowsiness degree of the person to be evaluated are determined, and then the attention parameter corresponding to the person to be evaluated is determined based on the facial orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter. Thus, attention detection can be performed by comprehensively considering parameters related to attention from different dimensions, so that the attention concentration degree of an evaluative person can be accurately evaluated, and the robustness is high.
Fig. 2 is a schematic structural diagram of an electronic device according to an embodiment of the present application. Referring to fig. 2, at the hardware level, the electronic device includes a processor, and optionally an internal bus, a network interface, and a memory. The Memory may include a Memory, such as a Random-Access Memory (RAM), and may further include a non-volatile Memory (non-volatile Memory), such as at least 1 disk Memory. Of course, the electronic device may also include hardware required for other services.
The processor, network interface, and memory may be interconnected by an internal bus, which may be an ISA (Industry Standard Architecture ) bus, a PCI (Peripheral Component Interconnect, peripheral component interconnect standard) bus, or EISA (Extended Industry Standard Architecture ) bus, among others. The buses may be classified as address buses, data buses, control buses, etc. For ease of illustration, only one bi-directional arrow is shown in FIG. 2, but not only one bus or type of bus.
And the memory is used for storing programs. In particular, the program may include program code including computer-operating instructions. The memory may include memory and non-volatile storage and provide instructions and data to the processor.
The processor reads the corresponding computer program from the nonvolatile memory into the memory and then runs the computer program to form the attention detection device based on the face orientation, the facial expression and the pupil tracking on a logic level. The processor is used for executing the programs stored in the memory and is specifically used for executing the following operations:
acquiring a target image frame containing a head portrait of an evaluated person;
Extracting a face image of the person to be evaluated from the target image frame;
carrying out normalization processing on the pixel values of the face image;
calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image, wherein the plurality of key point coordinates comprise eye contour coordinates;
determining the face orientation angle of the person to be evaluated according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates;
determining the eye pupil position of the person to be evaluated according to the eye region image;
determining a pupil deflection parameter of the subject based on the eye region image and the eye pupil position;
calculating pixel values of all pixel points of the face image after normalization processing as input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated, wherein the first expression parameter represents the positive/negative degree of the person to be evaluated, and the second expression parameter represents the wakefulness/drowsiness degree of the person to be evaluated;
And determining an attention parameter corresponding to the evaluative person based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, wherein the attention parameter characterizes the attention concentration degree of the evaluative person.
The method performed by the attention detection device based on facial orientation, facial expression and pupil tracking as disclosed in the embodiment shown in fig. 2 of the present application may be applied to or implemented by a processor. The processor may be an integrated circuit chip having signal processing capabilities. In implementation, the steps of the above method may be performed by integrated logic circuits of hardware in a processor or by instructions in the form of software. The processor may be a general-purpose processor, including a central processing unit (Central Processing Unit, CPU), a network processor (Network Processor, NP), etc.; but also digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), field programmable gate arrays (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components. The disclosed methods, steps, and logic blocks in one or more embodiments of the present application may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of a method disclosed in connection with one or more embodiments of the present application may be embodied directly in a hardware decoding processor or in a combination of hardware and software modules in a decoding processor. The software modules may be located in a random access memory, flash memory, read only memory, programmable read only memory, or electrically erasable programmable memory, registers, etc. as well known in the art. The storage medium is located in a memory, and the processor reads the information in the memory and, in combination with its hardware, performs the steps of the above method.
The electronic device may further execute the method of the embodiment shown in fig. 1, and implement the functions of the attention detection device of the embodiment shown in fig. 1 based on facial orientation, facial expression and pupil tracking, which are not described herein.
Of course, other implementations, such as a logic device or a combination of hardware and software, are not excluded from the electronic device of the present application, that is, the execution subject of the following processing flow is not limited to each logic unit, but may be hardware or a logic device.
The present embodiments also provide a computer readable storage medium storing one or more programs, the one or more programs comprising instructions, which when executed by a portable electronic device comprising a plurality of application programs, enable the portable electronic device to perform the method of the embodiment of fig. 1, and in particular to:
acquiring a target image frame containing a head portrait of an evaluated person;
extracting a face image of the person to be evaluated from the target image frame;
carrying out normalization processing on the pixel values of the face image;
calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image, wherein the plurality of key point coordinates comprise eye contour coordinates;
Determining the face orientation angle of the person to be evaluated according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates;
determining the eye pupil position of the person to be evaluated according to the eye region image;
determining a pupil deflection parameter of the subject based on the eye region image and the eye pupil position;
calculating pixel values of all pixel points of the face image after normalization processing as input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated, wherein the first expression parameter represents the positive/negative degree of the person to be evaluated, and the second expression parameter represents the wakefulness/drowsiness degree of the person to be evaluated;
and determining an attention parameter corresponding to the evaluative person based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, wherein the attention parameter characterizes the attention concentration degree of the evaluative person.
Fig. 3 is a schematic structural diagram of an attention detection device based on facial orientation, facial expression and pupil tracking according to an embodiment of the present application. Referring to fig. 3, in a software embodiment, the provided attention detection device based on facial orientation, facial expression and pupil tracking may include:
an acquisition module for acquiring a target image frame containing a head portrait of an evaluated person;
a first extraction module, configured to extract a face image of the person under evaluation from the target image frame;
the normalization module is used for carrying out normalization processing on the pixel values of the face image;
the first operation module is used for operating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image, wherein the plurality of key point coordinates comprise eye contour coordinates;
the first determining module is used for determining the face orientation angle of the person to be evaluated according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
the second extraction module is used for extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates;
The second determining module is used for determining the eye pupil position of the evaluated person according to the eye area image;
a third determining module for determining a pupil deflection parameter of the subject based on the eye region image and the eye pupil position;
the second operation module is used for operating the pixel values of each pixel point of the face image after normalization processing as the input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated, wherein the first expression parameter represents the positive/negative degree of the person to be evaluated, and the second expression parameter represents the wakefulness/drowsiness degree of the person to be evaluated;
and a fourth determining module, configured to determine an attention parameter corresponding to the person under evaluation based on the face orientation angle, the pupil deflection parameter, the first expression parameter, and the second expression parameter, where the attention parameter characterizes the attention concentration of the person under evaluation.
The foregoing describes specific embodiments of the present application. Other embodiments are within the scope of the following claims. In some cases, the actions or steps recited in the claims can be performed in a different order than in the embodiments and still achieve desirable results. In addition, the processes depicted in the accompanying figures do not necessarily require the particular order shown, or sequential order, to achieve desirable results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
In summary, the foregoing description is only a preferred embodiment of the present application, and is not intended to limit the scope of the present application. Any modification, equivalent replacement, improvement, etc. made within the spirit and principles of the present application should be included in the protection scope of the present application.
The system, apparatus, module or unit set forth in the above embodiments may be implemented in particular by a computer chip or entity, or by a product having a certain function. One typical implementation is a computer. In particular, the computer may be, for example, a personal computer, a laptop computer, a cellular telephone, a camera phone, a smart phone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
Computer readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of storage media for a computer include, but are not limited to, phase change memory (PRAM), static Random Access Memory (SRAM), dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), read Only Memory (ROM), electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium, which can be used to store information that can be accessed by a computing device. Computer-readable media, as defined herein, does not include transitory computer-readable media (transmission media), such as modulated data signals and carrier waves.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising one … …" does not exclude the presence of other like elements in a process, method, article or apparatus that comprises the element.
All embodiments in the application are described in a progressive manner, and identical and similar parts of all embodiments are mutually referred, so that each embodiment mainly describes differences from other embodiments. In particular, for system embodiments, since they are substantially similar to method embodiments, the description is relatively simple, as relevant to see a section of the description of method embodiments.
Claims (7)
1. A method of detecting attention based on facial orientation, facial expression, and pupil tracking, comprising:
Acquiring a target image frame containing a head portrait of an evaluated person;
extracting a face image of the person to be evaluated from the target image frame;
carrying out normalization processing on the pixel values of the face image;
calculating the pixel value of each pixel point of the face image after normalization processing as the input of a pre-trained face key point detection model to obtain a plurality of key point coordinates in the face image, wherein the plurality of key point coordinates comprise eye contour coordinates;
determining the face orientation angle of the person to be evaluated according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
extracting an eye region image of the person to be evaluated from the face image based on the eye contour coordinates;
determining the eye pupil position of the person to be evaluated according to the eye region image;
determining a pupil deflection parameter of the subject based on the eye region image and the eye pupil position;
calculating pixel values of all pixel points of the face image after normalization processing as input of a pre-trained facial emotion recognition model to obtain a first expression parameter and a second expression parameter corresponding to the person to be evaluated, wherein the first expression parameter represents the positive/negative degree of the person to be evaluated, and the second expression parameter represents the wakefulness/drowsiness degree of the person to be evaluated;
And determining an attention parameter corresponding to the evaluative person based on the face orientation angle, the pupil deflection parameter, the first expression parameter and the second expression parameter, wherein the attention parameter characterizes the attention concentration degree of the evaluative person.
2. The method of claim 1, wherein the target image frame is in RGB format, and the extracting the face image of the subject from the target image frame comprises:
sequentially converting the image channels of the target image frames into RGB;
scaling the target image frames after the image channels are sequentially converted into RGB to a first specified size;
normalizing the target image frames scaled to the first specified size;
calculating a matrix formed by pixel values of all pixel points of the target image frame subjected to normalization processing as input of a pre-trained face detection model to obtain face boundary point coordinates;
and extracting the face image from the target image frame based on the face boundary point coordinates.
3. The method of claim 2, wherein the face boundary point coordinates include first boundary point coordinates and second boundary point coordinates, the extracting the face image from the target image frame based on the face boundary point coordinates comprising:
And taking the first boundary point coordinates and the second boundary point coordinates as diagonal angles, and extracting the rectangular face image from the target image frame.
4. The method according to claim 1, wherein the method further comprises:
scaling the face image to a second specified size;
the normalizing the pixel value of the face image includes:
and carrying out normalization processing on the pixel values of the face image scaled to the second designated size.
5. The method of claim 1, wherein determining the face orientation angle of the subject based on the plurality of keypoint coordinates, the preset intra-camera parameter matrix, and the preset camera distortion parameter comprises:
determining a plurality of head three-dimensional key point reference coordinates corresponding to the plurality of key point coordinates one by one according to the plurality of key point coordinates, the preset camera internal parameter matrix and the preset camera distortion parameters;
determining a rotation vector and a transformation vector of a camera according to the preset camera internal parameter matrix, the preset camera distortion parameter, the target key point coordinates and the target head three-dimensional key point reference coordinates;
Converting the rotation vector into a rotation matrix;
splicing the rotation matrix with the transformation vector to obtain a posture matrix;
decomposing the gesture matrix to obtain the face orientation angle of the person to be evaluated;
the target key point coordinate is one of the plurality of key point coordinates, and the target head three-dimensional key point reference coordinate is a head three-dimensional key point reference coordinate corresponding to the target key point coordinate.
6. The method of claim 1, wherein said determining the eye pupil position of the subject from the eye region image comprises:
determining a first face image with pupil ratio in a preset ratio range in the face image corresponding to the target image frame, wherein the pupil ratio is the ratio of the pupil area to the eye area;
when the number of the first face images reaches a preset number, calculating the average value of the pupil ratio of each first face image;
finding out the target pupil duty ratio closest to the average value in the pupil duty ratios of the first face images;
selecting a first face image corresponding to the target pupil ratio as a target face image;
And taking the center of the pupil area of the eyes in the target face image as the pupil position.
7. The method of claim 6, wherein determining a first face image with a pupil ratio in a preset ratio range in the face image corresponding to the target image frame comprises:
based on eye contour coordinates in a face image corresponding to the target image frame, an externally connected rectangular area is obtained;
expanding the circumscribed rectangular area outwards to designate pixels;
performing bilateral filtering on the expanded circumscribed rectangular region, and then performing corrosion operation to obtain a corroded image;
performing binarization processing on the corroded image to obtain a binarized image;
inwards shrinking the specified pixels on the binarized image to obtain a shrunk image;
calculating the duty ratio of non-zero pixel values in the contracted image to obtain the pupil duty ratio of the face image;
and taking the face image with the pupil ratio in the range of the preset ratio as the first face image.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110310469.5A CN113052064B (en) | 2021-03-23 | 2021-03-23 | Attention detection method based on face orientation, facial expression and pupil tracking |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110310469.5A CN113052064B (en) | 2021-03-23 | 2021-03-23 | Attention detection method based on face orientation, facial expression and pupil tracking |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113052064A CN113052064A (en) | 2021-06-29 |
CN113052064B true CN113052064B (en) | 2024-04-02 |
Family
ID=76514596
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110310469.5A Active CN113052064B (en) | 2021-03-23 | 2021-03-23 | Attention detection method based on face orientation, facial expression and pupil tracking |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113052064B (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113283978B (en) * | 2021-05-06 | 2024-05-10 | 北京思图场景数据科技服务有限公司 | Financial risk assessment method based on biological basis, behavioral characteristics and business characteristics |
CN113838086B (en) * | 2021-08-23 | 2024-03-22 | 广东电网有限责任公司 | Attention assessment test method, device, electronic equipment and storage medium |
Citations (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000259833A (en) * | 1999-03-08 | 2000-09-22 | Toshiba Corp | Face image processor and processing method therefor |
JP2002282210A (en) * | 2001-03-27 | 2002-10-02 | Japan Science & Technology Corp | Method and apparatus for detecting visual axis |
JP2007268164A (en) * | 2006-03-31 | 2007-10-18 | National Univ Corp Shizuoka Univ | Visual axis movement detecting method and apparatus |
CN101593352A (en) * | 2009-06-12 | 2009-12-02 | 浙江大学 | Driving safety monitoring system based on face orientation and visual focus |
JP2010282339A (en) * | 2009-06-03 | 2010-12-16 | Seiko Epson Corp | Image processor for correcting position of pupil in eye, image processing method, image processing program and printer |
KR20110102073A (en) * | 2010-03-10 | 2011-09-16 | 홍익대학교 산학협력단 | Method for detecting liveness for face recognition system |
JP2012186821A (en) * | 2012-04-18 | 2012-09-27 | Toshiba Corp | Face image processing device, face image processing method, electronic still camera, digital image processing device and digital image processing method |
US8401248B1 (en) * | 2008-12-30 | 2013-03-19 | Videomining Corporation | Method and system for measuring emotional and attentional response to dynamic digital media content |
CN104013414A (en) * | 2014-04-30 | 2014-09-03 | 南京车锐信息科技有限公司 | Driver fatigue detecting system based on smart mobile phone |
WO2015154882A1 (en) * | 2014-04-11 | 2015-10-15 | The Eye Tribe Aps | Systems and methods of eye tracking calibration |
KR20160036375A (en) * | 2014-09-25 | 2016-04-04 | 백석대학교산학협력단 | Fast Eye Detection Method Using Block Contrast and Symmetry in Mobile Device |
WO2016143759A1 (en) * | 2015-03-06 | 2016-09-15 | 株式会社 脳機能研究所 | Emotion estimating device and emotion estimating method |
CN106228293A (en) * | 2016-07-18 | 2016-12-14 | 重庆中科云丛科技有限公司 | teaching evaluation method and system |
WO2017107957A1 (en) * | 2015-12-22 | 2017-06-29 | 中兴通讯股份有限公司 | Human face image retrieval method and apparatus |
US9775512B1 (en) * | 2014-03-19 | 2017-10-03 | Christopher W. Tyler | Binocular eye tracking from video frame sequences |
EP3225154A2 (en) * | 2016-03-31 | 2017-10-04 | Fujitsu Limited | Gaze detection apparatus and gaze detection method |
WO2018033143A1 (en) * | 2016-08-19 | 2018-02-22 | 北京市商汤科技开发有限公司 | Video image processing method, apparatus and electronic device |
CN107909057A (en) * | 2017-11-30 | 2018-04-13 | 广东欧珀移动通信有限公司 | Image processing method, device, electronic equipment and computer-readable recording medium |
WO2018142388A1 (en) * | 2017-02-05 | 2018-08-09 | Bioeye Ltd. | A method for pupil detection for cognitive monitoring, analysis, and biofeedback-based treatment and training |
CN108921061A (en) * | 2018-06-20 | 2018-11-30 | 腾讯科技(深圳)有限公司 | A kind of expression recognition method, device and equipment |
CN109271914A (en) * | 2018-09-07 | 2019-01-25 | 百度在线网络技术(北京)有限公司 | Detect method, apparatus, storage medium and the terminal device of sight drop point |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
CN109446892A (en) * | 2018-09-14 | 2019-03-08 | 杭州宇泛智能科技有限公司 | Human eye notice positioning method and system based on deep neural network |
CN109446880A (en) * | 2018-09-05 | 2019-03-08 | 广州维纳斯家居股份有限公司 | Intelligent subscriber participation evaluation method, device, intelligent elevated table and storage medium |
CN109902553A (en) * | 2019-01-03 | 2019-06-18 | 杭州电子科技大学 | A kind of face alignment method of the multi-angle based on facial pixel difference |
CN109902630A (en) * | 2019-03-01 | 2019-06-18 | 上海像我信息科技有限公司 | A kind of attention judgment method, device, system, equipment and storage medium |
CN110147744A (en) * | 2019-05-09 | 2019-08-20 | 腾讯科技(深圳)有限公司 | A kind of quality of human face image appraisal procedure, device and terminal |
WO2019184125A1 (en) * | 2018-03-30 | 2019-10-03 | 平安科技(深圳)有限公司 | Micro-expression-based risk identification method and device, equipment and medium |
CN110363133A (en) * | 2019-07-10 | 2019-10-22 | 广州市百果园信息技术有限公司 | A kind of method, apparatus, equipment and the storage medium of line-of-sight detection and video processing |
CN110705500A (en) * | 2019-10-12 | 2020-01-17 | 深圳创新奇智科技有限公司 | Attention detection method and system for personnel working image based on deep learning |
WO2020034902A1 (en) * | 2018-08-11 | 2020-02-20 | 昆山美卓智能科技有限公司 | Smart desk having status monitoring function, monitoring system server, and monitoring method |
WO2020042345A1 (en) * | 2018-08-28 | 2020-03-05 | 初速度(苏州)科技有限公司 | Method and system for acquiring line-of-sight direction of human eyes by means of single camera |
WO2020062960A1 (en) * | 2018-09-29 | 2020-04-02 | 北京市商汤科技开发有限公司 | Neural network training method and apparatus, gaze tracking method and apparatus, and electronic device |
CN110969061A (en) * | 2018-09-29 | 2020-04-07 | 北京市商汤科技开发有限公司 | Neural network training method, neural network training device, visual line detection method, visual line detection device and electronic equipment |
CN111046744A (en) * | 2019-11-21 | 2020-04-21 | 深圳云天励飞技术有限公司 | Method and device for detecting attention area, readable storage medium and terminal equipment |
CN111507592A (en) * | 2020-04-08 | 2020-08-07 | 山东大学 | Evaluation method for active modification behaviors of prisoners |
CN111563417A (en) * | 2020-04-13 | 2020-08-21 | 华南理工大学 | Pyramid structure convolutional neural network-based facial expression recognition method |
CN111680546A (en) * | 2020-04-26 | 2020-09-18 | 北京三快在线科技有限公司 | Attention detection method, attention detection device, electronic equipment and storage medium |
WO2020186883A1 (en) * | 2019-03-18 | 2020-09-24 | 北京市商汤科技开发有限公司 | Methods, devices and apparatuses for gaze area detection and neural network training |
CN111862287A (en) * | 2020-07-20 | 2020-10-30 | 广州市百果园信息技术有限公司 | Eye texture image generation method, texture mapping method, device and electronic equipment |
CN111914783A (en) * | 2020-08-10 | 2020-11-10 | 深圳市视美泰技术股份有限公司 | Method and device for determining human face deflection angle, computer equipment and medium |
CN111933275A (en) * | 2020-07-17 | 2020-11-13 | 兰州大学 | Depression evaluation system based on eye movement and facial expression |
WO2020231401A1 (en) * | 2019-05-13 | 2020-11-19 | Huawei Technologies Co., Ltd. | A neural network for head pose and gaze estimation using photorealistic synthetic data |
CN112149641A (en) * | 2020-10-23 | 2020-12-29 | 北京百度网讯科技有限公司 | Method, device, equipment and storage medium for monitoring driving state |
KR102198360B1 (en) * | 2019-10-24 | 2021-01-04 | 광운대학교 산학협력단 | Eye tracking system and method based on face images |
CN112329699A (en) * | 2020-11-19 | 2021-02-05 | 北京中科虹星科技有限公司 | Method for positioning human eye fixation point with pixel-level precision |
CN112381061A (en) * | 2020-12-04 | 2021-02-19 | 中国科学院大学 | Facial expression recognition method and system |
CN112446322A (en) * | 2020-11-24 | 2021-03-05 | 杭州网易云音乐科技有限公司 | Eyeball feature detection method, device, equipment and computer-readable storage medium |
CN112464865A (en) * | 2020-12-08 | 2021-03-09 | 北京理工大学 | Facial expression recognition method based on pixel and geometric mixed features |
WO2021047185A1 (en) * | 2019-09-12 | 2021-03-18 | 深圳壹账通智能科技有限公司 | Monitoring method and apparatus based on facial recognition, and storage medium and computer device |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI447658B (en) * | 2010-03-24 | 2014-08-01 | Ind Tech Res Inst | Facial expression capturing method and apparatus therewith |
US8408706B2 (en) * | 2010-12-13 | 2013-04-02 | Microsoft Corporation | 3D gaze tracker |
US10762335B2 (en) * | 2017-05-16 | 2020-09-01 | Apple Inc. | Attention detection |
US11393252B2 (en) * | 2019-05-01 | 2022-07-19 | Accenture Global Solutions Limited | Emotion sensing artificial intelligence |
-
2021
- 2021-03-23 CN CN202110310469.5A patent/CN113052064B/en active Active
Patent Citations (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000259833A (en) * | 1999-03-08 | 2000-09-22 | Toshiba Corp | Face image processor and processing method therefor |
JP2002282210A (en) * | 2001-03-27 | 2002-10-02 | Japan Science & Technology Corp | Method and apparatus for detecting visual axis |
JP2007268164A (en) * | 2006-03-31 | 2007-10-18 | National Univ Corp Shizuoka Univ | Visual axis movement detecting method and apparatus |
US8401248B1 (en) * | 2008-12-30 | 2013-03-19 | Videomining Corporation | Method and system for measuring emotional and attentional response to dynamic digital media content |
JP2010282339A (en) * | 2009-06-03 | 2010-12-16 | Seiko Epson Corp | Image processor for correcting position of pupil in eye, image processing method, image processing program and printer |
CN101593352A (en) * | 2009-06-12 | 2009-12-02 | 浙江大学 | Driving safety monitoring system based on face orientation and visual focus |
KR20110102073A (en) * | 2010-03-10 | 2011-09-16 | 홍익대학교 산학협력단 | Method for detecting liveness for face recognition system |
JP2012186821A (en) * | 2012-04-18 | 2012-09-27 | Toshiba Corp | Face image processing device, face image processing method, electronic still camera, digital image processing device and digital image processing method |
US9775512B1 (en) * | 2014-03-19 | 2017-10-03 | Christopher W. Tyler | Binocular eye tracking from video frame sequences |
WO2015154882A1 (en) * | 2014-04-11 | 2015-10-15 | The Eye Tribe Aps | Systems and methods of eye tracking calibration |
CN104013414A (en) * | 2014-04-30 | 2014-09-03 | 南京车锐信息科技有限公司 | Driver fatigue detecting system based on smart mobile phone |
KR20160036375A (en) * | 2014-09-25 | 2016-04-04 | 백석대학교산학협력단 | Fast Eye Detection Method Using Block Contrast and Symmetry in Mobile Device |
WO2016143759A1 (en) * | 2015-03-06 | 2016-09-15 | 株式会社 脳機能研究所 | Emotion estimating device and emotion estimating method |
WO2017107957A1 (en) * | 2015-12-22 | 2017-06-29 | 中兴通讯股份有限公司 | Human face image retrieval method and apparatus |
EP3225154A2 (en) * | 2016-03-31 | 2017-10-04 | Fujitsu Limited | Gaze detection apparatus and gaze detection method |
CN106228293A (en) * | 2016-07-18 | 2016-12-14 | 重庆中科云丛科技有限公司 | teaching evaluation method and system |
WO2018033143A1 (en) * | 2016-08-19 | 2018-02-22 | 北京市商汤科技开发有限公司 | Video image processing method, apparatus and electronic device |
WO2018142388A1 (en) * | 2017-02-05 | 2018-08-09 | Bioeye Ltd. | A method for pupil detection for cognitive monitoring, analysis, and biofeedback-based treatment and training |
CN107909057A (en) * | 2017-11-30 | 2018-04-13 | 广东欧珀移动通信有限公司 | Image processing method, device, electronic equipment and computer-readable recording medium |
WO2019184125A1 (en) * | 2018-03-30 | 2019-10-03 | 平安科技(深圳)有限公司 | Micro-expression-based risk identification method and device, equipment and medium |
CN108921061A (en) * | 2018-06-20 | 2018-11-30 | 腾讯科技(深圳)有限公司 | A kind of expression recognition method, device and equipment |
WO2020034902A1 (en) * | 2018-08-11 | 2020-02-20 | 昆山美卓智能科技有限公司 | Smart desk having status monitoring function, monitoring system server, and monitoring method |
CN109344693A (en) * | 2018-08-13 | 2019-02-15 | 华南理工大学 | A kind of face multizone fusion expression recognition method based on deep learning |
WO2020042345A1 (en) * | 2018-08-28 | 2020-03-05 | 初速度(苏州)科技有限公司 | Method and system for acquiring line-of-sight direction of human eyes by means of single camera |
CN109446880A (en) * | 2018-09-05 | 2019-03-08 | 广州维纳斯家居股份有限公司 | Intelligent subscriber participation evaluation method, device, intelligent elevated table and storage medium |
CN109271914A (en) * | 2018-09-07 | 2019-01-25 | 百度在线网络技术(北京)有限公司 | Detect method, apparatus, storage medium and the terminal device of sight drop point |
CN109446892A (en) * | 2018-09-14 | 2019-03-08 | 杭州宇泛智能科技有限公司 | Human eye notice positioning method and system based on deep neural network |
CN110969061A (en) * | 2018-09-29 | 2020-04-07 | 北京市商汤科技开发有限公司 | Neural network training method, neural network training device, visual line detection method, visual line detection device and electronic equipment |
WO2020062960A1 (en) * | 2018-09-29 | 2020-04-02 | 北京市商汤科技开发有限公司 | Neural network training method and apparatus, gaze tracking method and apparatus, and electronic device |
CN109902553A (en) * | 2019-01-03 | 2019-06-18 | 杭州电子科技大学 | A kind of face alignment method of the multi-angle based on facial pixel difference |
CN109902630A (en) * | 2019-03-01 | 2019-06-18 | 上海像我信息科技有限公司 | A kind of attention judgment method, device, system, equipment and storage medium |
WO2020186883A1 (en) * | 2019-03-18 | 2020-09-24 | 北京市商汤科技开发有限公司 | Methods, devices and apparatuses for gaze area detection and neural network training |
CN110147744A (en) * | 2019-05-09 | 2019-08-20 | 腾讯科技(深圳)有限公司 | A kind of quality of human face image appraisal procedure, device and terminal |
WO2020231401A1 (en) * | 2019-05-13 | 2020-11-19 | Huawei Technologies Co., Ltd. | A neural network for head pose and gaze estimation using photorealistic synthetic data |
CN110363133A (en) * | 2019-07-10 | 2019-10-22 | 广州市百果园信息技术有限公司 | A kind of method, apparatus, equipment and the storage medium of line-of-sight detection and video processing |
WO2021047185A1 (en) * | 2019-09-12 | 2021-03-18 | 深圳壹账通智能科技有限公司 | Monitoring method and apparatus based on facial recognition, and storage medium and computer device |
CN110705500A (en) * | 2019-10-12 | 2020-01-17 | 深圳创新奇智科技有限公司 | Attention detection method and system for personnel working image based on deep learning |
KR102198360B1 (en) * | 2019-10-24 | 2021-01-04 | 광운대학교 산학협력단 | Eye tracking system and method based on face images |
CN111046744A (en) * | 2019-11-21 | 2020-04-21 | 深圳云天励飞技术有限公司 | Method and device for detecting attention area, readable storage medium and terminal equipment |
CN111507592A (en) * | 2020-04-08 | 2020-08-07 | 山东大学 | Evaluation method for active modification behaviors of prisoners |
CN111563417A (en) * | 2020-04-13 | 2020-08-21 | 华南理工大学 | Pyramid structure convolutional neural network-based facial expression recognition method |
CN111680546A (en) * | 2020-04-26 | 2020-09-18 | 北京三快在线科技有限公司 | Attention detection method, attention detection device, electronic equipment and storage medium |
CN111933275A (en) * | 2020-07-17 | 2020-11-13 | 兰州大学 | Depression evaluation system based on eye movement and facial expression |
CN111862287A (en) * | 2020-07-20 | 2020-10-30 | 广州市百果园信息技术有限公司 | Eye texture image generation method, texture mapping method, device and electronic equipment |
CN111914783A (en) * | 2020-08-10 | 2020-11-10 | 深圳市视美泰技术股份有限公司 | Method and device for determining human face deflection angle, computer equipment and medium |
CN112149641A (en) * | 2020-10-23 | 2020-12-29 | 北京百度网讯科技有限公司 | Method, device, equipment and storage medium for monitoring driving state |
CN112329699A (en) * | 2020-11-19 | 2021-02-05 | 北京中科虹星科技有限公司 | Method for positioning human eye fixation point with pixel-level precision |
CN112446322A (en) * | 2020-11-24 | 2021-03-05 | 杭州网易云音乐科技有限公司 | Eyeball feature detection method, device, equipment and computer-readable storage medium |
CN112381061A (en) * | 2020-12-04 | 2021-02-19 | 中国科学院大学 | Facial expression recognition method and system |
CN112464865A (en) * | 2020-12-08 | 2021-03-09 | 北京理工大学 | Facial expression recognition method based on pixel and geometric mixed features |
Non-Patent Citations (2)
Title |
---|
Attention decrease detection based on video analysis in e-learning;Wang, Liying;《Transactions on Edutainment XIV》;20181231;166-179 * |
基于多维度信息融合的学生在线学习投入度研究;吴慧婷;《中国博士学位论文全文数据库 社会科学II辑》;20210115(第1期);H127-22 * |
Also Published As
Publication number | Publication date |
---|---|
CN113052064A (en) | 2021-06-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10395103B2 (en) | Object detection method, object detection apparatus, and program | |
US9552642B2 (en) | Apparatus and method for tracking object using feature descriptor, and apparatus and method for removing garbage feature | |
CN113052064B (en) | Attention detection method based on face orientation, facial expression and pupil tracking | |
CN111597884A (en) | Facial action unit identification method and device, electronic equipment and storage medium | |
CN110852311A (en) | Three-dimensional human hand key point positioning method and device | |
CN110751012B (en) | Target detection evaluation method and device, electronic equipment and storage medium | |
CN112348778B (en) | Object identification method, device, terminal equipment and storage medium | |
CN112085701A (en) | Face ambiguity detection method and device, terminal equipment and storage medium | |
CN115631112B (en) | Building contour correction method and device based on deep learning | |
CN112991349B (en) | Image processing method, device, equipment and storage medium | |
CN110232381B (en) | License plate segmentation method, license plate segmentation device, computer equipment and computer readable storage medium | |
CN113129298B (en) | Method for identifying definition of text image | |
CN108960247B (en) | Image significance detection method and device and electronic equipment | |
CN108960246B (en) | Binarization processing device and method for image recognition | |
CN110738204A (en) | Method and device for positioning certificate areas | |
EP2998928B1 (en) | Apparatus and method for extracting high watermark image from continuously photographed images | |
WO2020244076A1 (en) | Face recognition method and apparatus, and electronic device and storage medium | |
CN114255493A (en) | Image detection method, face detection device, face detection equipment and storage medium | |
CN113228105A (en) | Image processing method and device and electronic equipment | |
CN115841672A (en) | Character detection and identification method, device and equipment | |
JP7294454B2 (en) | Object detection method and object detection device | |
US10832076B2 (en) | Method and image processing entity for applying a convolutional neural network to an image | |
CN110852210A (en) | Method and device for generating vehicle weight recognition model | |
CN112308812A (en) | Method, terminal and storage medium for detecting picture definition | |
US9652878B2 (en) | Electronic device and method for adjusting page |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |