CN110543813B - Face image and gaze counting method and system based on scene - Google Patents

Face image and gaze counting method and system based on scene Download PDF

Info

Publication number
CN110543813B
CN110543813B CN201910660630.4A CN201910660630A CN110543813B CN 110543813 B CN110543813 B CN 110543813B CN 201910660630 A CN201910660630 A CN 201910660630A CN 110543813 B CN110543813 B CN 110543813B
Authority
CN
China
Prior art keywords
face
specified object
image
person
scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910660630.4A
Other languages
Chinese (zh)
Other versions
CN110543813A (en
Inventor
杨志明
其他发明人请求不公开姓名
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ideepwise Artificial Intelligence Robot Technology Beijing Co ltd
Original Assignee
Ideepwise Artificial Intelligence Robot Technology Beijing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ideepwise Artificial Intelligence Robot Technology Beijing Co ltd filed Critical Ideepwise Artificial Intelligence Robot Technology Beijing Co ltd
Priority to CN201910660630.4A priority Critical patent/CN110543813B/en
Publication of CN110543813A publication Critical patent/CN110543813A/en
Application granted granted Critical
Publication of CN110543813B publication Critical patent/CN110543813B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • G06V40/166Detection; Localisation; Normalisation using acquisition arrangements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Abstract

The invention provides a face image and gaze counting method and system based on a scene, wherein the face image and gaze counting method based on the scene comprises the following steps: acquiring an image of a specified object shot by a camera; preprocessing the image to obtain data adaptive to the input of a pre-established deep learning model; and inputting the data into a pre-trained deep learning model, and simultaneously outputting the face position, the positions of a plurality of face key points and the social attribute classification result of the person. The scene-based gaze technology method can obtain gaze counts based on a deep learning model, so that the interest degree of people in a specified object is obtained. The method can effectively judge the interest degree of each person for the object and the social attributes of the person, such as age, gender and the like, for the crowd within a certain range from the object in the actual scene, and combines the user image and the interest degree in the actual scene to provide an objective and effective data basis for subsequent information recommendation.

Description

Face image and gaze counting method and system based on scene
Technical Field
The invention relates to the field of image processing, in particular to a face image and gaze counting method and system based on a scene.
Background
The customer image refers to the appearance of a customer, namely a customer model, which is characterized by various labels according to collected customer information, including basic social attributes, living habits, consumption levels and the like of the customer. The method has wide application in various fields in real life, such as accurate marketing, namely, enterprises can form effective insights for users through analyzing the behavior data, and can predict future behaviors and decisions of the users according to the previous behaviors of the users.
Because of the diversity of the real scene, the eye counting method is easily influenced by factors such as illumination, visual angle, distance, face-shielded image motion blur and the like, so that the eye counting problem in the real scene is a very significant research problem; in addition, due to the special requirements of the client portrait data for the scene, the client portrait data is obtained more on the basis of the internet and the mobile terminal at present, and the user portrait obtaining in the actual scene is still a problem to be solved.
Disclosure of Invention
The present invention is directed to overcome the above technical drawbacks, and provides a method for depicting a human face in an actual scene, which utilizes deep learning, image understanding, and three-dimensional modeling technologies, and can analyze the gaze retention time of a person on a real object and the portrait of the corresponding person in the actual scene, and for a certain specified object, calculate the gaze retention time of each person on the object in a gaze counting manner, and finally estimate the degree of interest of the person with different customer portraits on the specified object, thereby analyzing the degree of interest of the person on the object and social attributes such as the age and gender of the person.
In order to achieve the above object, the present invention provides a face image method based on a scene, the method comprising:
acquiring an image of a specified object shot by a camera;
preprocessing the image to obtain data adaptive to the input of a pre-established deep learning model;
and inputting the data into a pre-trained deep learning model, and simultaneously outputting the face position, the positions of a plurality of face key points and the social attribute classification result of the person.
As an improvement of the above method, the building and training step of the deep learning model specifically includes:
establishing a deep learning model, wherein the deep learning model comprises four cascaded convolutional neural networks: the first stage of convolutional neural network, the second stage of convolutional neural network, the third stage of convolutional neural network and the fourth stage of convolutional neural network; the fourth-stage convolutional neural network is a multitask neural network;
and training the deep learning model by a deep learning training method based on a large amount of data acquired in a real scene to obtain the trained deep learning model.
As an improvement of the above method, the inputting data into a pre-trained deep learning model, and outputting face positions, positions of a plurality of face key points, and social attribute classification results of people specifically includes:
inputting data into a first-stage convolutional neural network and outputting a plurality of candidate regions;
inputting the candidate regions into a second-stage convolutional neural network, screening non-face candidate regions from the candidate regions, and outputting the remaining candidate regions;
inputting the remaining candidate regions into a third-level convolutional neural network, and screening out non-face candidate regions from the remaining candidate regions again to obtain a final candidate region and a plurality of face key points;
and inputting the final candidate region and the plurality of face key points into a fourth-level convolutional neural network, determining the face region, and simultaneously outputting face positions, the positions of the plurality of face key points and social attribute classification results of people.
As an improvement of the above method, the face key points include: two eyes, nose, two mouth corner edge points, and cheekbones; the number of the face key points is not less than 4.
As an improvement of the above method, the social attribute is age and gender.
The invention provides a face image system based on a scene, which comprises:
the image preprocessing module is used for preprocessing the image acquired by the camera to obtain data matched with the input of the pre-established deep learning model;
the candidate frame extraction module is used for inputting data into a first-stage convolutional neural network to generate a plurality of candidate areas;
the face preliminary screening module is used for inputting the candidate areas into a second-level convolutional neural network, screening most of the non-face candidate areas from the candidate areas and outputting the rest candidate areas;
the face fine screening module is used for inputting the remaining candidate regions into a third-level convolutional neural network, and screening out non-face candidate regions from the remaining candidate regions again to obtain a final candidate region and a plurality of face key points; and
and the classification module is used for inputting the final candidate region and the plurality of face key points into a fourth-level convolutional neural network, determining a face region, and simultaneously outputting face positions, the positions of the plurality of face key points and a classification result of the face region.
The invention provides a scene-based gaze counting method, which is realized based on the scene-based face image drawing method and comprises the following steps:
acquiring an image of a specified object shot by a camera;
judging whether the eye attention point of the person is on a specified object from the image based on the deep learning model;
and counting the stay time of the gaze on the specified object, thereby realizing gaze counting and obtaining the degree of interest of the person on the specified object.
As an improvement of the above method, when the camera is a color camera, the determining, from the image based on the deep learning model, whether the gaze point of interest of the person is on a specified object specifically includes:
step 101) obtaining a face position in an image and a plurality of coordinates of a face key point in the face position through a deep learning model;
step 102) establishing a face plane coordinate system by taking the centers of two eyes of a person as coordinate origin points, taking the right eye direction as the positive direction of an x axis and the direction of a nose as the negative direction of a y axis, and then obtaining the average position relation of key points of a face in the face plane coordinate system in reality through data statistics of a real scene;
step 103) calibrating the color camera through an off-line calibration tool kit to obtain a camera internal reference matrix K;
step 104) calculating projection mapping from a plane of the face in the world coordinate system to a plane of the face in the camera coordinate system by using the average position relation of the face in the real world coordinate system and a plurality of coordinates corresponding to key points of the face in the camera coordinate system to obtain a homography matrix H;
step 105) acquiring a rotation matrix R and a translation matrix T of a face plane relative to a plane where a specified object is located in reality through a homography matrix between planes and a camera internal reference matrix K to obtain a face orientation, and setting a normal direction passing through the centers of two eyes as a human eye watching direction;
step 106) calculating the intersection point of the gazing direction of the human eyes and the specified object, and determining whether the human eyes see the specified object by judging whether the intersection point is in the specified object.
As an improvement of the above method, when the camera is a depth camera, the determining whether the eye attention point of the person is on the specified object from the image specifically includes:
step 201) obtaining the position of a human face in an image and the position information of the central points of the human eye and the mouth through a deep learning model;
step 202) acquiring the position of a corresponding point in a real scene through information conversion, namely the relative spatial position of each human eye, each central point of the mouth and each depth camera; judging whether the central point is set as an original initialization value 0, and turning to step 203); otherwise, go to step 204);
step 203) supplementing by using surrounding neighborhood pixel points; if the number of the face is still 0 after the supplement, the depth camera is judged not to search the face position;
step 204) calculating a face plane and normal orientation according to the space coordinates of the center points of the eyes and the mouth of the person, and then setting the normal orientation passing through the centers of the eyes as the face orientation;
step 205) calculates the intersection point of the face orientation and the specified object, and determines whether the human eyes are looking at the specified object by judging whether the intersection point is in the specified object.
As an improvement of the above method, the counting the residence time of the eyes on the designated object includes: and counting the number of video frames aiming at staying at the specified object, and dividing the number of the video frames by the frame rate to obtain the staying time on the specified object.
As an improvement of the above method, the method further comprises: and associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person, and recommending related information in the region of interest of the person.
The invention provides a scene-based gaze counting system, comprising:
the image acquisition module is used for acquiring an image of a specified object shot by the camera;
the judging module is used for judging whether the attention point of the person is on the specified object or not from the image;
and the gaze counting module is used for counting the stay time of gaze on the specified object, thereby realizing gaze counting and obtaining the degree of interest of people on the specified object.
As an improvement of the above system, the system further comprises: and the association module is used for associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person and recommending related information in the region of interest of the person.
The invention has the beneficial effects that:
1. the method can effectively judge the interest degree of each person for the object and the social attributes of the person, such as age, gender and the like, for the crowd within a certain range from the object in the actual scene, and fills the gap of combining the client image and the interest degree in the actual scene at present;
2. the method can count the eye sight residence time of the person in the camera to the object and the related attributes of the person under the condition that the camera is arranged beside the specified object in the actual scene, so that the interest degree of the person with different attributes to the object is analyzed, and an objective and effective data supply basis is provided for subsequent information recommendation.
Drawings
FIG. 1 is a flow chart of a scene-based face imaging method of the present invention;
FIG. 2 is a flow chart of the present invention for determining whether a person is looking at a specified object based on a color camera;
FIG. 3 is a flow chart of the present invention for determining whether a person is looking at a specified object based on a depth camera.
Detailed Description
The following detailed description of the preferred embodiments of the present invention, taken in conjunction with the accompanying drawings, will make the advantages and features of the invention easier to understand by those skilled in the art, and thus will clearly and clearly define the scope of the invention.
The gaze counting and face portrait analysis technology is characterized in that related technologies such as deep learning, image understanding and three-dimensional modeling are utilized, a camera is arranged beside or behind a specified object, whether the gaze of a person in the camera is on the object is judged, when the camera shoots a plurality of persons, each person is judged respectively, and when the gaze is judged to be on the object, the stay time of the gaze on the object is counted to serve as a judgment standard for the interest degree of the person on the object. Meanwhile, the portrait attributes of the people are analyzed by utilizing a deep learning technology and are combined with eye count, so that the social attributes such as age, gender and the like of the people interested in the object are counted.
Example 1
Training is carried out on the basis of a large amount of data acquired in a real scene through a deep learning method, and the face position detection, the face key point detection and the gender and age classification are integrated into a multi-task deep learning model for prediction in a mode of combining the face detection and the key point detection. Finally, the position information of each face frame in the image and the classification results of the coordinate information, the gender and the age of a plurality of face key points can be obtained by inputting an actual scene image containing the face; the face key points comprise: two eyes, a nose, two mouth corner edge points, cheekbones and the like; the number of the face key points is more than or equal to 4.
As shown in fig. 1, embodiment 1 of the present invention provides a method for depicting a human face based on a scene, where the method includes:
acquiring an image of a specified object shot by a camera, and preprocessing the image to obtain a data format matched with the input of a deep learning model;
the deep learning model is 4 cascaded convolutional neural networks, can detect and position the face from coarse to fine, and simultaneously obtains the positions of a plurality of key points of the face and classifies the gender and the age of the person. The algorithm consists of four phases:
step S1) is to use shallow CNN (convolutional neural network) to generate a large number of candidate regions quickly;
step S2) is to screen out a large number of non-face candidate areas through complex CNN;
step S3) is to use the more complicated CNN to screen out a small number of non-face candidate regions again and output a plurality of key points at the same time;
step S4) is to determine the final face region with the more powerful CNN and output the location of multiple facial keypoints and gender, age classification for each face region simultaneously.
Each convolutional neural network may be one convolutional neural network or a cascaded convolutional neural network.
Example 2
Embodiment 2 of the present invention provides a face image system based on a scene, the system including:
the image preprocessing module is used for preprocessing the image acquired by the camera to obtain data matched with the input of the pre-established deep learning model;
the candidate frame extraction module is used for inputting data into a first-stage convolutional neural network to generate a plurality of candidate areas;
the face preliminary screening module is used for inputting the candidate areas into a second-level convolutional neural network, screening most of the non-face candidate areas from the candidate areas and outputting the rest candidate areas;
the face fine screening module is used for inputting the remaining candidate regions into a third-level convolutional neural network, and screening out non-face candidate regions from the remaining candidate regions again to obtain a final candidate region and a plurality of face key points; and
and the classification module is used for inputting the final candidate region and the plurality of face key points into a fourth-level convolutional neural network, determining a face region, and simultaneously outputting face positions, the positions of the plurality of face key points and a classification result of the face region.
Example 3
Embodiment 3 of the present invention provides a scene-based gaze counting method, in which a camera is placed beside a designated object, whether a gaze focus of a person is on the object is determined by using a visual image correlation technique, and a time spent on the object is counted to realize gaze counting. For the eye counting work, the following steps are mainly carried out:
step 1) acquiring image data through a camera;
the camera is a color camera or a depth camera;
step 2) judging whether a person looks at a specified object or not through image data;
as shown in figure 2 of the drawings, in which,
the method comprises the steps of judging whether a person looks at the billboard or not by taking the billboard in the elevator as an example according to the image data of the color camera
Step 2-1) simultaneously acquiring face frame coordinates in the image and position coordinates of a plurality of face key points (eyes, noses and other parts) in the image through a deep learning model;
and 2-2) taking the centers of two eyes of a person as coordinate origin points, setting the direction of the right eye as the positive direction of an x axis and the direction of a nose as the negative direction of a y axis, establishing a face plane coordinate system, and counting and acquiring the average position relation of a plurality of face key points (eyes, noses and the like) in the face plane coordinate system in reality through a large amount of real scene data.
Step 2-3) calibrating the color camera through an off-line calibration tool kit to obtain an internal reference matrix K of the camera;
step 2-4) calculating projection mapping from a plane of the face in the world coordinate system to a plane of the face in the camera coordinate system by using the average position relation of the face in the real world coordinate system and a plurality of corresponding face key point coordinates (eyes, noses and the like) in the camera coordinate system (namely in the image), and obtaining a homography matrix H;
step 2-5) acquiring a rotation matrix R and a translation matrix T of a face plane relative to a plane where a specified object is located in reality through a homography matrix between planes and a camera internal reference matrix to obtain a face orientation, and setting a normal direction passing through the centers of two eyes as a human eye watching direction;
and 2-6) calculating the intersection point of the gazing direction of the human eyes and the specified object, and determining whether the human eyes see the specified object by judging whether the intersection point is in the specified object.
As shown in fig. 3, the step of determining whether a person looks at a billboard specifically includes, for the image data of the depth camera, taking the billboard in the elevator as an example:
step 201) searching a human face frame in a color picture and position information of central points of human eyes and a mouth through a deep learning model;
step 202) acquiring the position of a corresponding point in an actual scene through information conversion, namely the relative spatial position of the eyes, the center point of the mouth and the depth camera of each face; judging whether the central point is set as an original initialization value 0, and turning to step 203); otherwise, go to step 204);
step 203) supplementing by using surrounding neighborhood pixel points; and if the number is still 0 after the supplement, determining that the face position is not searched by the depth camera.
Step 204) carrying out face orientation modeling through the obtained information of the eyes and the mouth, namely calculating a face plane and normal orientation according to the space coordinates of the eyes and the mouth, and then setting the normal direction passing through the center of the eyes as the face orientation;
step 205) calculating the intersection point of the face orientation and the plane of the billboard, and determining whether a person is looking at the billboard by judging whether the intersection point is in the plane of the billboard.
Particularly, when a plurality of faces exist in the picture, the same processing is performed on the faces with different ID information.
Step 3) counting the stay time of each person in the appointed object, thereby determining the interest degree of the person in the object;
for each person, counting the number of video frames of which the eyes stay at the specified object, and dividing the number by the frame rate to obtain the stay time on the specified object as the basis of the interest degree of the person on the object.
And 4) associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person, and recommending related information in the region of interest of the person.
Example 4
Embodiment 4 of the present invention provides a scene-based gaze counting system, including:
the image acquisition module is used for acquiring an image of a specified object shot by the camera;
the judging module is used for judging whether the attention point of the person is on the specified object or not from the image;
the gaze counting module is used for counting the stay time of gaze on the specified object, so that gaze counting is realized, and the degree of interest of people on the specified object is obtained;
and the association module is used for associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person and recommending related information in the region of interest of the person.
Finally, it should be noted that the above embodiments are only used for illustrating the technical solutions of the present invention and are not limited. Although the present invention has been described in detail with reference to the embodiments, it will be understood by those skilled in the art that various changes may be made and equivalents may be substituted without departing from the spirit and scope of the invention as defined in the appended claims.

Claims (9)

1. A scene-based gaze counting method, which is realized according to a scene-based face image method,
the face image method based on the scene comprises the following steps:
acquiring an image of a specified object shot by a camera;
preprocessing the image to obtain data adaptive to the input of a pre-established deep learning model;
inputting data into a pre-trained deep learning model, and simultaneously outputting face positions, positions of a plurality of face key points and social attribute classification results of people;
the eye count method comprises the following steps:
acquiring an image of a specified object shot by a camera;
judging whether the eye attention point of the person is on a specified object from the image based on the deep learning model;
counting the stay time of the gaze on the specified object, thereby realizing gaze counting and obtaining the degree of interest of the person on the specified object;
when the camera is a depth camera, the method for judging whether the eye attention point of the person is on the specified object from the image specifically comprises the following steps:
step 201) obtaining the position of a human face in an image and the position information of the central points of the human eye and the mouth through a deep learning model;
step 202) acquiring the position of a corresponding point in a real scene through information conversion, namely the relative spatial position of the central point of the human eye and mouth of each human face and a depth camera; judging whether the central point is set as an original initialization value: 0, go to step 203); otherwise, go to step 204);
step 203) supplementing by using surrounding neighborhood pixel points; if the number of the face is still 0 after the supplement, the depth camera is judged not to search the face position;
step 204) calculating a face plane and normal orientation according to the space coordinates of the center points of the eyes and the mouth of the person, and then setting the normal orientation passing through the centers of the eyes as the face orientation;
step 205) calculates the intersection point of the face orientation and the specified object, and determines whether the human eyes are looking at the specified object by judging whether the intersection point is in the specified object.
2. The scene-based gaze counting method of claim 1, wherein the steps of building and training the deep learning model specifically comprise:
establishing a deep learning model, wherein the deep learning model comprises four cascaded convolutional neural networks: the first stage of convolutional neural network, the second stage of convolutional neural network, the third stage of convolutional neural network and the fourth stage of convolutional neural network; the fourth-stage convolutional neural network is a multitask neural network;
and training the deep learning model by a deep learning training method based on a large amount of data acquired in a real scene to obtain the trained deep learning model.
3. The scene-based gaze counting method of claim 2, wherein the inputting of data into a pre-trained deep learning model and the outputting of face positions, positions of a plurality of face key points and social attribute classification results comprises:
inputting data into a first-stage convolutional neural network and outputting a plurality of candidate regions;
inputting the candidate regions into a second-stage convolutional neural network, screening non-face candidate regions from the candidate regions, and outputting the remaining candidate regions;
inputting the remaining candidate regions into a third-level convolutional neural network, and screening out non-face candidate regions from the remaining candidate regions again to obtain a final candidate region and a plurality of face key points;
and inputting the final candidate region and the plurality of face key points into a fourth-level convolutional neural network, determining the face region, and simultaneously outputting face positions, the positions of the plurality of face key points and social attribute classification results of people.
4. The scene-based gaze counting method of one of claims 1 to 3, wherein the face key points comprise: two eyes, nose, two mouth corner edge points, and cheekbones; the number of the face key points is not less than 4.
5. The scene-based gaze counting method of claim 1, wherein the social attributes are age and gender.
6. The scene-based gaze counting method according to claim 1, wherein the statistical gaze residence time on the designated object is as follows: and counting the number of video frames aiming at staying at the specified object, and dividing the number of the video frames by the frame rate to obtain the staying time on the specified object.
7. The scene-based gaze counting method of claim 1, further comprising: and associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person, and recommending related information in the region of interest of the person.
8. A scene-based eye count system, the system comprising:
the image acquisition module is used for acquiring an image of a specified object shot by the camera;
the judging module is used for judging whether the attention point of the person is on the specified object or not from the image;
the gaze counting module is used for counting the stay time of gaze on the specified object, so that gaze counting is realized, and the degree of interest of people on the specified object is obtained;
when the camera is a depth camera, the method for judging whether the eye attention point of the person is on the specified object from the image specifically comprises the following steps:
step 201) obtaining the position of a human face in an image and the position information of the central points of the human eye and the mouth through a deep learning model;
step 202) acquiring the position of a corresponding point in a real scene through information conversion, namely the relative spatial position of each human eye, each central point of the mouth and each depth camera; judging whether the central point is set as an original initialization value 0, and turning to step 203); otherwise, go to step 204);
step 203) supplementing by using surrounding neighborhood pixel points; if the number of the face is still 0 after the supplement, the depth camera is judged not to search the face position;
step 204) calculating a face plane and normal orientation according to the space coordinates of the center points of the eyes and the mouth of the person, and then setting the normal orientation passing through the centers of the eyes as the face orientation;
step 205) calculating the intersection point of the face orientation and the specified object, and determining whether the human eyes see the specified object by judging whether the intersection point is in the specified object;
the face image method based on the scene comprises the following steps:
acquiring an image of a specified object shot by a camera;
preprocessing the image to obtain data adaptive to the input of a pre-established deep learning model;
and inputting the data into a pre-trained deep learning model, and simultaneously outputting the face position, the positions of a plurality of face key points and the social attribute classification result of the person.
9. The scene-based gaze counting system of claim 8, further comprising: and the association module is used for associating the degree of interest of the person in the specified object with the classification result of the social attribute of the person and recommending related information in the region of interest of the person.
CN201910660630.4A 2019-07-22 2019-07-22 Face image and gaze counting method and system based on scene Active CN110543813B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910660630.4A CN110543813B (en) 2019-07-22 2019-07-22 Face image and gaze counting method and system based on scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910660630.4A CN110543813B (en) 2019-07-22 2019-07-22 Face image and gaze counting method and system based on scene

Publications (2)

Publication Number Publication Date
CN110543813A CN110543813A (en) 2019-12-06
CN110543813B true CN110543813B (en) 2022-03-15

Family

ID=68710273

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910660630.4A Active CN110543813B (en) 2019-07-22 2019-07-22 Face image and gaze counting method and system based on scene

Country Status (1)

Country Link
CN (1) CN110543813B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112488003A (en) * 2020-12-03 2021-03-12 深圳市捷顺科技实业股份有限公司 Face detection method, model creation method, device, equipment and medium
CN112540676B (en) * 2020-12-15 2021-06-18 广州舒勇五金制品有限公司 Projection system-based variable information display device

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105094337A (en) * 2015-08-19 2015-11-25 华南理工大学 Three-dimensional gaze estimation method based on irises and pupils
CN105912990A (en) * 2016-04-05 2016-08-31 深圳先进技术研究院 Face detection method and face detection device
CN108805910A (en) * 2018-06-01 2018-11-13 海信集团有限公司 More mesh Train-borne recorders, object detection method, intelligent driving system and automobile
CN109492514A (en) * 2018-08-28 2019-03-19 初速度(苏州)科技有限公司 A kind of method and system in one camera acquisition human eye sight direction
CN109874054A (en) * 2019-02-14 2019-06-11 深兰科技(上海)有限公司 A kind of advertisement recommended method and device

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101383055B (en) * 2008-09-18 2010-09-29 北京中星微电子有限公司 Three-dimensional human face constructing method and system
CN107193383B (en) * 2017-06-13 2020-04-07 华南师范大学 Secondary sight tracking method based on face orientation constraint
CN109753857A (en) * 2017-11-07 2019-05-14 北京虹图吉安科技有限公司 A kind of three-dimensional face identification apparatus and system based on photometric stereo visual imaging
CN109446892B (en) * 2018-09-14 2023-03-24 杭州宇泛智能科技有限公司 Human eye attention positioning method and system based on deep neural network
CN109902630B (en) * 2019-03-01 2022-12-13 上海像我信息科技有限公司 Attention judging method, device, system, equipment and storage medium

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105094337A (en) * 2015-08-19 2015-11-25 华南理工大学 Three-dimensional gaze estimation method based on irises and pupils
CN105912990A (en) * 2016-04-05 2016-08-31 深圳先进技术研究院 Face detection method and face detection device
CN108805910A (en) * 2018-06-01 2018-11-13 海信集团有限公司 More mesh Train-borne recorders, object detection method, intelligent driving system and automobile
CN109492514A (en) * 2018-08-28 2019-03-19 初速度(苏州)科技有限公司 A kind of method and system in one camera acquisition human eye sight direction
CN109874054A (en) * 2019-02-14 2019-06-11 深兰科技(上海)有限公司 A kind of advertisement recommended method and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于点特征的室内移动机器人主动视觉定位方法研究;田荣堂;《中国优秀博硕士学位论文全文数据库(硕士) 信息科技辑》;20170515(第05期);正文全文 *

Also Published As

Publication number Publication date
CN110543813A (en) 2019-12-06

Similar Documents

Publication Publication Date Title
Lebreton et al. GBVS360, BMS360, ProSal: Extending existing saliency prediction models from 2D to omnidirectional images
CN112418095B (en) Facial expression recognition method and system combined with attention mechanism
CN107545302B (en) Eye direction calculation method for combination of left eye image and right eye image of human eye
CN105472434B (en) It is implanted into method and system of the content into video display
CN110807364B (en) Modeling and capturing method and system for three-dimensional face and eyeball motion
EP3814985A1 (en) Video background subtraction using depth
WO2022156640A1 (en) Gaze correction method and apparatus for image, electronic device, computer-readable storage medium, and computer program product
CN109948447B (en) Character network relation discovery and evolution presentation method based on video image recognition
CN110210276A (en) A kind of motion track acquisition methods and its equipment, storage medium, terminal
CN105005777A (en) Face-based audio and video recommendation method and face-based audio and video recommendation system
CN106599800A (en) Face micro-expression recognition method based on deep learning
WO2014187223A1 (en) Method and apparatus for identifying facial features
CN110741377A (en) Face image processing method and device, storage medium and electronic equipment
CN111008971B (en) Aesthetic quality evaluation method of group photo image and real-time shooting guidance system
CN111507592A (en) Evaluation method for active modification behaviors of prisoners
CN113963032A (en) Twin network structure target tracking method fusing target re-identification
CN110163567A (en) Classroom roll calling system based on multitask concatenated convolutional neural network
CN110543813B (en) Face image and gaze counting method and system based on scene
CN113747640A (en) Intelligent central control method and system for digital exhibition hall lamplight
Kim A personal identity annotation overlay system using a wearable computer for augmented reality
JP2021026744A (en) Information processing device, image recognition method, and learning model generation method
RU2768797C1 (en) Method and system for determining synthetically modified face images on video
CN115862128A (en) Human body skeleton-based customer abnormal behavior identification method
Bonetto et al. Image processing issues in a social assistive system for the blind
Ando et al. A low complexity algorithm for eye detection and tracking in energy-constrained applications

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant