WO2019205633A1 - 眼睛状态的检测方法、电子设备、检测装置和计算机可读存储介质 - Google Patents
眼睛状态的检测方法、电子设备、检测装置和计算机可读存储介质 Download PDFInfo
- Publication number
- WO2019205633A1 WO2019205633A1 PCT/CN2018/118374 CN2018118374W WO2019205633A1 WO 2019205633 A1 WO2019205633 A1 WO 2019205633A1 CN 2018118374 W CN2018118374 W CN 2018118374W WO 2019205633 A1 WO2019205633 A1 WO 2019205633A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- eye
- feature points
- eye feature
- state
- target image
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title abstract description 14
- 238000012545 processing Methods 0.000 claims abstract description 21
- 238000010606 normalization Methods 0.000 claims abstract description 18
- 238000000034 method Methods 0.000 claims description 41
- 238000004590 computer program Methods 0.000 claims description 13
- 238000010586 diagram Methods 0.000 description 11
- 230000008569 process Effects 0.000 description 9
- 238000012549 training Methods 0.000 description 9
- 238000004891 communication Methods 0.000 description 4
- 210000000744 eyelid Anatomy 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 230000004397 blinking Effects 0.000 description 2
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/77—Determining position or orientation of objects or cameras using statistical methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/197—Matching; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30196—Human being; Person
- G06T2207/30201—Face
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
Definitions
- Embodiments of the present disclosure relate to a method of detecting an eye state, an electronic device, a detecting device, and a computer readable storage medium.
- the eye is the most important feature of the human face. It plays an extremely important role in computer vision research and application.
- the detection of eye state has always been the focus of researchers. On the basis of face recognition, the detection of eye state helps various smart devices to recognize the state of the human eye, and has broad application prospects in the field of fatigue detection and visual interaction, for example, driver's fatigue detection and invalid photos. filter.
- a method of detecting an eye state comprising: acquiring a target image; and locating a plurality of eye feature points in the target image to determine the plurality of eye features a position coordinate of the point; normalizing the position coordinates of the plurality of eye feature points to obtain processed position feature data; and determining a state of the eye in the target image based on the position feature data.
- the eye feature points include a left eye feature point and a right eye feature point; normalizing the position coordinates of the plurality of eye feature points, and obtaining the processed position feature data includes: determining the left Euclidean distance between the mean position coordinate of the eye feature point and the position coordinate mean value of the right eye feature point; normalizing the position coordinate of the eye feature point by using the Euclidean distance as a standard scale Processed feature data after processing.
- determining the state of the eye in the target image based on the location feature data includes: classifying the location feature data; and determining a state of the eye in the target image based on the classification result.
- determining the Euclidean distance between the position coordinate mean of the left eye feature point and the position coordinate mean of the right eye feature point includes:
- i is the i-th X-coordinate of the horizontal axis of the ocular feature points
- i is the Y coordinate of the longitudinal axis of the i-th eye feature points
- i is an integer in the range 1 to N, the first to 0.5N
- the eye feature points are the left eye feature points, the 0.5N+1th to the Nth eye feature points are the right eye feature points; the N is an even number; determining between E1 and Er based on El and Er European distance Ed.
- the position coordinates of the plurality of eye feature points are normalized by using the Euclidean distance as a standard scale, and the obtained position feature data includes:
- the N 12.
- the classifying the location feature data includes: classifying the location feature data by using a classifier.
- the classifying the location feature data further includes: training the classifier with the sample image to obtain a classifier parameter for detecting an eye state.
- training the classifier with a sample image to obtain a classifier parameter for detecting an eye state includes: obtaining a positive and negative sample image from a picture library; wherein an eye state in the positive sample image is a blink state, negative The eye state in the sample image is a closed eye state; the eye feature points are positioned on the positive and negative sample images, and the position coordinates of the plurality of eye feature points in the positive and negative sample images are determined; the positions of the plurality of eye feature points are determined The coordinates are normalized to obtain the processed position feature data; the positional feature data is used to train the classifier to obtain a classifier parameter for detecting the eye state.
- the positioning of the plurality of eye feature points in the target image includes: detecting whether the target image includes a human face; and when detecting that the target image includes a human face, Multiple eye feature points in the image are positioned.
- an electronic device comprising: at least one processor, at least one memory, and computer program instructions stored in the memory, when the computer program instructions are processed Performing: acquiring a target image; positioning a plurality of eye feature points in the target image to determine position coordinates of the plurality of eye feature points; performing position coordinates on the plurality of eye feature points
- the normalized processing obtains the processed position feature data; and determines the state of the eye in the target image based on the position feature data.
- the eye feature points include a left eye feature point and a right eye feature point; normalizing the position coordinates of the plurality of eye feature points, and obtaining the processed position feature data includes: determining the left Euclidean distance between the mean position coordinate of the eye feature point and the position coordinate mean of the right eye feature point; normalizing the position coordinates of the plurality of eye feature points by using the Euclidean distance as a standard scale , the processed position feature data is obtained.
- determining the state of the eye in the target image based on the location feature data includes: classifying the location feature data; and determining a state of the eye in the target image based on the classification result.
- determining the Euclidean distance between the position coordinate mean of the left eye feature point and the position coordinate mean of the right eye feature point includes:
- i is the i-th X-coordinate of the horizontal axis of the ocular feature points
- i is the Y coordinate of the longitudinal axis of the i-th eye feature points
- i is an integer in the range 1 to N, the first to 0.5N
- the eye feature points are the left eye feature points, and the 0.5N+1th to Nth eye feature points are the right eye feature points; the N is an even number;
- the Euclidean distance Ed between E1 and Er is determined based on El and Er.
- the position coordinates of the plurality of eye feature points are normalized by using the Euclidean distance as a standard scale, and the obtained position feature data includes:
- the classifying the location feature data includes: classifying the location feature data by using a classifier.
- the classifying the location feature data further includes: training the classifier with the sample image to obtain a classifier parameter for detecting an eye state.
- training the classifier with a sample image to obtain a classifier parameter for detecting an eye state includes: obtaining a positive and negative sample image from a picture library; wherein an eye state in the positive sample image is a blink state, negative The eye state in the sample image is a closed eye state; the eye feature points are positioned on the positive and negative sample images, and the position coordinates of the plurality of eye feature points in the positive and negative sample images are determined; the positions of the plurality of eye feature points are determined The coordinates are normalized to obtain the processed position feature data; the positional feature data is used to train the classifier to obtain a classifier parameter for detecting the eye state.
- an apparatus for detecting an eye state comprising: an acquisition unit configured to acquire a target image; and a positioning unit configured to target a plurality of eye feature points in the target image Performing positioning to determine position coordinates of the plurality of eye feature points; a normalization processing unit configured to normalize position coordinates of the plurality of eye feature points to obtain processed position features a data determining unit configured to determine a state of an eye in the target image based on the location feature data.
- a computer readable storage medium having stored thereon computer program instructions that, when executed by a processor, implements the methods described in the preceding embodiments.
- FIG. 1 is a schematic flowchart diagram of a method for detecting an eye state according to an embodiment of the present disclosure
- FIG. 2 is a schematic diagram of a training process of a classifier according to an embodiment of the present disclosure
- FIG. 3 is a schematic flowchart of a method for detecting an eye state according to an embodiment of the present disclosure
- FIG. 4 is a structural block diagram of an apparatus for detecting an eye state according to an embodiment of the present disclosure
- FIG. 5 is a schematic structural diagram of a computer device suitable for implementing an embodiment of the present disclosure according to an embodiment of the present disclosure.
- Embodiments of the present disclosure provide a method of detecting an eye state, wherein an eye state may include a blink state and a closed eye state.
- an eye state may also include a semi- ⁇ semi-closed state, a fatigue state, a blink state, and the like.
- the present disclosure describes the eye state detection only by taking the blink state and the closed eye state as an example.
- a schematic flowchart of a method for detecting an eye state according to an embodiment of the present disclosure includes the following steps 101-104 .
- the order of description of the above steps is merely an example of the embodiments of the present disclosure, and is not the only order, and other possible execution sequences are also conceivable by those skilled in the art in light of the present disclosure.
- Step 101 Acquire a target image.
- the target image may be acquired by a camera or may be received from other devices other than the camera.
- Step 102 Positioning a plurality of eye feature points in the target image to determine position coordinates of the plurality of eye feature points.
- a face detection technique refers to a method of searching for a given image using a certain strategy to determine whether a face is contained therein.
- a linear subspace method, a neural network method, or the like is used to detect whether a target image is included in a target image.
- the face region in the target image is determined, and then the plurality of eye feature points in the face region are positioned, thereby determining the position coordinates of the plurality of eye feature points.
- eye feature points can be located by an eye feature learning machine. For example, first, positive and negative samples of a plurality of eye feature points are acquired. For example, an image recognition algorithm is used to detect a plurality of images that may include eye features, and a positive sample and a sub-sample of a plurality of eye feature points are obtained.
- a positive sample is a sample that includes eye features, such as a sample that is similar to an eye feature but is not an eye feature.
- the face region in the target image or the target image is input into the trained learning machine, and the trained learning machine can automatically locate the eye feature points in the input face region and determine the positioning position at the target. The coordinate position on the image.
- the eye feature points include left eye feature points and right eye feature points, which may be, but are not limited to, edge points of the eye corner points and upper and lower eyelids.
- the number of feature points of the left and right eyes can be set according to the needs, based on the symmetrical characteristics of the human eye in the face, so the number of feature points of the left and right eyes can be the same.
- the algorithm used in the positioning of the left and right eye feature points in the embodiment of the present disclosure determines the number of left and right eye feature points as 12, that is, 6 left eye feature points and 6 right eye feature points, for example, respectively: left eye corner Point, right eye corner point, two edge points of the upper eyelid and two edge points of the lower eyelid.
- the number of left and right eye feature points may be other numbers depending on the detection accuracy. However, too many numbers may increase the amount of calculation, and the number is too small, which may lead to the problem of low positioning accuracy. Therefore, 12 feature points are selected in this example to balance the accuracy and the calculation amount.
- the position coordinates of the eye feature points may be, but are not limited to, located in the XY axis coordinate system.
- the coordinate system may take the upper left corner of the target image as the origin, the horizontal direction as the horizontal axis, that is, the X axis, and the vertical direction as the vertical axis, that is, the Y axis.
- Step 103 normalize the position coordinates of the plurality of eye feature points to obtain the processed position feature data. For example, the Euclidean distance between the position coordinate mean value of the left eye feature point and the position coordinate mean value of the right eye feature point is used as a standard scale, and the position coordinates of the plurality of eye feature points are normalized.
- the position coordinate mean of all eye feature points may be determined first.
- i is the i-th X-coordinate of the horizontal axis of the ocular feature points
- i is the Y coordinate of the longitudinal axis of the i-th eye feature points
- i is an integer in the range 1 to N, the first to 0.5N
- the eye feature points are left eye feature points, and the 0.5N+1 to Nth eye feature points are right eye feature points; N is an even number.
- calculate the Euclidean distance Ed of El and Er can be calculated, for example, by the following formula:
- ⁇ is the Euclidean distance between the point (x 2 , y 2 ) and the point (x 1 , y 1 ). Then the Euclidean distance between El and Er is:
- the new position coordinates of the obtained eye feature points are the position feature data obtained after the normalization process.
- Step 104 determining the state of the eye in the target image based on the location feature data. For example, the location feature data is classified to determine the state of the eye in the target image.
- the obtained location feature data can be classified by using a classifier.
- the embodiment of the present disclosure may further include:
- the classifier is trained using the sample image to obtain a classifier parameter for detecting the state of the eye.
- the training process of the classifier may be as shown in FIG. 2:
- the positive and negative sample images are obtained from the blinking closed eye sample image library; wherein the eye state in the positive sample image is the blink state, and the eye state in the negative sample image is the closed eye state.
- the position coordinates of the plurality of eye feature points are normalized to obtain the processed position feature data.
- the classifier is trained to obtain a classifier parameter for detecting the state of the eye.
- the classifier can classify the blinking state or the closed eye state. For example, if the position coordinates of the eye feature points are input into the classifier, the classifier determines whether the feature at the position coordinates is a blink state or a closed eye state.
- FIG. 3 is a schematic flowchart diagram of a method for detecting an eye state according to an embodiment of the present disclosure.
- the method for detecting the state of the eye specifically includes the following steps 301-305.
- the order of description of the above steps is only an example of the embodiments of the present disclosure, and is not the only order, and other possible execution sequences are also conceivable by those skilled in the art according to the content of the present disclosure.
- Step 301 Acquire a target image.
- Step 302 Detect whether a face is included in the target image.
- step 303 is performed.
- the process ends, and the process returns to step 301.
- Step 303 Positioning 12 eye feature points in the target image to determine position coordinates of 12 eye feature points.
- Step 304 normalize the position coordinates of the 12 eye feature points by using the Euclidean distance between the position coordinate mean value of the left eye feature point and the position coordinate mean value of the right eye feature point as a standard scale, and obtain the processed Location feature data.
- X i is the horizontal axis coordinate of the i-th eye feature point
- Y i is the vertical axis coordinate of the i-th eye feature point
- Step 305 classifying the obtained position feature data by using a classifier to determine the state of the eye in the target image.
- the detection method of the eye state determines the position coordinates of the plurality of eye feature points in the target image by acquiring the target image, and the position coordinate mean value of the left eye feature point and the position coordinate of the right eye feature point.
- the Euclidean distance between the mean values is used as a standard scale, and the position coordinates of the plurality of eye feature points are normalized to obtain the processed position feature data, and then the position feature data is classified to determine the state of the eye in the target image.
- the scheme can accurately detect the state of the eye in the target image, and because of the principle of normalization processing, it is not affected by the size and position of the eye region and the face in the target image, and has good robustness.
- FIG. 4 is a structural block diagram of an apparatus for detecting an eye condition according to an embodiment of the present disclosure, the apparatus includes:
- the obtaining unit 41 is configured to acquire a target image.
- the acquisition unit is, for example, a camera, a camera, or the like, and may be a program command that calls a target image.
- the positioning unit 42 is configured to locate a plurality of eye feature points in the target image to determine position coordinates of the plurality of eye feature points.
- eye feature points may include left eye feature points and right eye feature points.
- the normalization processing unit 43 normalizes the position coordinates of the plurality of eye feature points to obtain the processed position feature data.
- the position coordinates of the plurality of eye feature points may be normalized by using the Euclidean distance between the position coordinate mean value of the left eye feature point and the position coordinate mean value of the right eye feature point as a standard scale. deal with.
- the determining unit 44 is configured to determine a state of the eye in the target image based on the location feature data.
- the positioning unit 42, the normalization processing unit 43, and the determining unit 44 may be implemented by software, or may be implemented by hardware or firmware.
- it is implemented by a general purpose processor, a programmable logic circuit, and an integrated circuit.
- the positioning unit 42 is used, for example:
- the target image includes a human face
- a plurality of eye feature points in the target image are positioned.
- the normalization processing unit 43 is configured to:
- i is the i-th X-coordinate of the horizontal axis of the ocular feature points
- i is the Y coordinate of the longitudinal axis of the i-th eye feature points
- i is an integer in the range 1 to N, the first to 0.5N
- the eye feature points are the left eye feature points, and the 0.5N+1th to Nth eye feature points are the right eye feature points; the N is an even number;
- the N 12.
- the determining unit 44 is configured to:
- the location feature data is classified using a classifier.
- the apparatus further includes:
- a classifier training unit 45 is configured to train the classifier with the sample image to obtain a classifier parameter for detecting an eye state.
- embodiments of the present disclosure also provide a computer device suitable for implementing the embodiments of the present disclosure for implementing the method of the foregoing embodiments.
- the computer device includes a memory and a processor, the memory storing computer program instructions, and when the processor processes the program instructions, performing: acquiring a target image; and positioning a plurality of eye feature points in the target image to determine Position coordinates of the plurality of eye feature points; normalizing the position coordinates of the plurality of eye feature points to obtain processed position feature data; determining the target image based on the position feature data The state of the eye.
- the eye feature points include a left eye feature point and a right eye feature point; normalizing the position coordinates of the plurality of eye feature points, and obtaining the processed position feature data includes: determining the left Euclidean distance between the mean position coordinate of the eye feature point and the position coordinate mean value of the right eye feature point; normalizing the position coordinate of the eye feature point by using the Euclidean distance as a standard scale Processed feature data after processing.
- determining the state of the eye in the target image based on the location feature data includes: classifying the location feature data; and determining a state of the eye in the target image based on the classification result.
- determining the Euclidean distance between the position coordinate mean of the left eye feature point and the position coordinate mean of the right eye feature point includes:
- i is the i-th X-coordinate of the horizontal axis of the ocular feature points
- i is the Y coordinate of the longitudinal axis of the i-th eye feature points
- i is an integer in the range 1 to N, the first to 0.5N
- the eye feature points are the left eye feature points, the 0.5N+1th to the Nth eye feature points are the right eye feature points; the N is an even number; determining between E1 and Er based on El and Er European distance Ed.
- the position coordinates of the plurality of eye feature points are normalized by using the Euclidean distance as a standard scale, and the obtained position feature data includes:
- the N 12.
- the classifying the location feature data includes: classifying the location feature data by using a classifier.
- the classifying the location feature data further includes: training the classifier with the sample image to obtain a classifier parameter for detecting an eye state.
- training the classifier with a sample image to obtain a classifier parameter for detecting an eye state includes: obtaining a positive and negative sample image from a picture library; wherein an eye state in the positive sample image is a blink state, negative The eye state in the sample image is a closed eye state; the eye feature points are positioned on the positive and negative sample images, and the position coordinates of the plurality of eye feature points in the positive and negative sample images are determined; the positions of the plurality of eye feature points are determined The coordinates are normalized to obtain the processed position feature data; the positional feature data is used to train the classifier to obtain a classifier parameter for detecting the eye state.
- the positioning of the plurality of eye feature points in the target image includes: detecting whether the target image includes a human face; and when detecting that the target image includes a human face, Multiple eye feature points in the image are positioned.
- FIG. 5 a schematic structural diagram of a computer device suitable for implementing an embodiment of the present disclosure is provided in an embodiment of the present disclosure.
- the computer system includes a central processing unit (CPU) 501, which may be loaded according to a program stored in a read only memory (ROM) 502 or a program loaded from a storage portion 508 into a random access memory (RAM) 503. Perform various appropriate actions and processes.
- ROM read only memory
- RAM random access memory
- various programs and data required for the operation of the system 500 are also stored.
- the CPU 501, the ROM 502, and the RAM 503 are connected to each other through a bus 504.
- An input/output (I/O) interface 505 is also coupled to bus 504.
- the following components are connected to the I/O interface 505: an input portion 506 including a keyboard, a mouse, etc.; an output portion 507 including, for example, a cathode ray tube (CRT), a liquid crystal display (LCD), and the like, and a storage portion 508 including a hard disk or the like. And a communication portion 509 including a network interface card such as a LAN card, a modem, or the like. The communication section 509 performs communication processing via a network such as the Internet.
- Driver 310 is also coupled to I/O interface 505 as needed.
- a removable medium 511 such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory or the like is mounted on the drive 510 as needed so that a computer program read therefrom is installed into the storage portion 508 as needed.
- an embodiment of the present disclosure includes a computer program product comprising a computer program tangibly embodied on a machine readable medium, the computer program comprising program code for performing the methods of FIGS. 1-3.
- the computer program can be downloaded and installed from the network via the communication portion 509, and/or installed from the removable medium 511.
- each block of the flowchart or block diagrams can represent a module, a program segment, or a portion of code that includes one or more logic for implementing the specified.
- Functional executable instructions can also occur in a different order than that illustrated in the drawings. For example, two successively represented blocks may in fact be executed substantially in parallel, and they may sometimes be executed in the reverse order, depending upon the functionality involved.
- each block of the block diagrams and/or flowcharts, and combinations of blocks in the block diagrams and/or flowcharts can be implemented in a dedicated hardware-based system that performs the specified function or operation. Or it can be implemented by a combination of dedicated hardware and computer instructions.
- the units or modules described in the embodiments of the present disclosure may be implemented by software or by hardware.
- the described unit or module can also be provided in the processor.
- the names of these units or modules do not in any way constitute a limitation on the unit or module itself.
- the present disclosure further provides a computer readable storage medium, which may be a computer readable storage medium included in the apparatus described in the foregoing embodiments, or may exist separately, not A computer readable storage medium that is assembled into the device.
- the computer readable storage medium stores one or more programs that are used by one or more processors to perform the formula input methods described in this disclosure.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Ophthalmology & Optometry (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种眼睛状态的检测方法,包括:获取目标图像;对所述目标图像中的多个眼部特征点进行定位,以确定所述多个眼部特征点的位置坐标;对所述多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据;基于所述位置特征数据,确定所述目标图像中眼睛的状态。
- 根据权利要求1所述的方法,其中,所述眼部特征点包括左眼特征点和右眼特征点;对所述多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据包括:确定所述左眼特征点的位置坐标均值和所述右眼特征点的位置坐标均值之间的欧式距离;以所述欧式距离作为标准尺度,对所述眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据。
- 根据权利要求1或2所述的方法,其中,基于所述位置特征数据,确定所述目标图像中眼睛的状态包括:对所述位置特征数据进行分类;基于分类结果确定所述目标图像中眼睛的状态。
- 根据权利要求4或5所述的方法,其中,所述N=12。
- 根据权利要求3-6任一所述的方法,其中,所述对所述位置特征数据进行分类,包括:利用分类器对所述位置特征数据进行分类。
- 根据权利要求7所述的方法,其中,所述对所述位置特征数据进行分类还包括:利用样本图像对所述分类器进行训练,得到用于检测眼睛状态的分类器参数。
- 根据权利要求8所述的方法,其中,利用样本图像对所述分类器进行训练,得到用于检测眼睛状态的分类器参数包括:从图片库中获取正负样本图像;其中,正样本图像中的眼睛状态为睁眼状态,负样本图像中的眼睛状态为闭眼状态;对正负样本图像进行眼部特征点的定位,确定正负样本图像中的多个眼部特征点的位置坐标;对多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据;利用位置特征数据,对分类器进行训练,得到用于检测眼睛状态的分类器参数。
- 根据权利要求1所述的方法,其中,所述对所述目标图像中的多个眼部特征点进行定位,包括:检测所述目标图像中是否包含人脸;当检测出所述目标图像中包含人脸时,对所述目标图像中的多个眼部特征点进行定位。
- 一种电子设备,其中,包括:至少一个处理器、至少一个存储器以及存储在所述存储器中的计算机程序指令,当所述计算机程序指令被所述处理器执行:获取目标图像;对所述目标图像中的多个眼部特征点进行定位,以确定所述多个眼部特征点的位置坐标;对所述多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据;基于所述位置特征数据,确定所述目标图像中眼睛的状态。
- 根据权利要求11所述的电子设备,其中,所述眼部特征点包括左眼特征点和右眼特征点;对所述多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据包括:确定所述左眼特征点的位置坐标均值和所述右眼特征点的位置坐标均值之间的欧式距离;以所述欧式距离作为标准尺度,对所述多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据。
- 根据权利要求11或12所述的电子设备,其中,基于所述位置特征数据,确定所述目标图像中眼睛的状态包括:对所述位置特征数据进行分类;基于分类结果确定所述目标图像中眼睛的状态。
- 根据权利要求13-15任一所述的电子设备,其中,所述对所述位置特征数据进行分类,包括:利用分类器对所述位置特征数据进行分类。
- 根据权利要求16所述的电子设备,其中,所述对所述位置特征数据进行分类还包括:利用样本图像对所述分类器进行训练,得到用于检测眼睛状态的分类器参数。
- 根据权利要求17所述的电子设备,其中,利用样本图像对所述分类器进行训练,得到用于检测眼睛状态的分类器参数包括:从图片库中获取正负样本图像;其中,正样本图像中的眼睛状态为睁眼状态,负样本图像中的眼睛状态为闭眼状态;对正负样本图像进行眼部特征点的定位,确定正负样本图像中的多个眼部特征点的位置坐标;对多个眼部特征点的位置坐标进行归一化处理,得到处理后的位置特征数据;利用位置特征数据,对分类器进行训练,得到用于检测眼睛状态的分类器参数。
- 一种眼睛状态的检测装置,包括:获取单元,被配置为获取目标图像;定位单元,被配置为对所述目标图像中的多个眼部特征点进行定位,以确定所述多个眼部特征点的位置坐标;归一化处理单元,被配置为对所述多个眼部特征点的位置坐标进 行归一化处理,得到处理后的位置特征数据;确定单元,被配置为基于所述位置特征数据,确定所述目标图像中眼睛的状态。
- 一种计算机可读存储介质,其上存储有计算机程序指令,当所述计算机程序指令被处理器执行时实现如权利要求1-10中任一项所述的方法。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/473,491 US11386710B2 (en) | 2018-04-27 | 2018-11-30 | Eye state detection method, electronic device, detecting apparatus and computer readable storage medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810394919.1 | 2018-04-27 | ||
CN201810394919.1A CN108615014B (zh) | 2018-04-27 | 2018-04-27 | 一种眼睛状态的检测方法、装置、设备和介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019205633A1 true WO2019205633A1 (zh) | 2019-10-31 |
Family
ID=63661343
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/118374 WO2019205633A1 (zh) | 2018-04-27 | 2018-11-30 | 眼睛状态的检测方法、电子设备、检测装置和计算机可读存储介质 |
Country Status (3)
Country | Link |
---|---|
US (1) | US11386710B2 (zh) |
CN (1) | CN108615014B (zh) |
WO (1) | WO2019205633A1 (zh) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111178278A (zh) * | 2019-12-30 | 2020-05-19 | 上海商汤临港智能科技有限公司 | 视线方向确定方法、装置、电子设备及存储介质 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108615014B (zh) * | 2018-04-27 | 2022-06-21 | 京东方科技集团股份有限公司 | 一种眼睛状态的检测方法、装置、设备和介质 |
CN110555426A (zh) * | 2019-09-11 | 2019-12-10 | 北京儒博科技有限公司 | 视线检测方法、装置、设备及存储介质 |
CN110941333A (zh) * | 2019-11-12 | 2020-03-31 | 北京字节跳动网络技术有限公司 | 基于眼部动作的交互方法、装置、介质和电子设备 |
CN110866508B (zh) * | 2019-11-20 | 2023-06-27 | Oppo广东移动通信有限公司 | 识别目标对象的形态的方法、装置、终端及存储介质 |
CN112989890B (zh) * | 2019-12-17 | 2024-08-02 | 腾讯科技(深圳)有限公司 | 图像检测方法、装置及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101030316A (zh) * | 2007-04-17 | 2007-09-05 | 北京中星微电子有限公司 | 一种汽车安全驾驶监控系统和方法 |
CN106228293A (zh) * | 2016-07-18 | 2016-12-14 | 重庆中科云丛科技有限公司 | 教学评估方法及系统 |
CN106485191A (zh) * | 2015-09-02 | 2017-03-08 | 腾讯科技(深圳)有限公司 | 一种驾驶员疲劳状态检测方法及系统 |
CN107704805A (zh) * | 2017-09-01 | 2018-02-16 | 深圳市爱培科技术股份有限公司 | 疲劳驾驶检测方法、行车记录仪及存储装置 |
CN108615014A (zh) * | 2018-04-27 | 2018-10-02 | 京东方科技集团股份有限公司 | 一种眼睛状态的检测方法、装置、设备和介质 |
Family Cites Families (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100438841B1 (ko) * | 2002-04-23 | 2004-07-05 | 삼성전자주식회사 | 이용자 검증 및 데이터 베이스 자동 갱신 방법, 및 이를이용한 얼굴 인식 시스템 |
CN100389388C (zh) * | 2006-06-15 | 2008-05-21 | 北京中星微电子有限公司 | 基于人脸认证的屏幕保护方法及装置 |
JP4845755B2 (ja) * | 2007-01-30 | 2011-12-28 | キヤノン株式会社 | 画像処理装置、画像処理方法、プログラム及び記憶媒体 |
JP2009003730A (ja) * | 2007-06-22 | 2009-01-08 | Nintendo Co Ltd | 情報処理プログラムおよび情報処理装置 |
CN101339607B (zh) * | 2008-08-15 | 2012-08-01 | 北京中星微电子有限公司 | 人脸识别方法及系统、人脸识别模型训练方法及系统 |
JP5361368B2 (ja) * | 2008-12-22 | 2013-12-04 | 任天堂株式会社 | ゲームプログラム、ゲーム装置およびゲーム制御方法 |
JP5526233B2 (ja) * | 2010-06-28 | 2014-06-18 | 富士フイルム株式会社 | 立体視用画像撮影装置およびその制御方法 |
JP5682291B2 (ja) * | 2010-12-20 | 2015-03-11 | ソニー株式会社 | 補正値演算装置、複眼撮像装置、および、補正値演算装置の制御方法 |
CN102663361B (zh) * | 2012-04-01 | 2014-01-01 | 北京工业大学 | 一种面向整体特征分析的人脸图像可逆几何归一化方法 |
CN103793720B (zh) * | 2014-02-12 | 2017-05-31 | 北京海鑫科金高科技股份有限公司 | 一种眼睛定位方法和系统 |
CN106709404B (zh) * | 2015-11-16 | 2022-01-04 | 佳能株式会社 | 图像处理装置及图像处理方法 |
CN106529409B (zh) * | 2016-10-10 | 2019-08-09 | 中山大学 | 一种基于头部姿态的眼睛注视视角测定方法 |
WO2020121425A1 (ja) * | 2018-12-12 | 2020-06-18 | 三菱電機株式会社 | 状態判定装置、状態判定方法、及び状態判定プログラム |
CN110111418B (zh) * | 2019-05-15 | 2022-02-25 | 北京市商汤科技开发有限公司 | 创建脸部模型的方法、装置及电子设备 |
US20200401879A1 (en) * | 2019-06-19 | 2020-12-24 | LegInsight, LLC | Systems and methods for predicting whether experimental legislation will become enacted into law |
-
2018
- 2018-04-27 CN CN201810394919.1A patent/CN108615014B/zh active Active
- 2018-11-30 WO PCT/CN2018/118374 patent/WO2019205633A1/zh active Application Filing
- 2018-11-30 US US16/473,491 patent/US11386710B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101030316A (zh) * | 2007-04-17 | 2007-09-05 | 北京中星微电子有限公司 | 一种汽车安全驾驶监控系统和方法 |
CN106485191A (zh) * | 2015-09-02 | 2017-03-08 | 腾讯科技(深圳)有限公司 | 一种驾驶员疲劳状态检测方法及系统 |
CN106228293A (zh) * | 2016-07-18 | 2016-12-14 | 重庆中科云丛科技有限公司 | 教学评估方法及系统 |
CN107704805A (zh) * | 2017-09-01 | 2018-02-16 | 深圳市爱培科技术股份有限公司 | 疲劳驾驶检测方法、行车记录仪及存储装置 |
CN108615014A (zh) * | 2018-04-27 | 2018-10-02 | 京东方科技集团股份有限公司 | 一种眼睛状态的检测方法、装置、设备和介质 |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111178278A (zh) * | 2019-12-30 | 2020-05-19 | 上海商汤临港智能科技有限公司 | 视线方向确定方法、装置、电子设备及存储介质 |
CN111178278B (zh) * | 2019-12-30 | 2022-04-08 | 上海商汤临港智能科技有限公司 | 视线方向确定方法、装置、电子设备及存储介质 |
Also Published As
Publication number | Publication date |
---|---|
CN108615014B (zh) | 2022-06-21 |
CN108615014A (zh) | 2018-10-02 |
US11386710B2 (en) | 2022-07-12 |
US20210357617A1 (en) | 2021-11-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019205633A1 (zh) | 眼睛状态的检测方法、电子设备、检测装置和计算机可读存储介质 | |
US10684681B2 (en) | Neural network image processing apparatus | |
US11182592B2 (en) | Target object recognition method and apparatus, storage medium, and electronic device | |
US11295474B2 (en) | Gaze point determination method and apparatus, electronic device, and computer storage medium | |
CN107209849B (zh) | 眼睛跟踪 | |
WO2019232862A1 (zh) | 嘴巴模型训练方法、嘴巴识别方法、装置、设备及介质 | |
WO2019232866A1 (zh) | 人眼模型训练方法、人眼识别方法、装置、设备及介质 | |
JP3962803B2 (ja) | 頭部検出装置、頭部検出方法および頭部検出プログラム | |
WO2018219180A1 (zh) | 确定人脸图像质量的方法和装置、电子设备和计算机存储介质 | |
WO2019041519A1 (zh) | 目标跟踪装置、方法及计算机可读存储介质 | |
TWI754806B (zh) | 利用深度學習定位虹膜之裝置與方法 | |
WO2015165365A1 (zh) | 一种人脸识别方法及系统 | |
US7995807B2 (en) | Automatic trimming method, apparatus and program | |
CN110751043A (zh) | 基于人脸可见性的人脸识别方法、装置及存储介质 | |
US8090151B2 (en) | Face feature point detection apparatus and method of the same | |
WO2022021029A1 (zh) | 检测模型训练方法、装置、检测模型使用方法及存储介质 | |
CN112597872B (zh) | 视线角度估计方法和装置、存储介质、电子设备 | |
WO2011065952A1 (en) | Face recognition apparatus and methods | |
JP6071002B2 (ja) | 信頼度取得装置、信頼度取得方法および信頼度取得プログラム | |
JP6822482B2 (ja) | 視線推定装置、視線推定方法及びプログラム記録媒体 | |
WO2021143865A1 (zh) | 定位方法及装置、电子设备、计算机可读存储介质 | |
JP2023134688A (ja) | ビジョンシステムで画像内のパターンを検出及び分類するためのシステム及び方法 | |
WO2021208767A1 (zh) | 人脸轮廓修正方法、装置、设备及存储介质 | |
JP2007048172A (ja) | 情報分類装置 | |
US20140334694A1 (en) | Method for determining eye location on a frontal face digital image to validate the frontal face and determine points of reference |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18916913 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18916913 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18916913 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 14/05/2021) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18916913 Country of ref document: EP Kind code of ref document: A1 |