WO2017031901A1 - 人脸识别方法、装置及终端 - Google Patents
人脸识别方法、装置及终端 Download PDFInfo
- Publication number
- WO2017031901A1 WO2017031901A1 PCT/CN2015/099696 CN2015099696W WO2017031901A1 WO 2017031901 A1 WO2017031901 A1 WO 2017031901A1 CN 2015099696 W CN2015099696 W CN 2015099696W WO 2017031901 A1 WO2017031901 A1 WO 2017031901A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- face
- sub
- processed
- classifier
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 33
- 230000003044 adaptive effect Effects 0.000 claims description 30
- 238000012545 processing Methods 0.000 abstract description 10
- 238000010586 diagram Methods 0.000 description 19
- 238000004891 communication Methods 0.000 description 10
- 238000005516 engineering process Methods 0.000 description 7
- 230000003287 optical effect Effects 0.000 description 4
- 230000005236 sound signal Effects 0.000 description 4
- 238000012549 training Methods 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 230000006978 adaptation Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 229910044991 metal oxide Inorganic materials 0.000 description 2
- 150000004706 metal oxides Chemical class 0.000 description 2
- 210000000056 organ Anatomy 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/165—Detection; Localisation; Normalisation using facial parts and geometric relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/217—Validation; Performance evaluation; Active pattern learning techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
- G06V40/162—Detection; Localisation; Normalisation using pixel segmentation or colour matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Definitions
- the present disclosure relates to the field of image processing technologies, and in particular, to a face recognition method, apparatus, and terminal.
- a ratio between various organs such as a person's eyes and a nose can be used as a face feature.
- a ratio between various organs such as a person's eyes and a nose
- the face classifier may not recognize. The situation of the face.
- the present disclosure provides a face recognition method, apparatus, and terminal.
- a face recognition method comprising:
- a face in the original image is determined according to the result of the face recognition.
- the adding a pixel of a specified value to an edge region of the original image includes:
- a pixel of a specified value is added to the edge region where the face color pixel is present.
- the performing face recognition on the to-be-processed image includes:
- a pre-trained adaptive enhanced face classifier is used to determine whether each sub-image is a human face sub-image.
- the acquiring the multiple sub-images of the to-be-processed image includes:
- the image to be processed is traversed multiple times using a sliding window, and the image area covered by the sliding window at each position is used as a sub-image of the image to be processed each time traversing, wherein the traversing the image twice The size of the sliding window used for the image to be processed is different; or,
- the pre-trained adaptive enhanced face classifier is cascaded by a multi-level classifier. Determining whether each sub-image is a human face sub-image using a pre-trained adaptive enhanced face classifier includes:
- the output result of all the classifiers identifies that the sub-image is a face sub-image, it is determined that the sub-image is a face sub-image.
- the determining, according to the result of the face recognition, the face in the original image includes:
- a face in the original image is determined according to the position of the sub-image of the face sub-image in the original image.
- a face recognition device comprising:
- An adding module configured to add a pixel of a specified value to an edge region of the original image acquired by the acquiring module, to obtain an image to be processed
- An identification module configured to perform face recognition on the to-be-processed image obtained by the adding module
- a determining module configured to determine a face in the original image according to the result of the face recognition of the identification module.
- the adding module includes:
- a first acquiring unit configured to acquire pixel values of respective pixels of an edge region of the original image
- a first determining unit configured to determine, according to the pixel value of each pixel of the edge region and the preset face color pixel value obtained by the first acquiring unit, that an edge region of the face color pixel is present;
- an adding unit configured to add a pixel of a specified value to an edge region of the presence face color pixel determined by the first determining unit.
- the identifying module includes:
- a second acquiring unit configured to acquire a plurality of sub-images of the image to be processed
- a determining unit configured to determine, by using the pre-trained adaptive enhanced face classifier, whether each of the sub-images acquired by the second acquiring unit is a human face sub-image.
- the second acquiring unit is configured to use the sliding window to traverse the image to be processed multiple times, The image area covered by the sliding window at each position as a sub-image of the image to be processed at each traversal, wherein the size of the sliding window used for traversing the image to be processed twice is different; or
- the pre-trained adaptive enhanced face classifier is cascaded by a multi-level classifier. Determining, for determining, for any sub-image, starting from the first-level classifier of the pre-trained adaptive enhanced face classifier, determining whether the sub-image is a face sub-image until the adaptation The final classifier of the face classifier is enhanced; when the output results of all the classifiers identify that the sub-image is a face sub-image, the sub-image is determined to be a face sub-image.
- the determining module includes:
- a second determining unit configured to determine, when the sub-image of the face sub-image exists in the plurality of sub-images of the image to be processed, the position of the sub-image that is the face sub-image in the original image
- a third determining unit configured to determine, according to the location of the sub-image of the human face sub-image in the original image, The face in the original image.
- a terminal where the terminal includes:
- a memory for storing processor executable instructions
- processor is configured to:
- a face in the original image is determined according to the result of the face recognition.
- the image to be processed is subjected to face recognition to determine the face in the original image. Since a certain number of pixels are added to the edge region of the original image during face recognition, by adding this portion of the pixel, the original image is expanded to ensure that when the face is located at the edge region of the original image, the extension is performed.
- the sub-image including the human face can be found in the latter image, thereby ensuring that the face located in the edge region of the original image can be recognized, thereby improving the accuracy of the face recognition.
- FIG. 1 is a flowchart of a face recognition method according to an exemplary embodiment.
- FIG. 2 is a flowchart of a face recognition method according to an exemplary embodiment.
- FIG. 3 is a schematic diagram of an original image, according to an exemplary embodiment.
- FIG. 4 is a schematic diagram of an image to be processed, according to an exemplary embodiment.
- FIG. 5 is a schematic diagram of traversing a to-be-processed image using sliding windows of different sizes, according to an exemplary embodiment.
- FIG. 6 is a schematic diagram of a plurality of sizes of images to be processed, according to an exemplary embodiment.
- FIG. 7 is a schematic diagram of an Adaboost face classifier according to an exemplary embodiment.
- FIG. 8 is a block diagram of a face recognition device, according to an exemplary embodiment.
- FIG. 9 is a block diagram of an add module, according to an exemplary embodiment.
- FIG. 10 is a block diagram of an identification module, according to an exemplary embodiment.
- FIG. 11 is a block diagram of a determination module, according to an exemplary embodiment.
- FIG. 12 is a block diagram of a terminal, according to an exemplary embodiment.
- FIG. 1 is a flowchart of a face recognition method according to an exemplary embodiment, and a face recognition method is used in a terminal. As shown in FIG. 1 , the face recognition method provided by the embodiment of the present disclosure includes the following steps.
- step S101 the original image is acquired.
- step S102 pixels of a specified value are added to the edge region of the original image to obtain an image to be processed.
- step S103 face recognition is performed on the image to be processed.
- step S104 the face in the original image is determined based on the result of the face recognition.
- the method provided by the embodiment of the present disclosure performs face recognition on the image to be processed by adding pixels of a specified value to the edge region of the original image to obtain a face to be processed to determine a face in the original image. Since a certain number of pixels are added to the edge region of the original image during face recognition, by adding this portion of the pixel, the original image is expanded to ensure that when the face is located at the edge region of the original image, the extension is performed.
- the sub-image including the human face can be found in the latter image, thereby ensuring that the face located in the edge region of the original image can be recognized, thereby improving the accuracy of the face recognition.
- adding a pixel of a specified value to an edge region of the original image includes:
- a pixel of a specified value is added to an edge region where a face color pixel is present.
- the face to be processed is subjected to face recognition, including:
- a pre-trained adaptive enhanced face classifier is used to determine whether each sub-image is a human face sub-image.
- acquiring a plurality of sub-images of the image to be processed includes:
- the image to be processed is traversed multiple times using a sliding window, and the image area covered by the sliding window at each position is used as a sub-image of the image to be processed each time traversing, wherein the sliding window used for the image to be processed is traversed twice Different sizes; or,
- the image to be processed is multi-scaled to obtain a plurality of sizes of images to be processed; for any size of the image to be processed, the image to be processed of the size is cropped into image regions of a plurality of specified sizes, and each image region is taken as a sub-image image.
- the pre-trained adaptive enhanced face classifier is cascaded by a multi-level classifier, and the pre-trained adaptive enhanced face classifier is used to determine whether each sub-image is a face sub-image, including:
- any sub-image starting from the first-level classifier of the pre-trained adaptive enhanced face classifier, it is determined step by step whether the sub-image is a face sub-image until the last-level classifier of the adaptive enhanced face classifier;
- the output result of all the classifiers identifies that the sub-image is a face sub-image, it is determined that the sub-image is a face sub-image.
- determining a face in the original image based on the result of the face recognition includes:
- the face in the original image is determined based on the position of the sub-image of the face sub-image in the original image.
- FIG. 2 is a flowchart of a face recognition method according to an exemplary embodiment, and a face recognition method is used in a terminal. As shown in FIG. 2, the face recognition method provided by the embodiment of the present disclosure includes the following steps.
- step S201 the original image is acquired.
- the original image is an image that requires face recognition.
- Embodiments of the present disclosure need to identify whether a face is included in an original image, and if a face is included in the original image, which region of the original image the face is located in.
- an image can be read from any storage device as an original image; or an image can be downloaded from the Internet as an original image; an image can also be scanned using a scanner to obtain an original image, and An image taken by the camera is taken as the original image.
- step S202 pixels of a specified value are added to the edge region of the original image to obtain an image to be processed.
- the edge region of the original image is the outermost layer of pixels on each of the four sides of the original image.
- FIG. 3 shows a schematic diagram of an original image in which the face is located in the upper edge region of the original image.
- face recognition when the face is located in the edge area of the original image, when face recognition is performed, there may be cases where the face is not recognized.
- a pixel of a specified value is first added to an edge region of the original image.
- FIG. 4 shows a schematic diagram of an image to be processed, and the image to be processed in FIG. 4 is obtained by adding pixels to the four edge regions of the original image shown in FIG.
- the area with slashes in Figure 4 represents the added pixels.
- the specific values of the specified numerical values are not specifically limited in the embodiments of the present disclosure. When it is implemented, it can be set as needed. For example, two pixels, five pixels, ten pixels, and the like may be added to the periphery of each pixel of the outermost layer of pixels.
- all of the added pixels may have the same pixel value. That is, all pixels added have the same color.
- the same color may be white, black or other colors, and the like, which is not specifically limited in the embodiment of the present disclosure. Since pixels of the same color have the same pixel value, when the added pixels are pixels of the same color, it can be ensured that when the face to be processed is subjected to face recognition, when the pixels of a certain area of the image to be processed are found to have the same pixel value, Then, it can be determined that it is an increased pixel, so that it is not necessary to perform an excessive recognition process, and thus it is possible to have a relatively high recognition speed.
- pixels of a specified value may be added to all of the four edge regions of the original image.
- the number of pixels of a specified value added to each edge region can be different.
- the number of pixels added is different as the left edge area and the right edge area.
- the amount of calculation at the time of image recognition is increased.
- the edge region where the face may exist may be detected first, and the pixel may be added at the edge region where the face may exist.
- the pixel value of the face color pixel is usually a specific value or is within a certain value range, it is possible to detect whether the pixel value of each pixel of the edge region is the pixel value of the face color pixel. Determine if there is a possible face in the edge area.
- steps S2021 to S2023 when a pixel of a specified value is added to an edge region of the original image, including but not limited to, by the following steps S2021 to S2023:
- step S2021 the pixel values of the respective pixels of the edge region of the original image are acquired.
- the pixel values of the respective pixels of the edge region are acquired, it can be realized by determining the RGB values of the respective pixels.
- determining the RGB values of the respective pixels including but not limited to being implemented by a color sensor.
- step S2022 the edge region where the face color pixel is present is determined according to the pixel value of each pixel of the edge region and the preset face skin color pixel value.
- the pixel value of each pixel of the edge region may be compared with the preset face skin color pixel value, and the edge region of the face skin color pixel is determined according to the comparison result.
- the embodiment of the present disclosure is not limited. However, in order to ensure that the pixel is accurately recognized as a face skin color pixel, the first preset threshold may be set to be relatively small.
- the number of face skin color pixels may be the total number of pixels corresponding to the edge region according to all the pixels corresponding to the edge region.
- the ratio depends on. When the ratio is greater than the second preset threshold, it is determined that the edge region has a face color pixel; otherwise, it is determined that the edge region does not have a face skin pixel.
- the specific value of the second preset threshold can be set as needed.
- step S2023 pixels of a specified numerical value are added to the edge region where the face color pixel is present.
- the pixel of the specified value when the pixel is added to the edge region of the original image, the pixel of the specified value may be added only in the edge region where the face color pixel is present. For example, when the face skin color pixel exists in the upper edge region of the original image, the pixel of the specified value may be added only in the upper edge region to achieve a reduction in the amount of calculation of the image recognition.
- step S203 a plurality of sub-images of the image to be processed are acquired, and whether each sub-image is a human face sub-image is determined using a pre-trained adaptive enhanced face classifier.
- This step is a specific implementation of face recognition for the image to be processed.
- the embodiment of the present disclosure first acquires a plurality of sub-images of a to-be-processed image, and implements by determining whether each sub-image is a human face sub-image.
- the first way using the sliding window to traverse the image to be processed multiple times, the image area covered by the sliding window at each position is used as a sub-image of the image to be processed each time traversing, wherein the image to be processed is traversed twice or twice The size of the sliding window used is different.
- embodiments of the present disclosure will traverse the image to be processed separately using sliding windows of different sizes.
- sliding windows There are many types of sliding windows that can be used each time the image is to be processed. For example, when traversing the image to be processed a certain time, the size of the sliding window is 3*3; the next time the image is to be processed, the sliding The size of the window is 5*5 and so on.
- FIG. 5 shows a schematic diagram of traversing a to-be-processed image using sliding windows of different sizes. Each thick solid square in Figure 5 is a sliding window.
- the sliding window When traversing the image to be processed using a sliding window of any size, the sliding window will traverse in the horizontal direction (X direction) and the vertical direction (Y direction) of the image to be processed according to the specified step size, each in the X direction or Y
- the direction moves by one step and moves to a new position of the image to be processed, and each position defines an image range, and the image range defined by each position is a sub-image of the image to be processed.
- the specified step size can be one pixel, two pixels, and the like.
- the image to be processed is multi-scaled to obtain a plurality of sizes of the image to be processed; for any size of the image to be processed, the image to be processed of the size is cropped into image regions of a plurality of specified sizes, each of which will be Image area as A sub image.
- a size of the image to be processed is obtained.
- the image to be processed of the size may be cropped into a plurality of image regions, each image region having a size of a specified size.
- the specified size is 3*3 (pixel * pixel), 5*5, and the like.
- FIG. 6 a schematic diagram of a plurality of sizes of images to be processed is shown.
- (a) to (c) of Fig. 6 respectively show images of a size to be processed.
- (c) of FIG. 6 it shows a schematic diagram of cropping the image to be processed of the size.
- the rectangular frame surrounded by each thick solid line is a sub-image of the image to be processed of the size.
- each sub-image when determining whether each sub-image is a face sub-image, it can be implemented by a pre-trained face classifier.
- the pre-trained face classifier may be a support vector machine face classifier, a neural network face classifier, or an adaptive boost (Adaboost) face classifier or the like.
- Adaboost adaptive boost
- the pre-trained face classifier is used as an example of the Adaboost face classifier.
- the pre-trained Adaboost face classifier in the embodiment of the present disclosure is cascaded by a multi-level classifier.
- each classifier of the Adaboost face classifier is used to determine whether a sub-image is a face sub-image.
- the output of any classifier is "1" and "0".
- the output result is “1” to identify that the classifier determines that the sub-image is a face sub-image; the output result is “0” to identify that the classifier determines that the sub-image is not a face sub-image.
- Each classifier of the Adaboost face classifier is a strong classifier, and each strong classifier includes a plurality of weak classifiers.
- each classifier of the Adaboost face classifier it is implemented by training a plurality of weak classifiers included in the class classifier, and the output result of the class classifier is processed according to the data processing of all the weak classifiers included in the classifier. Decide.
- the manner of training the Adaboost face classifier and the manner of determining the output result of each classifier refer to the content of the existing Adaboost face classifier, which is not explained in detail in the embodiment of the present disclosure.
- the embodiment of the present disclosure is not specifically limited.
- the Adaboost face classifier may include more levels of classifiers, such as a 5-level classifier, an 8-level classifier, and the like.
- FIG. 7 which shows a schematic diagram of an Adaboost face classifier, each circular area in FIG. 7 represents a primary classifier.
- the pre-trained adaptive enhanced face classifier when using the pre-trained adaptive enhanced face classifier to determine whether each sub-image is a face sub-image, for any sub-image, starting from the first-level classifier of the pre-trained Adaboost face classifier, It is judged step by step whether the sub-image is a face sub-image until the last-level classifier of the Adaboost face classifier. When all the fractions When the output result of the classifier identifies that the sub-image is a face sub-image, the sub-image is determined to be a human face sub-image. When the output result of any classifier identifies that the sub-image is a non-human face sub-image, it is determined that the sub-image is a non-human face sub-image.
- the sub-image is input from the first-level classifier to the Adaboost face classifier, and when the first-level classifier determines that the sub-image is a face sub-image, the sub-image is input to the second level. a classifier, wherein the second level classifier determines whether the sub-image is a face sub-image, and so on, until the final classifier; when the first-level classifier determines that the sub-image is not a face sub-image, the first-level classifier obtains The next sub-image is identified and the next sub-image is identified.
- the face when the face is located in the edge region of the original image, it usually cannot recognize the face located in the edge region. However, when the face is occluded, it can correctly recognize the occluded face.
- the face located in the edge region is equivalent to being occluded by the added pixel, and therefore, the face region can be identified by the pre-trained face classifier. The face of the face can thus improve the accuracy of recognizing the face of the edge area.
- step S204 the face in the original image is determined based on the result of the face recognition.
- the sub-images belonging to the human face can be determined according to the result of the face recognition.
- the sub-image determined as the face sub-image is original
- the position in the image, and the face in the original image is determined based on the position of the sub-image of the face sub-image in the original image.
- the pixel values of each pixel of the sub-image may be extracted, and each of the original images is extracted.
- the pixel values of the pixels are further compared with the pixel values of the respective pixels in the original image.
- the position of the sub-image in the original image can be located.
- the face in the original image can be determined.
- the method provided by the embodiment of the present disclosure performs face recognition on the image to be processed by adding pixels of a specified value to the edge region of the original image to obtain a face to be processed to determine a face in the original image. Since a certain number of pixels are added to the edge region of the original image during face recognition, by adding this portion of the pixel, it is equivalent to the original image. Extending to ensure that when the face is located in the edge area of the original image, the sub-image including the face can be found in the expanded image, thereby ensuring that the face located in the edge area of the original image can be recognized, thereby improving the person The accuracy of face recognition.
- FIG. 8 is a block diagram of a face recognition device, according to an exemplary embodiment.
- the face recognition device includes an acquisition module 801, an addition module 802, an identification module 803, and a determination module 804. among them:
- the obtaining module 801 is configured to acquire an original image
- the adding module 802 is configured to add a pixel of a specified value to an edge region of the original image acquired by the obtaining module 801 to obtain an image to be processed;
- the identification module 803 is configured to perform face recognition on the to-be-processed image obtained by the adding module 802;
- the determining module 804 is configured to acquire a face in the original image according to the recognition result of the recognition module 803.
- the apparatus provided by the embodiment of the present disclosure performs face recognition on the image to be processed by adding pixels of a specified value to the edge region of the original image to obtain a face to be processed, to determine a face in the original image. Since a certain number of pixels are added to the edge region of the original image during face recognition, by adding this portion of the pixel, the original image is expanded to ensure that when the face is located at the edge region of the original image, the extension is performed.
- the sub-image including the human face can be found in the latter image, thereby ensuring that the face located in the edge region of the original image can be recognized, thereby improving the accuracy of the face recognition.
- the adding module 802 includes a first acquiring unit 8021, a first determining unit 8022, and an adding unit 8023. among them:
- the first obtaining unit 8021 is configured to acquire pixel values of respective pixels of an edge region of the original image
- the first determining unit 8022 is configured to determine, according to the pixel value of each pixel of the edge region and the preset face skin color pixel value acquired by the first acquiring unit, that the edge region of the face color pixel is present;
- the adding unit 8023 is configured to increase a pixel of a specified value in an edge region of the presence face color pixel determined by the first determining unit 8022.
- the identification module 803 includes:
- the second obtaining unit 8031 is configured to acquire a plurality of sub-images of the image to be processed
- the determining unit 8032 is configured to determine whether each of the sub-images acquired by the second acquiring unit is a human face sub-image using the pre-trained adaptive enhanced face classifier.
- the second obtaining unit 8031 is configured to traverse the image to be processed multiple times using a sliding window, and the image area covered by the sliding window at each position is used as a child of the image to be processed each time traversing An image in which the size of the sliding window used to traverse the image to be processed twice is different; or,
- the image to be processed is multi-scaled to obtain a plurality of sizes of images to be processed; for any size of the image to be processed, the image to be processed of the size is cropped into image regions of a plurality of specified sizes, and each image region is taken as a sub-image image.
- the pre-trained adaptive enhanced face classifier is cascaded by a multi-level classifier configured to pre-train adaptive adaptive face classification for any sub-image
- the first level classifier of the device starts to determine whether the sub-image is a face sub-image until the last-level classifier of the adaptive enhanced face classifier; when the output results of all the classifiers identify the sub-image as a face sub-image , determining that the sub-image is a human face sub-image.
- the determining module 804 includes a second determining unit 8041 and a third determining unit 8042. among them:
- the second determining unit 8041 is configured to determine, as a sub-image of the face sub-image among the plurality of sub-images of the image to be processed, a position of the sub-image of the face sub-image in the original image;
- the third determining unit 8042 is configured to determine a face in the original image based on the position of the sub-image of the face sub-image in the original image.
- FIG. 12 is a block diagram of a terminal 1200, which may be used to perform the face recognition method provided by the embodiment corresponding to FIG. 1 or FIG. 2, according to an exemplary embodiment.
- the terminal 1200 can be a mobile phone, a computer, a digital broadcast terminal, a messaging device, a game console, a tablet device, a medical device, a fitness device, a personal digital assistant, and the like.
- the terminal 1200 may include one or more of the following components: a processing component 1202, a memory 1204, a power component 1206, a multimedia component 1208, an audio component 1210, an I/O (Input/Output) interface 1212, and a sensor. Component 1214, and communication component 1216.
- Processing component 1202 typically controls the overall operations of terminal 1200, such as operations associated with display, telephone calls, data communications, camera operations, and recording operations.
- Processing component 1202 can include one or more processors 1220 to execute instructions to perform all or part of the steps described above.
- processing component 1202 can include one or more modules to facilitate interaction between component 1202 and other components.
- processing component 1202 can include a multimedia module to facilitate interaction between multimedia component 1208 and processing component 1202.
- the memory 1204 is configured to store various types of data to support operation at the terminal 1200. Examples of such data include instructions for any application or method operating on terminal 1200, contact data, phone book data, messages, pictures, videos, and the like.
- Memory 1204 can be of any type of volatile or non-volatile storage device or their Combined implementation, such as SRAM (Static Random Access Memory), EEPROM (Electrically-Erasable Programmable Read-Only Memory), EPROM (Erasable Programmable Read Only Memory, Erasable) In addition to programmable read only memory, PROM (Programmable Read-Only Memory), ROM (Read-Only Memory), magnetic memory, flash memory, magnetic disk or optical disk.
- SRAM Static Random Access Memory
- EEPROM Electrically-Erasable Programmable Read-Only Memory
- EPROM Erasable Programmable Read Only Memory, Erasable
- PROM Programmable Read-Only Memory
- ROM Read-Only Memory
- magnetic memory flash memory
- flash memory magnetic disk or
- Power component 1206 provides power to various components of terminal 1200.
- Power component 1206 can include a power management system, one or more power sources, and other components associated with generating, managing, and distributing power for terminal 1200.
- the multimedia component 1208 includes a screen between the terminal 1200 and the user that provides an output interface.
- the screen may include an LCD (Liquid Crystal Display) and a TP (Touch Panel). If the screen includes a touch panel, the screen can be implemented as a touch screen to receive input signals from the user.
- the touch panel includes one or more touch sensors to sense touches, slides, and gestures on the touch panel. The touch sensor can sense not only the boundaries of the touch or sliding action, but also the duration and pressure associated with the touch or slide operation.
- the multimedia component 1208 includes a front camera and/or a rear camera. When the terminal 1200 is in an operation mode such as a shooting mode or a video mode, the front camera and/or the rear camera can receive external multimedia data. Each front and rear camera can be a fixed optical lens system or have focal length and optical zoom capabilities.
- the audio component 1210 is configured to output and/or input an audio signal.
- the audio component 1210 includes a MIC (Microphone) that is configured to receive an external audio signal when the terminal 1200 is in an operational mode, such as a call mode, a recording mode, and a voice recognition mode.
- the received audio signal may be further stored in memory 1204 or transmitted via communication component 1216.
- audio component 1210 also includes a speaker for outputting an audio signal.
- the I/O interface 1212 provides an interface between the processing component 1202 and the peripheral interface module, which may be a keyboard, a click wheel, a button, or the like. These buttons may include, but are not limited to, a home button, a volume button, a start button, and a lock button.
- Sensor assembly 1214 includes one or more sensors for providing terminal 1200 with a status assessment of various aspects.
- the sensor component 1214 can detect the open/closed state of the terminal 1200, the relative positioning of the components, such as the display and the keypad of the terminal 1200, and the sensor component 1214 can also detect the location change of a component of the terminal 1200 or the terminal 1200, the user The presence or absence of contact with the terminal 1200, the orientation or acceleration/deceleration of the terminal 1200 and the temperature change of the terminal 1200.
- Sensor assembly 1214 can include a proximity sensor configured to detect the presence of nearby objects without any physical contact.
- Sensor component 1214 can also include a light sensor, such as CMOS (Complementary Metal Oxide Semiconductor, Complementary Metal Oxide) or CCD (Charge-coupled Device) image sensor for use in imaging applications.
- CMOS Complementary Metal Oxide Semiconductor, Complementary Metal Oxide
- CCD Charge-coupled Device
- the sensor assembly 1214 can also include an acceleration sensor, a gyro sensor, a magnetic sensor, a pressure sensor, or a temperature sensor.
- Communication component 1216 is configured to facilitate wired or wireless communication between terminal 1200 and other devices.
- the terminal 1200 can access a wireless network based on a communication standard such as WiFi, 2G or 3G, or a combination thereof.
- communication component 1216 receives broadcast signals or broadcast associated information from an external broadcast management system via a broadcast channel.
- the communication component 1216 further includes an NFC (Near Field Communication) module to facilitate short-range communication.
- the NFC module can be based on RFID (Radio Frequency Identification) technology, IrDA (Infra-red Data Association) technology, UWB (Ultra Wideband) technology, BT (Bluetooth) technology and Other technologies are implemented.
- the terminal 1200 may be configured by one or more ASICs (Application Specific Integrated Circuits), DSP (Digital Signal Processor), DSPD (Digital Signal Processor Device). Device), PLD (Programmable Logic Device), FPGA (Field Programmable Gate Array), controller, microcontroller, microprocessor or other electronic components are implemented to perform the above diagram 1 or the face recognition method provided by the embodiment corresponding to FIG. 2.
- ASICs Application Specific Integrated Circuits
- DSP Digital Signal Processor
- DSPD Digital Signal Processor Device
- PLD Programmable Logic Device
- FPGA Field Programmable Gate Array
- controller microcontroller, microprocessor or other electronic components are implemented to perform the above diagram 1 or the face recognition method provided by the embodiment corresponding to FIG. 2.
- non-transitory computer readable storage medium comprising instructions, such as a memory 1204 including instructions executable by processor 1220 of terminal 1200 to perform the above-described face recognition method.
- the non-transitory computer readable storage medium may be a ROM, a RAM (Random Access Memory), a CD-ROM (Compact Disc Read-Only Memory), a magnetic tape, a floppy disk, and optical data. Storage devices, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Oral & Maxillofacial Surgery (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Biology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Geometry (AREA)
- Image Analysis (AREA)
- Collating Specific Patterns (AREA)
- Image Processing (AREA)
- Studio Devices (AREA)
Abstract
Description
Claims (13)
- 一种人脸识别方法,其特征在于,所述方法包括:获取原始图像;在所述原始图像的边缘区域增加指定数值的像素,得到待处理图像;对所述待处理图像进行人脸识别;根据所述人脸识别的结果确定所述原始图像中的人脸。
- 根据权利要求1所述的方法,其特征在于,所述在所述原始图像的边缘区域增加指定数值的像素,包括:获取所述原始图像的边缘区域各个像素的像素值;根据所述边缘区域各个像素的像素值及预设人脸肤色像素值,确定存在人脸肤色像素的边缘区域;在所述存在人脸肤色像素的边缘区域增加指定数值的像素。
- 根据权利要求1或2所述的方法,其特征在于,所述对所述待处理图像进行人脸识别,包括:获取所述待处理图像的多个子图像;使用预先训练的自适应增强人脸分类器判断每个子图像是否为人脸子图像。
- 根据权利要求3所述的方法,其特征在于,所述获取所述待处理图像的多个子图像,包括:使用滑动窗口分多次遍历所述待处理图像,将每次遍历时所述滑动窗口在每个位置所覆盖的图像区域作为所述待处理图像的一个子图像,其中,任两次遍历所述待处理图像使用的滑动窗口的尺寸不同;或者,对所述待处理图像进行多次缩放,得到多个尺寸的待处理图像;对于任一尺寸的待处理图像,将所述尺寸的待处理图像裁剪为多个指定尺寸的图像区域,将每个图像区域作为一个子图像。
- 根据权利要求3所述的方法,其特征在于,所述预先训练的自适应增强人脸分类器由多级分类器级联而成,所述使用预先训练的自适应增强人脸分类器判断每个子图像是否为 人脸子图像,包括:对于任一子图像,从所述预先训练的自适应增强人脸分类器的第一级分类器开始,逐级判断所述子图像是否为人脸子图像,直至所述自适应增强人脸分类器的最后一级分类器;当所有级分类器的输出结果均标识所述子图像为人脸子图像时,确定所述子图像为人脸子图像。
- 根据权利要求3所述的方法,其特征在于,所述根据所述人脸识别的结果确定所述原始图像中的人脸,包括:在所述待处理图像的多个子图像中存在为人脸子图像的子图像时,确定所述为人脸子图像的子图像在所述原始图像中的位置;根据所述为人脸子图像的子图像在所述原始图像中的位置,确定所述原始图像中的人脸。
- 一种人脸识别装置,其特征在于,所述装置包括:获取模块,用于获取原始图像;增加模块,用于在所述获取模块获取的所述原始图像的边缘区域增加指定数值的像素,得到待处理图像;识别模块,用于对所述增加模块得到的所述待处理图像进行人脸识别;确定模块,用于根据所述识别模块的所述人脸识别的结果确定所述原始图像中的人脸。
- 根据权利要求7所述的装置,其特征在于,所述增加模块包括:第一获取单元,用于获取所述原始图像的边缘区域各个像素的像素值;第一确定单元,用于根据所述第一获取单元获取的所述边缘区域各个像素的像素值及预设人脸肤色像素值,确定存在人脸肤色像素的边缘区域;增加单元,用于在所述第一确定单元确定的所述存在人脸肤色像素的边缘区域增加指定数值的像素。
- 根据权利要求7或8所述的装置,其特征在于,所述识别模块包括:第二获取单元,用于获取所述待处理图像的多个子图像;判断单元,用于使用预先训练的自适应增强人脸分类器判断所述第二获取单元获取的每 个子图像是否为人脸子图像。
- 根据权利要求9所述的装置,其特征在于,,所述第二获取单元,用于使用滑动窗口分多次遍历所述待处理图像,将每次遍历时所述滑动窗口在每个位置所覆盖的图像区域作为所述待处理图像的一个子图像,其中,任两次遍历所述待处理图像使用的滑动窗口的尺寸不同;或者,对所述待处理图像进行多次缩放,得到多个尺寸的待处理图像;对于任一尺寸的待处理图像,将所述尺寸的待处理图像裁剪为多个指定尺寸的图像区域,将每个图像区域作为一个子图像。
- 根据权利要求9所述的装置,其特征在于,所述预先训练的自适应增强人脸分类器由多级分类器级联而成,所述判断单元,用于对于任一子图像,从所述预先训练的自适应增强人脸分类器的第一级分类器开始,逐级判断所述子图像是否为人脸子图像,直至所述自适应增强人脸分类器的最后一级分类器;当所有级分类器的输出结果均标识所述子图像为人脸子图像时,确定所述子图像为人脸子图像。
- 根据权利要求9所述的装置,其特征在于,所述确定模块包括:第二确定单元,用于在所述待处理图像的多个子图像中存在为人脸子图像的子图像时,确定所述为人脸子图像的子图像在所述原始图像中的位置;第三确定单元,用于根据所述为人脸子图像的子图像在所述原始图像中的位置,确定所述原始图像中的人脸。
- 一种终端,其特征在于,所述终端包括:处理器;用于存储处理器可执行指令的存储器;其中,所述处理器被配置为:获取原始图像;在所述原始图像的边缘区域增加指定数值的像素,得到待处理图像;对所述待处理图像进行人脸识别;根据所述人脸识别的结果确定所述原始图像中的人脸。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2016567408A JP6374986B2 (ja) | 2015-08-21 | 2015-12-30 | 顔認識方法、装置及び端末 |
KR1020167015669A KR20170033805A (ko) | 2015-08-21 | 2015-12-30 | 사람 얼굴 인식 방법, 장치 및 단말 |
RU2017102521A RU2664688C2 (ru) | 2015-08-21 | 2015-12-30 | Способ распознавания человеческих лиц, устройство и терминал |
MX2017008481A MX2017008481A (es) | 2015-08-21 | 2015-12-30 | Metodo, aparato y terminal para reconocimiento de la cara humana. |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510520457.X | 2015-08-21 | ||
CN201510520457.XA CN105095881B (zh) | 2015-08-21 | 2015-08-21 | 人脸识别方法、装置及终端 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2017031901A1 true WO2017031901A1 (zh) | 2017-03-02 |
Family
ID=54576269
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2015/099696 WO2017031901A1 (zh) | 2015-08-21 | 2015-12-30 | 人脸识别方法、装置及终端 |
Country Status (8)
Country | Link |
---|---|
US (1) | US10007841B2 (zh) |
EP (1) | EP3133527A1 (zh) |
JP (1) | JP6374986B2 (zh) |
KR (1) | KR20170033805A (zh) |
CN (1) | CN105095881B (zh) |
MX (1) | MX2017008481A (zh) |
RU (1) | RU2664688C2 (zh) |
WO (1) | WO2017031901A1 (zh) |
Families Citing this family (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9286509B1 (en) * | 2012-10-19 | 2016-03-15 | Google Inc. | Image optimization during facial recognition |
CN105095881B (zh) | 2015-08-21 | 2023-04-07 | 小米科技有限责任公司 | 人脸识别方法、装置及终端 |
EP3136289A1 (en) * | 2015-08-28 | 2017-03-01 | Thomson Licensing | Method and device for classifying an object of an image and corresponding computer program product and computer-readable medium |
CN106485567B (zh) * | 2016-09-14 | 2021-11-30 | 北京小米移动软件有限公司 | 物品推荐方法及装置 |
CN106372616B (zh) * | 2016-09-18 | 2019-08-30 | Oppo广东移动通信有限公司 | 人脸识别方法、装置及终端设备 |
CN106446884A (zh) * | 2016-09-19 | 2017-02-22 | 广东小天才科技有限公司 | 一种图像的快速截取的方法和装置 |
US10474880B2 (en) * | 2017-03-15 | 2019-11-12 | Nec Corporation | Face recognition using larger pose face frontalization |
CN108280420A (zh) * | 2018-01-19 | 2018-07-13 | 百度在线网络技术(北京)有限公司 | 用于处理图像的系统、方法和装置 |
CN109492550B (zh) * | 2018-10-25 | 2023-06-06 | 腾讯科技(深圳)有限公司 | 活体检测方法、装置及应用活体检测方法的相关系统 |
US10650564B1 (en) * | 2019-04-21 | 2020-05-12 | XRSpace CO., LTD. | Method of generating 3D facial model for an avatar and related device |
CN110401835B (zh) * | 2019-06-05 | 2021-07-02 | 西安万像电子科技有限公司 | 图像处理方法及装置 |
CN110248107A (zh) * | 2019-06-13 | 2019-09-17 | Oppo广东移动通信有限公司 | 图像处理方法和装置 |
CN110533002B (zh) * | 2019-09-06 | 2022-04-12 | 厦门久凌创新科技有限公司 | 基于人脸识别的大数据处理方法 |
JP2022522551A (ja) * | 2020-02-03 | 2022-04-20 | ベイジン センスタイム テクノロジー ディベロップメント カンパニー リミテッド | 画像処理方法及び装置、電子機器並びに記憶媒体 |
CN112132030B (zh) * | 2020-09-23 | 2024-05-28 | 湖南快乐阳光互动娱乐传媒有限公司 | 视频处理方法及装置、存储介质及电子设备 |
CN112966136B (zh) * | 2021-05-18 | 2021-09-07 | 武汉中科通达高新技术股份有限公司 | 一种人脸分类方法及装置 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070274573A1 (en) * | 2006-05-26 | 2007-11-29 | Canon Kabushiki Kaisha | Image processing method and image processing apparatus |
CN101488181A (zh) * | 2008-01-15 | 2009-07-22 | 华晶科技股份有限公司 | 多方向的人脸检测方法 |
CN102096802A (zh) * | 2009-12-11 | 2011-06-15 | 华为技术有限公司 | 人脸检测方法及装置 |
CN102270308A (zh) * | 2011-07-21 | 2011-12-07 | 武汉大学 | 一种基于五官相关aam模型的面部特征定位方法 |
CN105095881A (zh) * | 2015-08-21 | 2015-11-25 | 小米科技有限责任公司 | 人脸识别方法、装置及终端 |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4551839B2 (ja) * | 2004-08-16 | 2010-09-29 | キヤノン株式会社 | 撮像装置及び撮像装置の制御方法 |
EP1628465A1 (en) | 2004-08-16 | 2006-02-22 | Canon Kabushiki Kaisha | Image capture apparatus and control method therefor |
RU2295152C1 (ru) * | 2005-09-15 | 2007-03-10 | Роман Павлович Худеев | Способ распознавания лица человека по видеоизображению |
JP2008181439A (ja) * | 2007-01-26 | 2008-08-07 | Sanyo Electric Co Ltd | 顔検出装置及び方法並びに撮像装置 |
KR100973588B1 (ko) | 2008-02-04 | 2010-08-02 | 한국과학기술원 | 얼굴검출기의 부윈도우 설정방법 |
KR101105435B1 (ko) | 2009-04-14 | 2012-01-17 | 경북대학교 산학협력단 | 얼굴 검출과 얼굴 인지 방법 |
US9042599B2 (en) * | 2010-07-02 | 2015-05-26 | Intel Corporation | Techniques for face detection and tracking |
AU2013205535B2 (en) * | 2012-05-02 | 2018-03-15 | Samsung Electronics Co., Ltd. | Apparatus and method of controlling mobile terminal based on analysis of user's face |
-
2015
- 2015-08-21 CN CN201510520457.XA patent/CN105095881B/zh active Active
- 2015-12-30 JP JP2016567408A patent/JP6374986B2/ja active Active
- 2015-12-30 MX MX2017008481A patent/MX2017008481A/es unknown
- 2015-12-30 RU RU2017102521A patent/RU2664688C2/ru active
- 2015-12-30 WO PCT/CN2015/099696 patent/WO2017031901A1/zh active Application Filing
- 2015-12-30 KR KR1020167015669A patent/KR20170033805A/ko active Search and Examination
-
2016
- 2016-06-14 EP EP16174462.8A patent/EP3133527A1/en not_active Ceased
- 2016-06-28 US US15/195,368 patent/US10007841B2/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20070274573A1 (en) * | 2006-05-26 | 2007-11-29 | Canon Kabushiki Kaisha | Image processing method and image processing apparatus |
CN101488181A (zh) * | 2008-01-15 | 2009-07-22 | 华晶科技股份有限公司 | 多方向的人脸检测方法 |
CN102096802A (zh) * | 2009-12-11 | 2011-06-15 | 华为技术有限公司 | 人脸检测方法及装置 |
CN102270308A (zh) * | 2011-07-21 | 2011-12-07 | 武汉大学 | 一种基于五官相关aam模型的面部特征定位方法 |
CN105095881A (zh) * | 2015-08-21 | 2015-11-25 | 小米科技有限责任公司 | 人脸识别方法、装置及终端 |
Also Published As
Publication number | Publication date |
---|---|
JP6374986B2 (ja) | 2018-08-15 |
RU2017102521A (ru) | 2018-07-26 |
JP2017534090A (ja) | 2017-11-16 |
US10007841B2 (en) | 2018-06-26 |
CN105095881B (zh) | 2023-04-07 |
US20170053156A1 (en) | 2017-02-23 |
KR20170033805A (ko) | 2017-03-27 |
EP3133527A1 (en) | 2017-02-22 |
RU2664688C2 (ru) | 2018-08-21 |
RU2017102521A3 (zh) | 2018-07-26 |
MX2017008481A (es) | 2017-10-31 |
CN105095881A (zh) | 2015-11-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2017031901A1 (zh) | 人脸识别方法、装置及终端 | |
US9674395B2 (en) | Methods and apparatuses for generating photograph | |
WO2021031609A1 (zh) | 活体检测方法及装置、电子设备和存储介质 | |
RU2577188C1 (ru) | Способ, аппарат и устройство для сегментации изображения | |
US9959484B2 (en) | Method and apparatus for generating image filter | |
CN105631797B (zh) | 水印添加方法及装置 | |
WO2017088470A1 (zh) | 图像分类方法及装置 | |
WO2016011747A1 (zh) | 肤色调整方法和装置 | |
CN107944447B (zh) | 图像分类方法及装置 | |
RU2664003C2 (ru) | Способ и устройство для определения ассоциированного пользователя | |
WO2017128767A1 (zh) | 指纹模板录入方法及装置 | |
CN107944367B (zh) | 人脸关键点检测方法及装置 | |
WO2020181728A1 (zh) | 图像处理方法及装置、电子设备和存储介质 | |
CN105631803B (zh) | 滤镜处理的方法和装置 | |
WO2016192325A1 (zh) | 视频文件的标识处理方法及装置 | |
EP2998960A1 (en) | Method and device for video browsing | |
CN107730448B (zh) | 基于图像处理的美颜方法及装置 | |
CN109034150B (zh) | 图像处理方法及装置 | |
CN113409342A (zh) | 图像风格迁移模型的训练方法、装置及电子设备 | |
CN110111281A (zh) | 图像处理方法及装置、电子设备和存储介质 | |
KR20190111034A (ko) | 특징 이미지 획득 방법 및 디바이스, 및 사용자 인증 방법 | |
WO2020233201A1 (zh) | 图标位置确定方法和装置 | |
CN107507128B (zh) | 图像处理方法及设备 | |
CN108010009B (zh) | 一种去除干扰图像的方法及装置 | |
CN111507131B (zh) | 活体检测方法及装置、电子设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref document number: 20167015669 Country of ref document: KR Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2016567408 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2017102521 Country of ref document: RU Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 15902169 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2017/008481 Country of ref document: MX |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 15902169 Country of ref document: EP Kind code of ref document: A1 |