WO2023231400A1 - Method and apparatus for predicting facial angle, and device and readable storage medium - Google Patents

Method and apparatus for predicting facial angle, and device and readable storage medium Download PDF

Info

Publication number
WO2023231400A1
WO2023231400A1 PCT/CN2022/142276 CN2022142276W WO2023231400A1 WO 2023231400 A1 WO2023231400 A1 WO 2023231400A1 CN 2022142276 W CN2022142276 W CN 2022142276W WO 2023231400 A1 WO2023231400 A1 WO 2023231400A1
Authority
WO
WIPO (PCT)
Prior art keywords
angle
face
probabilities
detection window
face image
Prior art date
Application number
PCT/CN2022/142276
Other languages
French (fr)
Chinese (zh)
Inventor
何金辉
肖嵘
王孝宇
Original Assignee
青岛云天励飞科技有限公司
深圳云天励飞技术股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 青岛云天励飞科技有限公司, 深圳云天励飞技术股份有限公司 filed Critical 青岛云天励飞科技有限公司
Publication of WO2023231400A1 publication Critical patent/WO2023231400A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/77Determining position or orientation of objects or cameras using statistical methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions

Definitions

  • the present application belongs to the field of image recognition technology, and in particular relates to a face angle detection method, device, equipment and readable storage medium.
  • the facial image is first reconstructed into a three-dimensional image, and then the three-dimensional image is mapped into a two-dimensional image, and then based on the facial movement characteristics in the two-dimensional image Perform face pose prediction.
  • the angle can be corrected before recognition or not, to improve the accuracy of image recognition.
  • This application provides a face angle prediction method, device, equipment and readable storage medium, which can avoid the difficulty of predicting only the approximate posture of a face and the difficulty in accurately predicting the face angle, and can adapt to the relatively low accuracy of face angle prediction. High scene.
  • this application provides a face angle prediction method, including:
  • multiple angle probabilities of multiple angle types are determined.
  • the multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type.
  • the multiple angle types include yaw. angle, pitch and roll angles;
  • the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position is determined.
  • This application determines the multiple angle probabilities of multiple angle types through the facial features corresponding to the face area, and then determines the relative shooting position of the face in the face image to be measured based on the multiple angle probabilities of each angle type.
  • the predicted angle for each angle type Since the obtained multiple angle probabilities are angle probabilities corresponding to multiple angle intervals, the corresponding prediction angle is calculated through the angle probabilities corresponding to multiple angle intervals, ensuring the accuracy of the prediction angle and avoiding the situation where it is difficult to accurately predict the face angle. , which can be adapted to scenes with high accuracy in face angle prediction.
  • this application provides a face angle prediction device, which is used to perform the method in the above-mentioned first aspect or any possible implementation of the first aspect.
  • the device may include:
  • the acquisition module is used to obtain the face area of the face image to be tested
  • the first determination module is used to determine the facial features corresponding to the facial area
  • the second determination module is used to determine multiple angle probabilities of multiple angle types according to the facial features, and the multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type, said Multiple angle types including yaw, pitch, and roll angles;
  • the third determination module is configured to determine the predicted angle of each angle type of the human face in the face image to be measured relative to the shooting position based on multiple angle probabilities of each angle type.
  • the present application provides an electronic device, which includes a memory and a processor.
  • the memory is used to store instructions; the processor executes the instructions stored in the memory, so that the device performs the face angle prediction method in the first aspect or any possible implementation of the first aspect.
  • a computer-readable storage medium In a fourth aspect, a computer-readable storage medium is provided. Instructions are stored in the computer-readable storage medium. When the instructions are run on a computer, they cause the computer to execute the first aspect or any possible implementation of the first aspect. Face angle prediction method.
  • a fifth aspect provides a computer program product containing instructions that, when run on a device, cause the device to execute the face angle prediction method of the first aspect or any possible implementation of the first aspect.
  • Figure 1 is a schematic flowchart of a face angle prediction method provided by an embodiment of the present application
  • Figure 2 is a schematic flow chart of a face angle prediction method provided by an embodiment of the present application.
  • Figure 3 is a schematic flow chart of a face angle prediction method provided by an embodiment of the present application.
  • Figure 4 is a schematic flowchart of a face angle prediction method provided by an embodiment of the present application.
  • Figure 5 is a schematic structural diagram of a face angle prediction device provided by an embodiment of the present application.
  • FIG. 6 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
  • the term “if” may be interpreted as “when” or “once” or “in response to determining” or “in response to detecting” depending on the context. ". Similarly, the phrase “if determined” or “if [the described condition or event] is detected” may be interpreted, depending on the context, to mean “once determined” or “in response to a determination” or “once the [described condition or event] is detected ]” or “in response to detection of [the described condition or event]”.
  • This application provides a face angle prediction method, device, equipment and readable storage medium.
  • the method can be implemented through recognition equipment and applied in access control recognition, missing person search, case investigation, intelligent security and other scenarios.
  • the face angle includes three angle types of the face relative to the shooting position, and the three angle types are pitch angle, yaw angle and roll angle respectively.
  • the recognition device refers to the device used by users to predict face angles.
  • Identification devices can be access control devices, smartphones, desktop computers, laptops, tablets, wearable devices, handheld devices, vehicle-mounted devices, servers, etc.
  • the embodiments of this application do not place any restrictions on the specific types of identification devices.
  • the identification device may include display hardware, or may have an external display.
  • the recognition device can predict the face angle based on the face in the image, and determine whether the next action can be taken based on the predicted face angle. For example, when When the angle of the face is too large, a prompt message such as "The angle of the face is too large and cannot be recognized" will be displayed on the display screen of the access control device.
  • the face angle prediction method in the missing person search scenario, it is used to input the face image into the recognition device and predict the face angle corresponding to the face in the face image. The face obtained through prediction The angle determines whether the next step can be taken, for example, determining whether the recognition device can recognize the face, determining whether the face image matches the image in the missing person image library, and then determining whether it is a missing person.
  • the recognition device can be connected through communication with the surveillance camera, and the recognition device can predict the face angle corresponding to the face in the image by acquiring the image captured by the surveillance camera.
  • FIG. 1 shows a schematic flowchart of a face angle prediction method provided by an embodiment of the present application.
  • the face angle prediction method provided by this application can include:
  • the face image to be tested can be directly given by the user, or it can be extracted from video data collected by image collection equipment such as surveillance cameras and video cameras.
  • the face area refers to the area containing faces in the face image to be measured.
  • the face area is obtained by performing face detection on the face image to be tested, obtaining the first detection window, and intercepting the image in the first detection window.
  • the recognition device can perform an external expansion process on the first detection window to obtain an expanded second detection window, and intercept an area corresponding to the size of the second detection window as the face area.
  • the detection window refers to the wire frame from which the face in the face image to be detected can be extracted.
  • the face detection algorithm can be used to detect the face to be tested.
  • the face detection algorithm can be stored in the storage device.
  • the storage device can communicate with the recognition device, so that the recognition device can retrieve the face detection algorithm from the storage device.
  • This application does not limit the storage method and specific type of storage devices.
  • the YOLO (you only look once) algorithm is used for face detection.
  • the YOLO algorithm is an object recognition and positioning algorithm based on deep neural networks. Its biggest feature is its fast running speed.
  • the identification device is an access control device
  • the access control device includes a camera.
  • the camera captures a face image.
  • the access control device uses a face detection algorithm to detect the face image and obtain the face area of the face image to be measured.
  • the recognition device is a mobile phone
  • the mobile phone has a recognition applet.
  • the mobile phone communicates with the surveillance camera through the recognition applet to obtain images captured by the surveillance camera.
  • the recognition applet can use the face detection algorithm to detect faces in images captured by surveillance cameras and obtain the face area of the face image to be detected.
  • the recognition device can obtain the face area. Therefore, the recognition device can perform feature extraction on the face area and obtain the facial features corresponding to the face area.
  • the recognition device outputs the facial features by inputting the facial region into the backbone network of the facial angle recognition model.
  • the backbone network is used to extract facial features from facial images.
  • the backbone network is stored in advance in a storage device that communicates with the identification device.
  • the identification device is an access control device.
  • the camera captures the face image.
  • the access control device performs face detection on the face image through the face detection algorithm. After obtaining the face area of the face image, it calls the backbone network to detect the face. Feature extraction is performed on the region to obtain facial features.
  • the identification device is a mobile phone
  • the mobile phone has an identification applet.
  • the recognition applet performs face detection on the image through the face detection algorithm. After obtaining the face area corresponding to the image, it calls the backbone network to extract features of the face area to obtain the face features.
  • various angle types include yaw angle, pitch angle and roll angle.
  • Multiple angle probabilities for each angle type respectively correspond to multiple angle intervals for each angle type.
  • Multiple angle intervals refer to multiple angle intervals obtained by dividing the angle ranges of yaw angle, pitch angle and roll angle according to preset rules.
  • the prediction rule is to divide the angle range into intervals every 5 degrees.
  • the angle range of yaw angle and pitch angle is [-90, 90].
  • the angle range of yaw angle and pitch angle is divided into intervals every 5 degrees, resulting in 36 angle intervals respectively.
  • the 36 angle intervals of the yaw angle are [-90, -85), [-85, -80)... (80, 85], (85, 90].
  • the 36 angle intervals of the pitch angle are [-90, -85), [-85, -80)... (80, 85], (85, 90].
  • the angle range of the roll angle is [-180, 180].
  • the angle range of the roll angle is divided into intervals every 5 degrees, resulting in 72 angle intervals.
  • the 72 angle intervals of the roll angle are [-180, -175), [-175, -170), [-170, -165)... (165, 170], (170, 175], (175, 180 ].
  • the recognition device inputs the facial features into the fully connected classification network of the facial angle recognition model and outputs multiple angle probabilities for each angle type.
  • the fully connected classification network is used to predict the angle probabilities corresponding to the facial features in multiple angle intervals of each angle type.
  • the backbone network and fully connected classification network serve as face angle recognition models and are pre-stored in a storage device that communicates with the recognition device.
  • the fully connected classification network is connected to the output end of the backbone network.
  • the recognition device backbone network extracts the facial features of the face image to be tested, the facial features are input to the fully connected classification network for angle probability prediction.
  • the fully connected classification network includes three fully connected layers, namely the first fully connected layer, the second fully connected layer and the third fully connected layer.
  • the first fully connected layer and the second fully connected layer are connected to 36 nodes respectively, and the third fully connected layer is connected to 72 pivots.
  • the 36 nodes of the first fully connected layer and the second fully connected layer are used to predict the angular probability of the yaw angle and the pitch angle in their 36 angle intervals
  • the 72 nodes of the third fully connected layer are used It is used to predict the angle probability of roll angle in its 72 angle intervals.
  • the identification device is an access control device. After acquiring the facial features, the access control device predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features.
  • the identification device is a mobile phone
  • the mobile phone has an identification applet.
  • the recognition applet predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features.
  • the relative angle of the face in the face image to be measured is determined based on multiple angle probabilities, the number of angle intervals, and the intermediate angle of each angle interval. The angle of the shooting position.
  • the middle angle of each angle interval refers to every five angles being the middle angle in an angle interval. For example, if a certain angle range is [0, 5), then the corresponding intermediate angle is 2.5 degrees.
  • n represents the number of angle intervals
  • the identification device is an access control device
  • the access control device includes a display screen and a camera.
  • the access control device predicts the angular probability of the yaw angle and pitch angle in its 36 angle interval based on the facial features. Calculate the angle probability of each angle interval corresponding to the value of each angle interval, calculate the value corresponding to each interval based on the angle probability of the roll angle in its 72 angle intervals, and calculate the final face based on the value corresponding to each interval prediction angle.
  • the access control device can determine whether the next action can be taken based on the predicted angle of the face. For example, when the face angle is too large, a prompt message such as "The face angle is too large to be recognized" will be displayed on the display of the access control device. .
  • the identification device is a mobile phone
  • the mobile phone has an identification applet.
  • the recognition applet determines the corresponding facial features based on the facial features, predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features. Finally, calculate the corresponding value of each angle interval based on the angle probability of the yaw angle and pitch angle in its 36 angle intervals, calculate the corresponding value of each interval based on the angle probability of the roll angle in its 72 angle intervals, and Calculate the final predicted angle of the face based on the value corresponding to each interval.
  • the recognition applet can use the predicted angle of the face to determine whether the next action can be taken. For example, when the angle of the face is too large, the face image to be measured is corrected.
  • the face angle prediction method provided by this application obtains face features based on the face area of the face image to be measured, and then determines multiple angle probabilities for each of the multiple angle types based on the face features. Finally, based on each angle type The multiple angle probabilities are used to determine the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position. Therefore, for each angle type, since the obtained multiple angle probabilities are angle probabilities corresponding to multiple angle intervals, the predicted angle is calculated based on the angle probabilities corresponding to multiple angle intervals, ensuring the accuracy of the predicted angle.
  • the recognition device can obtain the maximum value of multiple angle probabilities of the roll angle.
  • the maximum value corresponds to the preset mapping angle interval
  • the calculated predicted angle of the roll angle is inaccurate.
  • the range of multiple angle intervals of the roll angle is [-180, 180].
  • the mapping angle interval includes multiple angle intervals corresponding to [-180, -90) or multiple angle intervals corresponding to (90, 180].
  • Figure 2 shows a schematic flow chart of a face angle prediction method provided by an embodiment of the present application.
  • the face angle prediction method provided by this application may include:
  • the maximum probability angle interval is the angle interval corresponding to the maximum value among multiple angle probabilities.
  • the maximum value among multiple angle probabilities for roll angle corresponds to the angle interval [-175, -170).
  • the roll angle of the face is considered to be larger.
  • the recognition device calculates the roll angle according to the method shown in S103 in Figure 1 The predicted angle is inaccurate.
  • the recognition device determines that the maximum probability angle interval corresponding to the roll angle is in the angle interval of [-90, 90], it can calculate the predicted angle of the roll angle according to the method shown in S103 in Figure 1.
  • the recognition device determines that the maximum probability angle interval corresponding to the roll angle is within the angle interval corresponding to [-180, -90) or the angle interval corresponding to (90, 180], the recognition device needs to determine the multiple angle probabilities of the roll angle. Perform linear mapping to obtain multiple angle probabilities after mapping.
  • the recognition device linearly maps multiple angle probabilities so that when the maximum probability angle interval corresponding to the roll angle is in the angle interval of [-180, 90) or (90, 180], the accurate angle value of the roll angle can be obtained .
  • linear mapping refers to:
  • the predicted angle of the roll angle is calculated directly according to the method shown in S103 in Figure 1, the predicted angle value obtained It may be -103.5 degrees. Obviously, -103.5 degrees is not in the angle interval [-175, -170), which is unreasonable.
  • the recognition device needs to first map multiple angle probabilities of the roll angle, for example, map the angle probability corresponding to the angle interval [-175, -170) to the angle interval (-10, -5], that is, Replace the angle probability corresponding to the angle interval [-175, -170) with the angle probability corresponding to the angle interval (-10, -5).
  • the calculated predicted angle value can be -5 degrees. Obviously -5 degrees is in the angle interval (-10, -5]. This is reasonable.
  • Inverse linear mapping refers to mapping the predicted angle value (mapping angle) calculated based on the angle probability corresponding to the angle interval (-10, -5) to the angle interval [-175, -170).
  • the angle value obtained based on S203 is -5 degrees
  • -5 degrees are de-reflected
  • the obtained angle value is -175 degrees, which is in the angle interval [-175, 170), which is reasonable.
  • the multiple angle probabilities are linearly mapped to obtain the mapped multiple angle probabilities, and then based on The mapping angle is determined based on the mapped multiple angle probabilities, the number of angle intervals of the roll angle and the middle angle of each angle interval, and the mapping angle is inversely mapped to predict the angle.
  • linear mapping is used to map the angle probabilities corresponding to the mapping angle interval, and then the mapping angle is calculated, and the predicted angle is calculated based on the mapping angle, so that more accuracy can be obtained prediction angle.
  • the recognition device when it determines the face area, it can expand the first detection window to obtain the second detection window, intercept the area corresponding to the size of the second detection window, and add the second detection window to the second detection window.
  • the area corresponding to the size of the two detection windows is determined as the face area.
  • the first detection window can be expanded to obtain more face information corresponding to the face image to be measured, ensuring that the final face angle accuracy is higher.
  • FIG. 3 shows a schematic flowchart of a face angle prediction method provided by an embodiment of the present application.
  • the face angle prediction method provided by this application may include:
  • face detection is performed on the face image to be tested, and the first detection window obtained is a rectangle.
  • the identification device can obtain the third detection window based on the length of the first detection window and the width as the side length and the center of the first detection window as the center.
  • face detection is performed on the face image to be tested, and the first detection window obtained is a square.
  • step S302 can be directly performed on the first detection window according to the expansion coefficient.
  • the first detection window obtained is a rectangle or a square. Whether the first detection window is rectangular or square is usually determined by the distance from the camera, the facial expression or movement, the angle of the human face, and other aspects.
  • the first detection window is a rectangle with a length of 60 pixels and a width of 40 pixels. Then, taking the center of the first detection window as the center and the length of the first detection window as the side length of the window, the third detection window obtained is a square with a side length of 60 pixels.
  • the preset expansion coefficient is 0.1.
  • the expansion coefficient can also be other values, such as 0.15, which can be set according to the actual situation, and will not be described in detail here.
  • the side length of the third detection window is 60 pixels, and the expansion coefficient is 0.1. Then, the third detection window is expanded, and the fourth detection window obtained is a square with a side length of 66 pixels.
  • the obtained fourth detection window may exceed the original face image to be tested, that is, the fourth detection window exceeds the corresponding side length of the face image to be tested.
  • the recognition device determines that the fourth detection window exceeds the side length corresponding to the face image to be measured, it removes the excess side length, and obtains the fifth detection window after removal.
  • the length of the face image to be measured is 90 pixels
  • the width is 60 pixels
  • the side length of the fourth detection window is 66 pixels.
  • the length of the fifth detection window obtained is 66 pixels and the width is 60 pixels.
  • S304 A window with the center of the fifth detection window as the center and the shorter side of the fifth detection window as the second detection window.
  • the length of the fifth detection window is 66 pixels and the width is 60 pixels. Then, taking the center of the fifth detection window as the center and the width of the fifth detection window as the side length, the obtained second detection window is a square with a side length of 60 pixels.
  • the identification device will take the center of the first detection window as the center and the long side of the first detection window as the third detection window.
  • each line of the third detection window will be The side lengths are all expanded to obtain the fourth detection window. Remove the side length of the fourth detection window that exceeds the corresponding side length of the face image to be tested, and obtain the fifth detection window.
  • the center of the fifth detection window will be the center.
  • the short side of the window is the second detection window.
  • the recognition device obtains a second detection window by expanding the first detection window, and the face area selected by the face image to be tested is larger and includes more face information. Facial features are extracted from the face area, and the facial features obtained are more accurate. By predicting the angle of the face with more accurate facial features, a more accurate prediction angle can be obtained.
  • this application also provides a generation process of a face angle recognition model including a backbone network and a fully connected classification network.
  • the recognition device when the recognition device obtains the facial features corresponding to the facial area, it obtains them through the backbone network in the facial angle recognition model.
  • the recognition device when the recognition device obtains the facial features corresponding to the face area, it obtains them through the fully connected classification network in the face angle recognition model.
  • the generation process of the face angle recognition model can be completed by a model generation device, or it can be generated by other feasible devices, which will not be described again here.
  • FIG. 4 shows a schematic flowchart of generating a face angle recognition model according to an embodiment of the present application.
  • the process of generating the face angle recognition model includes:
  • the sample face image set includes multiple frames of sample face images and the real angle corresponding to each angle type of the face in each frame of the sample face image relative to the shooting position.
  • the sample face image set at least includes a set of sample face images and real angles corresponding to each angle type of the face in the sample face image relative to the shooting position.
  • the sample face image set can be selected from an existing image data set (for example, the public data set 300W-LP), or it can be a face image captured by a camera in advance.
  • an existing image data set for example, the public data set 300W-LP
  • it can be a face image captured by a camera in advance.
  • the camera that captures face images can be a camera, a smartphone camera, a laptop camera, or a tablet camera.
  • the real angle corresponding to the sample face image can be obtained by using relevant sensors or by manual annotation.
  • S402 Perform data enhancement processing on each frame of the sample face image to obtain an enhanced sample face image.
  • Data enhancement processing may include one or a combination of random shearing, adding random noise, and color perturbation.
  • each frame of sample face image is randomly cut, random noise is added, and color perturbation is processed to obtain an enhanced sample face image.
  • the original face angle recognition model includes the original backbone network and the original fully connected classification network.
  • the output end of the original backbone network is connected to three fully connected layers of the original fully connected classification network.
  • the three fully connected layers are respectively the first original fully connected layer, the second original fully connected layer and the third original fully connected layer. connection layer.
  • the first original fully connected layer and the second original fully connected layer are connected to 36 nodes respectively, and the third original fully connected layer is connected to 72 pivots.
  • the 36 nodes of the first original fully connected layer and the second original fully connected layer are used to predict the angle probability of yaw angle and pitch angle in their 36 angle intervals respectively, and the 72 nodes of the third original fully connected layer are used for prediction.
  • the model generation device inputs the sample image set into the original backbone network, it outputs facial features.
  • the first original fully connected layer obtains the angles corresponding to the 36 angle intervals of the yaw angle based on the facial features. Probability.
  • the model generation device first calculates a loss function based on multiple angle probabilities of each angle type and the real angle corresponding to each angle type, and then adjusts the model parameters of the original angle recognition model through the loss function.
  • the above loss function is the cross entropy loss function.
  • the loss function can also be other types of loss functions, so I won’t go into details here.
  • the model generation device trains the original angle recognition model according to the loss function through an error backpropagation algorithm to obtain a trained face angle recognition model, and determines the trained face angle recognition model as the face angle Identify the model.
  • the model generation device when the model generation device generates the face angle recognition model, it first obtains a sample face image, and performs data enhancement processing on each frame of the sample face image to obtain an enhanced sample face image. Then input the enhanced sample face image into the original angle recognition model, output multiple angle probabilities for each angle type, and adjust the original angle probability based on the multiple angle probabilities for each angle type and the true angle corresponding to each angle type.
  • the model parameters of the angle recognition model determine the adjusted original angle recognition model as the face angle recognition model. Using preset rules to divide the angle ranges of the three angle types, and adjusting the original angle recognition model through the multiple angle probabilities corresponding to the three angle types and the real angles corresponding to the three angle types, more accurate predictions can be obtained Face angle recognition model for face angles.
  • this application also provides a face angle prediction device.
  • FIG. 5 shows a schematic block diagram of a face angle prediction device provided by an embodiment of the present application.
  • a face angle prediction device provided by an embodiment of the present application includes an acquisition module 501 , a first determination module 502 , a second determination module 503 and a third determination module 504 .
  • the acquisition module 501 is used to acquire the face area of the face image to be tested
  • the first determination module 502 is used to determine the facial features corresponding to the facial area
  • the second determination module 503 is used to determine multiple angle probabilities of multiple angle types based on the facial features.
  • the multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type, so
  • the various angle types include yaw angle, pitch angle and roll angle;
  • the third determination module 504 is configured to determine the predicted angle of each angle type of the human face in the face image to be measured relative to the shooting position based on multiple angle probabilities of each angle type.
  • the third determination module is specifically used for:
  • the angle of the face in the face image to be measured relative to the shooting position is determined based on multiple angle probabilities, the number of angle intervals, and the intermediate angle of each angle interval.
  • the third determination module 504 is specifically used to:
  • a maximum probability angle interval is determined, and the maximum probability angle interval is an angle interval corresponding to the maximum value among multiple angle probabilities;
  • the third determination module 504 is specifically used to:
  • the acquisition module 501 is specifically used to:
  • the method of obtaining the face area of the face image to be tested includes:
  • the acquisition module 501 is specifically used for:
  • each side length of the third detection window is expanded to obtain a fourth detection window
  • the first determination module 502 is specifically used to:
  • the backbone network is used to extract the face features in the face image
  • the facial features are input into the fully connected classification network of the face angle recognition model, and multiple angle probabilities of each angle type are output.
  • the fully connected classification network is used to predict the facial features in each angle type.
  • the angle probabilities corresponding to the multiple angle intervals of .
  • the model generation device is used for:
  • Obtain a sample face image set which includes a multi-frame sample face image and a true angle corresponding to each angle type of the face in each frame of the sample face image relative to the shooting position;
  • the original face angle recognition model includes the original backbone network and the original fully connected classification network;
  • the adjusted original angle recognition model is determined as the face angle recognition model.
  • the device 500 of the present application can be implemented through an application-specific integrated circuit (application-specific integrated circuit). integrated circuit (ASIC), or programmable logic device (PLD).
  • ASIC application-specific integrated circuit
  • PLD programmable logic device
  • the above PLD can be a complex programmable logical device (CPLD), a field-programmable gate array (field-programmable) gate array, FPGA), general array logic (generic array logic, GAL) or any combination thereof.
  • CPLD complex programmable logical device
  • FPGA field-programmable gate array
  • GAL general array logic
  • the face angle prediction method shown in Figure 1 can also be implemented through software.
  • the device 500 and its respective modules can also be software modules.
  • Figure 6 is a schematic structural diagram of an electronic device provided by this application.
  • the device 600 includes a processor 601, a memory 602, a communication interface 603 and a bus 604.
  • the processor 601, the memory 602, and the communication interface 603 communicate through the bus 604. Communication can also be achieved through other means such as wireless transmission.
  • the memory 602 is used to store instructions, and the processor 601 is used to execute the instructions stored in the memory 602.
  • the memory 602 stores program code 6021, and the processor 601 can call the program code 6021 stored in the memory 602 to execute the face angle prediction method shown in Figure 2.
  • the processor 601 may be a CPU, and the processor 601 may also be other general-purpose processors, digital signal processors (DSPs), application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or Other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
  • DSPs digital signal processors
  • ASICs application-specific integrated circuits
  • FPGAs field-programmable gate arrays
  • a general-purpose processor can be a microprocessor or any conventional processor, etc.
  • the memory 602 may include read-only memory and random access memory and provides instructions and data to the processor 601. Memory 602 may also include non-volatile random access memory.
  • the memory 602 may be volatile memory or non-volatile memory, or may include both volatile and non-volatile memory.
  • the non-volatile memory can be a read-only memory (read-only memory). memory, ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically erasable programmable read-only memory (electrically EPROM, EEPROM) or flash memory.
  • Volatile memory can be random access memory (random access memory (RAM), which serves as an external cache.
  • RAM static random access memory
  • DRAM dynamic random access memory
  • SDRAM synchronous dynamic random access memory
  • Double data rate synchronous dynamic random access memory double data date SDRAM, DDR SDRAM
  • enhanced synchronous dynamic random access memory enhanced SDRAM, ESDRAM
  • synchronous link dynamic random access memory direct memory bus random access memory
  • direct rambus RAM direct rambus RAM
  • bus 604 may also include a power bus, a control bus, a status signal bus, etc. However, for clarity of illustration, the various buses are labeled bus 604 in FIG. 6 .
  • the electronic device 600 may correspond to the device 500 in the present application, and may correspond to the device in the method shown in FIG. 1 of the present application.
  • the device 600 corresponds to the device in the method shown in FIG. 2 of the present application.
  • the above and other operations and/or functions of each module in the device 600 are respectively intended to implement the operating steps of the method performed by the device in Figure 2. For the sake of brevity, they will not be described again here.
  • This application also provides a computer-readable storage medium that stores a computer program.
  • the computer program is executed by a processor, the steps in each of the above method embodiments can be implemented.
  • the present application provides a computer program product.
  • the steps in each of the above method embodiments can be implemented when the electronic device is executed.
  • sequence number of each step in the above embodiment does not mean the order of execution.
  • the execution order of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of this application.
  • Module completion means dividing the internal structure of the above device into different functional units or modules to complete all or part of the functions described above.
  • Each functional unit and module in the embodiment can be integrated into one processing unit, or each unit can exist physically alone, or two or more units can be integrated into one unit.
  • the above-mentioned integrated unit can be hardware-based. It can also be implemented in the form of software functional units.
  • the specific names of each functional unit and module are only for the convenience of distinguishing each other and are not used to limit the scope of protection of the present application.
  • For the specific working processes of the units and modules in the above system please refer to the corresponding processes in the foregoing method embodiments, and will not be described again here.
  • the disclosed devices/network devices and methods can be implemented in other ways.
  • the device/network equipment embodiments described above are only illustrative.
  • the division of the above modules or units is only a logical function division. In actual implementation, there may be other division methods, such as multiple units or units. Components may be combined or may be integrated into another system, or some features may be ignored, or not implemented.
  • the coupling or direct coupling or communication connection between each other shown or discussed may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
  • the units described above as separate components may or may not be physically separated.
  • the components shown as units may or may not be physical units, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of this application.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • General Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The present application is applicable to the technical field of image recognition. Provided are a method and apparatus for predicting a facial angle, and a device and a readable storage medium. The method comprises: acquiring a facial area of a facial image to be subjected to detection; determining facial features corresponding to the facial area; according to the facial features, determining a plurality of angle probabilities for each angle type among a plurality of angle types, wherein the plurality of angle types comprise a yaw angle, a pitch angle, and a roll angle; and according to the plurality of angle probabilities for each angle type, determining a predicted angle of each angle type of a face in said facial image relative to a photographing position. Thus, in the present application, angle probabilities respectively corresponding to a plurality of angle intervals are acquired, and a predicted angle is calculated by means of angle probabilities respectively corresponding to the plurality of angle intervals, so that the accuracy of the predicted angle is ensured. The present application is applicable to a plurality of scenarios where accurate facial angles need to be determined.

Description

人脸角度预测方法、装置、设备及可读存储介质Face angle prediction method, device, equipment and readable storage medium 技术领域Technical field
本申请要求于2022年5月31日提交中国专利局,申请号为202210607682.7、发明名称为“人脸角度预测方法、装置、设备及可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application requests the priority of the Chinese patent application submitted to the China Patent Office on May 31, 2022, with the application number 202210607682.7 and the invention name "Facial Angle Prediction Method, Device, Equipment and Readable Storage Medium", and its entire content incorporated herein by reference.
本申请属于图像识别技术领域,尤其涉及一种人脸角度检测方法、装置、设备及可读存储介质。The present application belongs to the field of image recognition technology, and in particular relates to a face angle detection method, device, equipment and readable storage medium.
背景技术Background technique
随着人工智能的发展,人脸识别在各个行业的应用越来越多,而人脸的图像质量,对图像识别精度影响较大。其中,人脸的角度是影响图像识别精度的重要因素。With the development of artificial intelligence, face recognition is increasingly used in various industries, and the image quality of the face has a greater impact on the accuracy of image recognition. Among them, the angle of the face is an important factor affecting the accuracy of image recognition.
相关技术中,通过捕捉人脸在动作时的面部关键点特征,先将人脸图像重建成为三维图像,再将三维图像映射成二维图像,再根据二维图像中的人脸面部的动作特征进行人脸姿态预测。在确定人脸角度为大角度的人脸时,可以进行角度矫正后再进行识别或不进行识别,提升图像识别的精度。In related technologies, by capturing the facial key point characteristics of the human face in action, the facial image is first reconstructed into a three-dimensional image, and then the three-dimensional image is mapped into a two-dimensional image, and then based on the facial movement characteristics in the two-dimensional image Perform face pose prediction. When determining that the face angle is a large-angle face, the angle can be corrected before recognition or not, to improve the accuracy of image recognition.
然而,由于相关技术只能通过在人脸动作时的动作特征预测人脸在动作层面的姿态,难以对人脸角度准确预测,无法适用于需要预测准确度较高的人脸角度的场景。However, since the related technology can only predict the posture of the face at the action level through the action characteristics of the face when it is moving, it is difficult to accurately predict the face angle, and cannot be applied to scenes that require prediction of face angles with high accuracy.
技术解决方案Technical solutions
本申请提供了一种人脸角度预测方法、装置、设备及可读存储介质,可以避免仅能预测人脸的大致姿态,难以对人脸角度准确预测,可以适应人脸角度预测的准确度较高的场景。This application provides a face angle prediction method, device, equipment and readable storage medium, which can avoid the difficulty of predicting only the approximate posture of a face and the difficulty in accurately predicting the face angle, and can adapt to the relatively low accuracy of face angle prediction. High scene.
第一方面,本申请提供一种人脸角度预测方法,包括:In the first aspect, this application provides a face angle prediction method, including:
获取待测人脸图像的人脸区域;Obtain the face area of the face image to be tested;
确定所述人脸区域对应的人脸特征;Determine the facial features corresponding to the facial area;
根据所述人脸特征,确定多种角度类型各自的多个角度概率,每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间,所述多种角度类型包括偏航角、俯仰角和翻滚角;According to the facial features, multiple angle probabilities of multiple angle types are determined. The multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type. The multiple angle types include yaw. angle, pitch and roll angles;
根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。According to the multiple angle probabilities of each angle type, the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position is determined.
本申请通过人脸区域对应的人脸特征,确定多种角度类型各自的多个角度概率,再根据每种角度类型的多个角度概率,确定待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。由于获取的多个角度概率为多个角度区间对应的角度概率,通过多个角度区间对应角度概率计算对应的预测角度,保证了预测角度的准确率,避免了难以对人脸角度准确预测的情况,可以适应人脸角度预测的准确度较高的场景。This application determines the multiple angle probabilities of multiple angle types through the facial features corresponding to the face area, and then determines the relative shooting position of the face in the face image to be measured based on the multiple angle probabilities of each angle type. The predicted angle for each angle type. Since the obtained multiple angle probabilities are angle probabilities corresponding to multiple angle intervals, the corresponding prediction angle is calculated through the angle probabilities corresponding to multiple angle intervals, ensuring the accuracy of the prediction angle and avoiding the situation where it is difficult to accurately predict the face angle. , which can be adapted to scenes with high accuracy in face angle prediction.
第二方面,本申请提供了一种人脸角度预测装置,该装置用于执行上述第一方面或第一方面的任一可能的实现方式中的方法。具体地,该装置可以包括:In a second aspect, this application provides a face angle prediction device, which is used to perform the method in the above-mentioned first aspect or any possible implementation of the first aspect. Specifically, the device may include:
获取模块,用于获取待测人脸图像的人脸区域;The acquisition module is used to obtain the face area of the face image to be tested;
第一确定模块,用于确定所述人脸区域对应的人脸特征;The first determination module is used to determine the facial features corresponding to the facial area;
第二确定模块,用于根据所述人脸特征,确定多种角度类型各自的多个角度概率,每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间,所述多种角度类型包括偏航角、俯仰角和翻滚角;The second determination module is used to determine multiple angle probabilities of multiple angle types according to the facial features, and the multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type, said Multiple angle types including yaw, pitch, and roll angles;
第三确定模块,用于根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。The third determination module is configured to determine the predicted angle of each angle type of the human face in the face image to be measured relative to the shooting position based on multiple angle probabilities of each angle type.
第三方面,本申请提供了一种电子设备,该设备包括存储器与处理器。该存储器用于存储指令;该处理器执行该存储器存储的指令,使得该设备执行第一方面或第一方面的任一可能的实现方式中人脸角度预测方法。In a third aspect, the present application provides an electronic device, which includes a memory and a processor. The memory is used to store instructions; the processor executes the instructions stored in the memory, so that the device performs the face angle prediction method in the first aspect or any possible implementation of the first aspect.
第四方面,提供一种计算机可读存储介质,该计算机可读存储介质中存储有指令,当该指令在计算机上运行时,使得计算机执行第一方面或第一方面的任一可能的实现方式中人脸角度预测方法。In a fourth aspect, a computer-readable storage medium is provided. Instructions are stored in the computer-readable storage medium. When the instructions are run on a computer, they cause the computer to execute the first aspect or any possible implementation of the first aspect. Face angle prediction method.
第五方面,提供一种包含指令的计算机程序产品,当该指令在设备上运行时,使得设备执行第一方面或第一方面的任一可能的实现方式中人脸角度预测方法。A fifth aspect provides a computer program product containing instructions that, when run on a device, cause the device to execute the face angle prediction method of the first aspect or any possible implementation of the first aspect.
可以理解的是,上述第二方面至第五方面的有益效果可以参见上述第一方面中的相关描述,在此不再赘述。It can be understood that the beneficial effects of the above-mentioned second aspect to the fifth aspect can be referred to the relevant description in the above-mentioned first aspect, and will not be described again here.
附图说明Description of the drawings
为了更清楚地说明本申请中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to explain the technical solutions in this application more clearly, the drawings needed to be used in the embodiments or description of the prior art will be briefly introduced below. Obviously, the drawings in the following description are only some implementations of this application. For example, for those of ordinary skill in the art, other drawings can be obtained based on these drawings without exerting creative efforts.
图1是本申请一实施例提供的人脸角度预测方法的流程示意图;Figure 1 is a schematic flowchart of a face angle prediction method provided by an embodiment of the present application;
图2是本申请一实施例提供的人脸角度预测方法的流程示意图;Figure 2 is a schematic flow chart of a face angle prediction method provided by an embodiment of the present application;
图3是本申请一实施例提供的人脸角度预测方法的流程示意图;Figure 3 is a schematic flow chart of a face angle prediction method provided by an embodiment of the present application;
图4是本申请一实施例提供的人脸角度预测方法的流程示意图;Figure 4 is a schematic flowchart of a face angle prediction method provided by an embodiment of the present application;
图5是本申请一实施例提供的人脸角度预测装置的结构示意图;Figure 5 is a schematic structural diagram of a face angle prediction device provided by an embodiment of the present application;
图6是本申请一实施例提供的电子设备的结构示意图。FIG. 6 is a schematic structural diagram of an electronic device provided by an embodiment of the present application.
本发明的实施方式Embodiments of the invention
以下描述中,为了说明而不是为了限定,提出了诸如特定系统结构、技术之类的具体细节,以便透彻理解本申请。然而,本领域的技术人员应当清楚,在没有这些具体细节的其它实施例中也可以实现本申请。在其它情况中,省略对众所周知的系统、装置、电路以及方法的详细说明,以免不必要的细节妨碍本申请的描述。In the following description, specific details, such as specific system structures and technologies, are provided for purposes of explanation and not limitation, in order to provide a thorough understanding of the present application. However, it will be apparent to those skilled in the art that the present application may be practiced in other embodiments without these specific details. In other instances, detailed descriptions of well-known systems, devices, circuits, and methods are omitted so as not to obscure the description of the present application with unnecessary detail.
应当理解,当在本申请说明书和所附权利要求书中使用时,术语“包括”指示所描述特征、整体、步骤、操作、元素和/或组件的存在,但并不排除一个或多个其它特征、整体、步骤、操作、元素、组件和/或其集合的存在或添加。It will be understood that, when used in this specification and the appended claims, the term "comprising" indicates the presence of the described features, integers, steps, operations, elements and/or components but does not exclude one or more other The presence or addition of features, integers, steps, operations, elements, components and/or collections thereof.
还应当理解,在本申请说明书和所附权利要求书中使用的术语“和/或”是指相关联列出的项中的一个或多个的任何组合以及所有可能组合,并且包括这些组合。It will also be understood that the term "and/or" as used in this specification and the appended claims refers to and includes any and all possible combinations of one or more of the associated listed items.
如在本申请说明书和所附权利要求书中所使用的那样,术语“如果”可以依据上下文被解释为“当...时”或“一旦”或“响应于确定”或“响应于检测到”。类似地,短语“如果确定”或“如果检测到[所描述条件或事件]”可以依据上下文被解释为意指“一旦确定”或“响应于确定”或“一旦检测到[所描述条件或事件]”或“响应于检测到[所描述条件或事件]”。As used in this specification and the appended claims, the term "if" may be interpreted as "when" or "once" or "in response to determining" or "in response to detecting" depending on the context. ". Similarly, the phrase "if determined" or "if [the described condition or event] is detected" may be interpreted, depending on the context, to mean "once determined" or "in response to a determination" or "once the [described condition or event] is detected ]" or "in response to detection of [the described condition or event]".
另外,在本申请说明书和所附权利要求书的描述中,术语“第一”、“第二”、“第三”等仅用于区分描述,而不能理解为指示或暗示相对重要性。In addition, in the description of this application and the appended claims, the terms "first", "second", "third", etc. are only used to distinguish the description, and cannot be understood as indicating or implying relative importance.
在本申请说明书中描述的参考“一个实施例”或“一些实施例”等意味着在本申请的一个或多个实施例中包括结合该实施例描述的特定特征、结构或特点。由此,在本说明书中的不同之处出现的语句“在一个实施例中”、“在一些实施例中”、“在其他一些实施例中”、“在另外一些实施例中”等不是必然都参考相同的实施例,而是意味着“一个或多个但不是所有的实施例”,除非是以其他方式另外特别强调。术语“包括”、“包含”、“具有”及它们的变形都意味着“包括但不限于”,除非是以其他方式另外特别强调。Reference in this specification to "one embodiment" or "some embodiments" or the like means that a particular feature, structure or characteristic described in connection with the embodiment is included in one or more embodiments of the application. Therefore, the phrases "in one embodiment", "in some embodiments", "in other embodiments", "in other embodiments", etc. appearing in different places in this specification are not necessarily References are made to the same embodiment, but rather to "one or more but not all embodiments" unless specifically stated otherwise. The terms “including,” “includes,” “having,” and variations thereof all mean “including but not limited to,” unless otherwise specifically emphasized.
本申请提供一种人脸角度预测方法、装置、设备及可读存储介质,该方法可以通过识别设备实现,且应用于门禁识别、失踪人口查找、案件侦查、智能安防等场景中。This application provides a face angle prediction method, device, equipment and readable storage medium. The method can be implemented through recognition equipment and applied in access control recognition, missing person search, case investigation, intelligent security and other scenarios.
其中,人脸角度包括人脸相对于拍摄位置的三种角度类型,该三种角度类型分别为俯仰角、偏航角和翻滚角。Among them, the face angle includes three angle types of the face relative to the shooting position, and the three angle types are pitch angle, yaw angle and roll angle respectively.
其中,识别设备指的是用户进行人脸角度预测时的设备。识别设备可以为门禁设备、智能手机、台式电脑、笔记本电脑、平板电脑、可穿戴设备、手持设备、车载设备、服务器等。本申请实施例对识别设备的具体类型不作任何限制。Among them, the recognition device refers to the device used by users to predict face angles. Identification devices can be access control devices, smartphones, desktop computers, laptops, tablets, wearable devices, handheld devices, vehicle-mounted devices, servers, etc. The embodiments of this application do not place any restrictions on the specific types of identification devices.
识别设备可以包括显示屏硬件,或者外接有显示屏。The identification device may include display hardware, or may have an external display.
关于上述场景,下面进行举例说明人脸角度预测方法的应用:Regarding the above scenario, the following is an example to illustrate the application of the face angle prediction method:
1、在门禁识别场景下使用人脸角度预测方法时,识别设备可根据图像中的人脸对人脸角度进行预测,通过预测得到的人脸角度确定是否可以进行下一步的动作,例如,当人脸角度过大时,在门禁设备的显示屏上显示“人脸角度过大,无法识别”等提示信息。1. When using the face angle prediction method in an access control recognition scenario, the recognition device can predict the face angle based on the face in the image, and determine whether the next action can be taken based on the predicted face angle. For example, when When the angle of the face is too large, a prompt message such as "The angle of the face is too large and cannot be recognized" will be displayed on the display screen of the access control device.
2、在失踪人口查找场景下使用人脸角度预测方法时,用于将人脸图像输入到识别设备中,对人脸图像中的人脸对应的人脸角度进行预测,通过预测得到的人脸角度确定是否可以进行下一步的动作,例如,确定识别设备是否能够对人脸进行识别,以确定人脸图像是否与失踪人口图像库中的图像匹配,进而确定是否为失踪人口。2. When using the face angle prediction method in the missing person search scenario, it is used to input the face image into the recognition device and predict the face angle corresponding to the face in the face image. The face obtained through prediction The angle determines whether the next step can be taken, for example, determining whether the recognition device can recognize the face, determining whether the face image matches the image in the missing person image library, and then determining whether it is a missing person.
或者,识别装置可与监控摄像头通信连接,识别设备可通过获取监控摄像头拍摄的图像,对图像中的人脸对应的人脸角度进行预测。Alternatively, the recognition device can be connected through communication with the surveillance camera, and the recognition device can predict the face angle corresponding to the face in the image by acquiring the image captured by the surveillance camera.
基于上述场景描述,下面,以识别设备为例,结合附图和应用场景,对本申请实施例提供的人脸角度预测方法进行详细说明。Based on the above scenario description, the face angle prediction method provided by the embodiments of the present application will be described in detail below, taking the recognition device as an example and combining the drawings and application scenarios.
请参阅图1,图1示出了本申请一实施例提供的人脸角度预测方法的流程示意图。Please refer to FIG. 1 , which shows a schematic flowchart of a face angle prediction method provided by an embodiment of the present application.
如图1所示,本申请提供的人脸角度预测方法可以包括:As shown in Figure 1, the face angle prediction method provided by this application can include:
S101、获取待测人脸图像的人脸区域。S101. Obtain the face area of the face image to be measured.
待测人脸图像可以是用户直接给定的,也可以是从监控摄像头、摄像机等图像采集设备采集的视频数据中抽取的。The face image to be tested can be directly given by the user, or it can be extracted from video data collected by image collection equipment such as surveillance cameras and video cameras.
人脸区域指待测人脸图像中含有人脸的区域。The face area refers to the area containing faces in the face image to be measured.
在一些实施例中,人脸区域是通过对待测人脸图像进行人脸检测,获得第一检测窗口,对第一检测窗口中的图像进行截取得到的。In some embodiments, the face area is obtained by performing face detection on the face image to be tested, obtaining the first detection window, and intercepting the image in the first detection window.
可选的,识别设备可对第一检测窗口进行外扩处理,得到外扩后的第二检测窗口,截取第二检测窗口对应大小的区域作为人脸区域。Optionally, the recognition device can perform an external expansion process on the first detection window to obtain an expanded second detection window, and intercept an area corresponding to the size of the second detection window as the face area.
可以理解的是,检测窗口是指可提取待测人脸图像中人脸的线框。It can be understood that the detection window refers to the wire frame from which the face in the face image to be detected can be extracted.
其中,对待测人脸进行人脸检测可采用人脸检测算法。Among them, the face detection algorithm can be used to detect the face to be tested.
人脸检测算法可存储在存储设备中。The face detection algorithm can be stored in the storage device.
其中,存储设备可与识别设备进行通信,使得识别设备能够从存储设备中调取人脸检测算法。本申请对存储设备的存储方式和具体类型不做限定。The storage device can communicate with the recognition device, so that the recognition device can retrieve the face detection algorithm from the storage device. This application does not limit the storage method and specific type of storage devices.
在一些实施例中,采用YOLO(you only look once)算法进行人脸检测。YOLO算法是一种基于深度神经网络的对象识别和定位算法,其最大的特点是运行速度很快。In some embodiments, the YOLO (you only look once) algorithm is used for face detection. The YOLO algorithm is an object recognition and positioning algorithm based on deep neural networks. Its biggest feature is its fast running speed.
在一个具体的实施例中,假设识别设备为门禁设备,门禁设备包括摄像头。在人脸靠近门禁设备的摄像头时,摄像头拍摄到人脸图像,门禁设备通过人脸检测算法对人脸图像进行人脸检测,获取待测人脸图像的人脸区域。In a specific embodiment, it is assumed that the identification device is an access control device, and the access control device includes a camera. When a person's face is close to the camera of the access control device, the camera captures a face image. The access control device uses a face detection algorithm to detect the face image and obtain the face area of the face image to be measured.
在另一个具体的实施例中,假设识别设备为手机,手机上具有识别小程序,手机通过识别小程序与监控摄像头通信连接,获取监控摄像头拍摄的图像。识别小程序可通过人脸检测算法对监控摄像头拍摄的图像进行人脸检测,获取待测人脸图像的人脸区域。In another specific embodiment, it is assumed that the recognition device is a mobile phone, and the mobile phone has a recognition applet. The mobile phone communicates with the surveillance camera through the recognition applet to obtain images captured by the surveillance camera. The recognition applet can use the face detection algorithm to detect faces in images captured by surveillance cameras and obtain the face area of the face image to be detected.
S102、确定所述人脸区域对应的人脸特征。S102. Determine the facial features corresponding to the facial area.
基于S101,识别设备可获得人脸区域。从而,识别设备可对人脸区域进行特征提取,获得人脸区域对应的人脸特征。Based on S101, the recognition device can obtain the face area. Therefore, the recognition device can perform feature extraction on the face area and obtain the facial features corresponding to the face area.
在一些实施例中,识别设备通过将所述人脸区域输入到人脸角度识别模型的骨干网络中,输出所述人脸特征。In some embodiments, the recognition device outputs the facial features by inputting the facial region into the backbone network of the facial angle recognition model.
骨干网络用于提取人脸图像中的人脸特征。The backbone network is used to extract facial features from facial images.
其中,骨干网络预先存储在与识别设备通信的存储设备中。Among them, the backbone network is stored in advance in a storage device that communicates with the identification device.
在一个具体的实施例中,假设识别设备为门禁设备。在人脸靠近门禁设备的摄像头时,摄像头拍摄到人脸图像,门禁设备通过人脸检测算法对人脸图像进行人脸检测,获取人脸图像的人脸区域后,调取骨干网络对人脸区域进行特征提取,得到人脸特征。In a specific embodiment, it is assumed that the identification device is an access control device. When a person's face is close to the camera of the access control device, the camera captures the face image. The access control device performs face detection on the face image through the face detection algorithm. After obtaining the face area of the face image, it calls the backbone network to detect the face. Feature extraction is performed on the region to obtain facial features.
在另一个具体的实施例中,假设识别设备为手机,手机上具有识别小程序。识别小程序通过人脸检测算法对该图像进行人脸检测,获取图像对应的人脸区域后,调取骨干网络对人脸区域进行特征提取,得到人脸特征。In another specific embodiment, it is assumed that the identification device is a mobile phone, and the mobile phone has an identification applet. The recognition applet performs face detection on the image through the face detection algorithm. After obtaining the face area corresponding to the image, it calls the backbone network to extract features of the face area to obtain the face features.
S103、根据所述人脸特征,确定多种角度类型各自的多个角度概率。S103. Determine multiple angle probabilities of multiple angle types according to the facial features.
其中,多种角度类型包括偏航角、俯仰角和翻滚角。Among them, various angle types include yaw angle, pitch angle and roll angle.
每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间。Multiple angle probabilities for each angle type respectively correspond to multiple angle intervals for each angle type.
多个角度区间指对偏航角、俯仰角和翻滚角的角度范围按照预设规则进行划分得到的各自的多个角度区间。Multiple angle intervals refer to multiple angle intervals obtained by dividing the angle ranges of yaw angle, pitch angle and roll angle according to preset rules.
在一些实施例中,预测规则为对角度范围每5度划分一个区间。In some embodiments, the prediction rule is to divide the angle range into intervals every 5 degrees.
例如,偏航角和俯仰角的角度范围为[-90,90],对偏航角和俯仰角的角度范围每5度划分一个区间,分别得到36个角度区间。For example, the angle range of yaw angle and pitch angle is [-90, 90]. The angle range of yaw angle and pitch angle is divided into intervals every 5 degrees, resulting in 36 angle intervals respectively.
偏航角的36个角度区间分别为[-90,-85)、[-85,-80)……(80,85]、(85,90]。The 36 angle intervals of the yaw angle are [-90, -85), [-85, -80)... (80, 85], (85, 90].
俯仰角的36个角度区间分别为[-90,-85)、[-85,-80)……(80,85]、(85,90]。The 36 angle intervals of the pitch angle are [-90, -85), [-85, -80)... (80, 85], (85, 90].
例如,翻滚角的角度范围为[-180,180],对翻滚角的角度范围每5度划分一个区间,分别得到72个角度区间。For example, the angle range of the roll angle is [-180, 180]. The angle range of the roll angle is divided into intervals every 5 degrees, resulting in 72 angle intervals.
翻滚角的72个角度区间分别为[-180,-175)、[-175,-170)、[-170,-165)……(165,170]、(170,175]、(175,180]。The 72 angle intervals of the roll angle are [-180, -175), [-175, -170), [-170, -165)... (165, 170], (170, 175], (175, 180 ].
在一些实施例中,识别设备将所述人脸特征输入到所述人脸角度识别模型的全连接分类网络中,输出每种角度类型的多个角度概率。In some embodiments, the recognition device inputs the facial features into the fully connected classification network of the facial angle recognition model and outputs multiple angle probabilities for each angle type.
其中,全连接分类网络用于预测人脸特征在每种角度类型的多个角度区间分别对应的角度概率。Among them, the fully connected classification network is used to predict the angle probabilities corresponding to the facial features in multiple angle intervals of each angle type.
骨干网络和全连接分类网络作为人脸角度识别模型,预先存储在与识别设备通信的存储设备中。The backbone network and fully connected classification network serve as face angle recognition models and are pre-stored in a storage device that communicates with the recognition device.
具体地,全连接分类网络连接在骨干网络的输出端,当识别设备骨干网络提取到待测人脸图像的人脸特征后,将人脸特征输入至全连接分类网络进行角度概率预测。Specifically, the fully connected classification network is connected to the output end of the backbone network. After the recognition device backbone network extracts the facial features of the face image to be tested, the facial features are input to the fully connected classification network for angle probability prediction.
其中,全连接分类网络包括3个全连接层,分别为第一全连接层、第二全连接层和第三全连接层。第一全连接层和第二全连接层分别连接有36个节点,第三全连接层连接有72个支点。Among them, the fully connected classification network includes three fully connected layers, namely the first fully connected layer, the second fully connected layer and the third fully connected layer. The first fully connected layer and the second fully connected layer are connected to 36 nodes respectively, and the third fully connected layer is connected to 72 pivots.
可以理解的是,第一全连接层和第二全连接层的36个节点分别用于预测偏航角和俯仰角在其36个角度区间的角度概率,第三全连接层的72个节点用于预测翻滚角在其72个角度区间的角度概率。It can be understood that the 36 nodes of the first fully connected layer and the second fully connected layer are used to predict the angular probability of the yaw angle and the pitch angle in their 36 angle intervals, and the 72 nodes of the third fully connected layer are used It is used to predict the angle probability of roll angle in its 72 angle intervals.
因而,偏航角和俯仰角各自的多个角度概率具有36个,翻滚角的多个角度概率具有72个。Therefore, there are 36 multiple-angle probabilities for each of the yaw angle and the pitch angle, and there are 72 multiple-angle probabilities for the roll angle.
在一个具体的实施例中,假设识别设备为门禁设备。门禁设备在获取人脸特征后,根据人脸特征分别预测偏航角和俯仰角在其36个角度区间的角度概率,以及预测翻滚角在其72个角度区间的角度概率。In a specific embodiment, it is assumed that the identification device is an access control device. After acquiring the facial features, the access control device predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features.
在另一个具体的实施例中,假设识别设备为手机,手机上具有识别小程序。识别小程序在获取人脸特征后,根据人脸特征分别预测偏航角和俯仰角在其36个角度区间的角度概率,以及预测翻滚角在其72个角度区间的角度概率。In another specific embodiment, it is assumed that the identification device is a mobile phone, and the mobile phone has an identification applet. After obtaining the facial features, the recognition applet predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features.
S104、根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。S104. Based on the multiple angle probabilities of each angle type, determine the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position.
在一些实施例中,针对每种角度类型而言,根据多个角度概率、所述角度区间的数量和每个所述角度区间的中间角度,确定所述待测人脸图像中的人脸相对于拍摄位置的角度。In some embodiments, for each angle type, the relative angle of the face in the face image to be measured is determined based on multiple angle probabilities, the number of angle intervals, and the intermediate angle of each angle interval. The angle of the shooting position.
其中,每个角度区间的中间角度指每5个角度为一个角度区间中的中间的角度。例如,某个角度区间为[0,5),那么,对应的中间角度为2.5度。Among them, the middle angle of each angle interval refers to every five angles being the middle angle in an angle interval. For example, if a certain angle range is [0, 5), then the corresponding intermediate angle is 2.5 degrees.
上述确定待测人脸图像中的人脸相对于拍摄位置的角度的计算公式为:The above calculation formula for determining the angle of the face in the face image to be measured relative to the shooting position is:
每个角度类型的预测角度= Predicted angles for each angle type =
其中,n表示角度区间的数量, 表示第i个角度区间的中间角度, 表示第i个角度区间的角度概率。 Among them, n represents the number of angle intervals, Represents the middle angle of the i-th angle interval, Represents the angle probability of the i-th angle interval.
在一个具体的实施例中,假设识别设备为门禁设备,门禁设备包括显示屏和摄像头。门禁设备在根据人脸特征分别预测偏航角和俯仰角在其36个角度区间的角度概率,以及预测翻滚角在其72个角度区间的角度概率后,根据偏航角和俯仰角在其36个角度区间的角度概率计算对应的每个角度区间的值,根据翻滚角在其72个角度区间的角度概率计算对每个区间对应的值,并根据每个区间对应的值计算最终的人脸的预测角度。门禁设备可通过人脸的预测角度,确定是否可以进行下一步的动作,例如,当人脸角度过大时,在门禁设备的显示屏上显示“人脸角度过大,无法识别”等提示信息。In a specific embodiment, it is assumed that the identification device is an access control device, and the access control device includes a display screen and a camera. After predicting the angular probability of the yaw angle and the pitch angle in its 36 angle intervals based on the facial features, and predicting the angular probability of the roll angle in its 72 angle interval, the access control device predicts the angular probability of the yaw angle and pitch angle in its 36 angle interval based on the facial features. Calculate the angle probability of each angle interval corresponding to the value of each angle interval, calculate the value corresponding to each interval based on the angle probability of the roll angle in its 72 angle intervals, and calculate the final face based on the value corresponding to each interval prediction angle. The access control device can determine whether the next action can be taken based on the predicted angle of the face. For example, when the face angle is too large, a prompt message such as "The face angle is too large to be recognized" will be displayed on the display of the access control device. .
在另一个具体的实施例中,假设识别设备为手机,手机上具有识别小程序。识别小程序在根据人脸特征确定人脸特征对应的,根据人脸特征分别预测偏航角和俯仰角在其36个角度区间的角度概率,以及预测翻滚角在其72个角度区间的角度概率后,根据偏航角和俯仰角在其36个角度区间的角度概率计算对应的每个角度区间的值,根据翻滚角在其72个角度区间的角度概率计算对每个区间对应的值,并根据每个区间对应的值计算最终的人脸的预测角度。识别小程序可通过人脸的预测角度,确定是否可以进行下一步的动作,例如,当人脸角度过大时,对待测人脸图像进行校正。In another specific embodiment, it is assumed that the identification device is a mobile phone, and the mobile phone has an identification applet. The recognition applet determines the corresponding facial features based on the facial features, predicts the angle probabilities of the yaw angle and the pitch angle in its 36 angle intervals, and predicts the angle probability of the roll angle in its 72 angle intervals based on the facial features. Finally, calculate the corresponding value of each angle interval based on the angle probability of the yaw angle and pitch angle in its 36 angle intervals, calculate the corresponding value of each interval based on the angle probability of the roll angle in its 72 angle intervals, and Calculate the final predicted angle of the face based on the value corresponding to each interval. The recognition applet can use the predicted angle of the face to determine whether the next action can be taken. For example, when the angle of the face is too large, the face image to be measured is corrected.
本申请提供的人脸角度预测方法,通过根据待测人脸图像的人脸区域获取人脸特征,再根据人脸特征,确定多种角度类型各自的多个角度概率,最后根据每种角度类型的多个角度概率,确定待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。由此,对每种角度类型而言,由于获取的多个角度概率为多个角度区间对应的角度概率,通过多个角度区间对应角度概率计算预测角度,保证了预测角度的准确率。The face angle prediction method provided by this application obtains face features based on the face area of the face image to be measured, and then determines multiple angle probabilities for each of the multiple angle types based on the face features. Finally, based on each angle type The multiple angle probabilities are used to determine the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position. Therefore, for each angle type, since the obtained multiple angle probabilities are angle probabilities corresponding to multiple angle intervals, the predicted angle is calculated based on the angle probabilities corresponding to multiple angle intervals, ensuring the accuracy of the predicted angle.
基于上述图2所示S103实施例的描述,识别设备可获取翻滚角的多个角度概率的最大值,然而,当最大值对应预设的映射角度区间时,计算的翻滚角的预测角度不准确,可采用多种方式进行处理,保证获得更加准确的预测角度。Based on the description of the S103 embodiment shown in Figure 2 above, the recognition device can obtain the maximum value of multiple angle probabilities of the roll angle. However, when the maximum value corresponds to the preset mapping angle interval, the calculated predicted angle of the roll angle is inaccurate. , can be processed in a variety of ways to ensure a more accurate prediction angle.
其中,翻滚角的多个角度区间的范围为[-180,180]。Among them, the range of multiple angle intervals of the roll angle is [-180, 180].
映射角度区间包括[-180,-90)对应的多个角度区间或者(90,180]对应的多个角度区间。The mapping angle interval includes multiple angle intervals corresponding to [-180, -90) or multiple angle intervals corresponding to (90, 180].
下面,结合图2,详细介绍本申请的人脸角度预测方法的具体实现过程。Next, with reference to Figure 2, the specific implementation process of the face angle prediction method of this application is introduced in detail.
请参阅图2,图2示出了本申请一实施例提供的人脸角度预测方法的流程示意图。Please refer to Figure 2. Figure 2 shows a schematic flow chart of a face angle prediction method provided by an embodiment of the present application.
如图2所示,本申请提供的人脸角度预测方法可以包括:As shown in Figure 2, the face angle prediction method provided by this application may include:
S201、针对所述翻滚角而言,确定最大概率角度区间。S201. For the roll angle, determine the maximum probability angle interval.
最大概率角度区间为多个角度概率中的最大值对应的角度区间。The maximum probability angle interval is the angle interval corresponding to the maximum value among multiple angle probabilities.
例如,翻滚角的多个角度概率中的最大值对应于角度区间[-175,-170)。For example, the maximum value among multiple angle probabilities for roll angle corresponds to the angle interval [-175, -170).
S202、当所述最大概率角度区间处于预设的映射角度区间内时,对所述翻滚角的多个角度概率进行线性映射,得到映射后的多个角度概率。S202. When the maximum probability angle interval is within a preset mapping angle interval, linearly map multiple angle probabilities of the roll angle to obtain multiple mapped angle probabilities.
由于翻滚角的角度范围为[-180,180],即为一个圆周,基于翻滚角的这种周期性,在最大概率角度区间处于[-180,-90)对应的边缘的角度区间或者(90,180]对应的边缘的角度区间内时,认为人脸的翻滚角的角度较大。当人脸的翻滚角的角度较大时,识别设备按照图1中S103所示的方法计算翻滚角的预测角度不准确。Since the angle range of the roll angle is [-180, 180], which is a circle, based on the periodicity of the roll angle, the angle interval of the edge corresponding to the maximum probability angle interval is [-180, -90) or (90 , 180] is within the angle range of the corresponding edge, the roll angle of the face is considered to be larger. When the roll angle of the face is larger, the recognition device calculates the roll angle according to the method shown in S103 in Figure 1 The predicted angle is inaccurate.
因此,识别设备在确定翻滚角对应的最大概率角度区间处于[-90,90]的角度区间时,可以按照图1中S103所示的方法计算翻滚角的预测角度。Therefore, when the recognition device determines that the maximum probability angle interval corresponding to the roll angle is in the angle interval of [-90, 90], it can calculate the predicted angle of the roll angle according to the method shown in S103 in Figure 1.
而,识别设备在确定翻滚角对应的最大概率角度区间处于[-180,-90)对应的角度区间或者(90,180]对应的角度区间内时,识别设备需要对翻滚角的多个角度概率进行线性映射,得到映射后的多个角度概率。However, when the recognition device determines that the maximum probability angle interval corresponding to the roll angle is within the angle interval corresponding to [-180, -90) or the angle interval corresponding to (90, 180], the recognition device needs to determine the multiple angle probabilities of the roll angle. Perform linear mapping to obtain multiple angle probabilities after mapping.
识别设备通过对多个角度概率进行线性映射,以便于翻滚角对应的最大概率角度区间处于[-180,90)或(90,180]的角度区间时,也能得到精确的翻滚角的角度值。The recognition device linearly maps multiple angle probabilities so that when the maximum probability angle interval corresponding to the roll angle is in the angle interval of [-180, 90) or (90, 180], the accurate angle value of the roll angle can be obtained .
在一些实施例中,线性映射指:In some embodiments, linear mapping refers to:
将大于等于0度且小于等于180度的角度区间对应的角度概率替换为大于等于-180度且小于0度的角度区间对应的映射后的角度概率;Replace the angle probability corresponding to the angle interval greater than or equal to 0 degrees and less than or equal to 180 degrees with the mapped angle probability corresponding to the angle interval greater than or equal to -180 degrees and less than 0 degrees;
以及,将小于等于0度且大于等于-180度的角度区间对应的角度概率替换为小于等于180度且大于0度的角度区间对应的映射后的角度概率。And, replace the angle probabilities corresponding to the angle intervals of less than or equal to 0 degrees and greater than or equal to -180 degrees with the mapped angle probabilities corresponding to the angle intervals of less than or equal to 180 degrees and greater than 0 degrees.
举例说明,翻滚角的多个角度概率中的最大值对应于角度区间[-175,-170)时,如果直接按照图1中S103所示的方法计算翻滚角的预测角度,得到的预测角度值为可能为-103.5度,显然,-103.5度不处于角度区间[-175,-170)中,这是不合理的。For example, when the maximum value among the multiple angle probabilities of the roll angle corresponds to the angle interval [-175, -170), if the predicted angle of the roll angle is calculated directly according to the method shown in S103 in Figure 1, the predicted angle value obtained It may be -103.5 degrees. Obviously, -103.5 degrees is not in the angle interval [-175, -170), which is unreasonable.
因此,识别设备需要先对翻滚角的多个角度概率进行映射处理,例如,将角度区间[-175,-170)对应的角度概率映射到角度区间(-10,-5],也就是说,将角度区间[-175,-170)对应的角度概率替换为角度区间(-10,-5]对应的角度概率。Therefore, the recognition device needs to first map multiple angle probabilities of the roll angle, for example, map the angle probability corresponding to the angle interval [-175, -170) to the angle interval (-10, -5], that is, Replace the angle probability corresponding to the angle interval [-175, -170) with the angle probability corresponding to the angle interval (-10, -5).
S203、根据所述映射后的多个角度概率、所述翻滚角的角度区间的数量和每个角度区间的中间角度,确定映射角度。S203. Determine the mapping angle according to the mapped multiple angle probabilities, the number of angle intervals of the roll angle, and the intermediate angle of each angle interval.
举例说明,根据角度区间(-10,-5]对应的角度概率,计算得到的预测角度值(映射角度)可以为-5度,显然-5度处于角度区间(-10,-5]中,这是合理的。For example, according to the angle probability corresponding to the angle interval (-10, -5], the calculated predicted angle value (mapping angle) can be -5 degrees. Obviously -5 degrees is in the angle interval (-10, -5]. This is reasonable.
S204、对所述映射角度进行反线性映射,得到所述翻滚角的预测角度。S204: Perform inverse linear mapping on the mapping angle to obtain the predicted angle of the roll angle.
反线性映射指将根据角度区间(-10,-5]对应的角度概率,计算得到的预测角度值(映射角度)反映射到角度区间[-175,-170)中。Inverse linear mapping refers to mapping the predicted angle value (mapping angle) calculated based on the angle probability corresponding to the angle interval (-10, -5) to the angle interval [-175, -170).
举例说明,基于S203得到的角度值为-5度时,将-5度进行反映射,得到的角度值为-175度,处于角度区间[-175,170)中,这是合理的。For example, when the angle value obtained based on S203 is -5 degrees, -5 degrees are de-reflected, and the obtained angle value is -175 degrees, which is in the angle interval [-175, 170), which is reasonable.
本申请中,针对翻滚角而言,识别设备在确定多个角度概率的最大值对应于预设的映射区间时,对多个角度概率进行线性映射,得到映射后的多个角度概率,再根据映射后的多个角度概率、翻滚角的角度区间的数量和每个角度区间的中间角度,确定映射角度,并对映射角度进行反线性映射,预测角度。在确定多个角度概率的最大值对应于预设的映射区间时,借助线性映射将对应于映射角度区间的角度概率进行映射,再计算映射角度,并根据映射角度计算预测角度,可以获得更加精确的预测角度。In this application, for the roll angle, when the identification device determines that the maximum value of multiple angle probabilities corresponds to the preset mapping interval, the multiple angle probabilities are linearly mapped to obtain the mapped multiple angle probabilities, and then based on The mapping angle is determined based on the mapped multiple angle probabilities, the number of angle intervals of the roll angle and the middle angle of each angle interval, and the mapping angle is inversely mapped to predict the angle. When determining that the maximum value of multiple angle probabilities corresponds to the preset mapping interval, linear mapping is used to map the angle probabilities corresponding to the mapping angle interval, and then the mapping angle is calculated, and the predicted angle is calculated based on the mapping angle, so that more accuracy can be obtained prediction angle.
基于上述图1所示实施例S101的描述,识别设备在确定人脸区域时,可对第一检测窗口进行外扩处理,得到第二检测窗口,截取第二检测窗口对应大小的区域,将第二检测窗口对应大小的区域确定为人脸区域。Based on the description of the embodiment S101 shown in Figure 1 above, when the recognition device determines the face area, it can expand the first detection window to obtain the second detection window, intercept the area corresponding to the size of the second detection window, and add the second detection window to the second detection window. The area corresponding to the size of the two detection windows is determined as the face area.
下面,结合图3,详细介绍本申请的人脸角度预测方法的具体实现过程。Next, with reference to Figure 3, the specific implementation process of the face angle prediction method of this application is introduced in detail.
基于图1中S101的描述,可通过对第一检测窗口进行外扩处理,以获取待测人脸图像对应的更多的人脸信息,保证最终得到的人脸角度精确度更高。Based on the description of S101 in Figure 1, the first detection window can be expanded to obtain more face information corresponding to the face image to be measured, ensuring that the final face angle accuracy is higher.
请参阅图3,图3示出了本申请一实施例提供的人脸角度预测方法的流程示意图。Please refer to FIG. 3 , which shows a schematic flowchart of a face angle prediction method provided by an embodiment of the present application.
如图3所示,本申请提供的人脸角度预测方法可以包括:As shown in Figure 3, the face angle prediction method provided by this application may include:
S301、将以所述第一检测窗口的中心为中心,所述第一检测窗口的长边为边长的窗口作为第三检测窗口。S301. A window with the center of the first detection window as the center and the long side of the first detection window as the third detection window.
在一些实施例中,对待测人脸图像进行人脸检测,获取的第一检测窗口为长方形。In some embodiments, face detection is performed on the face image to be tested, and the first detection window obtained is a rectangle.
可以理解的是,当第一检测窗口为长方形,识别设备可根据第一检测窗口的长和宽中的长为边长,第一检测窗口的中心为中心,获取第三检测窗口。It can be understood that when the first detection window is a rectangle, the identification device can obtain the third detection window based on the length of the first detection window and the width as the side length and the center of the first detection window as the center.
在另一些实施例中,对待测人脸图像进行人脸检测,获取的第一检测窗口为正方形。In other embodiments, face detection is performed on the face image to be tested, and the first detection window obtained is a square.
可以理解的是,当第一检测窗口为正方形时,可直接根据外扩系数对第一检测窗口进行步骤S302。It can be understood that when the first detection window is a square, step S302 can be directly performed on the first detection window according to the expansion coefficient.
其中,对待测人脸图像进行人脸检测,获取的第一检测窗口为长方形或者正方形。第一检测窗口为长方形或者正方形通常由距离摄像头的远近、人脸表情或动作、人脸的角度等多个方面决定。Among them, when face detection is performed on the face image to be tested, the first detection window obtained is a rectangle or a square. Whether the first detection window is rectangular or square is usually determined by the distance from the camera, the facial expression or movement, the angle of the human face, and other aspects.
在一个具体的实施例中,假设第一检测窗口为长为60个像素点,宽为40个像素点的长方形。那么,以第一检测窗口的中心为中心,第一检测窗口的长为边长的窗口得到的第三检测窗口为边长为60个像素点的正方形。In a specific embodiment, it is assumed that the first detection window is a rectangle with a length of 60 pixels and a width of 40 pixels. Then, taking the center of the first detection window as the center and the length of the first detection window as the side length of the window, the third detection window obtained is a square with a side length of 60 pixels.
S302、根据预设外扩系数,对所述第三检测窗口的每条边长均进行外扩处理,得到第四检测窗口。S302. According to the preset expansion coefficient, perform an expansion process on each side length of the third detection window to obtain a fourth detection window.
在一些实施例中,预设外扩系数为0.1。当然,外扩系数也可以为其他数值,例如0.15,具体可根据实际情况进行设置,在此不多做赘述。In some embodiments, the preset expansion coefficient is 0.1. Of course, the expansion coefficient can also be other values, such as 0.15, which can be set according to the actual situation, and will not be described in detail here.
在一个具体的实施例中,假设第三检测窗口的边长为60个像素点,外扩系数为0.1。那么,对第三检测窗口进行外扩处理,得到的第四检测窗口为边长为66个像素点的正方形。In a specific embodiment, it is assumed that the side length of the third detection window is 60 pixels, and the expansion coefficient is 0.1. Then, the third detection window is expanded, and the fourth detection window obtained is a square with a side length of 66 pixels.
S303、去除所述第四检测窗口超过所述待测人脸图像对应的边长,得到第五检测窗口。S303. Eliminate the side length of the fourth detection window that exceeds the corresponding side length of the face image to be measured to obtain a fifth detection window.
可以理解的是,识别设备在对第三检测窗口外扩处理后,得到的第四检测窗口可能超出原待测人脸图像,即第四检测窗口超过所述待测人脸图像对应的边长。It can be understood that after the recognition device expands the third detection window, the obtained fourth detection window may exceed the original face image to be tested, that is, the fourth detection window exceeds the corresponding side length of the face image to be tested. .
识别设备在确定第四检测窗口超过待测人脸图像对应的边长时,对超过的边长进行去除,去除后得到第五检测窗口。When the recognition device determines that the fourth detection window exceeds the side length corresponding to the face image to be measured, it removes the excess side length, and obtains the fifth detection window after removal.
在一个具体的实施例中,假设待测人脸图像的长为90个像素点,宽为60个像素点,第四检测窗口的边长为66个像素点。去除超过待测人脸图像对应的边长,得到的第五检测窗口的长为66个像素点,宽为60个像素点。In a specific embodiment, it is assumed that the length of the face image to be measured is 90 pixels, the width is 60 pixels, and the side length of the fourth detection window is 66 pixels. After removing the side length corresponding to the face image to be tested, the length of the fifth detection window obtained is 66 pixels and the width is 60 pixels.
S304、将以所述第五检测窗口的中心为中心,所述第五检测窗口的短边为边长的窗口作为第二检测窗口。S304: A window with the center of the fifth detection window as the center and the shorter side of the fifth detection window as the second detection window.
在一个具体的实施例中,假设第五检测窗口的长为66个像素点,宽为60个像素点。那么,以第五检测窗口的中心为中心,第五检测窗口的宽为边长,得到的第二检测窗口为边长为60个像素点的正方形。In a specific embodiment, it is assumed that the length of the fifth detection window is 66 pixels and the width is 60 pixels. Then, taking the center of the fifth detection window as the center and the width of the fifth detection window as the side length, the obtained second detection window is a square with a side length of 60 pixels.
本申请中,识别设备将以第一检测窗口的中心为中心,第一检测窗口的长边为边长的窗口作为第三检测窗口,根据预设外扩系数,对第三检测窗口的每条边长均进行外扩处理,得到第四检测窗口,去除第四检测窗口超过待测人脸图像对应的边长,得到第五检测窗口,将以第五检测窗口的中心为中心,第五检测窗口的短边为边长的窗口作为第二检测窗口。识别设备通过对第一检测窗口外扩处理后得到的第二检测窗口,对待测人脸图像框选的人脸区域更大,包括的人脸信息更多,通过包括更多人脸信息的人脸区域提取人脸特征,得到的人脸特征更精确,通过更精确的人脸特征进行人脸的角度预测,可以获得准确度更高的预测角度。In this application, the identification device will take the center of the first detection window as the center and the long side of the first detection window as the third detection window. According to the preset expansion coefficient, each line of the third detection window will be The side lengths are all expanded to obtain the fourth detection window. Remove the side length of the fourth detection window that exceeds the corresponding side length of the face image to be tested, and obtain the fifth detection window. The center of the fifth detection window will be the center. The short side of the window is the second detection window. The recognition device obtains a second detection window by expanding the first detection window, and the face area selected by the face image to be tested is larger and includes more face information. Facial features are extracted from the face area, and the facial features obtained are more accurate. By predicting the angle of the face with more accurate facial features, a more accurate prediction angle can be obtained.
基于上述图1所示实施例的描述,本申请还提供了包括骨干网络和全连接分类网络的人脸角度识别模型的生成过程。Based on the above description of the embodiment shown in Figure 1, this application also provides a generation process of a face angle recognition model including a backbone network and a fully connected classification network.
下面,结合图4,详细介绍本申请的生成人脸角度识别模型的具体实现过程。Next, with reference to Figure 4, the specific implementation process of generating the face angle recognition model of this application is introduced in detail.
基于图1中S102的描述,识别设备在获取人脸区域对应的人脸特征时,通过人脸角度识别模型中的骨干网络获取。Based on the description of S102 in Figure 1, when the recognition device obtains the facial features corresponding to the facial area, it obtains them through the backbone network in the facial angle recognition model.
基于图1中S103的描述,识别设备在获取人脸区域对应的人脸特征时,通过人脸角度识别模型中的全连接分类网络获取。Based on the description of S103 in Figure 1, when the recognition device obtains the facial features corresponding to the face area, it obtains them through the fully connected classification network in the face angle recognition model.
其中,人脸角度识别模型的生成过程可以通过模型生成装置完成,也可以通过其他可行的装置生成,在此不再赘述。Among them, the generation process of the face angle recognition model can be completed by a model generation device, or it can be generated by other feasible devices, which will not be described again here.
请参阅图4,图4示出了本申请一实施例提供的生成人脸角度识别模型的流程示意图。Please refer to FIG. 4 , which shows a schematic flowchart of generating a face angle recognition model according to an embodiment of the present application.
如图4所示,生成所述人脸角度识别模型的过程,包括:As shown in Figure 4, the process of generating the face angle recognition model includes:
S401、获取样本人脸图像集。S401. Obtain a sample face image set.
样本人脸图像集包括多帧样本人脸图像以及每帧样本人脸图像中的人脸相对于拍摄位置的每种角度类型对应的真实角度。The sample face image set includes multiple frames of sample face images and the real angle corresponding to each angle type of the face in each frame of the sample face image relative to the shooting position.
可选地,样本人脸图像集至少包括一组样本人脸图像和样本人脸图像中的人脸相对于拍摄位置的每种角度类型对应的真实角度。Optionally, the sample face image set at least includes a set of sample face images and real angles corresponding to each angle type of the face in the sample face image relative to the shooting position.
样本人脸图像集可以从现有的图像数据集中选取得到(例如,公开数据集300W-LP),也可以预先通过摄像头拍摄得到的人脸图像。The sample face image set can be selected from an existing image data set (for example, the public data set 300W-LP), or it can be a face image captured by a camera in advance.
其中,通过摄像头拍摄人脸图像时,需要采用精度较高的摄像头对样本人脸从多个角度进行拍摄,以便于获得任意角度的样本人脸图像。Among them, when capturing face images through a camera, it is necessary to use a camera with higher precision to capture the sample face from multiple angles in order to obtain sample face images from any angle.
拍摄人脸图像的摄像头可以是相机、智能手机的摄像头、笔记本电脑的摄像头、平板电脑的摄像头。The camera that captures face images can be a camera, a smartphone camera, a laptop camera, or a tablet camera.
其中,样本人脸图像对应的真实角度可采用相关传感器获取,也可以由人工标注获得。Among them, the real angle corresponding to the sample face image can be obtained by using relevant sensors or by manual annotation.
S402、对每帧样本人脸图像进行数据增强处理,得到增强后的样本人脸图像。S402: Perform data enhancement processing on each frame of the sample face image to obtain an enhanced sample face image.
数据增强处理可包括随机剪切、增加随机噪声、颜色扰动中的一种或者多种的组合。Data enhancement processing may include one or a combination of random shearing, adding random noise, and color perturbation.
例如,对每帧样本人脸图像依次进行随机剪切、增加随机噪声、颜色扰动处理,得到增强后的样本人脸图像。For example, each frame of sample face image is randomly cut, random noise is added, and color perturbation is processed to obtain an enhanced sample face image.
S403、将所述增强后的样本人脸图像输入到原始角度识别模型中,输出每种角度类型的多个角度概率。S403. Input the enhanced sample face image into the original angle recognition model, and output multiple angle probabilities for each angle type.
原始人脸角度识别模型包括原始骨干网络和原始全连接分类网络。The original face angle recognition model includes the original backbone network and the original fully connected classification network.
其中,所述原始骨干网络的输出端分别与原始全连接分类网络的三个全连接层连接,三个全连接层分别为第一原始全连接层、第二原始全连接层和第三原始全连接层。Wherein, the output end of the original backbone network is connected to three fully connected layers of the original fully connected classification network. The three fully connected layers are respectively the first original fully connected layer, the second original fully connected layer and the third original fully connected layer. connection layer.
第一原始全连接层和第二原始全连接层分别连接有36个节点,第三原始全连接层连接有72个支点。The first original fully connected layer and the second original fully connected layer are connected to 36 nodes respectively, and the third original fully connected layer is connected to 72 pivots.
第一原始全连接层和第二原始全连接层的36个节点分别用于预测偏航角和俯仰角在其36个角度区间的角度概率,第三原始全连接层的72个节点用于预测翻滚角在其72个角度区间的角度概率。The 36 nodes of the first original fully connected layer and the second original fully connected layer are used to predict the angle probability of yaw angle and pitch angle in their 36 angle intervals respectively, and the 72 nodes of the third original fully connected layer are used for prediction. The angle probability of the roll angle in its 72 angle intervals.
以偏航角为例说明,在模型生成装置将样本图像集输入原始骨干网络后,输出人脸特征,第一原始全连接层根据人脸特征获取偏航角的36个角度区间分别对应的角度概率。Taking the yaw angle as an example, after the model generation device inputs the sample image set into the original backbone network, it outputs facial features. The first original fully connected layer obtains the angles corresponding to the 36 angle intervals of the yaw angle based on the facial features. Probability.
S404、根据每种角度类型的多个角度概率以及每种角度类型对应的真实角度,调整所述原始角度识别模型的模型参数。S404. Adjust the model parameters of the original angle recognition model according to the multiple angle probabilities of each angle type and the real angle corresponding to each angle type.
在一些实施例中,模型生成装置通过先根据每种角度类型的多个角度概率以及每种角度类型对应的真实角度计算损失函数,再通过损失函数调整原始角度识别模型的模型参数。In some embodiments, the model generation device first calculates a loss function based on multiple angle probabilities of each angle type and the real angle corresponding to each angle type, and then adjusts the model parameters of the original angle recognition model through the loss function.
其中,上述的损失函数为交叉熵损失函数。当然,损失函数也可以为其他类型的损失函数,在此不多做赘述。Among them, the above loss function is the cross entropy loss function. Of course, the loss function can also be other types of loss functions, so I won’t go into details here.
S405、将调整后的所述原始角度识别模型确定为所述人脸角度识别模型。S405. Determine the adjusted original angle recognition model as the face angle recognition model.
在一些实施例中,模型生成装置通过误差反向传播算法根据损失函数对原始角度识别模型进行训练,得到训练后的人脸角度识别模型,并将训练后的人脸角度识别模型确定为人脸角度识别模型。In some embodiments, the model generation device trains the original angle recognition model according to the loss function through an error backpropagation algorithm to obtain a trained face angle recognition model, and determines the trained face angle recognition model as the face angle Identify the model.
本申请中,模型生成装置生成人脸角度识别模型的过程中,先获取样本人脸图像,对每帧样本人脸图像进行数据增强处理,得到增强后的样本人脸图像。再将增强后的样本人脸图像输入到原始角度识别模型中,输出每种角度类型的多个角度概率,根据每种角度类型的多个角度概率以及每种角度类型对应的真实角度,调整原始角度识别模型的模型参数,将调整后的原始角度识别模型确定为人脸角度识别模型。借助预设规则对三种角度类型的角度范围进行划分,通过三种角度类型分别对应的多个角度概率与三种角度类型对应的真实角度对原始角度识别模型进行调整,可以得到更加精确的预测人脸角度的人脸角度识别模型。In this application, when the model generation device generates the face angle recognition model, it first obtains a sample face image, and performs data enhancement processing on each frame of the sample face image to obtain an enhanced sample face image. Then input the enhanced sample face image into the original angle recognition model, output multiple angle probabilities for each angle type, and adjust the original angle probability based on the multiple angle probabilities for each angle type and the true angle corresponding to each angle type. The model parameters of the angle recognition model determine the adjusted original angle recognition model as the face angle recognition model. Using preset rules to divide the angle ranges of the three angle types, and adjusting the original angle recognition model through the multiple angle probabilities corresponding to the three angle types and the real angles corresponding to the three angle types, more accurate predictions can be obtained Face angle recognition model for face angles.
对应于上述图1所示实施例所述的一种人脸角度预测方法,本申请还提供了一种人脸角度预测装置。Corresponding to the face angle prediction method described in the embodiment shown in Figure 1, this application also provides a face angle prediction device.
下面,结合图5,对本申请一实施例提供的人脸角度预测装置进行详细说明。Next, the face angle prediction device provided by an embodiment of the present application will be described in detail with reference to FIG. 5 .
请参阅图5,图5示出了本申请一实施例提供的人脸角度预测装置的示意性框图。Please refer to FIG. 5 , which shows a schematic block diagram of a face angle prediction device provided by an embodiment of the present application.
如图5所示,本申请一实施例提供的人脸角度预测装置,包括获取模块501、第一确定模块502、第二确定模块503和第三确定模块504。As shown in FIG. 5 , a face angle prediction device provided by an embodiment of the present application includes an acquisition module 501 , a first determination module 502 , a second determination module 503 and a third determination module 504 .
获取模块501,用于获取待测人脸图像的人脸区域;The acquisition module 501 is used to acquire the face area of the face image to be tested;
第一确定模块502,用于确定所述人脸区域对应的人脸特征;The first determination module 502 is used to determine the facial features corresponding to the facial area;
第二确定模块503,用于根据所述人脸特征,确定多种角度类型各自的多个角度概率,每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间,所述多种角度类型包括偏航角、俯仰角和翻滚角;The second determination module 503 is used to determine multiple angle probabilities of multiple angle types based on the facial features. The multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type, so The various angle types include yaw angle, pitch angle and roll angle;
第三确定模块504,用于根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。The third determination module 504 is configured to determine the predicted angle of each angle type of the human face in the face image to be measured relative to the shooting position based on multiple angle probabilities of each angle type.
在一些实施例中,504第三确定模块,具体用于:In some embodiments, 504 the third determination module is specifically used for:
针对每种角度类型而言,根据多个角度概率、所述角度区间的数量和每个所述角度区间的中间角度,确定所述待测人脸图像中的人脸相对于拍摄位置的角度。For each angle type, the angle of the face in the face image to be measured relative to the shooting position is determined based on multiple angle probabilities, the number of angle intervals, and the intermediate angle of each angle interval.
在一些实施例中,第三确定模块504,具体用于:In some embodiments, the third determination module 504 is specifically used to:
针对所述翻滚角而言,确定最大概率角度区间,所述最大概率角度区间为多个角度概率中的最大值对应的角度区间;For the roll angle, a maximum probability angle interval is determined, and the maximum probability angle interval is an angle interval corresponding to the maximum value among multiple angle probabilities;
当所述最大概率角度区间处于预设的映射角度区间内时,对所述翻滚角的多个角度概率进行线性映射,得到映射后的多个角度概率;When the maximum probability angle interval is within a preset mapping angle interval, perform linear mapping on multiple angle probabilities of the roll angle to obtain multiple mapped angle probabilities;
根据所述映射后的多个角度概率、所述翻滚角的角度区间的数量和每个角度区间的中间角度,确定映射角度;Determine the mapping angle according to the mapped multiple angle probabilities, the number of angle intervals of the roll angle, and the intermediate angle of each angle interval;
对所述映射角度进行反线性映射,得到所述翻滚角的预测角度。Perform inverse linear mapping on the mapping angle to obtain the predicted angle of the roll angle.
在一些实施例中,第三确定模块504,具体用于:In some embodiments, the third determination module 504 is specifically used to:
将大于等于0度且小于等于180度的角度区间对应的角度概率替换为大于等于-180度且小于0度的角度区间对应的映射后的角度概率;Replace the angle probability corresponding to the angle interval greater than or equal to 0 degrees and less than or equal to 180 degrees with the mapped angle probability corresponding to the angle interval greater than or equal to -180 degrees and less than 0 degrees;
将小于等于0度且大于等于-180度的角度区间对应的角度概率替换为小于等于180度且大于0度的角度区间对应的映射后的角度概率。Replace the angle probabilities corresponding to the angle intervals less than or equal to 0 degrees and greater than or equal to -180 degrees with the mapped angle probabilities corresponding to the angle intervals less than or equal to 180 degrees and greater than 0 degrees.
在一些实施例中,获取模块501,具体用于:In some embodiments, the acquisition module 501 is specifically used to:
所述获取待测人脸图像的人脸区域,包括:The method of obtaining the face area of the face image to be tested includes:
获取所述待测人脸图像;Obtain the face image to be tested;
对所述待测人脸图像进行人脸检测,得到第一检测窗口,所述第一检测窗口中包括所述待测人脸图像的至少部分人脸图像;Perform face detection on the face image to be tested to obtain a first detection window, where the first detection window includes at least part of the face image to be tested;
对所述第一检测窗口进行外扩处理,得到第二检测窗口;Perform external expansion processing on the first detection window to obtain a second detection window;
在所述待测人脸图像中,截取所述第二检测窗口对应大小的区域;In the face image to be tested, intercept an area corresponding to the size of the second detection window;
将所在一些实施例中,获取模块501,具体用于:In some embodiments, the acquisition module 501 is specifically used for:
将以所述第一检测窗口的中心为中心,所述第一检测窗口的长边为边长的窗口作为第三检测窗口;A window with the center of the first detection window as the center and the long side of the first detection window as the third detection window;
根据预设外扩系数,对所述第三检测窗口的每条边长均进行外扩处理,得到第四检测窗口;According to the preset expansion coefficient, each side length of the third detection window is expanded to obtain a fourth detection window;
去除所述第四检测窗口超过所述待测人脸图像对应的边长,得到第五检测窗口;Remove the side length corresponding to the face image to be measured that exceeds the fourth detection window to obtain a fifth detection window;
将以所述第五检测窗口的中心为中心,所述第五检测窗口的短边为边长的窗口作为第二检测窗口。A window with the center of the fifth detection window as the center and the shorter side of the fifth detection window as the second detection window.
在一些实施例中,第一确定模块502,具体用于:In some embodiments, the first determination module 502 is specifically used to:
将所述人脸区域输入到人脸角度识别模型的骨干网络中,输出所述人脸特征,所述骨干网络用于提取人脸图像中的人脸特征;Input the face area into the backbone network of the face angle recognition model and output the face features. The backbone network is used to extract the face features in the face image;
将所述人脸特征输入到所述人脸角度识别模型的全连接分类网络中,输出每种角度类型的多个角度概率,所述全连接分类网络用于预测人脸特征在每种角度类型的多个角度区间分别对应的角度概率。The facial features are input into the fully connected classification network of the face angle recognition model, and multiple angle probabilities of each angle type are output. The fully connected classification network is used to predict the facial features in each angle type. The angle probabilities corresponding to the multiple angle intervals of .
在一些实施例中,模型生成装置,用于:In some embodiments, the model generation device is used for:
获取样本人脸图像集,所述样本人脸图像集包括多帧样本人脸图像以及每帧样本人脸图像中的人脸相对于拍摄位置的每种角度类型对应的真实角度;Obtain a sample face image set, which includes a multi-frame sample face image and a true angle corresponding to each angle type of the face in each frame of the sample face image relative to the shooting position;
对每帧样本人脸图像进行数据增强处理,得到增强后的样本人脸图像;Perform data enhancement processing on each frame of sample face image to obtain an enhanced sample face image;
将所述增强后的样本人脸图像输入到原始角度识别模型中,输出每种角度类型的多个角度概率,所述原始人脸角度识别模型包括原始骨干网络和原始全连接分类网络;Input the enhanced sample face image into the original angle recognition model and output multiple angle probabilities for each angle type. The original face angle recognition model includes the original backbone network and the original fully connected classification network;
根据每种角度类型的多个角度概率以及每种角度类型对应的真实角度,调整所述原始角度识别模型的模型参数;Adjust the model parameters of the original angle recognition model according to the multiple angle probabilities of each angle type and the real angle corresponding to each angle type;
将调整后的所述原始角度识别模型确定为所述人脸角度识别模型。The adjusted original angle recognition model is determined as the face angle recognition model.
应理解的是,本申请的装置500可以通过专用集成电路(application-specific integrated circuit,ASIC)实现,或可编程逻辑器件(programmable logic device,PLD)实现,上述PLD可以是复杂程序逻辑器件(complex programmable logical device,CPLD),现场可编程门阵列(field-programmable gate array,FPGA),通用阵列逻辑(generic array logic,GAL)或其任意组合。也可以通过软件实现图1所示的人脸角度预测方法,当通过软件实现图1所示的人脸角度预测方法时,装置500及其各个模块也可以为软件模块。It should be understood that the device 500 of the present application can be implemented through an application-specific integrated circuit (application-specific integrated circuit). integrated circuit (ASIC), or programmable logic device (PLD). The above PLD can be a complex programmable logical device (CPLD), a field-programmable gate array (field-programmable) gate array, FPGA), general array logic (generic array logic, GAL) or any combination thereof. The face angle prediction method shown in Figure 1 can also be implemented through software. When the face angle prediction method shown in Figure 1 is implemented through software, the device 500 and its respective modules can also be software modules.
图6为本申请提供的一种电子设备的结构示意图。如图6所示,其中设备600包括处理器601、存储器602、通信接口603和总线604。其中,处理器601、存储器602、通信接口603通过总线604进行通信,也可以通过无线传输等其他手段实现通信。该存储器602用于存储指令,该处理器601用于执行该存储器602存储的指令。该存储器602存储程序代码6021,且处理器601可以调用存储器602 中存储的程序代码6021执行图2所示的人脸角度预测方法。Figure 6 is a schematic structural diagram of an electronic device provided by this application. As shown in Figure 6, the device 600 includes a processor 601, a memory 602, a communication interface 603 and a bus 604. Among them, the processor 601, the memory 602, and the communication interface 603 communicate through the bus 604. Communication can also be achieved through other means such as wireless transmission. The memory 602 is used to store instructions, and the processor 601 is used to execute the instructions stored in the memory 602. The memory 602 stores program code 6021, and the processor 601 can call the program code 6021 stored in the memory 602 to execute the face angle prediction method shown in Figure 2.
应理解,在本申请中,处理器601可以是CPU,处理器601还可以是其他通用处理器、数字信号处理器(DSP)、专用集成电路(ASIC)、现场可编程门阵列(FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者是任何常规的处理器等。It should be understood that in this application, the processor 601 may be a CPU, and the processor 601 may also be other general-purpose processors, digital signal processors (DSPs), application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or Other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general-purpose processor can be a microprocessor or any conventional processor, etc.
该存储器602可以包括只读存储器和随机存取存储器,并向处理器601提供指令和数据。存储器602还可以包括非易失性随机存取存储器。该存储器602可以是易失性存储器或非易失性存储器,或可包括易失性和非易失性存储器两者。其中,非易失性存储器可以是只读存储器(read-only memory,ROM)、可编程只读存储器(programmable ROM,PROM)、可擦除可编程只读存储器(erasable PROM,EPROM)、电可擦除可编程只读存储器(electrically EPROM,EEPROM)或闪存。易失性存储器可以是随机存取存储器(random access memory,RAM),其用作外部高速缓存。通过示例性但不是限制性说明,许多形式的RAM可用,例如静态随机存取存储器(static RAM,SRAM)、动态随机存取存储器(DRAM)、同步动态随机存取存储器(synchronous DRAM,SDRAM)、双倍数据速率同步动态随机存取存储器(double data date SDRAM,DDR SDRAM)、增强型同步动态随机存取存储器(enhanced SDRAM,ESDRAM)、同步连接动态随机存取存储器(synchlink DRAM,SLDRAM)和直接内存总线随机存取存储器(direct rambus RAM,DR RAM)。The memory 602 may include read-only memory and random access memory and provides instructions and data to the processor 601. Memory 602 may also include non-volatile random access memory. The memory 602 may be volatile memory or non-volatile memory, or may include both volatile and non-volatile memory. Among them, the non-volatile memory can be a read-only memory (read-only memory). memory, ROM), programmable read-only memory (programmable ROM, PROM), erasable programmable read-only memory (erasable PROM, EPROM), electrically erasable programmable read-only memory (electrically EPROM, EEPROM) or flash memory. Volatile memory can be random access memory (random access memory (RAM), which serves as an external cache. By way of illustration, but not limitation, many forms of RAM are available, such as static random access memory (SRAM), dynamic random access memory (DRAM), synchronous dynamic random access memory (synchronous DRAM, SDRAM), Double data rate synchronous dynamic random access memory (double data date SDRAM, DDR SDRAM), enhanced synchronous dynamic random access memory (enhanced SDRAM, ESDRAM), synchronous link dynamic random access memory (synchlink DRAM, SLDRAM) and direct memory bus random access memory (direct rambus RAM, DR RAM).
该总线604除包括数据总线之外,还可以包括电源总线、控制总线和状态信号总线等。但是为了清楚说明起见,在图6中将各种总线都标为总线604。In addition to a data bus, the bus 604 may also include a power bus, a control bus, a status signal bus, etc. However, for clarity of illustration, the various buses are labeled bus 604 in FIG. 6 .
应理解,根据本申请的电子设备600可对应于本申请中的装置500,并可以对应于本申请图1所示方法中的设备,当设备600对应于图2所示方法中的设备时,设备600中的各个模块的上述和其它操作和/或功能分别为了实现图2中的由设备执行的方法的操作步骤,为了简洁,在此不再赘述。It should be understood that the electronic device 600 according to the present application may correspond to the device 500 in the present application, and may correspond to the device in the method shown in FIG. 1 of the present application. When the device 600 corresponds to the device in the method shown in FIG. 2, The above and other operations and/or functions of each module in the device 600 are respectively intended to implement the operating steps of the method performed by the device in Figure 2. For the sake of brevity, they will not be described again here.
本申请还提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现可实现上述各个方法实施例中的步骤。This application also provides a computer-readable storage medium that stores a computer program. When the computer program is executed by a processor, the steps in each of the above method embodiments can be implemented.
本申请提供了一种计算机程序产品,当计算机程序产品在电子设备上运行时,使得电子设备执行时实现可实现上述各个方法实施例中的步骤。The present application provides a computer program product. When the computer program product is run on an electronic device, the steps in each of the above method embodiments can be implemented when the electronic device is executed.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请的实施过程构成任何限定。It should be understood that the sequence number of each step in the above embodiment does not mean the order of execution. The execution order of each process should be determined by its function and internal logic, and should not constitute any limitation on the implementation process of this application.
需要说明的是,上述装置/单元之间的信息交互、执行过程等内容,由于与本申请方法实施例基于同一构思,其具体功能及带来的技术效果,具体可参见方法实施例部分,此处不再赘述。It should be noted that the information interaction, execution process, etc. between the above-mentioned devices/units are based on the same concept as the method embodiments of the present application. For details of their specific functions and technical effects, please refer to the method embodiments section. No further details will be given.
所属领域的技术人员可以清楚地了解到,为了描述的方便和简洁,仅以上述各功能单元、模块的划分进行举例说明,实际应用中,可以根据需要而将上述功能分配由不同的功能单元、模块完成,即将上述装置的内部结构划分成不同的功能单元或模块,以完成以上描述的全部或者部分功能。实施例中的各功能单元、模块可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中,上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。另外,各功能单元、模块的具体名称也只是为了便于相互区分,并不用于限制本申请的保护范围。上述系统中单元、模块的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。Those skilled in the art can clearly understand that for the convenience and simplicity of description, only the division of the above functional units and modules is used as an example. In actual applications, the above functions can be allocated to different functional units and modules according to needs. Module completion means dividing the internal structure of the above device into different functional units or modules to complete all or part of the functions described above. Each functional unit and module in the embodiment can be integrated into one processing unit, or each unit can exist physically alone, or two or more units can be integrated into one unit. The above-mentioned integrated unit can be hardware-based. It can also be implemented in the form of software functional units. In addition, the specific names of each functional unit and module are only for the convenience of distinguishing each other and are not used to limit the scope of protection of the present application. For the specific working processes of the units and modules in the above system, please refer to the corresponding processes in the foregoing method embodiments, and will not be described again here.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above embodiments, each embodiment is described with its own emphasis. For parts that are not detailed or documented in a certain embodiment, please refer to the relevant descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各示例的单元及算法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本申请的范围。Those of ordinary skill in the art will appreciate that the units and algorithm steps of each example described in conjunction with the embodiments disclosed herein can be implemented with electronic hardware, or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the technical solution. Skilled artisans may implement the described functionality using different methods for each specific application, but such implementations should not be considered beyond the scope of this application.
在本申请所提供的实施例中,应该理解到,所揭露的装置/网络设备和方法,可以通过其它的方式实现。例如,以上所描述的装置/网络设备实施例仅仅是示意性的,例如,上述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通讯连接可以是通过一些接口,装置或单元的间接耦合或通讯连接,可以是电性,机械或其它的形式。In the embodiments provided in this application, it should be understood that the disclosed devices/network devices and methods can be implemented in other ways. For example, the device/network equipment embodiments described above are only illustrative. For example, the division of the above modules or units is only a logical function division. In actual implementation, there may be other division methods, such as multiple units or units. Components may be combined or may be integrated into another system, or some features may be ignored, or not implemented. On the other hand, the coupling or direct coupling or communication connection between each other shown or discussed may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
上述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本申请方案的目的。The units described above as separate components may or may not be physically separated. The components shown as units may or may not be physical units, that is, they may be located in one place, or they may be distributed to multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of this application.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-described embodiments are only used to illustrate the technical solutions of the present application, but not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, those of ordinary skill in the art should understand that they can still implement the above-mentioned implementations. The technical solutions described in the examples are modified, or some of the technical features are equivalently replaced; and these modifications or substitutions do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions in the embodiments of this application, and should be included in within the protection scope of this application.

Claims (10)

  1. 一种人脸角度预测方法,其特征在于,包括:A face angle prediction method, which is characterized by including:
    获取待测人脸图像的人脸区域;Obtain the face area of the face image to be tested;
    确定所述人脸区域对应的人脸特征;Determine the facial features corresponding to the facial area;
    根据所述人脸特征,确定多种角度类型各自的多个角度概率,每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间,所述多种角度类型包括偏航角、俯仰角和翻滚角;According to the facial features, multiple angle probabilities of multiple angle types are determined. The multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type. The multiple angle types include yaw. angle, pitch and roll angles;
    根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。According to the multiple angle probabilities of each angle type, the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position is determined.
  2. 如权利要求1所述的方法,其特征在于,所述根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度,包括:The method of claim 1, wherein the predicted angle of each angle type of the face in the face image to be measured relative to the shooting position is determined based on multiple angle probabilities of each angle type. ,include:
    针对每种角度类型而言,根据多个角度概率、所述角度区间的数量和每个所述角度区间的中间角度,确定所述待测人脸图像中的人脸相对于拍摄位置的角度。For each angle type, the angle of the face in the face image to be measured relative to the shooting position is determined based on multiple angle probabilities, the number of angle intervals, and the intermediate angle of each angle interval.
  3. 如权利要求2所述的方法,其特征在于,针对所述翻滚角而言,所述根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度,包括:The method of claim 2, wherein for the roll angle, the face in the face image to be measured is determined relative to the shooting position based on multiple angle probabilities of each angle type. Predicted angles for each angle type, including:
    确定最大概率角度区间,所述最大概率角度区间为多个角度概率中的最大值对应的角度区间;Determine the maximum probability angle interval, which is the angle interval corresponding to the maximum value among the multiple angle probabilities;
    当所述最大概率角度区间处于预设的映射角度区间内时,对所述翻滚角的多个角度概率进行线性映射,得到映射后的多个角度概率;When the maximum probability angle interval is within a preset mapping angle interval, perform linear mapping on multiple angle probabilities of the roll angle to obtain multiple mapped angle probabilities;
    根据所述映射后的多个角度概率、所述翻滚角的角度区间的数量和每个角度区间的中间角度,确定映射角度;Determine the mapping angle according to the mapped multiple angle probabilities, the number of angle intervals of the roll angle, and the intermediate angle of each angle interval;
    对所述映射角度进行反线性映射,得到所述翻滚角的预测角度。Perform inverse linear mapping on the mapping angle to obtain the predicted angle of the roll angle.
  4. 如权利要求3所述的方法,其特征在于,所述翻滚角的多个角度区间的范围为大于等于-180度且小于等于180度,所述映射角度区间为大于等于-180度且小于等于-90度或大于等于90度且小于等于180度的角度区间,所述对所述翻滚角的多个角度概率进行线性映射,得到映射后的角度概率,包括:The method according to claim 3, wherein the range of the multiple angle intervals of the roll angle is greater than or equal to -180 degrees and less than or equal to 180 degrees, and the range of the mapping angle intervals is greater than or equal to -180 degrees and less than or equal to 180 degrees. -90 degrees or greater than or equal to 90 degrees and less than or equal to 180 degrees in the angle interval. The multiple angular probabilities of the roll angle are linearly mapped to obtain the mapped angular probabilities, including:
    将大于等于0度且小于等于180度的角度区间对应的角度概率替换为大于等于-180度且小于0度的角度区间对应的映射后的角度概率;Replace the angle probability corresponding to the angle interval greater than or equal to 0 degrees and less than or equal to 180 degrees with the mapped angle probability corresponding to the angle interval greater than or equal to -180 degrees and less than 0 degrees;
    将小于等于0度且大于等于-180度的角度区间对应的角度概率替换为小于等于180度且大于0度的角度区间对应的映射后的角度概率。Replace the angle probabilities corresponding to the angle intervals less than or equal to 0 degrees and greater than or equal to -180 degrees with the mapped angle probabilities corresponding to the angle intervals less than or equal to 180 degrees and greater than 0 degrees.
  5. 如权利要求1-4任一项所述的方法,其特征在于,所述获取待测人脸图像的人脸区域,包括:The method according to any one of claims 1 to 4, characterized in that obtaining the face area of the face image to be measured includes:
    获取所述待测人脸图像;Obtain the face image to be tested;
    对所述待测人脸图像进行人脸检测,得到第一检测窗口,所述第一检测窗口中包括所述待测人脸图像的至少部分人脸图像;Perform face detection on the face image to be tested to obtain a first detection window, where the first detection window includes at least part of the face image to be tested;
    对所述第一检测窗口进行外扩处理,得到第二检测窗口;Perform external expansion processing on the first detection window to obtain a second detection window;
    在所述待测人脸图像中,截取所述第二检测窗口对应大小的区域;In the face image to be tested, intercept an area corresponding to the size of the second detection window;
    将所述第二检测窗口对应大小的区域确定为所述人脸区域。An area corresponding to the size of the second detection window is determined as the face area.
  6. 如权利要求5所述的方法,其特征在于,所述对所述第一检测窗口进行外扩处理,得到第二检测窗口,包括:The method according to claim 5, characterized in that, performing an external expansion process on the first detection window to obtain a second detection window includes:
    将以所述第一检测窗口的中心为中心,所述第一检测窗口的长边为边长的窗口作为第三检测窗口;A window with the center of the first detection window as the center and the long side of the first detection window as the third detection window;
    根据预设外扩系数,对所述第三检测窗口的每条边长均进行外扩处理,得到第四检测窗口;According to the preset expansion coefficient, each side length of the third detection window is expanded to obtain a fourth detection window;
    去除所述第四检测窗口超过所述待测人脸图像对应的边长,得到第五检测窗口;Remove the side length corresponding to the face image to be measured that exceeds the fourth detection window to obtain a fifth detection window;
    将以所述第五检测窗口的中心为中心,所述第五检测窗口的短边为边长的窗口作为第二检测窗口。A window with the center of the fifth detection window as the center and the shorter side of the fifth detection window as the second detection window.
  7. 如权利要求1-4任一项所述的方法,其特征在于,The method according to any one of claims 1-4, characterized in that,
    所述确定所述人脸区域对应的人脸特征,包括:Determining the facial features corresponding to the facial area includes:
    将所述人脸区域输入到人脸角度识别模型的骨干网络中,输出所述人脸特征,所述骨干网络用于提取人脸图像中的人脸特征;Input the face area into the backbone network of the face angle recognition model and output the face features. The backbone network is used to extract the face features in the face image;
    所述根据所述人脸特征,确定多种角度类型各自的多个角度概率,包括:Determining multiple angle probabilities for multiple angle types based on the facial features includes:
    将所述人脸特征输入到所述人脸角度识别模型的全连接分类网络中,输出每种角度类型的多个角度概率,所述全连接分类网络用于预测人脸特征在每种角度类型的多个角度区间分别对应的角度概率。The facial features are input into the fully connected classification network of the face angle recognition model, and multiple angle probabilities of each angle type are output. The fully connected classification network is used to predict the facial features in each angle type. The angle probabilities corresponding to the multiple angle intervals of .
  8. 如权利要求7所述的方法,其特征在于,生成所述人脸角度识别模型的过程,包括:The method of claim 7, wherein the process of generating the face angle recognition model includes:
    获取样本人脸图像集,所述样本人脸图像集包括多帧样本人脸图像以及每帧样本人脸图像中的人脸相对于拍摄位置的每种角度类型对应的真实角度;Obtain a sample face image set, which includes a multi-frame sample face image and a true angle corresponding to each angle type of the face in each frame of the sample face image relative to the shooting position;
    对每帧样本人脸图像进行数据增强处理,得到增强后的样本人脸图像;Perform data enhancement processing on each frame of sample face image to obtain an enhanced sample face image;
    将所述增强后的样本人脸图像输入到原始角度识别模型中,输出每种角度类型的多个角度概率,所述原始人脸角度识别模型包括原始骨干网络和原始全连接分类网络;Input the enhanced sample face image into the original angle recognition model and output multiple angle probabilities for each angle type. The original face angle recognition model includes the original backbone network and the original fully connected classification network;
    根据每种角度类型的多个角度概率以及每种角度类型对应的真实角度,调整所述原始角度识别模型的模型参数;Adjust the model parameters of the original angle recognition model according to the multiple angle probabilities of each angle type and the real angle corresponding to each angle type;
    将调整后的所述原始角度识别模型确定为所述人脸角度识别模型。The adjusted original angle recognition model is determined as the face angle recognition model.
  9. 一种人脸角度预测装置,其特征在于,包括:A face angle prediction device, characterized by including:
    获取模块,用于获取待测人脸图像的人脸区域;The acquisition module is used to obtain the face area of the face image to be tested;
    第一确定模块,用于确定所述人脸区域对应的人脸特征;The first determination module is used to determine the facial features corresponding to the facial area;
    第二确定模块,用于根据所述人脸特征,确定多种角度类型各自的多个角度概率,每种角度类型的多个角度概率分别对应于每种角度类型的多个角度区间,所述多种角度类型包括偏航角、俯仰角和翻滚角;The second determination module is used to determine multiple angle probabilities of multiple angle types according to the facial features, and the multiple angle probabilities of each angle type respectively correspond to multiple angle intervals of each angle type, said Multiple angle types including yaw, pitch, and roll angles;
    第三确定模块,用于根据每种角度类型的多个角度概率,确定所述待测人脸图像中的人脸相对于拍摄位置的每种角度类型的预测角度。The third determination module is configured to determine the predicted angle of each angle type of the human face in the face image to be measured relative to the shooting position based on multiple angle probabilities of each angle type.
  10. 一种电子设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至7任一项所述的方法。An electronic device includes a memory, a processor, and a computer program stored in the memory and executable on the processor, characterized in that when the processor executes the computer program, it implements claims 1 to 1 The method described in any one of 7.
PCT/CN2022/142276 2022-05-31 2022-12-27 Method and apparatus for predicting facial angle, and device and readable storage medium WO2023231400A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210607682.7 2022-05-31
CN202210607682.7A CN117197853A (en) 2022-05-31 2022-05-31 Face angle prediction method, device, equipment and readable storage medium

Publications (1)

Publication Number Publication Date
WO2023231400A1 true WO2023231400A1 (en) 2023-12-07

Family

ID=89003944

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/142276 WO2023231400A1 (en) 2022-05-31 2022-12-27 Method and apparatus for predicting facial angle, and device and readable storage medium

Country Status (2)

Country Link
CN (1) CN117197853A (en)
WO (1) WO2023231400A1 (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110647865A (en) * 2019-09-30 2020-01-03 腾讯科技(深圳)有限公司 Face gesture recognition method, device, equipment and storage medium
CN111814613A (en) * 2020-06-24 2020-10-23 浙江大华技术股份有限公司 Face recognition method, face recognition equipment and computer readable storage medium
CN112084856A (en) * 2020-08-05 2020-12-15 深圳市优必选科技股份有限公司 Face posture detection method and device, terminal equipment and storage medium
CN112818969A (en) * 2021-04-19 2021-05-18 南京烽火星空通信发展有限公司 Knowledge distillation-based face pose estimation method and system
CN112883918A (en) * 2021-03-22 2021-06-01 深圳市百富智能新技术有限公司 Face detection method and device, terminal equipment and computer readable storage medium
CN113971833A (en) * 2021-11-29 2022-01-25 成都新潮传媒集团有限公司 Multi-angle face recognition method and device, computer main equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110647865A (en) * 2019-09-30 2020-01-03 腾讯科技(深圳)有限公司 Face gesture recognition method, device, equipment and storage medium
CN111814613A (en) * 2020-06-24 2020-10-23 浙江大华技术股份有限公司 Face recognition method, face recognition equipment and computer readable storage medium
CN112084856A (en) * 2020-08-05 2020-12-15 深圳市优必选科技股份有限公司 Face posture detection method and device, terminal equipment and storage medium
CN112883918A (en) * 2021-03-22 2021-06-01 深圳市百富智能新技术有限公司 Face detection method and device, terminal equipment and computer readable storage medium
CN112818969A (en) * 2021-04-19 2021-05-18 南京烽火星空通信发展有限公司 Knowledge distillation-based face pose estimation method and system
CN113971833A (en) * 2021-11-29 2022-01-25 成都新潮传媒集团有限公司 Multi-angle face recognition method and device, computer main equipment and storage medium

Also Published As

Publication number Publication date
CN117197853A (en) 2023-12-08

Similar Documents

Publication Publication Date Title
WO2021057848A1 (en) Network training method, image processing method, network, terminal device and medium
CN109255352B (en) Target detection method, device and system
US10896518B2 (en) Image processing method, image processing apparatus and computer readable storage medium
WO2018119599A1 (en) Method and device for searching for person and communication system
WO2022027912A1 (en) Face pose recognition method and apparatus, terminal device, and storage medium.
CN111160178B (en) Image processing method and device, processor, electronic equipment and storage medium
WO2020215283A1 (en) Facial recognition method, processing chip and electronic device
WO2018082308A1 (en) Image processing method and terminal
WO2021169668A1 (en) Image processing method and related device
WO2017054442A1 (en) Image information recognition processing method and device, and computer storage medium
WO2021254134A1 (en) Privacy shielding processing method and apparatus, electronic device, and monitoring system
WO2021164269A1 (en) Attention mechanism-based disparity map acquisition method and apparatus
WO2020238556A1 (en) Configuration platform-based data transmission method, apparatus and computer device
WO2019033575A1 (en) Electronic device, face tracking method and system, and storage medium
WO2023124040A1 (en) Facial recognition method and apparatus
TW202011284A (en) Eye state detection system and method for operating an eye state detection system
WO2022262474A1 (en) Zoom control method and apparatus, electronic device, and computer-readable storage medium
WO2022017006A1 (en) Video processing method and apparatus, and terminal device and computer-readable storage medium
CN111080571A (en) Camera shielding state detection method and device, terminal and storage medium
WO2022061850A1 (en) Point cloud motion distortion correction method and device
JP2022545962A (en) Fog Recognition Method and Apparatus, Electronic Device, Storage Medium and Computer Program Product
CN112733901A (en) Structured action classification method and device based on federal learning and block chain
CN112689221A (en) Recording method, recording device, electronic device and computer readable storage medium
WO2021169625A1 (en) Method and apparatus for detecting reproduced network photograph, computer device, and storage medium
WO2022063321A1 (en) Image processing method and apparatus, device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22944701

Country of ref document: EP

Kind code of ref document: A1