WO2021051538A1 - Face detection method and apparatus, and terminal device - Google Patents

Face detection method and apparatus, and terminal device Download PDF

Info

Publication number
WO2021051538A1
WO2021051538A1 PCT/CN2019/117181 CN2019117181W WO2021051538A1 WO 2021051538 A1 WO2021051538 A1 WO 2021051538A1 CN 2019117181 W CN2019117181 W CN 2019117181W WO 2021051538 A1 WO2021051538 A1 WO 2021051538A1
Authority
WO
WIPO (PCT)
Prior art keywords
point cloud
cloud data
face
data
human body
Prior art date
Application number
PCT/CN2019/117181
Other languages
French (fr)
Chinese (zh)
Inventor
张国辉
李佼
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021051538A1 publication Critical patent/WO2021051538A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation

Definitions

  • This application belongs to the technical field of face detection, and particularly relates to a method, device and terminal device for face detection.
  • Face Detection refers to any given image, using a certain strategy to search for it to determine whether it contains a face, and if so, return information such as the position, size, and posture of the face.
  • Face Detection has important application value in content-based retrieval, digital video processing, and video detection.
  • face detection mainly includes two forms: 2D face detection and 3D face detection.
  • 2D face detection can be used to detect faces that appear in a plane image
  • 3D face detection can use a 3D camera to perform stereo imaging to identify the three-dimensional coordinate information of each point in the field of view.
  • the accuracy of 3D face detection analysis and judgment has been greatly improved compared to 2D face detection.
  • most of the 3D face detection in the prior art is implemented based on the projection of a 3D point cloud on a 2D image.
  • the face detection of the 3D point cloud is completed by the face detection of the RGB2D image. This method is compared Easily cracked.
  • the photo containing a face image is used to block one's face for face detection
  • the face detection algorithm based on RGB2D image the photo is also considered to be a face
  • the detection The bounding box of the face will be mapped to the 3D point cloud coordinates, so that the result of the face bounding box in the 3D point cloud can be output.
  • the photo containing the face image is only a plane in the point cloud taken by the 3D structured light camera, without any face information, and the detection result output according to this detection algorithm is actually wrong.
  • the embodiments of the present application provide a method, device and terminal device for face detection, so as to solve the problem that the 3D face detection algorithm implemented based on the projection of 3D point clouds on 2D images in the prior art is easily cracked. , The problem of lower security.
  • a method for face detection including:
  • each data point respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user
  • the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  • a face detection device including:
  • the collection module is used to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
  • the recognition module is used to recognize the nose tip position of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
  • a cropping module configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
  • the detection module is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes human face.
  • a terminal device including a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor.
  • the processor executes the computer-readable instructions The following steps are implemented in the method of face detection:
  • each data point respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user
  • the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  • a computer non-volatile readable storage medium stores computer readable instructions that, when executed by a processor, realize the human
  • the steps of the face detection method are as follows:
  • each data point respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user
  • the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  • the beneficial effects of the face detection method, device and terminal device provided by the embodiments of the present application are: by collecting the human body point cloud data of multiple sample users, it is possible to identify the data points of each sample user according to the coordinate value of each data point.
  • the face detection model obtained by the above training can be used to detect the point cloud data of the object to be detected, so as to identify whether the point cloud data of the object to be detected includes a human face.
  • 3D point cloud data sets by performing model training on existing open source data sets, a batch of 3D point cloud data sets with only face information can be obtained. These 3D point cloud data sets can be directly used for subsequent face detection without the need for RGB2D.
  • the image solves the problem that the 3D face detection algorithm implemented based on the projection of the 3D point cloud on the 2D image in the prior art is easily cracked, and the security of face detection is improved.
  • FIG. 1 is a schematic flowchart of steps of a method for face detection according to an embodiment of the present application
  • FIG. 2 is a schematic flowchart of the steps of another method for face detection according to an embodiment of the present application
  • Fig. 3 is a schematic diagram of a face detection apparatus according to an embodiment of the present application.
  • Fig. 4 is a schematic diagram of a terminal device according to an embodiment of the present application.
  • the terminal device can recognize whether the corresponding human face is included in the above-mentioned point cloud data by collecting the point cloud data of the detected object.
  • the human body point cloud data of multiple sample users can be collected for model training, and the corresponding face detection model can be constructed. , And then the face detection model completes the subsequent detection process.
  • human body point cloud data refers to 3D human body point cloud data.
  • 3D human body point cloud data is a kind of data that records the structure of the human body in the form of data points, and each data point contains three-dimensional coordinates. For example, it can be the coordinate values on the x, y, and z axes.
  • each data point may also contain other information such as gray scale, which is not limited in this embodiment.
  • the depth information of various parts of the human body can be obtained through a specific detection device or collection device. Then, these devices can automatically output 3D human body point cloud data based on the obtained depth information.
  • the above-mentioned equipment can be a depth camera, a depth camera, a depth sensor, or a lidar.
  • the depth camera is usually composed of an infrared projector and an infrared depth camera.
  • the infrared projector is mainly used to emit uniform infrared rays to the outside world and form an infrared speckle image on the target object (such as the human body or other objects).
  • the speckle image information is received by the infrared depth camera, and finally after forming the depth information of the target object, the infrared depth camera can output the point cloud data of the target object by analyzing and processing the formed depth information.
  • the human body point cloud data of the sample user is the sample data that needs to be collected in advance for subsequent model training, and the sample data can be obtained by 3D shooting of multiple different users through devices such as a depth camera.
  • the human body point cloud data of multiple sample users can also be directly extracted from some databases storing human body point cloud data.
  • This embodiment does not limit the collection method of the human body point cloud data used as the sample data.
  • S102 According to the coordinate value of each data point, respectively identify the nose tip position of the face in the human body point cloud data of each sample user;
  • the collected human body point cloud data may include a full-body point cloud or a half-body point cloud, and so on.
  • the detection object of the detection model obtained by collecting these human body point cloud data for model training is certain, that is, whether the point cloud data includes a human face is detected. Therefore, after obtaining the full-body point cloud or half-body point cloud data of the sample user, in order to reduce the amount of data processing for subsequent model training, the face point cloud data can be cropped from the human body point cloud data.
  • the human nose is basically in the center of the human face. Therefore, in order to cut out the face point cloud data from the collected human body point cloud data, it is possible to first identify the approximate position of the nose tip of each sample user in the respective human body point cloud data.
  • the position after determining the position of the nose tip of the human face, the position can be used as the origin, and data of a certain length can be cropped in various directions of the coordinate axis, so as to obtain the face point cloud data.
  • the position of the nose tip of the face can be used as the center point of the sphere, and a sphere with the center point as the center of the sphere and a specific value as the radius is cut out from the human body point cloud data, and the data contained in the sphere is regarded as the face point Cloud data.
  • the above length and value can be determined by those skilled in the art based on empirical values, which is not limited in this embodiment.
  • S104 Generate a face detection model by performing model training on the face point cloud data of the multiple sample users;
  • the face point cloud data After obtaining the face point cloud data of each sample user, the face point cloud data can be used as sample data for model training to obtain a face detection model.
  • the face detection model can be obtained by inputting the face point cloud data of each sample user mentioned above into a preset three-dimensional point cloud network model for training.
  • the aforementioned three-dimensional point cloud network model may be a PointNet++ model.
  • the PointNet++ model is a deep learning multi-classification framework model based on 3D point cloud design. This model can be used to classify objects presented in the 3D point cloud.
  • the generated face detection model can obtain the correlation information between the face point cloud data of each sample user.
  • the associated information can be used for subsequent face detection.
  • the object to be detected may be a user or other object to be detected.
  • the point cloud data of the object to be detected can also be collected by depth cameras and other equipment.
  • the collected point cloud data of the object to be detected can be input to the face detection model generated in step S104, and the model performs the point cloud data detection on the current point cloud data that needs to be detected according to the correlation information between the face point cloud data obtained by training. Recognition, so as to output the information of whether the face is included in the point cloud data.
  • the face detection model can judge the points to be detected by comparing the correlation information between the current point cloud data to be detected and the correlation information between the face point cloud data obtained through model training. Whether the cloud data includes face point cloud data.
  • the position of the nose tip of the face in the human body point cloud data of each sample user can be identified according to the coordinate value of each data point, so as to be based on the human body point cloud data.
  • the position of the tip of the face and nose can further cut out the face point cloud data from the human point cloud data as sample data for model training to generate a face detection model; so that when the point cloud data of the object to be detected is received, the above can be used
  • the face detection model obtained by training detects the point cloud data of the object to be detected to identify whether the point cloud data of the object to be detected includes a face.
  • 3D point cloud data sets by performing model training on existing open source data sets, a batch of 3D point cloud data sets with only face information can be obtained. These 3D point cloud data sets can be directly used for subsequent face detection without the need for RGB2D.
  • the image solves the problem that the 3D face detection algorithm implemented based on the projection of the 3D point cloud on the 2D image in the prior art is easily cracked, and the security of face detection is improved.
  • FIG. 2 there is shown a schematic flow chart of the steps of another face detection method according to an embodiment of the present application, which may specifically include the following steps:
  • the human body point cloud data of multiple sample users is the sample data for subsequent model training, and the corresponding face detection model can be generated by performing model training on the sample data.
  • the human body point cloud data of the sample user can be collected through equipment such as a depth camera, a depth camera, a depth sensor, or a lidar.
  • the collected human body point cloud data may include a whole body point cloud or a half body point cloud.
  • these data points include coordinate values in a three-dimensional coordinate system, and the information embodied by these data points can characterize the specific human body structure .
  • S202 Preprocessing the human body point cloud data of the multiple sample users
  • the training error is reduced.
  • the human body point cloud data can also be preprocessed.
  • the preprocessing of human point cloud data can include denoising processing and normalization processing.
  • the collected human body point cloud data will have some noise, such as some outliers. You can denoise the human body point cloud data of multiple sample users to filter out these outliers and remove the noise for subsequent follow-up. Identify the impact.
  • the human body point cloud data can be normalized to the human body point cloud data with preset specifications by performing scale transformation on the coordinate values of each data point after denoising.
  • the size of different human point cloud data may be different.
  • the area covered by some point cloud data is 3*3*3, while the area covered by some point cloud data is 6*6*6. Therefore, all human body point cloud data can be normalized to get The processed human body point cloud data with the same specifications.
  • the human body point cloud data since the human body point cloud data includes the coordinate value of each data point, a cube containing all the data points in all the human body point cloud data can be generated according to each coordinate value, and then the coordinate value of each data point can be calculated.
  • the scale transformation method normalizes each data point to a data point in a cube of the same specification.
  • S203 According to the origin and direction of the preset coordinate system, identify the position of the data point corresponding to the maximum value of the coordinate value on the horizontal axis or the vertical axis of the coordinate system in the human body point cloud data as the nose tip position of the human face;
  • the face point cloud data can be cropped from the human body point cloud data, and the face point cloud data As the positive sample data for subsequent training.
  • the human nose is basically in the center of the human face. Therefore, in order to cut out the face point cloud data from the collected human body point cloud data, it is possible to first identify the approximate position of the nose tip of each sample user in the respective human body point cloud data.
  • the position corresponding to the maximum value on the horizontal axis or the vertical axis can be selected as the nose tip position in the constructed coordinate system. Whether the position of the data point corresponding to the maximum value of the horizontal axis or the maximum value of the vertical axis is used as the position of the nose tip of the human face can be specifically determined according to the directions of the horizontal axis and the vertical axis of the coordinate system.
  • the shape of the human body is approximately symmetrical.
  • a plane (second plane) can be determined first, by which the human body point cloud can be divided into two parts, and the number of point clouds in the left and right parts is approximately equal.
  • the center point of the human body point cloud data can be determined, and the center point can be used as the origin of the coordinate system to be constructed.
  • the horizontal axis and the vertical axis of the coordinate system can be constructed based on the origin, so that another plane (the first plane) formed by the horizontal axis and the vertical axis is parallel to the horizontal plane and is parallel to the second plane.
  • the plane is vertical. In this way, the horizontal or vertical axis of the coordinate system is parallel to the second plane.
  • the vertical axis of the above-mentioned coordinate system is perpendicular to the second plane, and the position corresponding to the maximum value on the horizontal axis can be used as the position of the nose tip of the face; if the above-mentioned coordinate system The vertical axis of is parallel to the second plane, and the horizontal axis of the coordinate system is perpendicular to the second plane. At this time, the position corresponding to the maximum value on the vertical axis can be used as the position of the nose tip of the human face. It should be noted that the maximum value may be the maximum absolute value of the coordinate value.
  • S204 Constructing a three-dimensional coordinate system with the position of the nose tip of the face as the origin, and obtaining face point cloud data by extracting multiple data points within a preset length in each direction of the three-dimensional coordinate system;
  • the position after determining the position of the nose tip of the human face, the position can be used as the origin, and data of a certain length can be cropped in various directions of the coordinate axis, so as to obtain the face point cloud data.
  • the position of the nose tip of a human face can be determined as the origin to construct a three-dimensional coordinate system, and then starting from the origin, data points within a certain length range in each direction of the coordinate axis can be extracted respectively, and the human body point cloud data can be "faced”. Obtain face point cloud data.
  • S205 Generate a face detection model by performing model training on the face point cloud data of the multiple sample users, and the face detection model stores one of the data points in the face point cloud data obtained after training. Sparsity data between;
  • the face point cloud data After obtaining the face point cloud data of each sample user, the face point cloud data can be used as positive sample data for model training to obtain a face detection model.
  • the face point cloud data of multiple sample users can be input into the preset 3D point cloud network model PointNet++ for model training for the first time. Then, by configuring the fully connected layer of the PointNet++ model into two layers, a two-class face detection model is generated.
  • the PointNet++ model is a deep learning multi-classification framework model based on 3D point cloud design, this model can be used to classify objects in the data presented by the 3D point cloud. Therefore, in the embodiment of the present application, by modifying the output result of the aforementioned PointNet++ model to a two-class classification, it is possible to classify whether the detected object is a human face. That is, the detected object is recognized through the PointNet++ model, and the corresponding output result is a human face or not a human face.
  • the fully connected layer of the PointNet++ model can be configured to output two types of results, and the pre-collected sample set can be trained to realize the classification of faces and non-faces.
  • the collected face point cloud data as positive sample data into the PointNet++ model for training.
  • some non-face point clouds can also be collected
  • the data is trained as negative sample data.
  • the PointNet++ model can obtain the sparseness between face point cloud data and non-face point cloud data by training the sample set data.
  • the sparsity data of the face point cloud data can be used to indicate the location of each data point in the face point cloud data and the relative positional relationship between the various data points.
  • the output result of the two-classified PointNet++ model since the output result of the two-classified PointNet++ model only includes two cases, one is a human face, and the other is not a human face. Therefore, after collecting the point cloud data of the object to be detected, pass Input the point cloud data into the PointNet++ model, you can directly identify whether the point cloud data is a face.
  • the above-mentioned pre-generated two-class face detection model can be used to obtain the sparsity between each data point in the point cloud data of the object to be detected; Calculating the similarity between the sparsity of each data point and the sparsity data between the data points stored in the face detection model can identify whether the point cloud data of the object to be detected includes a face.
  • the face detection model can output corresponding detection results in real time.
  • a batch of 3D point cloud data sets with only face information can be obtained.
  • the trained 3D point cloud data can be directly used for face detection without the need to resort to RGB2D images, which improves the security of face detection.
  • FIG. 3 a schematic diagram of a face detection apparatus according to an embodiment of the present application is shown, which may specifically include the following modules:
  • the collection module 301 is configured to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
  • the recognition module 302 is configured to recognize the position of the nose tip of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
  • the cropping module 303 is configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
  • the generating module 304 is configured to generate a face detection model by performing model training on the face point cloud data of the multiple sample users;
  • the detection module 305 is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected is Including human faces.
  • the device may further include the following modules:
  • a denoising module configured to perform denoising processing on the human body point cloud data of the multiple sample users
  • the normalization module is used to normalize the human body point cloud data into target point cloud data with preset specifications by performing proportional transformation on the coordinate values of the respective data points after denoising.
  • the identification module 302 may specifically include the following sub-modules:
  • the nose tip position recognition sub-module of the face is used to identify the data point corresponding to the maximum value of the coordinate value on the horizontal or vertical axis of the coordinate system in the human body point cloud data according to the origin and direction of the preset coordinate system
  • the position is the position of the nose tip of the human face; wherein the origin of the coordinate system is the center point of the human body point cloud data, and the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane,
  • the second plane is used to divide the human body point cloud data into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
  • the cropping module 303 may specifically include the following sub-modules:
  • the face point cloud data extraction sub-module is used to construct a three-dimensional coordinate system with the nose tip position of the face as the origin, and obtain a face by extracting multiple data points within a preset length in each direction of the three-dimensional coordinate system Point cloud data.
  • the generating module 304 may specifically include the following sub-modules:
  • the model training sub-module is used to input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
  • the model configuration sub-module is used to generate a two-class face detection model by configuring the fully connected layer of the three-dimensional point cloud network model into two layers, and the face detection model stores the face obtained after training The sparsity data between each data point in the point cloud data.
  • the detection module 305 may specifically include the following sub-modules:
  • a sparsity acquisition sub-module for acquiring the sparsity between various data points in the point cloud data of the object to be detected by using the two-class face detection model
  • a similarity calculation sub-module for calculating the similarity between the sparsity between the various data points and the sparsity data between the various data points stored in the face detection model;
  • the face detection sub-module is configured to recognize that the point cloud data of the object to be detected includes a face when the similarity exceeds a preset threshold.
  • the description is relatively simple, and for related parts, please refer to the description of the method embodiment part.
  • the terminal device 400 of this embodiment includes a processor 410, a memory 420, and computer-readable instructions 421 stored in the memory 420 and running on the processor 410.
  • the processor 410 executes the computer-readable instructions 421
  • the steps in the various embodiments of the above-mentioned face detection method are implemented, for example, steps S101 to S105 shown in FIG. 1.
  • the processor 410 executes the computer-readable instructions 421
  • the functions of the modules/units in the foregoing device embodiments, such as the functions of the modules 301 to 305 shown in FIG. 3, are implemented.
  • the computer-readable instructions 421 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 420 and executed by the processor 410.
  • the one or more modules/units may be a series of computer-readable instruction segments capable of completing specific functions, and the instruction segments may be used to describe the execution process of the computer-readable instructions 421 in the terminal device 400.
  • the computer-readable instructions 421 can be divided into a collection module, an identification module, a cropping module, a generation module, and a detection module. The specific functions of each module are as follows:
  • the collection module is used to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
  • the recognition module is used to recognize the nose tip position of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
  • a cropping module configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
  • the detection module is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes human face.
  • the terminal device 400 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
  • the terminal device 400 may include, but is not limited to, a processor 410 and a memory 420.
  • FIG. 4 is only an example of the terminal device 400, and does not constitute a limitation on the terminal device 400. It may include more or less components than shown in the figure, or combine certain components, or different components.
  • the terminal device 400 may also include input and output devices, network access devices, buses, and so on.
  • the processor 410 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
  • the memory 420 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400.
  • the memory 420 may also be an external storage device of the terminal device 400, such as a plug-in hard disk equipped on the terminal device 400, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD). Card, Flash Card, etc.
  • the memory 420 may also include both an internal storage unit of the terminal device 400 and an external storage device.
  • the memory 420 is used to store the computer-readable instructions 421 and other instructions and data required by the terminal device 400.
  • the memory 420 may also be used to temporarily store data that has been output or will be output.
  • Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Processing (AREA)

Abstract

A face detection method and apparatus, and a terminal device, wherein same fall within the technical field of face detection. The method comprises: collecting body point cloud data of a plurality of sample users, wherein the body point cloud data comprises a plurality of data points, and all the data points respectively have corresponding coordinate values (S101); respectively identifying the nose tip position on the face in the body point cloud data of each sample user according to the coordinate values of all the data points (S102); cutting face point cloud data out of the body point cloud data on the basis of the nose tip positions on the faces (S103); generating a face detection model by means of carrying out model training on the face point cloud data of the plurality of sample users (S104); and when point cloud data of an object to be detected is received, carrying out detection on the point cloud data of said object by means of the face detection model, and identifying whether the point cloud data of said object comprises a face (S105). The problem in the art of a 3D face detection algorithm being easily cracked is solved, thereby improving the security of face detection.

Description

一种人脸检测的方法、装置及终端设备Method, device and terminal equipment for face detection
本申请申明享有2019年09月18日递交的申请号为201910882002.0、名称为“一种人脸检测的方法、装置及终端设备”中国专利申请的优先权,该中国专利申请的整体内容以参考的方式结合在本申请中。This application affirms that it enjoys the priority of the Chinese patent application with the application number 201910882002.0 filed on September 18, 2019, entitled "A method, device and terminal device for face detection", and the entire content of the Chinese patent application is by reference The method is incorporated in this application.
技术领域Technical field
本申请属于人脸检测技术领域,特别是涉及一种人脸检测的方法、装置及终端设备。This application belongs to the technical field of face detection, and particularly relates to a method, device and terminal device for face detection.
背景技术Background technique
人脸检测(Face Detection)是指对于任意给定的图像,采用一定的策略对其进行搜索以确定其中是否含有人脸,如果是则返回人脸的位置、大小和姿态等信息。随着技术的发展,人脸检测在基于内容的检索、数字视频处理、视频检测等方面有着重要的应用价值。Face Detection (Face Detection) refers to any given image, using a certain strategy to search for it to determine whether it contains a face, and if so, return information such as the position, size, and posture of the face. With the development of technology, face detection has important application value in content-based retrieval, digital video processing, and video detection.
目前,人脸检测主要包括2D人脸检测和3D人脸检测两种形式。其中,2D人脸检测可以用于对平面图像中出现的人脸进行检测,而3D人脸检测则能够通过3D摄像头立体成像,识别视野内空间每个点位的三维坐标信息。由于机器获取的信息多了,3D人脸检测分析判断的准确性相较于2D人脸检测有了极大的提升。但是,现有技术中的3D人脸检测大多都是基于3D点云在2D图像上的投影来实现的,通过对RGB2D图像的人脸检测来完成对3D点云的人脸检测,该方法比较容易被破解。At present, face detection mainly includes two forms: 2D face detection and 3D face detection. Among them, 2D face detection can be used to detect faces that appear in a plane image, and 3D face detection can use a 3D camera to perform stereo imaging to identify the three-dimensional coordinate information of each point in the field of view. As the machine obtains more information, the accuracy of 3D face detection analysis and judgment has been greatly improved compared to 2D face detection. However, most of the 3D face detection in the prior art is implemented based on the projection of a 3D point cloud on a 2D image. The face detection of the 3D point cloud is completed by the face detection of the RGB2D image. This method is compared Easily cracked.
例如,在3D结构光摄像头前,用一张包含人脸图像的照片挡住自己的脸进行人脸检测时,在基于RGB2D图像的人脸检测算法中,也认为这张照片是人脸,那么检测出该脸的边界框会被映射到3D点云坐标上,从而可以输出人脸边界框在3D点云中的结果。但是,包含人脸图像的这张照片在3D结构光摄像头拍摄的点云中仅仅是一个平面,没有任何的人脸信息,按照这种检测算法输出的检测结果实际上是错误的。For example, in front of a 3D structured light camera, when a photo containing a face image is used to block one's face for face detection, in the face detection algorithm based on RGB2D image, the photo is also considered to be a face, then the detection The bounding box of the face will be mapped to the 3D point cloud coordinates, so that the result of the face bounding box in the 3D point cloud can be output. However, the photo containing the face image is only a plane in the point cloud taken by the 3D structured light camera, without any face information, and the detection result output according to this detection algorithm is actually wrong.
发明概述Summary of the invention
技术问题technical problem
有鉴于此,本申请实施例提供了一种人脸检测的方法、装置及终端设备,以解决现有技术中基于3D点云在2D图像上的投影来实现的3D人脸检测算法容易被破解,安全性较低的问题。In view of this, the embodiments of the present application provide a method, device and terminal device for face detection, so as to solve the problem that the 3D face detection algorithm implemented based on the projection of 3D point clouds on 2D images in the prior art is easily cracked. , The problem of lower security.
问题的解决方案The solution to the problem
技术解决方案Technical solutions
为解决上述技术问题,本申请实施例采用的技术方案是:In order to solve the above technical problems, the technical solutions adopted in the embodiments of this application are:
第一方面,提供了一种人脸检测的方法,包括:In the first aspect, a method for face detection is provided, including:
采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
第二方面,提供了一种人脸检测的装置,包括:In a second aspect, a face detection device is provided, including:
采集模块,用于采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;The collection module is used to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
识别模块,用于根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;The recognition module is used to recognize the nose tip position of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
裁剪模块,用于基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;A cropping module, configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
生成模块,用于通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;A generating module for generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
检测模块,用于在接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。The detection module is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes human face.
第三方面,提供了一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,所述处理器执行所述计算机可读指令时实现上述人脸检测的方法的如下步骤:In a third aspect, a terminal device is provided, including a memory, a processor, and computer-readable instructions stored in the memory and executable on the processor. When the processor executes the computer-readable instructions The following steps are implemented in the method of face detection:
采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
第四方面,提供了一种计算机非易失性可读存储介质,所述计算机非易失性可读存储介质存储有计算机可读指令,所述计算机可读指令被处理器执行时实现上述人脸检测的方法的如下步骤:In a fourth aspect, a computer non-volatile readable storage medium is provided, and the computer non-volatile readable storage medium stores computer readable instructions that, when executed by a processor, realize the human The steps of the face detection method are as follows:
采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
本申请实施例提供的人脸检测的方法、装置及终端设备的有益效果在于:通过采集多个样本用户的人体点云数据,可以根据各个数据点的坐标值,分别识别出在各个样本用户的人体点云数据中的人脸鼻尖位置,从而基于鼻尖位置,能够从人体点云数据中裁剪出人脸点云数据作为样本数据用于模型训练,生成人脸检测模型;使得在接收到待检测对象的点云数据时,可以采用上述训练获得的人脸检测模型对待检测对象的点云数据进行检测,以识别出待检测对象的点 云数据中是否包括人脸。本实施例通过对现有的开源数据集进行模型训练,可以获得一批仅有人脸信息的3D点云数据集,这些3D点云数据集可以直接被用作后续的人脸检测,无需借助RGB2D图像,解决了现有技术中基于3D点云在2D图像上的投影来实现的3D人脸检测算法容易被破解的问题,提高了人脸检测的安全性。The beneficial effects of the face detection method, device and terminal device provided by the embodiments of the present application are: by collecting the human body point cloud data of multiple sample users, it is possible to identify the data points of each sample user according to the coordinate value of each data point. The position of the nose tip of the face in the human body point cloud data, so that based on the nose tip position, the face point cloud data can be cut out from the human body point cloud data as sample data for model training to generate a face detection model; In the case of the point cloud data of the object, the face detection model obtained by the above training can be used to detect the point cloud data of the object to be detected, so as to identify whether the point cloud data of the object to be detected includes a human face. In this embodiment, by performing model training on existing open source data sets, a batch of 3D point cloud data sets with only face information can be obtained. These 3D point cloud data sets can be directly used for subsequent face detection without the need for RGB2D. The image solves the problem that the 3D face detection algorithm implemented based on the projection of the 3D point cloud on the 2D image in the prior art is easily cracked, and the security of face detection is improved.
发明的有益效果The beneficial effects of the invention
对附图的简要说明Brief description of the drawings
附图说明Description of the drawings
图1是本申请一个实施例的一种人脸检测的方法的步骤流程示意图;FIG. 1 is a schematic flowchart of steps of a method for face detection according to an embodiment of the present application;
图2是本申请一个实施例的另一种人脸检测的方法的步骤流程示意图;FIG. 2 is a schematic flowchart of the steps of another method for face detection according to an embodiment of the present application;
图3是本申请一个实施例的一种人脸检测的装置的示意图;Fig. 3 is a schematic diagram of a face detection apparatus according to an embodiment of the present application;
图4是本申请一个实施例的一种终端设备的示意图。Fig. 4 is a schematic diagram of a terminal device according to an embodiment of the present application.
发明实施例Invention embodiment
本发明的实施方式Embodiments of the present invention
参照图1,示出了本申请一个实施例的一种人脸检测的方法的步骤流程示意图,具体可以包括如下步骤:1, there is shown a schematic flow chart of the steps of a face detection method according to an embodiment of the present application, which may specifically include the following steps:
S101、采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;S101. Collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value.
需要说明的是,本方法可以应用于终端设备中。该终端设备通过采集被检测对象的点云数据,可以识别出上述点云数据中是否包括相应的人脸。It should be noted that this method can be applied to terminal equipment. The terminal device can recognize whether the corresponding human face is included in the above-mentioned point cloud data by collecting the point cloud data of the detected object.
在本申请实施例中,为了实现对被检测对象的点云数据中是否包括人脸进行检测,首先可以通过采集多个样本用户的人体点云数据进行模型训练,构建出相应的人脸检测模型,进而由该人脸检测模型完成后续的检测过程。In the embodiment of the present application, in order to detect whether the point cloud data of the detected object includes a face, firstly, the human body point cloud data of multiple sample users can be collected for model training, and the corresponding face detection model can be constructed. , And then the face detection model completes the subsequent detection process.
通常,人体点云数据即是指3D人体点云数据。3D人体点云数据是以数据点的形式记录人体结构的一种数据,每一个数据点均包含有三维坐标。例如,可以是x、y、z轴上的坐标值。当然,每一个数据点也还可以包含有灰度等其他信息,本实施例对此不作限定。Generally, human body point cloud data refers to 3D human body point cloud data. 3D human body point cloud data is a kind of data that records the structure of the human body in the form of data points, and each data point contains three-dimensional coordinates. For example, it can be the coordinate values on the x, y, and z axes. Of course, each data point may also contain other information such as gray scale, which is not limited in this embodiment.
在具体实现中,可以通过特定的检测设备或采集设备获取人体各个部位的深度信息。然后,这些设备可以基于得到的深度信息自动输出3D人体点云数据。通常,上述设备可以是深度摄像机、深度照相机、深度传感器或激光雷达等设备。In a specific implementation, the depth information of various parts of the human body can be obtained through a specific detection device or collection device. Then, these devices can automatically output 3D human body point cloud data based on the obtained depth information. Generally, the above-mentioned equipment can be a depth camera, a depth camera, a depth sensor, or a lidar.
以深度摄像机为例。深度摄像机通常由红外投影机和红外深度摄像机构成,其中,红外投影机主要用于向外界发射均匀的红外线,并在目标对象(如人体或其他物体)上形成红外散斑图像,目标物体反射得到的散斑图像信息由红外深度摄像机接收,最后在形成目标对象的深度信息后,红外深度摄像机通过对形成的深度信息进行分析处理,可以输出目标对象的点云数据。Take the depth camera as an example. The depth camera is usually composed of an infrared projector and an infrared depth camera. The infrared projector is mainly used to emit uniform infrared rays to the outside world and form an infrared speckle image on the target object (such as the human body or other objects). The speckle image information is received by the infrared depth camera, and finally after forming the depth information of the target object, the infrared depth camera can output the point cloud data of the target object by analyzing and processing the formed depth information.
在本申请实施例中,样本用户的人体点云数据即是为了用于后续进行模型训练而需要预先采集的样本数据,样本数据可以通过深度摄像机等设备对多个不同的用户进行3D拍摄获得。In the embodiment of the present application, the human body point cloud data of the sample user is the sample data that needs to be collected in advance for subsequent model training, and the sample data can be obtained by 3D shooting of multiple different users through devices such as a depth camera.
当然,多个样本用户的人体点云数据也可以直接从一些存储有人体点云数据的数据库中提取获得,本实施例对用作样本数据的人体点云数据的采集方式不作限定。Of course, the human body point cloud data of multiple sample users can also be directly extracted from some databases storing human body point cloud data. This embodiment does not limit the collection method of the human body point cloud data used as the sample data.
S102、根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;S102: According to the coordinate value of each data point, respectively identify the nose tip position of the face in the human body point cloud data of each sample user;
在本申请实施例中,采集得到的人体点云数据可以包括全身点云或半身点云等等。由于采集这些人体点云数据进行模型训练获得的检测模型的检测对象是确定的,即检测的是点云数据中是否包括人脸。因此,在获得样本用户的全身点云或半身点云数据后,为了减少后续模型训练的数据处理量,可以从这些人体点云数据中裁剪出人脸点云数据。In the embodiment of the present application, the collected human body point cloud data may include a full-body point cloud or a half-body point cloud, and so on. The detection object of the detection model obtained by collecting these human body point cloud data for model training is certain, that is, whether the point cloud data includes a human face is detected. Therefore, after obtaining the full-body point cloud or half-body point cloud data of the sample user, in order to reduce the amount of data processing for subsequent model training, the face point cloud data can be cropped from the human body point cloud data.
通常,人的鼻子基本上处于人脸的居中位置。因此,为了从采集得到的人体点云数据中裁剪出人脸点云数据,可以首先识别出各个样本用户的鼻尖在各自的人体点云数据中的大致位置。Generally, the human nose is basically in the center of the human face. Therefore, in order to cut out the face point cloud data from the collected human body point cloud data, it is possible to first identify the approximate position of the nose tip of each sample user in the respective human body point cloud data.
S103、基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;S103: Based on the position of the nose tip of the face, crop the face point cloud data from the human body point cloud data;
在本申请实施例中,在确定人脸鼻尖的位置后,可以以该位置为原点,通过在坐标轴的各个方向上剪裁出一定长度的数据,从而可以得到人脸点云数据。In the embodiment of the present application, after determining the position of the nose tip of the human face, the position can be used as the origin, and data of a certain length can be cropped in various directions of the coordinate axis, so as to obtain the face point cloud data.
或者,也可以以人脸鼻尖位置作为球体的中心点,通过在人体点云数据中裁剪出以该中心点为球心,特定数值为半径的球体,将该球体中包含的数据作为人脸点云数据。上述长度和数值均可以由本领域技术人员根据经验值确定,本实施例对此不作限定。Alternatively, the position of the nose tip of the face can be used as the center point of the sphere, and a sphere with the center point as the center of the sphere and a specific value as the radius is cut out from the human body point cloud data, and the data contained in the sphere is regarded as the face point Cloud data. The above length and value can be determined by those skilled in the art based on empirical values, which is not limited in this embodiment.
S104、通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;S104: Generate a face detection model by performing model training on the face point cloud data of the multiple sample users;
在得到各个样本用户的人脸点云数据后,便可以以这些人脸点云数据作为样本数据进行模型训练,获得人脸检测模型。After obtaining the face point cloud data of each sample user, the face point cloud data can be used as sample data for model training to obtain a face detection model.
在本申请实施例中,可以通过将上述各个样本用户的人脸点云数据输入预置的三维点云网络模型进行训练中,获得人脸检测模型。上述三维点云网络模型可以是PointNet++模型。In the embodiment of the present application, the face detection model can be obtained by inputting the face point cloud data of each sample user mentioned above into a preset three-dimensional point cloud network model for training. The aforementioned three-dimensional point cloud network model may be a PointNet++ model.
PointNet++模型是基于3D点云设计的深度学习多分类框架模型,可以利用该模型来对3D点云呈现的数据进行物体分类。The PointNet++ model is a deep learning multi-classification framework model based on 3D point cloud design. This model can be used to classify objects presented in the 3D point cloud.
在具体实现中,采用PointNet++模型对作为样本数据的各个样本用户的人脸点云数据进行训练后,生成的人脸检测模型可以获得各个样本用户的人脸点云数据之间的关联信息,这些关联信息可以用于后续的人脸检测。In the specific implementation, after the PointNet++ model is used to train the face point cloud data of each sample user as the sample data, the generated face detection model can obtain the correlation information between the face point cloud data of each sample user. The associated information can be used for subsequent face detection.
S105、当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。S105. When the point cloud data of the object to be detected is received, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes a human face.
在本申请实施例中,待检测对象可以待检测的用户或其他物体。待检测对象的点云数据同样可以通过深度摄像机等设备采集得到。In the embodiment of the present application, the object to be detected may be a user or other object to be detected. The point cloud data of the object to be detected can also be collected by depth cameras and other equipment.
采集得到的待检测对象的点云数据可以被输入至步骤S104中生成的人脸检测模型中,由模型根据训练获得的人脸点云数据之间的关联信息对当前需要检测的点云数据进行识别,从而输出这些点云数据中是否包括人脸的信息。The collected point cloud data of the object to be detected can be input to the face detection model generated in step S104, and the model performs the point cloud data detection on the current point cloud data that needs to be detected according to the correlation information between the face point cloud data obtained by training. Recognition, so as to output the information of whether the face is included in the point cloud data.
在具体实现中,人脸检测模型可以通过比较当前待检测的点云数据之间的关联信息与经模型训练获得的人脸点云数据之间的关联信息是否具有相似性,判断待检测的点云数据是否包括人脸点云数据。In specific implementation, the face detection model can judge the points to be detected by comparing the correlation information between the current point cloud data to be detected and the correlation information between the face point cloud data obtained through model training. Whether the cloud data includes face point cloud data.
在本申请实施例中,通过采集多个样本用户的人体点云数据,可以根据各个数 据点的坐标值,分别识别出在各个样本用户的人体点云数据中的人脸鼻尖位置,从而基于人脸鼻尖位置,能够进一步地从人体点云数据中裁剪出人脸点云数据作为样本数据用于模型训练,生成人脸检测模型;使得在接收到待检测对象的点云数据时,可以采用上述训练获得的人脸检测模型对待检测对象的点云数据进行检测,以识别出待检测对象的点云数据中是否包括人脸。本实施例通过对现有的开源数据集进行模型训练,可以获得一批仅有人脸信息的3D点云数据集,这些3D点云数据集可以直接被用作后续的人脸检测,无需借助RGB2D图像,解决了现有技术中基于3D点云在2D图像上的投影来实现的3D人脸检测算法容易被破解的问题,提高了人脸检测的安全性。In the embodiment of the present application, by collecting the human body point cloud data of multiple sample users, the position of the nose tip of the face in the human body point cloud data of each sample user can be identified according to the coordinate value of each data point, so as to be based on the human body point cloud data. The position of the tip of the face and nose can further cut out the face point cloud data from the human point cloud data as sample data for model training to generate a face detection model; so that when the point cloud data of the object to be detected is received, the above can be used The face detection model obtained by training detects the point cloud data of the object to be detected to identify whether the point cloud data of the object to be detected includes a face. In this embodiment, by performing model training on existing open source data sets, a batch of 3D point cloud data sets with only face information can be obtained. These 3D point cloud data sets can be directly used for subsequent face detection without the need for RGB2D. The image solves the problem that the 3D face detection algorithm implemented based on the projection of the 3D point cloud on the 2D image in the prior art is easily cracked, and the security of face detection is improved.
参照图2,示出了本申请一个实施例的另一种人脸检测的方法的步骤流程示意图,具体可以包括如下步骤:Referring to FIG. 2, there is shown a schematic flow chart of the steps of another face detection method according to an embodiment of the present application, which may specifically include the following steps:
S201、采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;S201. Collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value.
在本申请实施例中,多个样本用户的人体点云数据即是后续用于模型训练的样本数据,通过对样本数据进行模型训练,可以生成相应的人脸检测模型。In the embodiment of the present application, the human body point cloud data of multiple sample users is the sample data for subsequent model training, and the corresponding face detection model can be generated by performing model training on the sample data.
在具体实现中,可以通过深度摄像机、深度照相机、深度传感器或激光雷达等设备采集样本用户的人体点云数据。采集得到的人体点云数据可以包括全身点云或半身点云。当然,无论是全身点云或半深点云,均包括有多个数据点,这些数据点包含有三维坐标系下的坐标值,通过这些数据点所体现出的信息,可以表征具体的人体结构。In a specific implementation, the human body point cloud data of the sample user can be collected through equipment such as a depth camera, a depth camera, a depth sensor, or a lidar. The collected human body point cloud data may include a whole body point cloud or a half body point cloud. Of course, whether it is a full-body point cloud or a semi-deep point cloud, there are multiple data points, these data points include coordinate values in a three-dimensional coordinate system, and the information embodied by these data points can characterize the specific human body structure .
S202、对所述多个样本用户的人体点云数据进行预处理;S202: Preprocessing the human body point cloud data of the multiple sample users;
在本申请实施例中,为了减少后续模型训练时的数据处理量,减少训练误差。在采集得到人体点云数据后,还可以对这些人体点云数据作预处理。对人体点云数据的预处理可以包括去噪处理和归一化处理。In the embodiment of the present application, in order to reduce the amount of data processing during subsequent model training, the training error is reduced. After the human body point cloud data is collected, the human body point cloud data can also be preprocessed. The preprocessing of human point cloud data can include denoising processing and normalization processing.
通常,采集的人体点云数据都会存在一些噪点,例如一些离群的点,可以通过对多个样本用户的人体点云数据进行去噪处理,将这些离群的点过滤掉,去除噪点对后续识别的影响。Generally, the collected human body point cloud data will have some noise, such as some outliers. You can denoise the human body point cloud data of multiple sample users to filter out these outliers and remove the noise for subsequent follow-up. Identify the impact.
然后,在进行归一化处理时,可以通过对去噪后的各个数据点的坐标值作比例 变换,将人体点云数据归一化为具有预设规格的人体点云数据。Then, during the normalization process, the human body point cloud data can be normalized to the human body point cloud data with preset specifications by performing scale transformation on the coordinate values of each data point after denoising.
通常,不同的人体点云数据的规格大小可能是不同的。例如,某些点云数据覆盖的区域为3*3*3,而另外一些点云数据覆盖的区域为6*6*6,因此,可以对全部的人体点云数据进行归一化处理,得到规格相同的处理后的人体点云数据。Generally, the size of different human point cloud data may be different. For example, the area covered by some point cloud data is 3*3*3, while the area covered by some point cloud data is 6*6*6. Therefore, all human body point cloud data can be normalized to get The processed human body point cloud data with the same specifications.
在具体实现中,由于人体点云数据中包括各个数据点的坐标值,因此可以根据各个坐标值生成一个包含全部人体点云数据中全部数据点的立方体,然后通过对各个数据点的坐标值作比例变换的方式,将各个数据点归一化为相同规格的立方体中的数据点。In specific implementation, since the human body point cloud data includes the coordinate value of each data point, a cube containing all the data points in all the human body point cloud data can be generated according to each coordinate value, and then the coordinate value of each data point can be calculated. The scale transformation method normalizes each data point to a data point in a cube of the same specification.
S203、根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人脸鼻尖位置;S203: According to the origin and direction of the preset coordinate system, identify the position of the data point corresponding to the maximum value of the coordinate value on the horizontal axis or the vertical axis of the coordinate system in the human body point cloud data as the nose tip position of the human face;
在本申请实施例中,在获得样本用户的人体点云数据后,为了减少后续模型训练的数据处理量,可以从这些人体点云数据中裁剪出人脸点云数据,以人脸点云数据作为后续训练的正样本数据。In this embodiment of the application, after obtaining the human body point cloud data of the sample user, in order to reduce the amount of data processing for subsequent model training, the face point cloud data can be cropped from the human body point cloud data, and the face point cloud data As the positive sample data for subsequent training.
通常,人的鼻子基本上处于人脸的居中位置。因此,为了从采集得到的人体点云数据中裁剪出人脸点云数据,可以首先识别出各个样本用户的鼻尖在各自的人体点云数据中的大致位置。Generally, the human nose is basically in the center of the human face. Therefore, in order to cut out the face point cloud data from the collected human body point cloud data, it is possible to first identify the approximate position of the nose tip of each sample user in the respective human body point cloud data.
由于人体点云数据是一种立体的三维数据,可以在构建出的坐标系中选择横轴或纵轴上的最大值所对应的位置作为人脸鼻尖位置。对于究竟是横轴最大值还是纵轴最大值对应的数据点位置作为人脸鼻尖位置,可以根据该坐标系横轴和纵轴的方向具体确定。Since the human body point cloud data is a three-dimensional three-dimensional data, the position corresponding to the maximum value on the horizontal axis or the vertical axis can be selected as the nose tip position in the constructed coordinate system. Whether the position of the data point corresponding to the maximum value of the horizontal axis or the maximum value of the vertical axis is used as the position of the nose tip of the human face can be specifically determined according to the directions of the horizontal axis and the vertical axis of the coordinate system.
一般而言,人体形状是近似于左右对称的。在采集得到人体点云数据后,可以首先确定出一个平面(第二平面),通过该第二平面可以将人体点云划分为左右两部分,并使得左右两部分的点云数量大致相等。然后,可以根据各个数据点的坐标值,确定出人体点云数据的中心点,并以该中心点作为待构建的坐标系的原点。在确定出坐标系的原点后,可以基于该原点构建出坐标系的横轴和纵轴,使得由上述横轴和纵轴构成的另一平面(第一平面)平行于水平面并且与上述第二平面垂直。这样,坐标系的横轴或纵轴便与第二平面平行。Generally speaking, the shape of the human body is approximately symmetrical. After the human body point cloud data is collected, a plane (second plane) can be determined first, by which the human body point cloud can be divided into two parts, and the number of point clouds in the left and right parts is approximately equal. Then, according to the coordinate value of each data point, the center point of the human body point cloud data can be determined, and the center point can be used as the origin of the coordinate system to be constructed. After the origin of the coordinate system is determined, the horizontal axis and the vertical axis of the coordinate system can be constructed based on the origin, so that another plane (the first plane) formed by the horizontal axis and the vertical axis is parallel to the horizontal plane and is parallel to the second plane. The plane is vertical. In this way, the horizontal or vertical axis of the coordinate system is parallel to the second plane.
如果上述坐标系的横轴与第二平面平行,则该坐标系的纵轴垂直与第二平面, 此时可以将横轴上的最大值所对应的位置作为人脸鼻尖位置;如果上述坐标系的纵轴与第二平面平行,则该坐标系的横轴垂直与第二平面,此时可以将纵轴上的最大值所对应的位置作为人脸鼻尖位置。需要说明的是,该最大值可以是坐标值绝对值的最大值。If the horizontal axis of the above-mentioned coordinate system is parallel to the second plane, the vertical axis of the above-mentioned coordinate system is perpendicular to the second plane, and the position corresponding to the maximum value on the horizontal axis can be used as the position of the nose tip of the face; if the above-mentioned coordinate system The vertical axis of is parallel to the second plane, and the horizontal axis of the coordinate system is perpendicular to the second plane. At this time, the position corresponding to the maximum value on the vertical axis can be used as the position of the nose tip of the human face. It should be noted that the maximum value may be the maximum absolute value of the coordinate value.
S204、以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸点云数据;S204: Constructing a three-dimensional coordinate system with the position of the nose tip of the face as the origin, and obtaining face point cloud data by extracting multiple data points within a preset length in each direction of the three-dimensional coordinate system;
在本申请实施例中,在确定人脸鼻尖的位置后,可以以该位置为原点,通过在坐标轴的各个方向上剪裁出一定长度的数据,从而可以得到人脸点云数据。In the embodiment of the present application, after determining the position of the nose tip of the human face, the position can be used as the origin, and data of a certain length can be cropped in various directions of the coordinate axis, so as to obtain the face point cloud data.
例如,可以以确定出的人脸鼻尖位置为原点构建三维坐标系,然后从原点出发,分别提取出坐标轴各个方向上一定长度范围内的数据点,进行人体点云数据的“抠脸”,得到人脸点云数据。For example, the position of the nose tip of a human face can be determined as the origin to construct a three-dimensional coordinate system, and then starting from the origin, data points within a certain length range in each direction of the coordinate axis can be extracted respectively, and the human body point cloud data can be "faced". Obtain face point cloud data.
S205、通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据;S205. Generate a face detection model by performing model training on the face point cloud data of the multiple sample users, and the face detection model stores one of the data points in the face point cloud data obtained after training. Sparsity data between;
在得到各个样本用户的人脸点云数据后,便可以以这些人脸点云数据作为正样本数据进行模型训练,获得人脸检测模型。After obtaining the face point cloud data of each sample user, the face point cloud data can be used as positive sample data for model training to obtain a face detection model.
在具体实现中,可以首秀将多个样本用户的人脸点云数据输入预置的三维点云网络模型PointNet++中进行模型训练。然后通过将PointNet++模型的全连接层配置为两层,生成二分类的人脸检测模型。In the specific implementation, the face point cloud data of multiple sample users can be input into the preset 3D point cloud network model PointNet++ for model training for the first time. Then, by configuring the fully connected layer of the PointNet++ model into two layers, a two-class face detection model is generated.
由于PointNet++模型是基于3D点云设计的深度学习多分类框架模型,利用该模型可以实现对3D点云呈现的数据进行物体分类。因此,在本申请实施例中,通过将上述PointNet++模型的输出结果修改为二分类,可以实现对被检测对象是否为人脸进行分类。即,通过PointNet++模型对被检测对象进行识别,相应的输出结果为人脸或不是人脸。Since the PointNet++ model is a deep learning multi-classification framework model based on 3D point cloud design, this model can be used to classify objects in the data presented by the 3D point cloud. Therefore, in the embodiment of the present application, by modifying the output result of the aforementioned PointNet++ model to a two-class classification, it is possible to classify whether the detected object is a human face. That is, the detected object is recognized through the PointNet++ model, and the corresponding output result is a human face or not a human face.
在具体实现中,可以通过将PointNet++模型的全连接层配置为输出结果为两类,并对预先采集的样本集进行训练即可实现对人脸和非人脸的分类。In specific implementation, the fully connected layer of the PointNet++ model can be configured to output two types of results, and the pre-collected sample set can be trained to realize the classification of faces and non-faces.
当然,为了提高模型后续识别的准确性,在将采集得到的人脸点云数据作为正样本数据输入PointNet++模型中进行训练,生成人脸检测模型的过程中,还可以 采集一些非人脸点云数据作为负样本数据进行训练。Of course, in order to improve the accuracy of subsequent recognition of the model, input the collected face point cloud data as positive sample data into the PointNet++ model for training. In the process of generating the face detection model, some non-face point clouds can also be collected The data is trained as negative sample data.
由于进行模型训练的上述样本集包括有人脸点云数据和非人脸点云数据,PointNet++模型可以通过对样本集进行训练获得人脸点云数据和非人脸点云数据各自之间的稀疏度数据。其中,人脸点云数据的稀疏度数据可以用于表示人脸点云数据中各个数据点所在的位置,以及各个数据点之间的相对位置关系。在后续的识别过程中,可以通过比较待检测的点云数据与人脸点云数据的稀疏度数据之间的相似度,判断出待检测的点云数据是否为人脸点云数据。Since the above sample set for model training includes face point cloud data and non-face point cloud data, the PointNet++ model can obtain the sparseness between face point cloud data and non-face point cloud data by training the sample set data. Among them, the sparsity data of the face point cloud data can be used to indicate the location of each data point in the face point cloud data and the relative positional relationship between the various data points. In the subsequent recognition process, it is possible to determine whether the point cloud data to be detected is face point cloud data by comparing the similarity between the point cloud data to be detected and the sparsity data of the face point cloud data.
S206、当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。S206: When the point cloud data of the object to be detected is received, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes a human face.
在本申请实施例中,由于二分类的PointNet++模型的输出结果只包括两种情况,一种是人脸,另一种不是人脸,因此,在采集到待检测对象的点云数据后,通过将点云数据输入至PointNet++模型中,可以直接识别出该点云数据是否为人脸。In the embodiment of this application, since the output result of the two-classified PointNet++ model only includes two cases, one is a human face, and the other is not a human face. Therefore, after collecting the point cloud data of the object to be detected, pass Input the point cloud data into the PointNet++ model, you can directly identify whether the point cloud data is a face.
在具体实现中,在采集到待检测对象的点云数据时,可以采用采用上述预先生成的二分类的人脸检测模型获取待检测对象的点云数据中各个数据点之间的稀疏度;通过计算各个数据点之间的稀疏度与人脸检测模型中存储的各个数据点之间的稀疏度数据的相似度,可以识别待检测对象的点云数据中是否包括人脸。In specific implementation, when the point cloud data of the object to be detected is collected, the above-mentioned pre-generated two-class face detection model can be used to obtain the sparsity between each data point in the point cloud data of the object to be detected; Calculating the similarity between the sparsity of each data point and the sparsity data between the data points stored in the face detection model can identify whether the point cloud data of the object to be detected includes a face.
一般地,若上述相似度超过一定阈值,则可以判定待检测对象的点云数据中包括人脸,否则,则不包括人脸。人脸检测模型可以实时地输出相应的检测结果。Generally, if the aforementioned similarity exceeds a certain threshold, it can be determined that the point cloud data of the object to be detected includes a human face, otherwise, it does not include a human face. The face detection model can output corresponding detection results in real time.
在本申请实施例中,通过将深度学习多分类框架模型PointNet++修改为二分类,并对现有的开源数据集进行预处理,可以得到一批仅有人脸信息的3D点云数据集,在采用上述3D点云数据集进行模型训练后,可以直接利用训练后的3D点云数据来进行人脸检测,无需借助RGB2D图像,提高了人脸检测的安全性。In the embodiment of this application, by modifying the deep learning multi-classification framework model PointNet++ to two classifications, and preprocessing the existing open source data sets, a batch of 3D point cloud data sets with only face information can be obtained. After the above-mentioned 3D point cloud data set is trained on the model, the trained 3D point cloud data can be directly used for face detection without the need to resort to RGB2D images, which improves the security of face detection.
需要说明的是,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本申请实施例的实 施过程构成任何限定。It should be noted that the size of the sequence number of each step in the above embodiment does not mean the order of execution. The execution sequence of each process should be determined by its function and internal logic, and should not constitute any implementation process of the embodiments of this application. limited.
参照图3,示出了本申请一个实施例的一种人脸检测的装置的示意图,具体可以包括如下模块:Referring to FIG. 3, a schematic diagram of a face detection apparatus according to an embodiment of the present application is shown, which may specifically include the following modules:
采集模块301,用于采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;The collection module 301 is configured to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
识别模块302,用于根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;The recognition module 302 is configured to recognize the position of the nose tip of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
裁剪模块303,用于基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;The cropping module 303 is configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
生成模块304,用于通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;The generating module 304 is configured to generate a face detection model by performing model training on the face point cloud data of the multiple sample users;
检测模块305,用于在接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。The detection module 305 is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected is Including human faces.
在本申请实施例中,所述装置还可以包括如下模块:In the embodiment of the present application, the device may further include the following modules:
去噪模块,用于对所述多个样本用户的人体点云数据进行去噪处理;A denoising module, configured to perform denoising processing on the human body point cloud data of the multiple sample users;
归一化模块,用于通过对去噪后的所述各个数据点的坐标值作比例变换,将所述人体点云数据归一化为具有预设规格的目标点云数据。The normalization module is used to normalize the human body point cloud data into target point cloud data with preset specifications by performing proportional transformation on the coordinate values of the respective data points after denoising.
在本申请实施例中,所述识别模块302具体可以包括如下子模块:In the embodiment of the present application, the identification module 302 may specifically include the following sub-modules:
人脸鼻尖位置识别子模块,用于根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人脸鼻尖位置;其中,所述坐标系的原点为所述人体点云数据的中心点,所述坐标系的横轴和纵轴构成的第一平面平行于水平面且垂直于第二平面,所述第二平面用于将所述人体点云数据划分为两部分,所述坐标系的横轴或纵轴与所述第二平面平行。The nose tip position recognition sub-module of the face is used to identify the data point corresponding to the maximum value of the coordinate value on the horizontal or vertical axis of the coordinate system in the human body point cloud data according to the origin and direction of the preset coordinate system The position is the position of the nose tip of the human face; wherein the origin of the coordinate system is the center point of the human body point cloud data, and the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane, The second plane is used to divide the human body point cloud data into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
在本申请实施例中,所述裁剪模块303具体可以包括如下子模块:In the embodiment of the present application, the cropping module 303 may specifically include the following sub-modules:
人脸点云数据提取子模块,用于以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸 点云数据。The face point cloud data extraction sub-module is used to construct a three-dimensional coordinate system with the nose tip position of the face as the origin, and obtain a face by extracting multiple data points within a preset length in each direction of the three-dimensional coordinate system Point cloud data.
在本申请实施例中,所述生成模块304具体可以包括如下子模块:In the embodiment of the present application, the generating module 304 may specifically include the following sub-modules:
模型训练子模块,用于将所述多个样本用户的人脸点云数据输入预置的三维点云网络模型中进行模型训练;The model training sub-module is used to input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
模型配置子模块,用于通过将所述三维点云网络模型的全连接层配置为两层,生成二分类的人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据。The model configuration sub-module is used to generate a two-class face detection model by configuring the fully connected layer of the three-dimensional point cloud network model into two layers, and the face detection model stores the face obtained after training The sparsity data between each data point in the point cloud data.
在本申请实施例中,所述检测模块305具体可以包括如下子模块:In the embodiment of the present application, the detection module 305 may specifically include the following sub-modules:
稀疏度获取子模块,用于采用所述二分类的人脸检测模型获取所述待检测对象的点云数据中各个数据点之间的稀疏度;A sparsity acquisition sub-module for acquiring the sparsity between various data points in the point cloud data of the object to be detected by using the two-class face detection model;
相似度计算子模块,用于计算所述各个数据点之间的稀疏度与所述人脸检测模型中存储的各个数据点之间的稀疏度数据的相似度;A similarity calculation sub-module for calculating the similarity between the sparsity between the various data points and the sparsity data between the various data points stored in the face detection model;
人脸检测子模块,用于在所述相似度超过预设阈值时,识别所述待检测对象的点云数据中包括人脸。The face detection sub-module is configured to recognize that the point cloud data of the object to be detected includes a face when the similarity exceeds a preset threshold.
对于装置实施例而言,由于其与方法实施例基本相似,所以描述得比较简单,相关之处参见方法实施例部分的说明即可。As for the device embodiment, since it is basically similar to the method embodiment, the description is relatively simple, and for related parts, please refer to the description of the method embodiment part.
参照图4,示出了本申请一个实施例的一种终端设备的示意图。如图4所示,本实施例的终端设备400包括:处理器410、存储器420以及存储在所述存储器420中并可在所述处理器410上运行的计算机可读指令421。所述处理器410执行所述计算机可读指令421时实现上述人脸检测的方法各个实施例中的步骤,例如图1所示的步骤S101至S105。或者,所述处理器410执行所述计算机可读指令421时实现上述各装置实施例中各模块/单元的功能,例如图3所示模块301至305的功能。Referring to FIG. 4, a schematic diagram of a terminal device according to an embodiment of the present application is shown. As shown in FIG. 4, the terminal device 400 of this embodiment includes a processor 410, a memory 420, and computer-readable instructions 421 stored in the memory 420 and running on the processor 410. When the processor 410 executes the computer-readable instructions 421, the steps in the various embodiments of the above-mentioned face detection method are implemented, for example, steps S101 to S105 shown in FIG. 1. Alternatively, when the processor 410 executes the computer-readable instructions 421, the functions of the modules/units in the foregoing device embodiments, such as the functions of the modules 301 to 305 shown in FIG. 3, are implemented.
示例性的,所述计算机可读指令421可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器420中,并由所述处理器410执行,以完成本申请。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机可读指令段,该指令段可以用于描述所述计算机可读指令421在所述终端设备400中的执行过程。例如,所述计算机可读指令421可以被分割成采集模块、 识别模块、裁剪模块、生成模块和检测模块,各模块具体功能如下:Exemplarily, the computer-readable instructions 421 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 420 and executed by the processor 410. To complete this application. The one or more modules/units may be a series of computer-readable instruction segments capable of completing specific functions, and the instruction segments may be used to describe the execution process of the computer-readable instructions 421 in the terminal device 400. For example, the computer-readable instructions 421 can be divided into a collection module, an identification module, a cropping module, a generation module, and a detection module. The specific functions of each module are as follows:
采集模块,用于采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;The collection module is used to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
识别模块,用于根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;The recognition module is used to recognize the nose tip position of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
裁剪模块,用于基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;A cropping module, configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
生成模块,用于通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;A generating module for generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
检测模块,用于在接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。The detection module is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes human face.
所述终端设备400可以是桌上型计算机、笔记本、掌上电脑及云端服务器等计算设备。所述终端设备400可包括,但不仅限于,处理器410、存储器420。本领域技术人员可以理解,图4仅仅是终端设备400的一种示例,并不构成对终端设备400的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述终端设备400还可以包括输入输出设备、网络接入设备、总线等。The terminal device 400 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server. The terminal device 400 may include, but is not limited to, a processor 410 and a memory 420. Those skilled in the art can understand that FIG. 4 is only an example of the terminal device 400, and does not constitute a limitation on the terminal device 400. It may include more or less components than shown in the figure, or combine certain components, or different components. For example, the terminal device 400 may also include input and output devices, network access devices, buses, and so on.
所述处理器410可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,DSP)、专用集成电路(Application Specific Integrated Circuit,ASIC)、现成可编程门阵列(Field-Programmable Gate Array,FPGA)或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 410 may be a central processing unit (Central Processing Unit, CPU), or other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc. The general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
所述存储器420可以是所述终端设备400的内部存储单元,例如终端设备400的硬盘或内存。所述存储器420也可以是所述终端设备400的外部存储设备,例如所述终端设备400上配备的插接式硬盘,智能存储卡(Smart Media Card,SMC),安全数字(Secure Digital,SD)卡,闪存卡(Flash Card)等等。进一步地,所 述存储器420还可以既包括所述终端设备400的内部存储单元也包括外部存储设备。所述存储器420用于存储所述计算机可读指令421以及所述终端设备400所需的其他指令和数据。所述存储器420还可以用于暂时地存储已经输出或者将要输出的数据。The memory 420 may be an internal storage unit of the terminal device 400, such as a hard disk or a memory of the terminal device 400. The memory 420 may also be an external storage device of the terminal device 400, such as a plug-in hard disk equipped on the terminal device 400, a smart memory card (Smart Media Card, SMC), or a Secure Digital (SD). Card, Flash Card, etc. Further, the memory 420 may also include both an internal storage unit of the terminal device 400 and an external storage device. The memory 420 is used to store the computer-readable instructions 421 and other instructions and data required by the terminal device 400. The memory 420 may also be used to temporarily store data that has been output or will be output.
本领域普通技术人员可以理解,实现上述实施例方法中的全部或部分流程,是可以通过计算机可读指令来指令相关的硬件来完成,所述的计算机可读指令可存储于一非易失性计算机可读取存储介质中,该计算机可读指令在执行时,可包括如上述各方法的实施例的流程。其中,本申请所提供的各实施例中所使用的对存储器、存储、数据库或其它介质的任何引用,均可包括非易失性和/或易失性存储器。非易失性存储器可包括只读存储器(ROM)、可编程ROM(PROM)、电可编程ROM(EPROM)、电可擦除可编程ROM(EEPROM)或闪存。易失性存储器可包括随机存取存储器(RAM)或者外部高速缓冲存储器。作为说明而非局限,RAM以多种形式可得,诸如静态RAM(SRAM)、动态RAM(DRAM)、同步DRAM(SDRAM)、双数据率SDRAM(DDRSDRAM)、增强型SDRAM(ESDRAM)、同步链路(Synchlink)DRAM(SLDRAM)、存储器总线(Rambus)直接RAM(RDRAM)、直接存储器总线动态RAM(DRDRAM)、以及存储器总线动态RAM(RDRAM)等。A person of ordinary skill in the art can understand that the implementation of all or part of the processes in the methods of the above-mentioned embodiments can be accomplished by instructing relevant hardware through computer-readable instructions, and the computer-readable instructions can be stored in a non-volatile In a computer-readable storage medium, when the computer-readable instructions are executed, they may include the processes of the above-mentioned method embodiments. Wherein, any reference to memory, storage, database or other media used in the embodiments provided in this application may include non-volatile and/or volatile memory. Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory. Volatile memory may include random access memory (RAM) or external cache memory. As an illustration and not a limitation, RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.
以上所述实施例仅用以说明本申请的技术方案,而非对其限制;尽管参照前述实施例对本申请进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本申请各实施例技术方案的精神和范围,均应包含在本申请的保护范围之内。The above-mentioned embodiments are only used to illustrate the technical solutions of the present application, not to limit them; although the present application has been described in detail with reference to the foregoing embodiments, a person of ordinary skill in the art should understand that it can still implement the foregoing The technical solutions recorded in the examples are modified, or some of the technical features are equivalently replaced; these modifications or replacements do not cause the essence of the corresponding technical solutions to deviate from the spirit and scope of the technical solutions of the embodiments of the application, and should be included in Within the scope of protection of this application.

Claims (20)

  1. 一种人脸检测的方法,其特征在于,包括:A method for face detection, which is characterized in that it includes:
    采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
    根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
    基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
    通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
    当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  2. 根据权利要求1所述的方法,其特征在于,在所述采集多个样本用户的人体点云数据的步骤后,还包括:The method according to claim 1, characterized in that, after the step of collecting human body point cloud data of a plurality of sample users, the method further comprises:
    对所述多个样本用户的人体点云数据进行去噪处理;Denoising processing on the human body point cloud data of the multiple sample users;
    通过对去噪后的所述各个数据点的坐标值作比例变换,将所述人体点云数据归一化为具有预设规格的人体点云数据。The human body point cloud data is normalized to human body point cloud data with preset specifications by performing proportional transformation on the coordinate values of the respective data points after denoising.
  3. 根据权利要求1所述的方法,其特征在于,所述根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置的步骤包括:The method according to claim 1, wherein the step of respectively identifying the position of the nose tip of the face in the human body point cloud data of each sample user according to the coordinate value of each data point comprises:
    根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人脸鼻尖位置;其中,所述坐标系的原点为所述人体点云数据的中心点,所述坐标系的横轴和纵轴构成的第一平面平行于水平面且垂直于第二平面,所述第二平面用于将所述人体点云数据划分为两部分,所述坐标系的横轴或纵轴与所述第二平面平行。According to the origin and direction of the preset coordinate system, the position of the data point corresponding to the maximum value of the coordinate value on the horizontal axis or the vertical axis of the coordinate system in the human point cloud data is identified as the nose tip position of the human face; The origin of the coordinate system is the center point of the human body point cloud data, the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane, and the second plane is used to transfer the The human body point cloud data is divided into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
  4. 根据权利要求3所述的方法,其特征在于,所述基于所述人脸鼻尖 位置,从所述人体点云数据中裁剪出人脸点云数据的步骤包括:The method according to claim 3, wherein the step of cropping face point cloud data from the human body point cloud data based on the position of the nose tip of the face comprises:
    以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸点云数据。A three-dimensional coordinate system is constructed with the position of the nose tip of the human face as the origin, and a plurality of data points within a preset length in each direction of the three-dimensional coordinate system are extracted to obtain the face point cloud data.
  5. 根据权利要求1所述的方法,其特征在于,所述通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型的步骤包括:The method according to claim 1, wherein the step of generating a face detection model by performing model training on the face point cloud data of the multiple sample users comprises:
    将所述多个样本用户的人脸点云数据输入预置的三维点云网络模型中进行模型训练;Input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
    通过将所述三维点云网络模型的全连接层配置为两层,生成二分类的人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据。By configuring the fully connected layer of the three-dimensional point cloud network model into two layers, a two-class face detection model is generated, and the face detection model stores each data point in the face point cloud data obtained after training Sparsity data between.
  6. 根据权利要求5所述的方法,其特征在于,所述采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸的步骤包括:The method of claim 5, wherein the face detection model is used to detect the point cloud data of the object to be detected, and to identify whether the point cloud data of the object to be detected includes a human face The steps include:
    采用所述二分类的人脸检测模型获取所述待检测对象的点云数据中各个数据点之间的稀疏度;Acquiring the sparsity between each data point in the point cloud data of the object to be detected by using the two-class face detection model;
    计算所述各个数据点之间的稀疏度与所述人脸检测模型中存储的各个数据点之间的稀疏度数据的相似度;Calculating the similarity between the sparsity between the various data points and the sparsity data between the various data points stored in the face detection model;
    若所述相似度超过预设阈值,则识别所述待检测对象的点云数据中包括人脸。If the similarity exceeds a preset threshold, the point cloud data identifying the object to be detected includes a human face.
  7. 一种人脸检测的装置,其特征在于,包括:A face detection device, which is characterized in that it comprises:
    采集模块,用于采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;The collection module is used to collect human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
    识别模块,用于根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;The recognition module is used to recognize the nose tip position of the face in the human body point cloud data of each sample user according to the coordinate value of each data point;
    裁剪模块,用于基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;A cropping module, configured to crop the face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
    生成模块,用于通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;A generating module for generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
    检测模块,用于在接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。The detection module is configured to, when receiving the point cloud data of the object to be detected, use the face detection model to detect the point cloud data of the object to be detected, and identify whether the point cloud data of the object to be detected includes human face.
  8. 根据权利要求7所述的装置,其特征在于,所述识别模块包括:The device according to claim 7, wherein the identification module comprises:
    人脸鼻尖位置识别子模块,用于根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人脸鼻尖位置;其中,所述坐标系的原点为所述人体点云数据的中心点,所述坐标系的横轴和纵轴构成的第一平面平行于水平面且垂直于第二平面,所述第二平面用于将所述人体点云数据划分为两部分,所述坐标系的横轴或纵轴与所述第二平面平行。The nose tip position recognition sub-module of the face is used to identify the data point corresponding to the maximum value of the coordinate value on the horizontal or vertical axis of the coordinate system in the human body point cloud data according to the origin and direction of the preset coordinate system The position is the position of the nose tip of the human face; wherein the origin of the coordinate system is the center point of the human body point cloud data, and the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane, The second plane is used to divide the human body point cloud data into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
  9. 根据权利要求8所述的装置,其特征在于,所述裁剪模块包括:The device according to claim 8, wherein the cutting module comprises:
    人脸点云数据提取子模块,用于以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸点云数据。The face point cloud data extraction sub-module is used to construct a three-dimensional coordinate system with the nose tip position of the face as the origin, and obtain a face by extracting multiple data points within a preset length in each direction of the three-dimensional coordinate system Point cloud data.
  10. 根据权利要求7所述的装置,其特征在于,所述生成模块包括:The device according to claim 7, wherein the generating module comprises:
    模型训练子模块,用于将所述多个样本用户的人脸点云数据输入预置的三维点云网络模型中进行模型训练;The model training sub-module is used to input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
    模型配置子模块,用于通过将所述三维点云网络模型的全连接层配置为两层,生成二分类的人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据。The model configuration sub-module is used to generate a two-class face detection model by configuring the fully connected layer of the three-dimensional point cloud network model into two layers, and the face detection model stores the face obtained after training The sparsity data between each data point in the point cloud data.
  11. 一种终端设备,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机可读指令,其特征在于,所述处理器执行所述计算机可读指令时实现如下步骤:A terminal device, comprising a memory, a processor, and computer-readable instructions stored in the memory and running on the processor, wherein the processor executes the computer-readable instructions as follows step:
    采集多个样本用户的人体点云数据,所述人体点云数据包括多个 数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
    根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
    基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
    通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
    当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  12. 根据权利要求11所述的终端设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The terminal device according to claim 11, wherein the processor further implements the following steps when executing the computer-readable instruction:
    根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人脸鼻尖位置;其中,所述坐标系的原点为所述人体点云数据的中心点,所述坐标系的横轴和纵轴构成的第一平面平行于水平面且垂直于第二平面,所述第二平面用于将所述人体点云数据划分为两部分,所述坐标系的横轴或纵轴与所述第二平面平行。According to the origin and direction of the preset coordinate system, the position of the data point corresponding to the maximum value of the coordinate value on the horizontal or vertical axis of the coordinate system in the human point cloud data is identified as the nose tip position of the human face; The origin of the coordinate system is the center point of the human body point cloud data, the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane, and the second plane is used to transfer the The human body point cloud data is divided into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
  13. 根据权利要求12所述的终端设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The terminal device according to claim 12, wherein the processor further implements the following steps when executing the computer-readable instruction:
    以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸点云数据。A three-dimensional coordinate system is constructed with the position of the nose tip of the human face as the origin, and a plurality of data points within a preset length in each direction of the three-dimensional coordinate system are extracted to obtain the face point cloud data.
  14. 根据权利要求11所述的终端设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The terminal device according to claim 11, wherein the processor further implements the following steps when executing the computer-readable instruction:
    将所述多个样本用户的人脸点云数据输入预置的三维点云网络模型中进行模型训练;Input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
    通过将所述三维点云网络模型的全连接层配置为两层,生成二分 类的人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据。By configuring the fully connected layer of the three-dimensional point cloud network model into two layers, a two-class face detection model is generated, and the face detection model stores each data point in the face point cloud data obtained after training Sparsity data between.
  15. 根据权利要求14所述的终端设备,其特征在于,所述处理器执行所述计算机可读指令时还实现如下步骤:The terminal device according to claim 14, wherein the processor further implements the following steps when executing the computer-readable instruction:
    采用所述二分类的人脸检测模型获取所述待检测对象的点云数据中各个数据点之间的稀疏度;Acquiring the sparsity between each data point in the point cloud data of the object to be detected by using the two-class face detection model;
    计算所述各个数据点之间的稀疏度与所述人脸检测模型中存储的各个数据点之间的稀疏度数据的相似度;Calculating the similarity between the sparsity between the various data points and the sparsity data between the various data points stored in the face detection model;
    若所述相似度超过预设阈值,则识别所述待检测对象的点云数据中包括人脸。If the similarity exceeds a preset threshold, the point cloud data identifying the object to be detected includes a human face.
  16. 一种计算机非易失性可读存储介质,所述计算机非易失性可读存储介质存储有计算机可读指令,其特征在于,所述计算机可读指令被处理器执行时实现如下步骤:A computer non-volatile readable storage medium, the computer non-volatile readable storage medium storing computer readable instructions, wherein the computer readable instructions are executed by a processor to implement the following steps:
    采集多个样本用户的人体点云数据,所述人体点云数据包括多个数据点,各个数据点分别具有相应的坐标值;Collecting human body point cloud data of multiple sample users, where the human body point cloud data includes multiple data points, and each data point has a corresponding coordinate value;
    根据所述各个数据点的坐标值,分别识别各个样本用户的人体点云数据中的人脸鼻尖位置;According to the coordinate value of each data point, respectively identify the position of the nose tip of the face in the human body point cloud data of each sample user;
    基于所述人脸鼻尖位置,从所述人体点云数据中裁剪出人脸点云数据;Cropping out face point cloud data from the human body point cloud data based on the position of the nose tip of the face;
    通过对所述多个样本用户的人脸点云数据进行模型训练,生成人脸检测模型;Generating a face detection model by performing model training on the face point cloud data of the multiple sample users;
    当接收到待检测对象的点云数据时,采用所述人脸检测模型对所述待检测对象的点云数据进行检测,识别所述待检测对象的点云数据中是否包括人脸。When the point cloud data of the object to be detected is received, the face detection model is used to detect the point cloud data of the object to be detected, and it is recognized whether the point cloud data of the object to be detected includes a human face.
  17. 根据权利要求16所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被处理器执行时还实现如下步骤:The computer non-volatile readable storage medium according to claim 16, wherein the computer readable instruction further implements the following steps when being executed by the processor:
    根据预设的坐标系的原点和方向,识别所述人体点云数据中在所述坐标系的横轴或纵轴上的坐标值最大值对应的数据点位置为人 脸鼻尖位置;其中,所述坐标系的原点为所述人体点云数据的中心点,所述坐标系的横轴和纵轴构成的第一平面平行于水平面且垂直于第二平面,所述第二平面用于将所述人体点云数据划分为两部分,所述坐标系的横轴或纵轴与所述第二平面平行。According to the origin and direction of the preset coordinate system, the position of the data point corresponding to the maximum value of the coordinate value on the horizontal axis or the vertical axis of the coordinate system in the human point cloud data is identified as the nose tip position of the human face; The origin of the coordinate system is the center point of the human body point cloud data, the first plane formed by the horizontal axis and the vertical axis of the coordinate system is parallel to the horizontal plane and perpendicular to the second plane, and the second plane is used to transfer the The human body point cloud data is divided into two parts, and the horizontal axis or the vertical axis of the coordinate system is parallel to the second plane.
  18. 根据权利要求17所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被处理器执行时还实现如下步骤:The computer non-volatile readable storage medium according to claim 17, wherein the computer readable instruction further implements the following steps when being executed by the processor:
    以所述人脸鼻尖位置为原点构建三维坐标系,通过提取在所述三维坐标系的各个方向上预设长度内的多个数据点,获得人脸点云数据。A three-dimensional coordinate system is constructed with the position of the nose tip of the human face as the origin, and a plurality of data points within a preset length in each direction of the three-dimensional coordinate system are extracted to obtain the face point cloud data.
  19. 根据权利要求16所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被处理器执行时还实现如下步骤:The computer non-volatile readable storage medium according to claim 16, wherein the computer readable instruction further implements the following steps when being executed by the processor:
    将所述多个样本用户的人脸点云数据输入预置的三维点云网络模型中进行模型训练;Input the face point cloud data of the multiple sample users into a preset three-dimensional point cloud network model for model training;
    通过将所述三维点云网络模型的全连接层配置为两层,生成二分类的人脸检测模型,所述人脸检测模型存储有训练后获得的所述人脸点云数据中各个数据点之间的稀疏度数据。By configuring the fully connected layer of the three-dimensional point cloud network model into two layers, a two-class face detection model is generated, and the face detection model stores each data point in the face point cloud data obtained after training Sparsity data between.
  20. 根据权利要求19所述的计算机非易失性可读存储介质,其特征在于,所述计算机可读指令被处理器执行时还实现如下步骤:The computer non-volatile readable storage medium of claim 19, wherein the computer readable instruction further implements the following steps when being executed by the processor:
    采用所述二分类的人脸检测模型获取所述待检测对象的点云数据中各个数据点之间的稀疏度;Acquiring the sparsity between each data point in the point cloud data of the object to be detected by using the two-class face detection model;
    计算所述各个数据点之间的稀疏度与所述人脸检测模型中存储的各个数据点之间的稀疏度数据的相似度;Calculating the similarity between the sparsity between the various data points and the sparsity data between the various data points stored in the face detection model;
    若所述相似度超过预设阈值,则识别所述待检测对象的点云数据中包括人脸。If the similarity exceeds a preset threshold, the point cloud data identifying the object to be detected includes a human face.
PCT/CN2019/117181 2019-09-18 2019-11-11 Face detection method and apparatus, and terminal device WO2021051538A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910882002.0A CN110717406B (en) 2019-09-18 2019-09-18 Face detection method and device and terminal equipment
CN201910882002.0 2019-09-18

Publications (1)

Publication Number Publication Date
WO2021051538A1 true WO2021051538A1 (en) 2021-03-25

Family

ID=69210571

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/117181 WO2021051538A1 (en) 2019-09-18 2019-11-11 Face detection method and apparatus, and terminal device

Country Status (2)

Country Link
CN (1) CN110717406B (en)
WO (1) WO2021051538A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7380633B2 (en) * 2021-04-01 2023-11-15 トヨタ自動車株式会社 Monitoring device, monitoring method and monitoring system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105654035A (en) * 2015-12-21 2016-06-08 湖南拓视觉信息技术有限公司 Three-dimensional face recognition method and data processing device applying three-dimensional face recognition method
CN105809113A (en) * 2016-03-01 2016-07-27 湖南拓视觉信息技术有限公司 Three-dimensional human face identification method and data processing apparatus using the same
CN106096555A (en) * 2016-06-15 2016-11-09 湖南拓视觉信息技术有限公司 The method and apparatus of three dimensional face detection
CN109670487A (en) * 2019-01-30 2019-04-23 汉王科技股份有限公司 A kind of face identification method, device and electronic equipment
CN110175529A (en) * 2019-04-30 2019-08-27 东南大学 A kind of three-dimensional face features' independent positioning method based on noise reduction autoencoder network

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105701448B (en) * 2015-12-31 2019-08-09 湖南拓视觉信息技术有限公司 Three-dimensional face point cloud nose detection method and the data processing equipment for applying it
CN110020578A (en) * 2018-01-10 2019-07-16 广东欧珀移动通信有限公司 Image processing method, device, storage medium and electronic equipment

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105654035A (en) * 2015-12-21 2016-06-08 湖南拓视觉信息技术有限公司 Three-dimensional face recognition method and data processing device applying three-dimensional face recognition method
CN105809113A (en) * 2016-03-01 2016-07-27 湖南拓视觉信息技术有限公司 Three-dimensional human face identification method and data processing apparatus using the same
CN106096555A (en) * 2016-06-15 2016-11-09 湖南拓视觉信息技术有限公司 The method and apparatus of three dimensional face detection
CN109670487A (en) * 2019-01-30 2019-04-23 汉王科技股份有限公司 A kind of face identification method, device and electronic equipment
CN110175529A (en) * 2019-04-30 2019-08-27 东南大学 A kind of three-dimensional face features' independent positioning method based on noise reduction autoencoder network

Also Published As

Publication number Publication date
CN110717406B (en) 2024-04-09
CN110717406A (en) 2020-01-21

Similar Documents

Publication Publication Date Title
WO2021051539A1 (en) Face recognition method and apparatus, and terminal device
US10509985B2 (en) Method and apparatus for security inspection
US10049262B2 (en) Method and system for extracting characteristic of three-dimensional face image
WO2021139324A1 (en) Image recognition method and apparatus, computer-readable storage medium and electronic device
WO2020199479A1 (en) Human motion recognition method and device
US20190147245A1 (en) Three-dimensional object detection for autonomous robotic systems using image proposals
TW202006602A (en) Three-dimensional living-body face detection method, face authentication recognition method, and apparatuses
WO2015165365A1 (en) Facial recognition method and system
CN110866466B (en) Face recognition method, device, storage medium and server
CN108629336B (en) Face characteristic point identification-based color value calculation method
Li Analysis of object detection performance based on Faster R-CNN
WO2015070764A1 (en) Face positioning method and device
CN110889355B (en) Face recognition verification method, face recognition verification system and storage medium
TW200910223A (en) Image processing apparatus and image processing method
WO2021031817A1 (en) Emotion recognition method and device, computer device, and storage medium
CN109815823B (en) Data processing method and related product
CN112101208A (en) Feature series fusion gesture recognition method and device for elderly people
CN112686191B (en) Living body anti-counterfeiting method, system, terminal and medium based on three-dimensional information of human face
CN111488943A (en) Face recognition method and device
CN113298158A (en) Data detection method, device, equipment and storage medium
WO2021203718A1 (en) Method and system for facial recognition
CN112200056A (en) Face living body detection method and device, electronic equipment and storage medium
CN111368581A (en) Face recognition method based on TOF camera module, face recognition device and electronic equipment
CN108875488B (en) Object tracking method, object tracking apparatus, and computer-readable storage medium
CN110969101A (en) Face detection and tracking method based on HOG and feature descriptor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19945753

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19945753

Country of ref document: EP

Kind code of ref document: A1