WO2019127273A1 - Multi-person face detection method, apparatus, server, system, and storage medium - Google Patents

Multi-person face detection method, apparatus, server, system, and storage medium Download PDF

Info

Publication number
WO2019127273A1
WO2019127273A1 PCT/CN2017/119569 CN2017119569W WO2019127273A1 WO 2019127273 A1 WO2019127273 A1 WO 2019127273A1 CN 2017119569 W CN2017119569 W CN 2017119569W WO 2019127273 A1 WO2019127273 A1 WO 2019127273A1
Authority
WO
WIPO (PCT)
Prior art keywords
video image
training
face
network
face information
Prior art date
Application number
PCT/CN2017/119569
Other languages
French (fr)
Chinese (zh)
Inventor
李恒
刘光军
Original Assignee
深圳市锐明技术股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳市锐明技术股份有限公司 filed Critical 深圳市锐明技术股份有限公司
Priority to PCT/CN2017/119569 priority Critical patent/WO2019127273A1/en
Priority to CN201780002310.9A priority patent/CN108351967A/en
Publication of WO2019127273A1 publication Critical patent/WO2019127273A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation

Definitions

  • the present invention relates to the field of video information processing technologies, and in particular, to a multi-face detection method, apparatus, server, system, and storage medium.
  • vehicle face detection mostly focuses on the aspect of passenger flow statistics on passenger face detection, and has achieved certain results in face recognition on vehicles. Under normal circumstances, passenger flow statistics and analysis can be completed.
  • the embodiment of the invention provides a multi-face detection method, device, server, system and storage medium, which can identify the face information of multiple passengers and upload them to a designated platform server for the flow rate statistics of the designated platform server. And analysis, to avoid the problem of face missing detection when multiple passengers get on the train at the same time.
  • a multi-face detection method including:
  • the identified individual face information is uploaded to a designated platform server.
  • DeepID network is pre-trained by the following steps:
  • training group samples including a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
  • the DeepID network includes three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner.
  • the video image of the passenger on the target vehicle when loading the vehicle includes:
  • the captured image is the video image.
  • a multi-face detection device including:
  • a video image acquisition module configured to collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
  • a dimension adjustment module configured to adjust a data dimension of the video image to a specified data dimension
  • a face recognition module configured to input the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtain each face information that is recognized by the DeepID network from the video image;
  • the information uploading module is configured to upload the identified individual face information to a designated platform server.
  • DeepID network is pre-trained by the following modules:
  • a training sample collection module configured to pre-collect a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
  • a face information marking module configured to pre-mark standard face information corresponding to each first video image in the training group sample
  • a first adjustment module configured to adjust a data dimension of the first video image to a specified data dimension
  • a network training module configured to input the first video image with the data dimension adjusted as input to the DeepID network, and obtain the training face information that is obtained by the DeepID network from the first video image;
  • a network parameter adjustment module configured to use the training face information as a target, and adjust network parameters of the DeepID network to minimize the obtained training face information and standard face information corresponding to the training group sample. Error between
  • the training completion module is configured to determine that the DeepID network training is completed if the error satisfies a preset condition.
  • a platform server comprising a memory, a processor, and a computer program stored in the memory and operable on the processor, the processor implementing the computer program to implement the above The steps of the face detection method.
  • a multi-face detection system including a camera, an in-vehicle intelligent control terminal, and the platform server described above;
  • the camera is installed at a designated position of the target vehicle for capturing a video image of a passenger on the target vehicle when getting on the vehicle;
  • the in-vehicle intelligent control terminal is configured to upload a video image captured by the camera to the platform server.
  • a computer readable storage medium storing a computer program, the computer program being executed by a processor to implement the steps of the multi-face detection method.
  • a video image when a passenger on a target vehicle gets on the vehicle is collected, the video image includes a face image of the boarding passenger; then, the data dimension of the video image is adjusted to a specified data dimension; And inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image; and finally, each of the identified
  • the personal face information is uploaded to a designated platform server, so that the platform server performs passenger flow statistics and analysis on the target vehicle.
  • the multiple IDs can be accurately detected by pre-training the DeepID network, and the face information of each passenger is identified and uploaded to the designated platform server.
  • the passenger flow statistics and analysis are carried out for the designated platform server, which avoids the problem of face missing detection when multiple passengers get on the train at the same time.
  • FIG. 1 is a flowchart of an embodiment of a multi-face detection method according to an embodiment of the present invention
  • FIG. 2 is a schematic flowchart of step 101 of a multi-face detection method in an application scenario according to an embodiment of the present invention
  • FIG. 3 is a schematic flowchart of pre-training a DeepID network in an application scenario according to an embodiment of the present invention
  • FIG. 4 is a schematic flowchart of testing whether a DeepID network is trained in an application scenario in a multi-face detection method according to an embodiment of the present invention
  • FIG. 5 is a structural diagram of an embodiment of a multi-face detecting device according to an embodiment of the present invention.
  • FIG. 6 is a schematic diagram of a platform server according to an embodiment of the present invention.
  • the embodiment of the invention provides a multi-face detection method, device, server, system and storage medium, which are used for solving the problem of face missing detection when multiple passengers get on the vehicle at the same time.
  • an embodiment of a multi-face detection method includes:
  • a plurality of cameras can be installed at appropriate positions on the target vehicle, the cameras being aligned with the positions of the passengers on the front and rear doors of the target vehicle.
  • the cameras can capture the video image containing the face image of the passenger, so that the executor of the embodiment can collect the videos. image.
  • the execution subject of the embodiment may be an in-vehicle intelligent terminal, a system, or a platform server installed on a vehicle.
  • the following is uniformly expressed as an execution subject.
  • the foregoing step 101 may include:
  • step 201 Detect whether the door of the target vehicle is open or the target vehicle is in the station, and if yes, proceed to step 202, and if not, continue to wait;
  • the camera when the camera is controlled to start capturing, it is also detected whether there is a face in the captured image. It can be understood that when the door of the target vehicle is opened or the target vehicle is pitted, not every time a passenger gets on the bus, there may be no face in the image captured by the camera. In order to avoid that the image of the face does not occupy the computing resources of the execution subject, it is possible to detect whether the image has a human face, and if so, determine that the image is a video image, and then hand it to the execution subject for subsequent processing.
  • the data dimension of the video image needs to be adjusted so that the data dimension meets the requirements of the DeepID network. If the original data dimension of the video image is high, PCA (Principal) can be used. Component Analysis (Principal Component Analysis) and other methods perform pre-dimension reduction processing; conversely, if the original data dimension of the video image is low, the video image may be subjected to up-dimensional processing; finally, the data dimension of the video image is equal to the specified data. Dimensions.
  • the video image with the adjusted data dimension is input as input to the pre-trained DeepID network, and each face information identified by the DeepID network from the video image is obtained;
  • the video image may be input as input to the pre-trained DeepID network, and each face information identified by the DeepID network from the video image is obtained.
  • the DeepID network is obtained by training a large number of training samples in advance, and can perform face recognition on the face information in the video image, and can accurately identify each face in the current video image, thereby Output the corresponding face information.
  • the pre-training process and network structure for the DeepID network will be described in detail below.
  • the identified individual face information may be uploaded to a designated platform server, so that the platform server Passenger flow statistics and analysis are performed on the target vehicle.
  • these face information can be uploaded to the public security server for criminal monitoring and searching, etc., and the application scenarios are extremely extensive.
  • the DeepID network can be pre-trained by the following steps:
  • the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
  • the first video image with the data dimension adjusted is input as an input to the DeepID network, and each training face information that is obtained by the DeepID network from the first video image is obtained.
  • the training face information is used as a target, and network parameters of the DeepID network are adjusted to minimize an error between the obtained training face information and standard face information corresponding to the training group sample.
  • a plurality of video images for training that is, the first video image described above, need to be collected in advance.
  • the face information described in this embodiment may include, but is not limited to, identity information of a person, facial feature information of a face, and the like.
  • the present embodiment adopts a supervised learning mode, that is, step 302 is performed to mark the standard face information of each sample.
  • step 302 is performed to mark the standard face information of each sample.
  • the information corresponding to different faces belongs to the inter-class information, so that such information should be separated as much as possible; the information corresponding to the same personal face belongs to the intra-class information, so Information is as aggregated as possible. That is to achieve the effect of compact and inter-class separation in the class, which is beneficial to the subsequent training of DeepID network, so that the training effect is better.
  • step 303 is the same as the content of step 102, and details are not described herein again.
  • the adopted DeepID network may include three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner.
  • step 304 before inputting the data dimension-adjusted first video image, first initialize network parameters of the DeepID network, and extract grayscale features, LBP features, and gradient features of each image data according to the input first video image.
  • the traditional single gray-scale feature method has poor stability and limited ability to describe facial features
  • the LBP feature can better describe the texture of the face in the image, and has good robustness to different illumination situations
  • the gradient feature is capable of extracting contour and direction information that facilitates distinguishing between different types of face images.
  • each sub-CNN needs to be convoluted and pooled separately, and the specific number of groups is determined according to the data processing situation, but the convolution and the number of pooled groups of the three sub-CNNs are the same.
  • the network structure of the three sub-CNNs is the same except that the initial filter parameters may be different, and the maximum pooling mode is adopted in all three sub-CNN networks.
  • the data in each group has the same label form, so there is no influence on the recognition and judgment of subsequent results.
  • the process of convolution and pooling is the process of dimension reduction and feature extraction of image data. Through this step, three characteristics can be obtained. Form of face data in the form of a label. Where convolution and pooling are not strictly combined, the last layer of this step may be a convolutional layer or a pooled layer.
  • the result of the above process is input into the feature fusion layer of the DeepID network, that is, the fully connected layer, and the feature fusion is performed.
  • a node sufficient to represent the number of face features can be extracted as a feature of the face image by linear or non-linear activation.
  • each training sample is divided into corresponding numbers according to the specific number of training sets. If the previous layer of the full connection layer is a convolution layer, it is directly connected to the full connection layer. If the previous layer of the full connection layer is a pool layer, the common information of the joint pool layer and the previous convolution layer is used as a full connection. The input of the layer, so that the DeepID feature can be obtained more comprehensively. Finally, the DeepID network outputs the training face information identified in each of the first video images as training samples.
  • the training face information is used as a target, and the standard face information corresponding to the training face information and the training group sample is calculated. If the error does not meet the preset condition, it is necessary to adjust the network parameters of the DeepID network according to the calculated error, such as the hidden layer parameter of the DeepID network, and try to make the training face information and the standard face of the subsequent training output as much as possible. The error between the information is minimized.
  • the hidden units of the DeepID network can be respectively calculated with errors, and the corresponding hidden units are adjusted according to the errors.
  • the preset condition may be determined when training a specific DeepID network, for example, the setting error is less than a specific threshold, and the specific threshold may be a percentage value, and the smaller the specific threshold, the last The more stable the DeepID network is, the higher the recognition accuracy is.
  • a test group sample different from the training group sample may be prepared to test and verify the DeepID network.
  • a test group sample may be collected in advance, the test set sample including a plurality of second video images for testing, the second video images each including a face image of a plurality of boarding passengers; and then, pre-marked Standard face information corresponding to each second video image in the test group sample.
  • the multi-face detection method may further include:
  • the second video image with the data dimension adjusted is input as an input to the DeepID network, and each test face information that is obtained by the DeepID network from the second video image is obtained.
  • step 404 determining whether the test error is less than a preset error threshold, if not, proceeding to step 405, and if so, executing step 306;
  • steps 401-402 are similar to the contents of steps 303-304, and the principles are basically the same, and are not described herein again.
  • the test error between the standard face information corresponding to the test group sample is calculated, and the degree of training completion of the DeepID network is evaluated by the test error. Since the test test sample is different from the training set sample, it is more unfamiliar to the DeepID network, so the evaluation effect will be better than the training phase.
  • test error of the test is not less than the preset error threshold, it indicates that the DeepID network still does not meet the actual use requirement, and the training is still not completed, so that it can be determined that the DeepID network is not trained.
  • the next training is started. If the test error is less than the preset error threshold, the DeepID network has met the actual usage requirement, and the training is completed. Step 306 is performed to determine that the DeepID network training is completed.
  • the face information can also be applied to different application scenarios, such as being applied to the public security system for searching of the target person.
  • a GPS positioning module may be installed on the target vehicle, so that the execution body can acquire real-time positioning information of the target vehicle in real time.
  • the execution body uploads the obtained face information to the public security server, so that the public security server compares the face information with the target person information stored in the public security server; if the positioning request is received, the entity acquires The real-time positioning information of the target vehicle, the positioning request is initiated by the public security server after the comparison is successful; finally, the real-time positioning information is uploaded to the public security server.
  • the public security system can use these face information to quickly locate the target person (such as criminals, terrorists), carry out arrests or other law enforcement activities, and realize the closed loop of the public security system.
  • the obtained face information can be registered on the platform server, so that the platform server can perform face matching and searching later.
  • the execution body may upload the face information and the corresponding video image to the designated FTP server when the network connection is normal, and when the network is disconnected, first These face information and corresponding video images are cached in local storage.
  • the platform server may periodically extract face information from the FTP server, and query whether the face information is the same as the face information of the registered person identity. If the same, the face information belonging to the same person is updated to the identity of the person. If not the same, register a new person identity for the new face information and the corresponding video image. It can be seen that after the platform server accumulates a large number of personnel identity and face information, the data of the platform server can be applied to multiple fields such as access control, blacklist monitoring, and face photo search.
  • a video image when a passenger on the target vehicle gets on the vehicle is collected, the video image includes a face image of the boarding passenger; then, the data dimension of the video image is adjusted to a specified data dimension; then, Inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image; and finally, identifying the respective individuals
  • the face information is uploaded to a designated platform server, so that the platform server performs passenger flow statistics and analysis on the target vehicle.
  • the DeepID network in the case that multiple passengers get on the train at the same time, the DeepID network can be accurately pre-trained to realize accurate detection of multiple faces, and the face information of each passenger is identified and uploaded to the designated platform server, For the designated platform server to carry out passenger flow statistics and analysis, avoiding the problem of face missing detection when multiple passengers get on the train at the same time.
  • the multi-face detection method provided by the invention also has the following advantages: the non-contact type of face information is collected, and the capture can be performed remotely without human contact, without invasiveness; the concealment is strong, and the person who does not need to be captured is not required. Cooperate, it can be captured in the moment of getting on the train, which is not easy to attract the attention of passengers; the equipment is simple and versatile, the equipment cost is low, and no special capture equipment is needed. Only the camera can be realized on the hardware.
  • a multi-face detection method has been mainly described above, and a multi-face detection device will be described in detail below.
  • FIG. 5 is a structural diagram showing an embodiment of a multi-face detecting apparatus according to an embodiment of the present invention.
  • a multi-face detection device includes:
  • a video image acquisition module 501 configured to collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
  • a dimension adjustment module 502 configured to adjust a data dimension of the video image to a specified data dimension
  • the face recognition module 503 is configured to input the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtain each face information that is recognized by the DeepID network from the video image;
  • the information uploading module 504 is configured to upload the identified individual face information to a designated platform server.
  • DeepID network can be pre-trained by the following modules:
  • a training sample collection module configured to pre-collect a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
  • a face information marking module configured to pre-mark standard face information corresponding to each first video image in the training group sample
  • a first adjustment module configured to adjust a data dimension of the first video image to a specified data dimension
  • a network training module configured to input the first video image with the data dimension adjusted as input to the DeepID network, and obtain the training face information that is obtained by the DeepID network from the first video image;
  • a network parameter adjustment module configured to use the training face information as a target, and adjust network parameters of the DeepID network to minimize the obtained training face information and standard face information corresponding to the training group sample. Error between
  • the training completion module is configured to determine that the DeepID network training is completed if the error satisfies a preset condition.
  • the DeepID network may include three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner.
  • the video image collection module may include:
  • a vehicle detecting unit configured to detect whether a door of the target vehicle is open or the target vehicle is pitted
  • a capture unit configured to control a camera installed at a specified location on the target vehicle to start capturing if the vehicle detecting unit detects that the door of the target vehicle is open or the target vehicle is in the station;
  • an image determining unit configured to determine that the captured image is the video image if a human face exists in the captured image.
  • the multi-face detecting device may further include:
  • the public security module is configured to upload the identified face information to the public security server, so that the public security server compares the face information with the target person information stored in the public security server;
  • a positioning information acquiring module configured to acquire real-time positioning information of the target vehicle if the positioning request is received, where the positioning request is initiated by the public security server after the comparison is successful;
  • the positioning information uploading module is configured to upload the real-time positioning information to the public security server.
  • FIG. 6 is a schematic diagram of a platform server according to an embodiment of the present invention.
  • the platform server 6 of this embodiment includes a processor 60, a memory 61, and a computer program 62 stored in the memory 61 and operable on the processor 60, for example, performing the above-described multi-face.
  • the procedure for the detection method When the processor 60 executes the computer program 62, the steps in the embodiments of the various multi-face detection methods described above are implemented, such as steps 101 to 104 shown in FIG.
  • the processor 60 when executing the computer program 62, implements the functions of the modules/units in the various apparatus embodiments described above, such as the functions of the modules 501 through 504 shown in FIG.
  • the computer program 62 can be partitioned into one or more modules/units that are stored in the memory 61 and executed by the processor 60 to complete this invention.
  • the one or more modules/units may be a series of computer program instruction segments capable of performing a particular function, the instruction segments being used to describe the execution of the computer program 62 in the platform server 6.
  • the platform server 6 may be a computing device such as a local server or a cloud server.
  • the platform server may include, but is not limited to, a processor 60, a memory 61. It will be understood by those skilled in the art that FIG. 6 is merely an example of the platform server 6, and does not constitute a limitation of the platform server 6, and may include more or less components than those illustrated, or combine some components, or different components.
  • the platform server may further include an input/output device, a network access device, a bus, and the like.
  • the processor 60 can be a central processing unit (Central Processing Unit, CPU), can also be other general-purpose processors, digital signal processors (DSP), application specific integrated circuits (Application Specific Integrated Circuit (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, discrete hardware components, etc.
  • the general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
  • the memory 61 may be an internal storage unit of the platform server 6, such as a hard disk or a memory of the platform server 6.
  • the memory 61 may also be an external storage device of the platform server 6, such as a plug-in hard disk provided on the platform server 6, a smart memory card (SMC), and a secure digital (SD). Card, flash card (Flash Card) and so on.
  • the memory 61 may also include both an internal storage unit of the platform server 6 and an external storage device.
  • the memory 61 is used to store the computer program and other programs and data required by the platform server.
  • the memory 61 can also be used to temporarily store data that has been output or is about to be output.
  • the present invention also provides a multi-face detection system including a camera, an in-vehicle intelligent control terminal, and the above-described platform server.
  • the camera is installed at a designated position of the target vehicle for capturing a video image of a passenger on the target vehicle when the vehicle is boarded; the vehicle intelligent control terminal is configured to upload a video image captured by the camera to The platform server.
  • modules, units, and/or method steps of the various embodiments described in connection with the embodiments disclosed herein can be implemented in electronic hardware or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the solution. A person skilled in the art can use different methods for implementing the described functions for each particular application, but such implementation should not be considered to be beyond the scope of the present invention.
  • the disclosed system, apparatus, and method may be implemented in other manners.
  • the device embodiments described above are merely illustrative.
  • the division of the unit is only a logical function division.
  • there may be another division manner for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed.
  • the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
  • each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit.
  • the above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
  • the integrated unit if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium. Based on such understanding, the present invention implements all or part of the processes in the foregoing embodiments, and may also be completed by a computer program to instruct related hardware.
  • the computer program may be stored in a computer readable storage medium. The steps of the various method embodiments described above may be implemented when the program is executed by the processor.
  • the computer program comprises computer program code, which may be in the form of source code, object code form, executable file or some intermediate form.
  • the computer readable medium can include any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard drive, a magnetic disk, an optical disk, a computer memory, a read only memory (ROM, Read-Only) Memory), random access memory (RAM, Random) Access Memory), electrical carrier signals, telecommunications signals, and software distribution media.
  • ROM Read Only memory
  • RAM Random Access Memory

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

A multi-person face detection method, apparatus, server, system, and storage medium, used for solving the problem of missed facial detection when multiple passengers board a vehicle. the multi-person facial detection method comprises: collecting a video image when passengers board a target vehicle, the video image comprising a facial image of the boarding passengers; adjusting the data dimension of the video image to a specific data dimension; feeding the video image having undergone data dimension adjustment as an input into a pre-trained DeepID network and obtaining individual facial information recognised by the DeepID network from the video image; and uploading the recognised individual facial information to a specified platform server.

Description

一种多人脸检测方法、装置、服务器、系统及存储介质Multi-face detection method, device, server, system and storage medium 技术领域Technical field
本发明涉及视频信息处理技术领域,尤其涉及一种多人脸检测方法、装置、服务器、系统及存储介质。The present invention relates to the field of video information processing technologies, and in particular, to a multi-face detection method, apparatus, server, system, and storage medium.
背景技术Background technique
目前关于车辆人脸检测的技术多围绕关于乘客人脸检测的客流统计的方面进行,在车辆上人脸识别方面取得了一定的成果,在一般情况下可以完成车辆的客流统计和分析。At present, the technology of vehicle face detection mostly focuses on the aspect of passenger flow statistics on passenger face detection, and has achieved certain results in face recognition on vehicles. Under normal circumstances, passenger flow statistics and analysis can be completed.
然而,当多位乘客同时上车时,现有技术的人脸检测往往容易漏检一个甚至多个人脸,导致客流统计不准确的问题。However, when a plurality of passengers get on the train at the same time, the face detection of the prior art tends to easily miss one or more faces, resulting in inaccurate passenger flow statistics.
技术问题technical problem
本发明实施例提供了一种多人脸检测方法、装置、服务器、系统及存储介质,能够识别出多个乘客各自的人脸信息并上传至指定平台服务器,以供指定的平台服务器进行客流统计和分析,避免多位乘客同时上车的情况下发生人脸漏检的问题。The embodiment of the invention provides a multi-face detection method, device, server, system and storage medium, which can identify the face information of multiple passengers and upload them to a designated platform server for the flow rate statistics of the designated platform server. And analysis, to avoid the problem of face missing detection when multiple passengers get on the train at the same time.
技术解决方案Technical solution
第一方面,提供了一种多人脸检测方法,包括:In a first aspect, a multi-face detection method is provided, including:
采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;Collecting a video image of a passenger on the target vehicle when the vehicle is on the vehicle, the video image including a face image of the passenger on board;
将所述视频图像的数据维度调整为指定数据维度;Adjusting a data dimension of the video image to a specified data dimension;
将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;And inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image;
将识别得到的所述各个人脸信息上传至指定的平台服务器。The identified individual face information is uploaded to a designated platform server.
进一步地,所述DeepID网络通过以下步骤预先训练得到:Further, the DeepID network is pre-trained by the following steps:
预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;Pre-collecting training group samples, the training group samples including a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;Pre-marking standard face information corresponding to each first video image in the training group sample;
将所述第一视频图像的数据维度调整为指定数据维度;Adjusting a data dimension of the first video image to a specified data dimension;
将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;And inputting the first video image with the data dimension adjusted as input to the DeepID network, and obtaining each training face information that is obtained by the DeepID network from the first video image;
将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;Adjusting, by the training face information, a network parameter of the DeepID network to minimize an error between the obtained training face information and standard face information corresponding to the training group sample;
若所述误差满足预设条件,则确定所述DeepID网络训练完成。If the error satisfies the preset condition, it is determined that the DeepID network training is completed.
进一步地,所述DeepID网络包括三个子卷积神经网络,所述三个子卷积神经网络的网络结构相同,且均采用最大池化方式。Further, the DeepID network includes three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner.
进一步地,所述采集目标车辆上乘客上车时的视频图像包括:Further, the video image of the passenger on the target vehicle when loading the vehicle includes:
检测是否所述目标车辆的车门打开或者所述目标车辆进站;Detecting whether the door of the target vehicle is open or the target vehicle is pitted;
若检测到所述目标车辆的车门打开或者所述目标车辆进站,则控制所述目标车辆上安装在指定位置的摄像头开始抓拍;If it is detected that the door of the target vehicle is open or the target vehicle is in the station, controlling the camera installed at the designated position on the target vehicle to start capturing;
若抓拍到的图像中存在人脸,则确定所述抓拍到的图像为所述视频图像。If there is a human face in the captured image, it is determined that the captured image is the video image.
进一步地,还包括:Further, it also includes:
将识别得到的所述各个人脸信息上传至公安服务器,以便公安服务器将所述各个人脸信息与公安服务器中存储的目标人物信息进行比对;And uploading the identified individual face information to the public security server, so that the public security server compares the respective face information with the target person information stored in the public security server;
若接收到定位请求,则获取所述目标车辆的实时定位信息,所述定位请求由所述公安服务器在比对成功之后发起;Obtaining real-time positioning information of the target vehicle if the positioning request is received, where the positioning request is initiated by the public security server after the comparison is successful;
将所述实时定位信息上传至公安服务器。Uploading the real-time positioning information to the public security server.
第二方面,提供了一种多人脸检测装置,包括:In a second aspect, a multi-face detection device is provided, including:
视频图像采集模块,用于采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;a video image acquisition module, configured to collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
维度调整模块,用于将所述视频图像的数据维度调整为指定数据维度;a dimension adjustment module, configured to adjust a data dimension of the video image to a specified data dimension;
人脸识别模块,用于将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;a face recognition module, configured to input the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtain each face information that is recognized by the DeepID network from the video image;
信息上传模块,用于将识别得到的所述各个人脸信息上传至指定的平台服务器。The information uploading module is configured to upload the identified individual face information to a designated platform server.
进一步地,所述DeepID网络通过以下模块预先训练得到:Further, the DeepID network is pre-trained by the following modules:
训练样本收集模块,用于预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;a training sample collection module, configured to pre-collect a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
人脸信息标记模块,用于预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;a face information marking module, configured to pre-mark standard face information corresponding to each first video image in the training group sample;
第一调整模块,用于将所述第一视频图像的数据维度调整为指定数据维度;a first adjustment module, configured to adjust a data dimension of the first video image to a specified data dimension;
网络训练模块,用于将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;a network training module, configured to input the first video image with the data dimension adjusted as input to the DeepID network, and obtain the training face information that is obtained by the DeepID network from the first video image;
网络参数调整模块,用于将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;a network parameter adjustment module, configured to use the training face information as a target, and adjust network parameters of the DeepID network to minimize the obtained training face information and standard face information corresponding to the training group sample. Error between
训练完成模块,用于若所述误差满足预设条件,则确定所述DeepID网络训练完成。The training completion module is configured to determine that the DeepID network training is completed if the error satisfies a preset condition.
第三方面,提供了一种平台服务器,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,所述处理器执行所述计算机程序时实现上述的多人脸检测方法的步骤。In a third aspect, a platform server is provided, comprising a memory, a processor, and a computer program stored in the memory and operable on the processor, the processor implementing the computer program to implement the above The steps of the face detection method.
第四方面,提供了一种多人脸检测系统,包括摄像头、车载智能控制终端,以及上述的平台服务器;In a fourth aspect, a multi-face detection system is provided, including a camera, an in-vehicle intelligent control terminal, and the platform server described above;
所述摄像头安装在所述目标车辆的指定位置,用于抓拍所述目标车辆上乘客上车时的视频图像;The camera is installed at a designated position of the target vehicle for capturing a video image of a passenger on the target vehicle when getting on the vehicle;
所述车载智能控制终端用于将所述摄像头抓拍到的视频图像上传至所述平台服务器。The in-vehicle intelligent control terminal is configured to upload a video image captured by the camera to the platform server.
第五方面,提供了一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时实现上述多人脸检测方法的步骤。In a fifth aspect, a computer readable storage medium is provided, the computer readable storage medium storing a computer program, the computer program being executed by a processor to implement the steps of the multi-face detection method.
有益效果Beneficial effect
从以上技术方案可以看出,本发明实施例具有以下优点:It can be seen from the above technical solutions that the embodiments of the present invention have the following advantages:
本发明实施例中,首先,采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;然后,将所述视频图像的数据维度调整为指定数据维度;接着,将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;最后,将识别得到的所述各个人脸信息上传至指定平台服务器,以便所述平台服务器对所述目标车辆进行客流统计和分析。可见,在本发明实施例中,可以在多乘客同时上车的情况下,通过预训练完成DeepID网络实现多人脸的准确检测,识别出这些乘客各自的人脸信息并上传至指定平台服务器,以供指定的平台服务器进行客流统计和分析,避免了多位乘客同时上车的情况下发生人脸漏检的问题。In the embodiment of the present invention, first, a video image when a passenger on a target vehicle gets on the vehicle is collected, the video image includes a face image of the boarding passenger; then, the data dimension of the video image is adjusted to a specified data dimension; And inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image; and finally, each of the identified The personal face information is uploaded to a designated platform server, so that the platform server performs passenger flow statistics and analysis on the target vehicle. It can be seen that, in the embodiment of the present invention, the multiple IDs can be accurately detected by pre-training the DeepID network, and the face information of each passenger is identified and uploaded to the designated platform server. The passenger flow statistics and analysis are carried out for the designated platform server, which avoids the problem of face missing detection when multiple passengers get on the train at the same time.
附图说明DRAWINGS
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本发明的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动性的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings used in the embodiments or the description of the prior art will be briefly described below. It is obvious that the drawings in the following description are only the present invention. For some embodiments, other drawings may be obtained from those of ordinary skill in the art in light of the inventive workability.
图1为本发明实施例中一种多人脸检测方法一个实施例流程图;1 is a flowchart of an embodiment of a multi-face detection method according to an embodiment of the present invention;
图2为本发明实施例中一种多人脸检测方法步骤101在一个应用场景下的流程示意图;2 is a schematic flowchart of step 101 of a multi-face detection method in an application scenario according to an embodiment of the present invention;
图3为本发明实施例中一种多人脸检测方法在一个应用场景下预先训练DeepID网络的流程示意图;3 is a schematic flowchart of pre-training a DeepID network in an application scenario according to an embodiment of the present invention;
图4为本发明实施例中一种多人脸检测方法在一个应用场景下测试DeepID网络是否训练完成的流程示意图;4 is a schematic flowchart of testing whether a DeepID network is trained in an application scenario in a multi-face detection method according to an embodiment of the present invention;
图5为本发明实施例中一种多人脸检测装置一个实施例结构图;FIG. 5 is a structural diagram of an embodiment of a multi-face detecting device according to an embodiment of the present invention; FIG.
图6为本发明一实施例提供的平台服务器的示意图。FIG. 6 is a schematic diagram of a platform server according to an embodiment of the present invention.
本发明的实施方式Embodiments of the invention
本发明实施例提供了一种多人脸检测方法、装置、服务器、系统及存储介质,用于解决多位乘客同时上车的情况下发生人脸漏检的问题。The embodiment of the invention provides a multi-face detection method, device, server, system and storage medium, which are used for solving the problem of face missing detection when multiple passengers get on the vehicle at the same time.
为使得本发明的发明目的、特征、优点能够更加的明显和易懂,下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,下面所描述的实施例仅仅是本发明一部分实施例,而非全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其它实施例,都属于本发明保护的范围。In order to make the object, the features and the advantages of the present invention more obvious and easy to understand, the technical solutions in the embodiments of the present invention will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the present invention. The described embodiments are only a part of the embodiments of the invention, and not all of the embodiments. All other embodiments obtained by a person of ordinary skill in the art based on the embodiments of the present invention without creative efforts are within the scope of the present invention.
请参阅图1,本发明实施例中一种多人脸检测方法一个实施例包括:Referring to FIG. 1, an embodiment of a multi-face detection method according to an embodiment of the present invention includes:
101、采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;101. Collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
在本实施例中,可以在目标车辆上的合适位置安装若干个摄像头,这些摄像头对准目标车辆的前后门的乘客上下车的位置。当目标车辆到站停稳,前后门打开时,乘客从前门或后门上车,此时这些摄像头可以抓拍到包含乘客的人脸图像的视频图像,从而本实施例的执行主体可以采集到这些视频图像。In the present embodiment, a plurality of cameras can be installed at appropriate positions on the target vehicle, the cameras being aligned with the positions of the passengers on the front and rear doors of the target vehicle. When the target vehicle arrives at the station and the front and rear doors are opened, the passengers get on the front door or the rear door. At this time, the cameras can capture the video image containing the face image of the passenger, so that the executor of the embodiment can collect the videos. image.
需要说明的是,本实施例的执行主体具体可以是安装在车辆上的车载智能终端、系统或者平台服务器,为便于描述,下面统一表述为执行主体。It should be noted that the execution subject of the embodiment may be an in-vehicle intelligent terminal, a system, or a platform server installed on a vehicle. For convenience of description, the following is uniformly expressed as an execution subject.
进一步地,还可以对摄像头抓拍控制和图像筛选。比如,当前后门关闭后一段时间内,摄像头停止抓拍工作,等待下一次前后门的打开再进行抓拍,这样可以节省目标车辆的电量。在一个具体应用场景下,如图2所示,上述步骤101可以包括:Further, camera capture control and image screening are also possible. For example, after the current back door is closed, the camera stops capturing work, waiting for the next front and rear doors to open and then capture, which can save the target vehicle's power. In a specific application scenario, as shown in FIG. 2, the foregoing step 101 may include:
201、检测是否所述目标车辆的车门打开或者所述目标车辆进站,若是,则执行步骤202,若否,则继续等待;201. Detect whether the door of the target vehicle is open or the target vehicle is in the station, and if yes, proceed to step 202, and if not, continue to wait;
202、控制所述目标车辆上安装在指定位置的摄像头开始抓拍;202. Control a camera installed in the designated location on the target vehicle to start capturing;
203、若抓拍到的图像中存在人脸,则确定所述抓拍到的图像为所述视频图像。203. If there is a human face in the captured image, determine that the captured image is the video image.
对于上述步骤201~203,在控制摄像头什么时候开始抓拍的同时,还检测抓拍到的图像中是否存在人脸。可以理解的是,目标车辆的车门打开或者目标车辆进站时,并不是每次都有乘客上车的,所以摄像头抓拍到的图像中可能并不存在人脸。为了避免不存在人脸的图像占用执行主体的运算资源,可以检测该图像是否存在人脸,若是,在确定该图像为视频图像,从而交给执行主体进行后续处理。For the above steps 201-203, when the camera is controlled to start capturing, it is also detected whether there is a face in the captured image. It can be understood that when the door of the target vehicle is opened or the target vehicle is pitted, not every time a passenger gets on the bus, there may be no face in the image captured by the camera. In order to avoid that the image of the face does not occupy the computing resources of the execution subject, it is possible to detect whether the image has a human face, and if so, determine that the image is a video image, and then hand it to the execution subject for subsequent processing.
102、将所述视频图像的数据维度调整为指定数据维度;102. Adjust a data dimension of the video image to a specified data dimension.
可以理解的是,在将视频图像投入到DeepID网络进行识别之前,需要调整视频图像的数据维度,使其数据维度符合DeepID网络的要求。如果该视频图像原先的数据维度较高,则可以采用PCA(Principal Component Analysis,主成分分析)等方法进行预降维处理;反之,如果该视频图像原先的数据维度较低,则可以对该视频图像进行升维处理;最终使得该视频图像的数据维度等于指定数据维度。It can be understood that before the video image is put into the DeepID network for identification, the data dimension of the video image needs to be adjusted so that the data dimension meets the requirements of the DeepID network. If the original data dimension of the video image is high, PCA (Principal) can be used. Component Analysis (Principal Component Analysis) and other methods perform pre-dimension reduction processing; conversely, if the original data dimension of the video image is low, the video image may be subjected to up-dimensional processing; finally, the data dimension of the video image is equal to the specified data. Dimensions.
103、将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;103. The video image with the adjusted data dimension is input as input to the pre-trained DeepID network, and each face information identified by the DeepID network from the video image is obtained;
在调整数据维度之后,可以将该视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息。After adjusting the data dimension, the video image may be input as input to the pre-trained DeepID network, and each face information identified by the DeepID network from the video image is obtained.
可以理解的是,该DeepID网络是预先经过大量的训练样本训练完成得到的,可以对视频图像中的人脸信息进行分类识别,对当前视频图像中的各个人脸可以做出准确的识别,从而输出对应的人脸信息。关于该DeepID网络的预训练过程和网络结构将在下面内容中进行详细描述。It can be understood that the DeepID network is obtained by training a large number of training samples in advance, and can perform face recognition on the face information in the video image, and can accurately identify each face in the current video image, thereby Output the corresponding face information. The pre-training process and network structure for the DeepID network will be described in detail below.
104、将识别得到的所述各个人脸信息上传至指定的平台服务器。104. Upload the identified individual face information to a specified platform server.
在本实施例中,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息之后,则可以将识别得到的所述各个人脸信息上传至指定的平台服务器,以便所述平台服务器对所述目标车辆进行客流统计和分析。除此之外,还可以将这些人脸信息上传给公安服务器进行罪犯监控和搜索等,应用场景极其广泛。In this embodiment, after obtaining the face information identified by the DeepID network from the video image, the identified individual face information may be uploaded to a designated platform server, so that the platform server Passenger flow statistics and analysis are performed on the target vehicle. In addition, these face information can be uploaded to the public security server for criminal monitoring and searching, etc., and the application scenarios are extremely extensive.
下面将对该DeepID网络的预训练过程和网络结果进行详细介绍。如图3所示,所述DeepID网络可以通过以下步骤预先训练得到:The pre-training process and network results of the DeepID network will be described in detail below. As shown in FIG. 3, the DeepID network can be pre-trained by the following steps:
301、预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;301, pre-collecting a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
302、预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;302: Pre-marking standard face information corresponding to each first video image in the training group sample;
303、将所述第一视频图像的数据维度调整为指定数据维度;303. Adjust a data dimension of the first video image to a specified data dimension.
304、将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;304. The first video image with the data dimension adjusted is input as an input to the DeepID network, and each training face information that is obtained by the DeepID network from the first video image is obtained.
305、将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;305. The training face information is used as a target, and network parameters of the DeepID network are adjusted to minimize an error between the obtained training face information and standard face information corresponding to the training group sample.
306、若所述误差满足预设条件,则确定所述DeepID网络训练完成。306. If the error meets a preset condition, determine that the DeepID network training is completed.
对于上述步骤301和302,在训练DeepID网络之前,需要预先收集用于训练的多个视频图像,即上述的第一视频图像。这些第一视频图像的数据量越大,对DeepID网络的训练效果就越好。For the above steps 301 and 302, before training the DeepID network, a plurality of video images for training, that is, the first video image described above, need to be collected in advance. The larger the amount of data of these first video images, the better the training effect on the DeepID network.
在收集到这些训练组样本之后,还需要标记这些训练组样本中各个第一视频图像对应的标准人脸信息,即在每个第一视频图像中存在哪些人脸,这些人脸分别对应的人员信息是什么。需要说明的是,本实施例中所述的人脸信息可以包括但不限于人员的身份信息、人脸的脸部特征信息等。After collecting the training group samples, it is also required to mark the standard face information corresponding to each of the first video images in the training group samples, that is, which faces exist in each of the first video images, and the corresponding faces of the faces are respectively What is the information. It should be noted that the face information described in this embodiment may include, but is not limited to, identity information of a person, facial feature information of a face, and the like.
另外,对于本实施例中的DeepID网络,本实施例采用的是有监督的学习方式,即执行步骤302标记各个样本的标准人脸信息。标记每张人脸对应的信息标签时,对于不同的人脸所对应的信息属于类间信息,要使这类信息尽量分离;对于同一个人脸所对应的信息属于类内信息,要使这类信息尽量聚合。即达到类内紧凑、类间分离的效果,这样有利于后续对DeepID网络训练,使得训练效果更好。In addition, for the DeepID network in this embodiment, the present embodiment adopts a supervised learning mode, that is, step 302 is performed to mark the standard face information of each sample. When marking the information label corresponding to each face, the information corresponding to different faces belongs to the inter-class information, so that such information should be separated as much as possible; the information corresponding to the same personal face belongs to the intra-class information, so Information is as aggregated as possible. That is to achieve the effect of compact and inter-class separation in the class, which is beneficial to the subsequent training of DeepID network, so that the training effect is better.
上述步骤303与步骤102的内容相同,此处不再赘述。The foregoing step 303 is the same as the content of step 102, and details are not described herein again.
本实施例中,采用的DeepID网络可以包括三个子卷积神经网络,所述三个子卷积神经网络的网络结构相同,且均采用最大池化方式。对于步骤304,在将数据维度调整后的第一视频图像输入之前,先初始化该DeepID网络的网络参数,并根据输入的第一视频图像提取得到各个图像数据的灰度特征、LBP特征和梯度特征。可以理解的是,传统的单一采用灰度特征方式稳定性差、对人脸特征描述能力有限;LBP特征能够较好地描述图像中人脸的纹理,对不同光照情况具有很好的鲁棒性;梯度特征能够提取有利于区分不同类别人脸图像的轮廓和方向信息。这三种特征可以分别作为三个子CNN(卷积神经网络)的输入。In this embodiment, the adopted DeepID network may include three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner. For step 304, before inputting the data dimension-adjusted first video image, first initialize network parameters of the DeepID network, and extract grayscale features, LBP features, and gradient features of each image data according to the input first video image. . It can be understood that the traditional single gray-scale feature method has poor stability and limited ability to describe facial features; the LBP feature can better describe the texture of the face in the image, and has good robustness to different illumination situations; The gradient feature is capable of extracting contour and direction information that facilitates distinguishing between different types of face images. These three features can be used as inputs to three sub-CNNs (convolutional neural networks), respectively.
在步骤304训练DeepID网络的过程中,需要对对每个子CNN分别进行卷积与池化,具体组数要根据数据处理情况而定,但三个子CNN的卷积与池化组数要相同,除初始滤波器参数可能不同外,三个子CNN的网络结构均相同,三个子CNN网络中均采用最大池化方式。每组中的数据都具有相同的标签形式,因此对后续结果的识别判断没有影响,卷积与池化的进行过程即对图像数据的降维与特征提取过程,通过此步骤可以得到三种特征形式的带标签形式人脸数据。其中,卷积与池化非严格组合,此步骤最后一层可以为卷积层或池化层。In the process of training the DeepID network in step 304, each sub-CNN needs to be convoluted and pooled separately, and the specific number of groups is determined according to the data processing situation, but the convolution and the number of pooled groups of the three sub-CNNs are the same. The network structure of the three sub-CNNs is the same except that the initial filter parameters may be different, and the maximum pooling mode is adopted in all three sub-CNN networks. The data in each group has the same label form, so there is no influence on the recognition and judgment of subsequent results. The process of convolution and pooling is the process of dimension reduction and feature extraction of image data. Through this step, three characteristics can be obtained. Form of face data in the form of a label. Where convolution and pooling are not strictly combined, the last layer of this step may be a convolutional layer or a pooled layer.
将上述过程处理后的结果输入DeepID网络的特征融合层,即全连接层,进行特征融合。可以采用线性或非线性激活的方式提取足够表征人脸特征个数的节点作为人脸图像的提取特征,最后根据具体的训练集人数,在标签层将各个训练样本分为对应的数目。若全连接层前一层是卷积层,则直接与全连接层相连,若全连接层前一层是池化层,则联合池化层与上一的卷积层的共同信息作为全连接层的输入,这样可以更全面地获取DeepID特征。最后,DeepID网络输出各个作为训练样本的第一视频图像中识别得到的训练人脸信息。The result of the above process is input into the feature fusion layer of the DeepID network, that is, the fully connected layer, and the feature fusion is performed. A node sufficient to represent the number of face features can be extracted as a feature of the face image by linear or non-linear activation. Finally, each training sample is divided into corresponding numbers according to the specific number of training sets. If the previous layer of the full connection layer is a convolution layer, it is directly connected to the full connection layer. If the previous layer of the full connection layer is a pool layer, the common information of the joint pool layer and the previous convolution layer is used as a full connection. The input of the layer, so that the DeepID feature can be obtained more comprehensively. Finally, the DeepID network outputs the training face information identified in each of the first video images as training samples.
对于上述步骤305和306,每次训练中,在得到训练人脸信息之后,将所述训练人脸信息作为目标,计算所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差,如果该误差未满足预设条件,则需要根据计算出来的误差调整该DeepID网络的网络参数,比如DeepID网络的隐层参数,尽量使得后续训练输出的训练人脸信息与标准人脸信息之间的误差最小化。其中,在计算误差时,为了提高调整网络参数的精度和效率,可以对DeepID网络的隐单元分别计算出误差,并根据这些误差来调整对应的隐单元。For the above steps 305 and 306, in each training, after the training face information is obtained, the training face information is used as a target, and the standard face information corresponding to the training face information and the training group sample is calculated. If the error does not meet the preset condition, it is necessary to adjust the network parameters of the DeepID network according to the calculated error, such as the hidden layer parameter of the DeepID network, and try to make the training face information and the standard face of the subsequent training output as much as possible. The error between the information is minimized. In the calculation error, in order to improve the accuracy and efficiency of adjusting the network parameters, the hidden units of the DeepID network can be respectively calculated with errors, and the corresponding hidden units are adjusted according to the errors.
如果计算出来的误差满足预设条件,则可以确定所述DeepID网络训练完成。其中,对于步骤306所述的预设条件,该预设条件可以在训练具体的DeepID网络时确定,比如设定误差小于特定阈值,该特定阈值可以是一个百分比数值,特定阈值越小,则最后训练完成得到的DeepID网络越稳定,识别精度越高。If the calculated error satisfies the preset condition, it may be determined that the DeepID network training is completed. For the preset condition described in step 306, the preset condition may be determined when training a specific DeepID network, for example, the setting error is less than a specific threshold, and the specific threshold may be a percentage value, and the smaller the specific threshold, the last The more stable the DeepID network is, the higher the recognition accuracy is.
对于上述步骤306,为了更进一步验证该DeepID网络的训练完成程度,还可以准备一套不同于训练组样本的测试组样本对该DeepID网络进行测试、检验。在测试之前,可以预先收集测试组样本,所述测试组样本包括用于测试的多个第二视频图像,所述第二视频图像均包括多个上车乘客的人脸图像;然后,预先标记所述测试组样本中各个第二视频图像对应的标准人脸信息。如图4所示,在确定所述DeepID网络训练完成之前,所述多人脸检测方法还可以包括:For the above step 306, in order to further verify the training completion degree of the DeepID network, a test group sample different from the training group sample may be prepared to test and verify the DeepID network. Before testing, a test group sample may be collected in advance, the test set sample including a plurality of second video images for testing, the second video images each including a face image of a plurality of boarding passengers; and then, pre-marked Standard face information corresponding to each second video image in the test group sample. As shown in FIG. 4, before determining that the DeepID network training is completed, the multi-face detection method may further include:
401、将所述第二视频图像的数据维度调整为指定数据维度;401. Adjust a data dimension of the second video image to a specified data dimension.
402、将数据维度调整后的所述第二视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第二视频图像中识别得到的各个测试人脸信息;402. The second video image with the data dimension adjusted is input as an input to the DeepID network, and each test face information that is obtained by the DeepID network from the second video image is obtained.
403、计算所述测试人脸信息与所述测试组样本对应的标准人脸信息之间的测试误差;403. Calculate a test error between the test face information and the standard face information corresponding to the test group sample.
404、判断所述测试误差是否小于预设的误差阈值,若否,则执行步骤405,若是,则执行步骤306;404, determining whether the test error is less than a preset error threshold, if not, proceeding to step 405, and if so, executing step 306;
405、确定所述DeepID网络未训练完成,开始下一次训练。405. Determine that the DeepID network is not trained to complete, and start the next training.
上述步骤401~402与步骤303~304的内容相似,原理基本相同,此处不再赘述。The foregoing steps 401-402 are similar to the contents of steps 303-304, and the principles are basically the same, and are not described herein again.
对于上述步骤403,在得到测试人脸信息之后,计算其与测试组样本对应的标准人脸信息之间的测试误差,通过测试误差来评估该DeepID网络的训练完成程度。由于测试用的测试组样本有别于训练组样本,其对于该DeepID网络来说更为陌生,因此评估的效果也会好于训练阶段的评估效果。For the above step 403, after the test face information is obtained, the test error between the standard face information corresponding to the test group sample is calculated, and the degree of training completion of the DeepID network is evaluated by the test error. Since the test test sample is different from the training set sample, it is more unfamiliar to the DeepID network, so the evaluation effect will be better than the training phase.
对于步骤404~405,若本次测试的测试误差不小于预设的误差阈值,则说明该DeepID网络仍未满足实际使用的需求,训练仍未完成,从而可以确定所述DeepID网络未训练完成,开始下一次训练;反之,若所述测试误差小于预设的误差阈值,则说明该DeepID网络已满足实际使用的需求,训练完成,执行步骤306确定所述DeepID网络训练完成。For the steps 404-405, if the test error of the test is not less than the preset error threshold, it indicates that the DeepID network still does not meet the actual use requirement, and the training is still not completed, so that it can be determined that the DeepID network is not trained. The next training is started. If the test error is less than the preset error threshold, the DeepID network has met the actual usage requirement, and the training is completed. Step 306 is performed to determine that the DeepID network training is completed.
优选地,本实施例中,在将视频图像投入DeepID网络识别得到各个人脸信息之后,还可以将这些人脸信息应用至各个不同的应用场景下,比如应用于公安系统进行目标人物的搜索中。具体地,目标车辆上可以安装有GPS定位模块,从而执行主体可以实时获取到目标车辆的实时定位信息。首先,执行主体将识别得到的所述各个人脸信息上传至公安服务器,以便公安服务器将所述各个人脸信息与公安服务器中存储的目标人物信息进行比对;若接收到定位请求,则获取所述目标车辆的实时定位信息,所述定位请求由所述公安服务器在比对成功之后发起;最后,将所述实时定位信息上传至公安服务器。可见,公安系统可以利用这些人脸信息快速定位目标人物(例如犯罪分子、恐怖分子)的位置,实施抓捕或者其它执法行为,实现了公安系统的管理闭环。Preferably, in this embodiment, after the video image is put into the DeepID network to identify each face information, the face information can also be applied to different application scenarios, such as being applied to the public security system for searching of the target person. . Specifically, a GPS positioning module may be installed on the target vehicle, so that the execution body can acquire real-time positioning information of the target vehicle in real time. First, the execution body uploads the obtained face information to the public security server, so that the public security server compares the face information with the target person information stored in the public security server; if the positioning request is received, the entity acquires The real-time positioning information of the target vehicle, the positioning request is initiated by the public security server after the comparison is successful; finally, the real-time positioning information is uploaded to the public security server. It can be seen that the public security system can use these face information to quickly locate the target person (such as criminals, terrorists), carry out arrests or other law enforcement activities, and realize the closed loop of the public security system.
另外,还可以将得到的人脸信息在平台服务器上进行人脸注册,便于平台服务器在后续进行人脸比对和查找。具体地,执行主体在获取到DeepID网络输出的各个人脸信息之后,可以在网络连接正常时将这些人脸信息和对应的视频图像上传到指定的FTP服务器,而在网络断开时,先将这些人脸信息和对应的视频图像缓存在本地存储中。平台服务器可以定期从FTP服务器上提取人脸信息,并查询这些人脸信息与已注册的人员身份的人脸信息是否相同,若相同,则将属于同一人员的人脸信息更新至该人员身份下;若不相同,则为新的人脸信息和对应的视频图像注册一个新的人员身份。可知,在平台服务器累积大量的人员身份和人脸信息之后,该平台服务器的数据可以应用在出入控制、黑名单监控、人脸照片搜索等多个领域上。In addition, the obtained face information can be registered on the platform server, so that the platform server can perform face matching and searching later. Specifically, after acquiring the face information output by the DeepID network, the execution body may upload the face information and the corresponding video image to the designated FTP server when the network connection is normal, and when the network is disconnected, first These face information and corresponding video images are cached in local storage. The platform server may periodically extract face information from the FTP server, and query whether the face information is the same as the face information of the registered person identity. If the same, the face information belonging to the same person is updated to the identity of the person. If not the same, register a new person identity for the new face information and the corresponding video image. It can be seen that after the platform server accumulates a large number of personnel identity and face information, the data of the platform server can be applied to multiple fields such as access control, blacklist monitoring, and face photo search.
本实施例中,首先,采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;然后,将所述视频图像的数据维度调整为指定数据维度;接着,将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;最后,将识别得到的所述各个人脸信息上传至指定平台服务器,以便所述平台服务器对所述目标车辆进行客流统计和分析。可见,在本实施例中,可以在多乘客同时上车的情况下,通过预训练完成DeepID网络实现多人脸的准确检测,识别出这些乘客各自的人脸信息并上传至指定平台服务器,以供指定的平台服务器进行客流统计和分析,避免了多位乘客同时上车的情况下发生人脸漏检的问题。In this embodiment, first, a video image when a passenger on the target vehicle gets on the vehicle is collected, the video image includes a face image of the boarding passenger; then, the data dimension of the video image is adjusted to a specified data dimension; then, Inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image; and finally, identifying the respective individuals The face information is uploaded to a designated platform server, so that the platform server performs passenger flow statistics and analysis on the target vehicle. It can be seen that in the embodiment, in the case that multiple passengers get on the train at the same time, the DeepID network can be accurately pre-trained to realize accurate detection of multiple faces, and the face information of each passenger is identified and uploaded to the designated platform server, For the designated platform server to carry out passenger flow statistics and analysis, avoiding the problem of face missing detection when multiple passengers get on the train at the same time.
另外,采用本发明提供的多人脸检测方法还具有以下优点:采用非接触式采集人脸信息,无需通过人为接触,可远程实施抓拍,没有侵犯性;隐蔽性强,不需要被抓拍人员的配合,可在上车瞬间进行抓拍,不容易引起乘客的注意;实现设备简单通用、设备成本低,不需专用的抓拍设备,硬件上仅需摄像头即可实现。In addition, the multi-face detection method provided by the invention also has the following advantages: the non-contact type of face information is collected, and the capture can be performed remotely without human contact, without invasiveness; the concealment is strong, and the person who does not need to be captured is not required. Cooperate, it can be captured in the moment of getting on the train, which is not easy to attract the attention of passengers; the equipment is simple and versatile, the equipment cost is low, and no special capture equipment is needed. Only the camera can be realized on the hardware.
应理解,上述实施例中各步骤的序号的大小并不意味着执行顺序的先后,各过程的执行顺序应以其功能和内在逻辑确定,而不应对本发明实施例的实施过程构成任何限定。It should be understood that the size of the sequence of the steps in the above embodiments does not imply a sequence of executions, and the order of execution of the processes should be determined by its function and internal logic, and should not be construed as limiting the implementation of the embodiments of the present invention.
上面主要描述了一种多人脸检测方法,下面将对一种多人脸检测装置进行详细描述。A multi-face detection method has been mainly described above, and a multi-face detection device will be described in detail below.
图5示出了本发明实施例中一种多人脸检测装置一个实施例结构图。FIG. 5 is a structural diagram showing an embodiment of a multi-face detecting apparatus according to an embodiment of the present invention.
本实施例中,一种多人脸检测装置包括:In this embodiment, a multi-face detection device includes:
视频图像采集模块501,用于采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;a video image acquisition module 501, configured to collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
维度调整模块502,用于将所述视频图像的数据维度调整为指定数据维度;a dimension adjustment module 502, configured to adjust a data dimension of the video image to a specified data dimension;
人脸识别模块503,用于将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;The face recognition module 503 is configured to input the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtain each face information that is recognized by the DeepID network from the video image;
信息上传模块504,用于将识别得到的所述各个人脸信息上传至指定的平台服务器。The information uploading module 504 is configured to upload the identified individual face information to a designated platform server.
进一步地,所述DeepID网络可以通过以下模块预先训练得到:Further, the DeepID network can be pre-trained by the following modules:
训练样本收集模块,用于预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;a training sample collection module, configured to pre-collect a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
人脸信息标记模块,用于预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;a face information marking module, configured to pre-mark standard face information corresponding to each first video image in the training group sample;
第一调整模块,用于将所述第一视频图像的数据维度调整为指定数据维度;a first adjustment module, configured to adjust a data dimension of the first video image to a specified data dimension;
网络训练模块,用于将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;a network training module, configured to input the first video image with the data dimension adjusted as input to the DeepID network, and obtain the training face information that is obtained by the DeepID network from the first video image;
网络参数调整模块,用于将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;a network parameter adjustment module, configured to use the training face information as a target, and adjust network parameters of the DeepID network to minimize the obtained training face information and standard face information corresponding to the training group sample. Error between
训练完成模块,用于若所述误差满足预设条件,则确定所述DeepID网络训练完成。The training completion module is configured to determine that the DeepID network training is completed if the error satisfies a preset condition.
进一步地,所述DeepID网络可以包括三个子卷积神经网络,所述三个子卷积神经网络的网络结构相同,且均采用最大池化方式。Further, the DeepID network may include three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt the maximum pooling manner.
进一步地,所述视频图像采集模块可以包括:Further, the video image collection module may include:
车辆检测单元,用于检测是否所述目标车辆的车门打开或者所述目标车辆进站;a vehicle detecting unit, configured to detect whether a door of the target vehicle is open or the target vehicle is pitted;
抓拍单元,用于若所述车辆检测单元检测到所述目标车辆的车门打开或者所述目标车辆进站,则控制所述目标车辆上安装在指定位置的摄像头开始抓拍;a capture unit, configured to control a camera installed at a specified location on the target vehicle to start capturing if the vehicle detecting unit detects that the door of the target vehicle is open or the target vehicle is in the station;
图像确定单元,用于若抓拍到的图像中存在人脸,则确定所述抓拍到的图像为所述视频图像。And an image determining unit, configured to determine that the captured image is the video image if a human face exists in the captured image.
进一步地,所述多人脸检测装置还可以包括:Further, the multi-face detecting device may further include:
上传公安模块,用于将识别得到的所述各个人脸信息上传至公安服务器,以便公安服务器将所述各个人脸信息与公安服务器中存储的目标人物信息进行比对;The public security module is configured to upload the identified face information to the public security server, so that the public security server compares the face information with the target person information stored in the public security server;
定位信息获取模块,用于若接收到定位请求,则获取所述目标车辆的实时定位信息,所述定位请求由所述公安服务器在比对成功之后发起;a positioning information acquiring module, configured to acquire real-time positioning information of the target vehicle if the positioning request is received, where the positioning request is initiated by the public security server after the comparison is successful;
定位信息上传模块,用于将所述实时定位信息上传至公安服务器。The positioning information uploading module is configured to upload the real-time positioning information to the public security server.
图6是本发明一实施例提供的平台服务器的示意图。如图6所示,该实施例的平台服务器6包括:处理器60、存储器61以及存储在所述存储器61中并可在所述处理器60上运行的计算机程序62,例如执行上述多人脸检测方法的程序。所述处理器60执行所述计算机程序62时实现上述各个多人脸检测方法实施例中的步骤,例如图1所示的步骤101至104。或者,所述处理器60执行所述计算机程序62时实现上述各装置实施例中各模块/单元的功能,例如图5所示模块501至504的功能。FIG. 6 is a schematic diagram of a platform server according to an embodiment of the present invention. As shown in FIG. 6, the platform server 6 of this embodiment includes a processor 60, a memory 61, and a computer program 62 stored in the memory 61 and operable on the processor 60, for example, performing the above-described multi-face. The procedure for the detection method. When the processor 60 executes the computer program 62, the steps in the embodiments of the various multi-face detection methods described above are implemented, such as steps 101 to 104 shown in FIG. Alternatively, the processor 60, when executing the computer program 62, implements the functions of the modules/units in the various apparatus embodiments described above, such as the functions of the modules 501 through 504 shown in FIG.
示例性的,所述计算机程序62可以被分割成一个或多个模块/单元,所述一个或者多个模块/单元被存储在所述存储器61中,并由所述处理器60执行,以完成本发明。所述一个或多个模块/单元可以是能够完成特定功能的一系列计算机程序指令段,该指令段用于描述所述计算机程序62在所述平台服务器6中的执行过程。Illustratively, the computer program 62 can be partitioned into one or more modules/units that are stored in the memory 61 and executed by the processor 60 to complete this invention. The one or more modules/units may be a series of computer program instruction segments capable of performing a particular function, the instruction segments being used to describe the execution of the computer program 62 in the platform server 6.
所述平台服务器6可以是本地服务器、云端服务器等计算设备。所述平台服务器可包括,但不仅限于,处理器60、存储器61。本领域技术人员可以理解,图6仅仅是平台服务器6的示例,并不构成对平台服务器6的限定,可以包括比图示更多或更少的部件,或者组合某些部件,或者不同的部件,例如所述平台服务器还可以包括输入输出设备、网络接入设备、总线等。The platform server 6 may be a computing device such as a local server or a cloud server. The platform server may include, but is not limited to, a processor 60, a memory 61. It will be understood by those skilled in the art that FIG. 6 is merely an example of the platform server 6, and does not constitute a limitation of the platform server 6, and may include more or less components than those illustrated, or combine some components, or different components. For example, the platform server may further include an input/output device, a network access device, a bus, and the like.
所述处理器60可以是中央处理单元(Central Processing Unit,CPU),还可以是其他通用处理器、数字信号处理器 (Digital Signal Processor,DSP)、专用集成电路 (Application Specific Integrated Circuit,ASIC)、现成可编程门阵列 (Field-Programmable Gate Array,FPGA) 或者其他可编程逻辑器件、分立门或者晶体管逻辑器件、分立硬件组件等。通用处理器可以是微处理器或者该处理器也可以是任何常规的处理器等。The processor 60 can be a central processing unit (Central Processing Unit, CPU), can also be other general-purpose processors, digital signal processors (DSP), application specific integrated circuits (Application Specific Integrated Circuit (ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, discrete hardware components, etc. The general purpose processor may be a microprocessor or the processor or any conventional processor or the like.
所述存储器61可以是所述平台服务器6的内部存储单元,例如平台服务器6的硬盘或内存。所述存储器61也可以是所述平台服务器6的外部存储设备,例如所述平台服务器6上配备的插接式硬盘,智能存储卡(Smart Media Card, SMC),安全数字(Secure Digital, SD)卡,闪存卡(Flash Card)等。进一步地,所述存储器61还可以既包括所述平台服务器6的内部存储单元也包括外部存储设备。所述存储器61用于存储所述计算机程序以及所述平台服务器所需的其他程序和数据。所述存储器61还可以用于暂时地存储已经输出或者将要输出的数据。The memory 61 may be an internal storage unit of the platform server 6, such as a hard disk or a memory of the platform server 6. The memory 61 may also be an external storage device of the platform server 6, such as a plug-in hard disk provided on the platform server 6, a smart memory card (SMC), and a secure digital (SD). Card, flash card (Flash Card) and so on. Further, the memory 61 may also include both an internal storage unit of the platform server 6 and an external storage device. The memory 61 is used to store the computer program and other programs and data required by the platform server. The memory 61 can also be used to temporarily store data that has been output or is about to be output.
本发明还提供了一种多人脸检测系统,该多人脸检测系统包括摄像头、车载智能控制终端,以及上述的平台服务器。其中,所述摄像头安装在所述目标车辆的指定位置,用于抓拍所述目标车辆上乘客上车时的视频图像;所述车载智能控制终端用于将所述摄像头抓拍到的视频图像上传至所述平台服务器。The present invention also provides a multi-face detection system including a camera, an in-vehicle intelligent control terminal, and the above-described platform server. The camera is installed at a designated position of the target vehicle for capturing a video image of a passenger on the target vehicle when the vehicle is boarded; the vehicle intelligent control terminal is configured to upload a video image captured by the camera to The platform server.
所属领域的技术人员可以清楚地了解到,为描述的方便和简洁,上述描述的系统,装置和单元的具体工作过程,可以参考前述方法实施例中的对应过程,在此不再赘述。A person skilled in the art can clearly understand that for the convenience and brevity of the description, the specific working process of the system, the device and the unit described above can refer to the corresponding process in the foregoing method embodiment, and details are not described herein again.
在上述实施例中,对各个实施例的描述都各有侧重,某个实施例中没有详述或记载的部分,可以参见其它实施例的相关描述。In the above embodiments, the descriptions of the various embodiments are different, and the parts that are not detailed or described in a certain embodiment can be referred to the related descriptions of other embodiments.
本领域普通技术人员可以意识到,结合本文中所公开的实施例描述的各实施例的模块、单元和/或方法步骤,能够以电子硬件、或者计算机软件和电子硬件的结合来实现。这些功能究竟以硬件还是软件方式来执行,取决于技术方案的特定应用和设计约束条件。专业技术人员可以对每个特定的应用来使用不同方法来实现所描述的功能,但是这种实现不应认为超出本发明的范围。Those of ordinary skill in the art will appreciate that the modules, units, and/or method steps of the various embodiments described in connection with the embodiments disclosed herein can be implemented in electronic hardware or a combination of computer software and electronic hardware. Whether these functions are performed in hardware or software depends on the specific application and design constraints of the solution. A person skilled in the art can use different methods for implementing the described functions for each particular application, but such implementation should not be considered to be beyond the scope of the present invention.
在本申请所提供的几个实施例中,应该理解到,所揭露的系统,装置和方法,可以通过其它的方式实现。例如,以上所描述的装置实施例仅仅是示意性的,例如,所述单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。另一点,所显示或讨论的相互之间的耦合或直接耦合或通信连接可以是通过一些接口,装置或单元的间接耦合或通信连接,可以是电性,机械或其它的形式。In the several embodiments provided by the present application, it should be understood that the disclosed system, apparatus, and method may be implemented in other manners. For example, the device embodiments described above are merely illustrative. For example, the division of the unit is only a logical function division. In actual implementation, there may be another division manner, for example, multiple units or components may be combined or Can be integrated into another system, or some features can be ignored or not executed. In addition, the mutual coupling or direct coupling or communication connection shown or discussed may be an indirect coupling or communication connection through some interface, device or unit, and may be in an electrical, mechanical or other form.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施例方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, may be located in one place, or may be distributed to multiple network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of the embodiment.
另外,在本发明各个实施例中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, each functional unit in each embodiment of the present invention may be integrated into one processing unit, or each unit may exist physically separately, or two or more units may be integrated into one unit. The above integrated unit can be implemented in the form of hardware or in the form of a software functional unit.
所述集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明实现上述实施例方法中的全部或部分流程,也可以通过计算机程序来指令相关的硬件来完成,所述的计算机程序可存储于一计算机可读存储介质中,该计算机程序在被处理器执行时,可实现上述各个方法实施例的步骤。其中,所述计算机程序包括计算机程序代码,所述计算机程序代码可以为源代码形式、对象代码形式、可执行文件或某些中间形式等。所述计算机可读介质可以包括:能够携带所述计算机程序代码的任何实体或装置、记录介质、U盘、移动硬盘、磁碟、光盘、计算机存储器、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、电载波信号、电信信号以及软件分发介质等。需要说明的是,所述计算机可读介质包含的内容可以根据司法管辖区内立法和专利实践的要求进行适当的增减,例如在某些司法管辖区,根据立法和专利实践,计算机可读介质不包括电载波信号和电信信号。The integrated unit, if implemented in the form of a software functional unit and sold or used as a standalone product, may be stored in a computer readable storage medium. Based on such understanding, the present invention implements all or part of the processes in the foregoing embodiments, and may also be completed by a computer program to instruct related hardware. The computer program may be stored in a computer readable storage medium. The steps of the various method embodiments described above may be implemented when the program is executed by the processor. Wherein, the computer program comprises computer program code, which may be in the form of source code, object code form, executable file or some intermediate form. The computer readable medium can include any entity or device capable of carrying the computer program code, a recording medium, a USB flash drive, a removable hard drive, a magnetic disk, an optical disk, a computer memory, a read only memory (ROM, Read-Only) Memory), random access memory (RAM, Random) Access Memory), electrical carrier signals, telecommunications signals, and software distribution media. It should be noted that the content contained in the computer readable medium may be appropriately increased or decreased according to the requirements of legislation and patent practice in a jurisdiction, for example, in some jurisdictions, according to legislation and patent practice, computer readable media Does not include electrical carrier signals and telecommunication signals.
以上所述,以上实施例仅用以说明本发明的技术方案,而非对其限制;尽管参照前述实施例对本发明进行了详细的说明,本领域的普通技术人员应当理解:其依然可以对前述各实施例所记载的技术方案进行修改,或者对其中部分技术特征进行等同替换;而这些修改或者替换,并不使相应技术方案的本质脱离本发明各实施例技术方案的精神和范围。The above embodiments are only used to illustrate the technical solutions of the present invention, and are not intended to be limiting; although the present invention has been described in detail with reference to the foregoing embodiments, those skilled in the art will understand that The technical solutions described in the embodiments are modified, or the equivalents of the technical features are replaced by the equivalents of the technical solutions of the embodiments of the present invention.

Claims (10)

  1. 一种多人脸检测方法,其特征在于,包括:A multi-face detection method, comprising:
    采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;Collecting a video image of a passenger on the target vehicle when the vehicle is on the vehicle, the video image including a face image of the passenger on board;
    将所述视频图像的数据维度调整为指定数据维度;Adjusting a data dimension of the video image to a specified data dimension;
    将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;And inputting the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtaining each face information identified by the DeepID network from the video image;
    将识别得到的所述各个人脸信息上传至指定的平台服务器。The identified individual face information is uploaded to a designated platform server.
  2. 根据权利要求1所述的多人脸检测方法,其特征在于,所述DeepID网络通过以下步骤预先训练得到:The multi-face detection method according to claim 1, wherein the DeepID network is pre-trained by the following steps:
    预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;Pre-collecting training group samples, the training group samples including a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
    预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;Pre-marking standard face information corresponding to each first video image in the training group sample;
    将所述第一视频图像的数据维度调整为指定数据维度;Adjusting a data dimension of the first video image to a specified data dimension;
    将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;And inputting the first video image with the data dimension adjusted as input to the DeepID network, and obtaining each training face information that is obtained by the DeepID network from the first video image;
    将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;Adjusting, by the training face information, a network parameter of the DeepID network to minimize an error between the obtained training face information and standard face information corresponding to the training group sample;
    若所述误差满足预设条件,则确定所述DeepID网络训练完成。If the error satisfies the preset condition, it is determined that the DeepID network training is completed.
  3. 根据权利要求2所述的多人脸检测方法,其特征在于,所述DeepID网络包括三个子卷积神经网络,所述三个子卷积神经网络的网络结构相同,且均采用最大池化方式。The multi-face detection method according to claim 2, wherein the DeepID network comprises three sub-convolution neural networks, and the network structures of the three sub-convolution neural networks are the same, and both adopt a maximum pooling manner.
  4. 根据权利要求1所述的多人脸检测方法,其特征在于,所述采集目标车辆上乘客上车时的视频图像包括:The multi-face detection method according to claim 1, wherein the video image when the passenger on the target vehicle is on the vehicle includes:
    检测是否所述目标车辆的车门打开或者所述目标车辆进站;Detecting whether the door of the target vehicle is open or the target vehicle is pitted;
    若检测到所述目标车辆的车门打开或者所述目标车辆进站,则控制所述目标车辆上安装在指定位置的摄像头开始抓拍;If it is detected that the door of the target vehicle is open or the target vehicle is in the station, controlling the camera installed at the designated position on the target vehicle to start capturing;
    若抓拍到的图像中存在人脸,则确定所述抓拍到的图像为所述视频图像。If there is a human face in the captured image, it is determined that the captured image is the video image.
  5. 根据权利要求1至4中任一项所述的多人脸检测方法,其特征在于,还包括:The multi-face detection method according to any one of claims 1 to 4, further comprising:
    将识别得到的所述各个人脸信息上传至公安服务器,以便公安服务器将所述各个人脸信息与公安服务器中存储的目标人物信息进行比对;And uploading the identified individual face information to the public security server, so that the public security server compares the respective face information with the target person information stored in the public security server;
    若接收到定位请求,则获取所述目标车辆的实时定位信息,所述定位请求由所述公安服务器在比对成功之后发起;Obtaining real-time positioning information of the target vehicle if the positioning request is received, where the positioning request is initiated by the public security server after the comparison is successful;
    将所述实时定位信息上传至公安服务器。Uploading the real-time positioning information to the public security server.
  6. 一种多人脸检测装置,其特征在于,包括:A multi-face detection device, comprising:
    视频图像采集模块,用于采集目标车辆上乘客上车时的视频图像,所述视频图像包括上车乘客的人脸图像;a video image acquisition module, configured to collect a video image of a passenger on the target vehicle when the vehicle is on the vehicle, where the video image includes a face image of the passenger on board;
    维度调整模块,用于将所述视频图像的数据维度调整为指定数据维度;a dimension adjustment module, configured to adjust a data dimension of the video image to a specified data dimension;
    人脸识别模块,用于将数据维度调整后的所述视频图像作为输入投入至预训练完成的DeepID网络,得到所述DeepID网络从所述视频图像中识别得到的各个人脸信息;a face recognition module, configured to input the video image with the adjusted data dimension as input to the pre-trained DeepID network, and obtain each face information that is recognized by the DeepID network from the video image;
    信息上传模块,用于将识别得到的所述各个人脸信息上传至指定的平台服务器。The information uploading module is configured to upload the identified individual face information to a designated platform server.
  7. 根据权利要求6所述的多人脸检测装置,其特征在于,所述DeepID网络通过以下模块预先训练得到:The multi-face detection device according to claim 6, wherein the DeepID network is pre-trained by the following modules:
    训练样本收集模块,用于预先收集训练组样本,所述训练组样本包括用于训练的多个第一视频图像,所述第一视频图像均包括多个上车乘客的人脸图像;a training sample collection module, configured to pre-collect a training group sample, where the training group sample includes a plurality of first video images for training, the first video images each including a face image of a plurality of boarding passengers;
    人脸信息标记模块,用于预先标记所述训练组样本中各个第一视频图像对应的标准人脸信息;a face information marking module, configured to pre-mark standard face information corresponding to each first video image in the training group sample;
    第一调整模块,用于将所述第一视频图像的数据维度调整为指定数据维度;a first adjustment module, configured to adjust a data dimension of the first video image to a specified data dimension;
    网络训练模块,用于将数据维度调整后的所述第一视频图像作为输入投入至DeepID网络,得到所述DeepID网络从所述第一视频图像中识别得到的各个训练人脸信息;a network training module, configured to input the first video image with the data dimension adjusted as input to the DeepID network, and obtain the training face information that is obtained by the DeepID network from the first video image;
    网络参数调整模块,用于将所述训练人脸信息作为目标,调整所述DeepID网络的网络参数,以最小化得到的所述训练人脸信息与所述训练组样本对应的标准人脸信息之间的误差;a network parameter adjustment module, configured to use the training face information as a target, and adjust network parameters of the DeepID network to minimize the obtained training face information and standard face information corresponding to the training group sample. Error between
    训练完成模块,用于若所述误差满足预设条件,则确定所述DeepID网络训练完成。The training completion module is configured to determine that the DeepID network training is completed if the error satisfies a preset condition.
  8. 一种平台服务器,包括存储器、处理器以及存储在所述存储器中并可在所述处理器上运行的计算机程序,其特征在于,所述处理器执行所述计算机程序时实现如权利要求1至5中任一项所述多人脸检测方法的步骤。A platform server comprising a memory, a processor, and a computer program stored in the memory and operable on the processor, wherein the processor executes the computer program as claimed in claim 1 The step of the multi-face detection method according to any one of 5.
  9. 一种多人脸检测系统,其特征在于,包括摄像头、车载智能控制终端,以及如权利要求8所述的平台服务器;A multi-face detection system, comprising: a camera, an in-vehicle intelligent control terminal, and the platform server according to claim 8;
    所述摄像头安装在所述目标车辆的指定位置,用于抓拍所述目标车辆上乘客上车时的视频图像;The camera is installed at a designated position of the target vehicle for capturing a video image of a passenger on the target vehicle when getting on the vehicle;
    所述车载智能控制终端用于将所述摄像头抓拍到的视频图像上传至所述平台服务器。The in-vehicle intelligent control terminal is configured to upload a video image captured by the camera to the platform server.
  10. 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,其特征在于,所述计算机程序被处理器执行时实现如权利要求1至5中任一项所述多人脸检测方法的步骤。A computer readable storage medium storing a computer program, wherein the computer program is executed by a processor to implement the multi-face detection according to any one of claims 1 to 5. The steps of the method.
PCT/CN2017/119569 2017-12-28 2017-12-28 Multi-person face detection method, apparatus, server, system, and storage medium WO2019127273A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
PCT/CN2017/119569 WO2019127273A1 (en) 2017-12-28 2017-12-28 Multi-person face detection method, apparatus, server, system, and storage medium
CN201780002310.9A CN108351967A (en) 2017-12-28 2017-12-28 A kind of plurality of human faces detection method, device, server, system and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2017/119569 WO2019127273A1 (en) 2017-12-28 2017-12-28 Multi-person face detection method, apparatus, server, system, and storage medium

Publications (1)

Publication Number Publication Date
WO2019127273A1 true WO2019127273A1 (en) 2019-07-04

Family

ID=62960786

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2017/119569 WO2019127273A1 (en) 2017-12-28 2017-12-28 Multi-person face detection method, apparatus, server, system, and storage medium

Country Status (2)

Country Link
CN (1) CN108351967A (en)
WO (1) WO2019127273A1 (en)

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110555421A (en) * 2019-09-09 2019-12-10 南京创维信息技术研究院有限公司 Monitoring system and monitoring method
CN110706169A (en) * 2019-09-26 2020-01-17 深圳市半冬科技有限公司 Star portrait optimization method and device and storage device
CN111027397A (en) * 2019-11-14 2020-04-17 上海交通大学 Method, system, medium and device for detecting comprehensive characteristic target in intelligent monitoring network
CN111046818A (en) * 2019-12-18 2020-04-21 天地伟业技术有限公司 Face recognition service method based on PCIE smart card development
CN111079566A (en) * 2019-11-28 2020-04-28 深圳市信义科技有限公司 Large-scale face recognition model optimization system
CN111078804A (en) * 2019-12-09 2020-04-28 武汉数文科技有限公司 Information association method, system and computer terminal
CN111127574A (en) * 2019-12-26 2020-05-08 北京奇艺世纪科技有限公司 Video image editing method and device, electronic equipment and readable storage medium
CN111339831A (en) * 2020-01-23 2020-06-26 深圳市大拿科技有限公司 Lighting lamp control method and system
CN111382678A (en) * 2020-02-25 2020-07-07 浙江大学 Tourist bus passenger flow statistical algorithm based on improved CNN network
CN111507268A (en) * 2020-04-17 2020-08-07 浙江大华技术股份有限公司 Alarm method and device, storage medium and electronic device
CN111582654A (en) * 2020-04-14 2020-08-25 五邑大学 Service quality evaluation method and device based on deep cycle neural network
CN111598865A (en) * 2020-05-14 2020-08-28 上海锘科智能科技有限公司 Hand-foot-and-mouth disease detection method, device and system based on thermal infrared and RGB (red, green and blue) double-taking
CN111814761A (en) * 2020-08-24 2020-10-23 国网湖南省电力有限公司 Intelligent safety monitoring method for energy storage power station
CN111832378A (en) * 2019-08-13 2020-10-27 北京嘀嘀无限科技发展有限公司 Method and device for identifying vehicle overtaking
CN112052780A (en) * 2020-09-01 2020-12-08 北京嘀嘀无限科技发展有限公司 Face verification method, device and system and storage medium
CN112069931A (en) * 2020-08-20 2020-12-11 深圳数联天下智能科技有限公司 State report generation method and state monitoring system
CN112069887A (en) * 2020-07-31 2020-12-11 深圳市优必选科技股份有限公司 Face recognition method, face recognition device, terminal equipment and storage medium
CN112132022A (en) * 2020-09-22 2020-12-25 平安科技(深圳)有限公司 Face snapshot framework, face snapshot method, device, equipment and storage medium
CN112580538A (en) * 2020-12-23 2021-03-30 平安银行股份有限公司 Customer service personnel scheduling method, device, equipment and storage medium
CN112631896A (en) * 2020-12-02 2021-04-09 武汉旷视金智科技有限公司 Equipment performance testing method and device, storage medium and electronic equipment
CN112633143A (en) * 2020-12-21 2021-04-09 杭州海康威视数字技术股份有限公司 Image processing system, method, head-mounted device, processing device, and storage medium
CN112835954A (en) * 2021-01-26 2021-05-25 浙江大华技术股份有限公司 Method, device and equipment for determining target service object
CN112949577A (en) * 2021-03-29 2021-06-11 杭州海康威视数字技术股份有限公司 Information association method, device, server and storage medium
CN112949667A (en) * 2019-12-09 2021-06-11 北京金山云网络技术有限公司 Image recognition method, system, electronic device and storage medium
CN113158923A (en) * 2021-04-27 2021-07-23 华录智达科技股份有限公司 Bus transfer reminding system based on face recognition
CN113496155A (en) * 2020-03-20 2021-10-12 北京京东振世信息技术有限公司 Information processing method, device, equipment and computer readable medium
CN113743688A (en) * 2020-05-27 2021-12-03 鸿富锦精密电子(天津)有限公司 Quality control method and device, computer device and storage medium
CN114476891A (en) * 2022-03-02 2022-05-13 北京云迹科技股份有限公司 Method, device, equipment and storage medium for preventing tailing in elevator
CN115631553A (en) * 2022-10-11 2023-01-20 中国第一汽车股份有限公司 Vehicle unlocking method, vehicle unlocking device and vehicle

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109325513B (en) * 2018-08-01 2021-06-25 中国计量大学 Image classification network training method based on massive single-class images
CN113743236A (en) * 2021-08-11 2021-12-03 交控科技股份有限公司 Passenger portrait analysis method, device, electronic equipment and computer readable storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130294642A1 (en) * 2012-05-01 2013-11-07 Hulu Llc Augmenting video with facial recognition
CN105426869A (en) * 2015-12-15 2016-03-23 重庆凯泽科技有限公司 Face recognition system and recognition method based on railway security check
CN105678250A (en) * 2015-12-31 2016-06-15 北京小孔科技有限公司 Face identification method in video and face identification device in video
CN105809178A (en) * 2014-12-31 2016-07-27 中国科学院深圳先进技术研究院 Population analyzing method based on human face attribute and device
CN106339673A (en) * 2016-08-19 2017-01-18 中山大学 ATM identity authentication method based on face recognition

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106415594B (en) * 2014-06-16 2020-01-10 北京市商汤科技开发有限公司 Method and system for face verification
JP2017161890A (en) * 2016-03-08 2017-09-14 パナソニックIpマネジメント株式会社 Obstacle detection device and monitor device
CN107509063A (en) * 2017-09-30 2017-12-22 珠海芯桥科技有限公司 A kind of vehicle-mounted suspect's monitoring system based on recognition of face
CN107483902A (en) * 2017-09-30 2017-12-15 珠海芯桥科技有限公司 A kind of passenger on public transport Tracking monitoring system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130294642A1 (en) * 2012-05-01 2013-11-07 Hulu Llc Augmenting video with facial recognition
CN105809178A (en) * 2014-12-31 2016-07-27 中国科学院深圳先进技术研究院 Population analyzing method based on human face attribute and device
CN105426869A (en) * 2015-12-15 2016-03-23 重庆凯泽科技有限公司 Face recognition system and recognition method based on railway security check
CN105678250A (en) * 2015-12-31 2016-06-15 北京小孔科技有限公司 Face identification method in video and face identification device in video
CN106339673A (en) * 2016-08-19 2017-01-18 中山大学 ATM identity authentication method based on face recognition

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111832378A (en) * 2019-08-13 2020-10-27 北京嘀嘀无限科技发展有限公司 Method and device for identifying vehicle overtaking
CN110555421A (en) * 2019-09-09 2019-12-10 南京创维信息技术研究院有限公司 Monitoring system and monitoring method
CN110706169A (en) * 2019-09-26 2020-01-17 深圳市半冬科技有限公司 Star portrait optimization method and device and storage device
CN111027397A (en) * 2019-11-14 2020-04-17 上海交通大学 Method, system, medium and device for detecting comprehensive characteristic target in intelligent monitoring network
CN111027397B (en) * 2019-11-14 2023-05-12 上海交通大学 Comprehensive feature target detection method, system, medium and equipment suitable for intelligent monitoring network
CN111079566B (en) * 2019-11-28 2023-05-02 深圳市信义科技有限公司 Large-scale face recognition model optimization system
CN111079566A (en) * 2019-11-28 2020-04-28 深圳市信义科技有限公司 Large-scale face recognition model optimization system
CN111078804B (en) * 2019-12-09 2024-03-15 武汉数文科技有限公司 Information association method, system and computer terminal
CN111078804A (en) * 2019-12-09 2020-04-28 武汉数文科技有限公司 Information association method, system and computer terminal
CN112949667A (en) * 2019-12-09 2021-06-11 北京金山云网络技术有限公司 Image recognition method, system, electronic device and storage medium
CN111046818A (en) * 2019-12-18 2020-04-21 天地伟业技术有限公司 Face recognition service method based on PCIE smart card development
CN111127574A (en) * 2019-12-26 2020-05-08 北京奇艺世纪科技有限公司 Video image editing method and device, electronic equipment and readable storage medium
CN111339831A (en) * 2020-01-23 2020-06-26 深圳市大拿科技有限公司 Lighting lamp control method and system
CN111339831B (en) * 2020-01-23 2023-08-18 深圳市大拿科技有限公司 Lighting lamp control method and system
CN111382678A (en) * 2020-02-25 2020-07-07 浙江大学 Tourist bus passenger flow statistical algorithm based on improved CNN network
CN111382678B (en) * 2020-02-25 2023-04-18 浙江大学 Tourist bus passenger flow statistical algorithm based on improved CNN network
CN113496155B (en) * 2020-03-20 2023-09-29 北京京东振世信息技术有限公司 Method, apparatus, device and computer readable medium for information processing
CN113496155A (en) * 2020-03-20 2021-10-12 北京京东振世信息技术有限公司 Information processing method, device, equipment and computer readable medium
CN111582654A (en) * 2020-04-14 2020-08-25 五邑大学 Service quality evaluation method and device based on deep cycle neural network
CN111582654B (en) * 2020-04-14 2023-03-28 五邑大学 Service quality evaluation method and device based on deep cycle neural network
CN111507268B (en) * 2020-04-17 2024-02-20 浙江华感科技有限公司 Alarm method and device, storage medium and electronic device
CN111507268A (en) * 2020-04-17 2020-08-07 浙江大华技术股份有限公司 Alarm method and device, storage medium and electronic device
CN111598865A (en) * 2020-05-14 2020-08-28 上海锘科智能科技有限公司 Hand-foot-and-mouth disease detection method, device and system based on thermal infrared and RGB (red, green and blue) double-taking
CN111598865B (en) * 2020-05-14 2023-05-16 上海锘科智能科技有限公司 Hand-foot-mouth disease detection method, device and system based on thermal infrared and RGB double-shooting
CN113743688A (en) * 2020-05-27 2021-12-03 鸿富锦精密电子(天津)有限公司 Quality control method and device, computer device and storage medium
CN113743688B (en) * 2020-05-27 2023-10-20 富联精密电子(天津)有限公司 Quality control method, quality control device, computer device and storage medium
CN112069887B (en) * 2020-07-31 2023-12-29 深圳市优必选科技股份有限公司 Face recognition method, device, terminal equipment and storage medium
CN112069887A (en) * 2020-07-31 2020-12-11 深圳市优必选科技股份有限公司 Face recognition method, face recognition device, terminal equipment and storage medium
CN112069931A (en) * 2020-08-20 2020-12-11 深圳数联天下智能科技有限公司 State report generation method and state monitoring system
CN111814761A (en) * 2020-08-24 2020-10-23 国网湖南省电力有限公司 Intelligent safety monitoring method for energy storage power station
CN112052780A (en) * 2020-09-01 2020-12-08 北京嘀嘀无限科技发展有限公司 Face verification method, device and system and storage medium
CN112132022B (en) * 2020-09-22 2023-09-29 平安科技(深圳)有限公司 Face snapshot architecture and face snapshot method, device, equipment and storage medium thereof
CN112132022A (en) * 2020-09-22 2020-12-25 平安科技(深圳)有限公司 Face snapshot framework, face snapshot method, device, equipment and storage medium
CN112631896A (en) * 2020-12-02 2021-04-09 武汉旷视金智科技有限公司 Equipment performance testing method and device, storage medium and electronic equipment
CN112631896B (en) * 2020-12-02 2024-04-05 武汉旷视金智科技有限公司 Equipment performance test method and device, storage medium and electronic equipment
CN112633143A (en) * 2020-12-21 2021-04-09 杭州海康威视数字技术股份有限公司 Image processing system, method, head-mounted device, processing device, and storage medium
CN112633143B (en) * 2020-12-21 2023-09-05 杭州海康威视数字技术股份有限公司 Image processing system, method, head-mounted device, processing device, and storage medium
CN112580538A (en) * 2020-12-23 2021-03-30 平安银行股份有限公司 Customer service personnel scheduling method, device, equipment and storage medium
CN112580538B (en) * 2020-12-23 2024-03-26 平安银行股份有限公司 Customer service personnel scheduling method, device, equipment and storage medium
CN112835954A (en) * 2021-01-26 2021-05-25 浙江大华技术股份有限公司 Method, device and equipment for determining target service object
CN112835954B (en) * 2021-01-26 2023-05-23 浙江大华技术股份有限公司 Method, device and equipment for determining target service object
CN112949577A (en) * 2021-03-29 2021-06-11 杭州海康威视数字技术股份有限公司 Information association method, device, server and storage medium
CN113158923A (en) * 2021-04-27 2021-07-23 华录智达科技股份有限公司 Bus transfer reminding system based on face recognition
CN113158923B (en) * 2021-04-27 2022-09-06 华录智达科技股份有限公司 Bus transfer reminding system based on face recognition
CN114476891A (en) * 2022-03-02 2022-05-13 北京云迹科技股份有限公司 Method, device, equipment and storage medium for preventing tailing in elevator
CN115631553A (en) * 2022-10-11 2023-01-20 中国第一汽车股份有限公司 Vehicle unlocking method, vehicle unlocking device and vehicle

Also Published As

Publication number Publication date
CN108351967A (en) 2018-07-31

Similar Documents

Publication Publication Date Title
WO2019127273A1 (en) Multi-person face detection method, apparatus, server, system, and storage medium
CN104732601B (en) Automatic high-recognition-rate attendance checking device and method based on face recognition technology
CN108229376B (en) Method and device for detecting blinking
CN104778474B (en) A kind of classifier construction method and object detection method for target detection
TW202006602A (en) Three-dimensional living-body face detection method, face authentication recognition method, and apparatuses
CN110609920A (en) Pedestrian hybrid search method and system in video monitoring scene
CN108549852B (en) Specific scene downlink person detector automatic learning method based on deep network enhancement
TWI712980B (en) Claim information extraction method and device, and electronic equipment
CN109829381A (en) A kind of dog only identifies management method, device, system and storage medium
CN105160318A (en) Facial expression based lie detection method and system
CN107590473B (en) Human face living body detection method, medium and related device
CN109948616A (en) Image detecting method, device, electronic equipment and computer readable storage medium
CN109886222A (en) Face identification method, neural network training method, device and electronic equipment
Wang et al. A benchmark for clothes variation in person re‐identification
US10423817B2 (en) Latent fingerprint ridge flow map improvement
Chandran et al. Missing child identification system using deep learning and multiclass SVM
WO2019144416A1 (en) Information processing method and system, cloud processing device and computer program product
CN105184229A (en) Online learning based real-time pedestrian detection method in dynamic scene
CN109409250A (en) A kind of across the video camera pedestrian of no overlap ken recognition methods again based on deep learning
CN110619280B (en) Vehicle re-identification method and device based on deep joint discrimination learning
CN105184236A (en) Robot-based face identification system
CN112766065A (en) Mobile terminal examinee identity authentication method, device, terminal and storage medium
CN109711232A (en) Deep learning pedestrian recognition methods again based on multiple objective function
CN113743160A (en) Method, apparatus and storage medium for biopsy
CN111507289A (en) Video matching method, computer device and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17936258

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17936258

Country of ref document: EP

Kind code of ref document: A1