CN114038035A - Artificial intelligence recognition device based on big data - Google Patents

Artificial intelligence recognition device based on big data Download PDF

Info

Publication number
CN114038035A
CN114038035A CN202111303955.0A CN202111303955A CN114038035A CN 114038035 A CN114038035 A CN 114038035A CN 202111303955 A CN202111303955 A CN 202111303955A CN 114038035 A CN114038035 A CN 114038035A
Authority
CN
China
Prior art keywords
face
data
dimensional
artificial intelligence
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111303955.0A
Other languages
Chinese (zh)
Inventor
赵鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202111303955.0A priority Critical patent/CN114038035A/en
Publication of CN114038035A publication Critical patent/CN114038035A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2135Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Abstract

The invention belongs to the technical field of face recognition, and discloses an artificial intelligence recognition device based on big data, which comprises: the three-dimensional laser scanning module acquires face information data through a dimensional micro-scanner; simultaneously, preprocessing the three-dimensional face point cloud; the three-dimensional image preprocessing module is used for analyzing the two aspects of extracting the characteristics of the visual two-dimensional multimedia target image and generating the change characteristics of the three-dimensional multimedia visual target image to extract the characteristics of the three-dimensional multimedia visual target image; the face model training module is used for calling an OpenCV internal function to obtain a trained model based on an OpenCV technology; the interactive recognition module is used for inputting face information data and recognizing the face information data; and the remote face recognition module is used for carrying out data processing and face recognition in a LabVIEW development environment and finally feeding back the recognized identity information to the slave machine. The invention can make the computer technology develop towards intellectualization and automation.

Description

Artificial intelligence recognition device based on big data
Technical Field
The invention belongs to the technical field of face recognition; relates to an artificial intelligence recognition device based on big data.
Background
The intelligent human face recognition technology is characterized in that key parts of a human face are recognized and analyzed, expressions of a person are collected and stored in a database, and then the key parts of the human face and the expressions are recognized through comparison and analysis. Currently, the identification of strange faces by human beings is not completely accurate, and especially for some people with similar long faces, although the people can be distinguished, how to identify the faces is difficult to describe. That is more difficult for the vision system image capture device.
The problem of low recognition rate exists in the current face intelligent recognition technology.
Disclosure of Invention
In order to solve the technical problems, the invention provides an artificial intelligent identification device based on big data.
The invention is realized by the following steps: big data-based artificial intelligence recognition device, the big data-based artificial intelligence recognition device comprising:
the three-dimensional laser scanning module acquires face information data through the micro-dimensional scanner, and the face information data usually contains redundant information of hair and shoulders irrelevant to face recognition besides the face. In order to reduce the operation amount and improve the recognition rate, the three-dimensional face point cloud is preprocessed at the same time; the pretreatment comprises the following steps: face cutting and posture correction;
the three-dimensional image preprocessing module is used for extracting the three-dimensional multimedia visual target image characteristics by analyzing the two aspects of the extraction of the visual two-dimensional multimedia visual target image characteristics and the generation of the three-dimensional multimedia visual target image change characteristics; constructing a sparse representation algorithm-based model, limiting the minimum value of a target function by using gradient projection, optimizing the gradient direction to obtain a sparsely represented identification visual image, and realizing artificial intelligent identification of the three-dimensional multimedia visual image;
the face model training module is used for providing the algorithm in OpenCV2.4.9 based on the OpenCV technology, and only needs to classify faces in a database according to different people, then respectively read in the faces and call OpenCV internal functions to obtain a trained model;
the interactive recognition module is used for inputting face information data and recognizing the face information data;
the remote face recognition module is used for transmitting data to the host computer through a wireless channel by using the NI USRP equipment; the host receives the data of the slave from the connected NI USRP, performs data processing and face recognition in a LabVIEW development environment, and finally feeds back the recognized identity information to the slave.
Preferably, the three-dimensional laser scanning module includes:
the face cutting unit is used for determining the position of a nose tip quickly according to the fact that the nose tip has the maximum value on the face of the human face according to the geometric constraint of the human face, then drawing a sphere by taking the nose tip as a center, and the face area contained in the sphere is the effective area of the face of the human face;
the gesture correction unit is used for performing Principal Component Analysis (PCA) on the cut face point cloud, taking a feature vector corresponding to the maximum feature value as a Y axis of a new coordinate system, taking a feature vector corresponding to the minimum feature value as a Z axis, and establishing a right-hand coordinate system, wherein the coordinate system is called a posture coordinate system PCS; converting the cut human face point cloud into the PCS by taking the nose tip point as an original point of the PCS to finish the human face posture correction; all the human faces are converted into the front postures by establishing the PCS of the human face point cloud, and then the human face point cloud is converted into the same resolution ratio.
Preferably, the three-dimensional image preprocessing module includes:
the two-dimensional multimedia target image feature extraction unit is used for finding out a two-dimensional coordinate point by utilizing a visual target three-dimensional contour image and setting a multimedia geometric projection model capable of showing a perspective conversion model in a real environment;
and the three-dimensional multimedia visual target image unit is used for extracting three-dimensional multimedia visual dynamic image characteristics through optimization conversion of a nonlinear algorithm.
Preferably, the remote face recognition module adopts a coherent loss of a twin neural network: the device consists of 6 convolutional layers, 4 pooling layers and 1 full-connection layer. The 5 x 5 convolution kernel is split into two layers of 3 x 3 convolution kernels.
Preferably, the big data-based artificial intelligence recognition method for operating the big data-based artificial intelligence recognition apparatus includes:
the method comprises the steps that a micro-dimensional optical scanner obtains face information data, wherein the face information data comprise a face part and redundant information such as hair and shoulders which are irrelevant to face identification; simultaneously preprocessing the three-dimensional face point cloud; the pretreatment comprises the following steps: face cutting and posture correction;
extracting three-dimensional multimedia visual target image characteristics by analyzing the two aspects of visual two-dimensional multimedia target image characteristic extraction and three-dimensional multimedia visual target image change characteristic generation; constructing a sparse representation algorithm-based model, limiting the minimum value of a target function by utilizing gradient projection, optimizing the gradient direction to obtain a sparsely represented identification visual image, and realizing artificial intelligence identification of the three-dimensional multimedia visual image;
based on the OpenCV technology, the algorithm is provided in OpenCV2.4.9, and trained models can be obtained only by classifying human faces in a database according to different people, respectively reading the human faces and calling OpenCV internal functions;
inputting face information data for recognition;
the NI USRP equipment is used for sending data to the host through a wireless channel; the host receives the data of the slave from the connected NI USRP, performs data processing and face recognition in a LabVIEW development environment, and finally feeds back the recognized identity information to the slave.
Preferably, the PCA dimension reduction of the artificial intelligence recognition method based on big data comprises the following specific steps:
(1) removing a classification label of the data, and taking the d-dimensional data after removal as a sample;
(2) calculating a mean vector of d dimensions, the mean of each dimension vector of all data;
(3) calculating a dispersion matrix or covariance matrix of all data;
(4) calculating a feature value e1, e 2.., ed and a corresponding feature vector lambda1, lambda 2.., lambda d;
(5) sorting the eigenvectors in a descending order according to the magnitude of the eigenvalues, selecting the first k largest eigenvectors to form a matrix W with dimensions d x k, wherein each column represents one eigenvector;
(6) transforming the sample data into a new subspace by using the eigenvector matrix W of d × K:
y=WT×x
where x is a vector in d x 1 dimensions representing one sample and y is a vector in K x 1 dimensions in the new subspace.
Preferably, the wireless channel of the remote face recognition based on the artificial intelligence recognition method of big data is realized by adopting a wireless communication system, and the information transmission is completed by adopting an asynchronous communication mode; the data is transmitted in blocks, and check bits are added when the data is transmitted; after the data to be sent is coded, a bit code is formed, then the data is divided into frames, and the data sequence in the frames is check, synchronous bits, the sequence of the data packet in the whole information sequence and data bits.
Preferably, the remote face recognition module of the artificial intelligence recognition method based on big data specifically includes:
(1) coherent loss using a twin neural network: the device consists of 6 convolutional layers, 4 pooling layers and 1 full-connection layer. Splitting a 5 × 5 convolution kernel into two layers of 3 × 3 convolution kernels, increasing the depth of a network, and taking P-RELU as an activation function without increasing the calculated amount:
Figure BDA0003339447190000051
(2) the whole network is trained by using CASIA-Webface:
firstly organizing training data, then performing feature learning on a network, learning to discrete Features by using the Microsoft max loss + center loss, and finally performing feature extraction or classification through a trained model;
in order to make the training result better, the types with too few samples are removed firstly, only 10010 people with the most images in 10575 are selected, and image cleaning is performed at the same time, so that the identity of the people in the test set is ensured not to be repeated. The data set is balanced by randomly selecting the same number of sheets in each class, and the data of the image is enhanced by horizontal turning; carrying out face key point detection and face normalization on the balanced data set, and then training a network after unifying the size 112 x 96;
(3) the feature comparison is carried out by adopting a method for calculating cosine similarity, the larger the cosine value is, the more similar the two human faces are, and the cosine formula is as follows:
Figure BDA0003339447190000052
if vectorization is performed on both sides b and c, the cosine values for calculating two vector angles can be written as:
Figure BDA0003339447190000053
and a feature extraction stage, namely extracting depth feature vectors from the test image through a trained network, and splicing and representing feature graphs of the original graph and a horizontal turnover graph of the original graph into representation of a modified graph.
Preferably, the feature extraction of the remote face recognition module of the artificial intelligence recognition method based on big data includes the steps of:
obtaining a plurality of local areas according to the segmented face images;
comparing one pixel point in each region with a pixel point in a neighborhood, if the gray value of the pixel point in the neighborhood is larger than that of the central pixel point, marking the position of the pixel point in the neighborhood as 1, and otherwise marking as 0, thereby obtaining the LBP value of the central pixel point;
calculating histograms of all neighborhoods, namely calculating the frequency of each number, and then normalizing the histograms;
and connecting the calculated histograms of each neighborhood into a feature vector, wherein the feature vector is a vector for describing LBP texture features of the whole image.
Preferably, the intelligent processing terminal carries the artificial intelligence recognition device based on big data; the intelligent processing terminal comprises: computer, panel, cell-phone.
In summary, the invention has the advantages and technical effects that: the invention converts all human faces to the front postures by establishing the PCS of the human face point cloud, and then converts the human face point cloud to the same resolution. PCA dimension reduction can map high-dimensional vectors to low-dimensional vectors while preserving principal component information of the vectors well. The method comprises the steps of preprocessing three-dimensional face point cloud; the pretreatment comprises the following steps: face cutting and posture correction; the calculation amount is reduced, the recognition rate is improved, and the subsequent processing of the face information data is facilitated.
The invention has the advantages of higher accuracy of visual image recognition, reduced recognition time and ensured stability of recognition rate. The wireless channel of the remote face recognition module provided by the invention is realized by adopting a wireless communication system, and the information transmission is finished by adopting an asynchronous communication mode. The data is transmitted in blocks, and check bits are added when the data is transmitted, so that the accuracy of data transmission is ensured. In order to make the training result better, the invention firstly removes the types with too few samples, only selects 10010 people with the most images in 10575, and simultaneously cleans the images, thereby ensuring that the number of the people in the test set is not repeated.
The application of the artificial intelligent recognition technology is a great revolution in technological innovation, and the computer technology can be developed towards intellectualization and automation. The research on the artificial intelligent identification technology improves the technical content, breaks through the bottleneck existing in the application and further reaches the world leading level.
Drawings
FIG. 1 is a schematic structural diagram of an artificial intelligence identification device based on big data according to the present invention;
FIG. 2 is a schematic structural diagram of a three-dimensional laser scanning module according to the present invention;
FIG. 3 is a schematic structural diagram of a three-dimensional image preprocessing module according to the present invention;
FIG. 4 is a flow chart of wireless communication of the remote face recognition module of the present invention;
FIG. 5 is a schematic diagram of a remote face recognition module of the present invention;
FIG. 6 is a flow chart of feature extraction for the remote face recognition module of the present invention;
description of reference numerals:
1: a three-dimensional laser scanning module;
2: a three-dimensional image preprocessing module;
3: a face model training module;
4: an interaction identification module;
5: a remote face recognition module;
1-1: a face cutting unit;
1-2: an attitude correction unit;
2-1: a two-dimensional multimedia target image feature extraction unit;
2-2: and a three-dimensional multimedia visual target image unit.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The technical solution of the present invention is further described with reference to the following specific examples.
Examples
As shown in fig. 1, the artificial intelligence recognition apparatus based on big data provided by the present invention includes: the system comprises a three-dimensional laser scanning module 1, a three-dimensional image preprocessing module 2, a face model training module 3, an interactive recognition module 4 and a remote face recognition module 5.
The three-dimensional laser scanning module 1 acquires face information data through a micro-dimensional scanner, and the face information data usually contains redundant information such as hair and shoulders which are irrelevant to face recognition besides a face. In order to reduce the operation amount and improve the recognition rate, the three-dimensional face point cloud is preprocessed at the same time; the pretreatment comprises the following steps: face cutting and posture correction.
The three-dimensional image preprocessing module 2 is used for extracting the three-dimensional multimedia visual target image characteristics by analyzing the two aspects of the extraction of the visual two-dimensional multimedia visual target image characteristics and the generation of the three-dimensional multimedia visual target image change characteristics; and constructing a sparse representation algorithm-based model, limiting the minimum value of a target function by using gradient projection, optimizing the gradient direction to obtain a sparsely represented identification visual image, and realizing artificial intelligent identification of the three-dimensional multimedia visual image.
The face model training module 3 is used for providing the algorithm in opencv2.4.9 based on the OpenCV technology, and the trained model can be obtained only by classifying faces in a database according to different people, respectively reading in the faces and calling an OpenCV internal function.
And the interactive identification module 4 is used for inputting the face information data and identifying.
The remote face recognition module 5 is used for transmitting data to the host computer through a wireless channel by using NI USRP equipment; the host receives the data of the slave from the connected NI USRP, performs data processing and face recognition in a LabVIEW development environment, and finally feeds back the recognized identity information to the slave.
As shown in fig. 2, the three-dimensional laser scanning module 1 provided by the present invention includes:
the face cutting unit 1-1 is used for determining the position of a nose tip quickly according to the fact that the nose tip has the maximum value on the face of the human face according to the geometric constraint of the human face, then drawing a sphere by taking the nose tip as a center and taking the radius of the sphere as the center, and the face area contained in the sphere is the effective area of the face of the human face.
The pose correction unit 1-2 is configured to perform Principal Component Analysis (PCA) on the cut face point cloud, use a feature vector corresponding to the maximum feature value as a Y axis of a new coordinate system, use a feature vector corresponding to the minimum feature value as a Z axis, and establish a right-hand coordinate system, which is called a Pose Coordinate System (PCS). And converting the point cloud of the cut human face into the PCS by taking the nose tip point as the origin of the PCS to finish the correction of the human face posture. All the faces are converted to the frontal pose by establishing a PCS of the face point cloud, which is then converted to the same resolution.
PCA dimension reduction can map high-dimensional vectors to low-dimensional vectors while preserving the principal component information of the vectors well.
The PCA dimension reduction comprises the following specific steps:
(1) removing a classification label (label) of the data, and taking the d-dimensional data after removal as a sample;
(2) calculating a d-dimensional mean vector (i.e., the mean of each-dimensional vector of all data);
(3) calculating a dispersion matrix (or covariance matrix) for all data;
(4) calculating a feature value (e1, e 2.., ed) and a corresponding feature vector (lambda1, lambda 2.., lambda d);
(5) sorting the eigenvectors in descending order according to the magnitude of the eigenvalue, selecting the first k largest eigenvectors to form a d x k dimensional matrix W (wherein each column represents an eigenvector)
(6) Transforming the sample data into a new subspace by using the eigenvector matrix W of d × K:
y=WT×x
where x is a vector in d x 1 dimensions representing one sample and y is a vector in K x 1 dimensions in the new subspace.
The method comprises the steps of preprocessing three-dimensional face point cloud; the pretreatment comprises the following steps: face cutting and posture correction; the calculation amount is reduced, the recognition rate is improved, and the subsequent processing of the face information data is facilitated.
As shown in fig. 3, the three-dimensional image preprocessing module 2 provided by the present invention includes:
the two-dimensional multimedia target image feature extraction unit 2-1 is used for finding a two-dimensional coordinate point by utilizing a visual target three-dimensional contour image and setting a multimedia geometric projection model capable of showing a perspective conversion model in a real environment;
and the three-dimensional multimedia visual target image unit 2-2 is used for extracting the characteristics of the three-dimensional multimedia visual dynamic image through the optimization conversion of a nonlinear algorithm.
The invention has the advantages of higher accuracy of visual image recognition, reduced recognition time and ensured stability of recognition rate.
As shown in fig. 4, the wireless channel of the remote face recognition module 5 provided by the present invention is implemented by using a wireless communication system, and the information transmission is completed by using an asynchronous communication mode. The data is transmitted in blocks, and check bits are added when the data is transmitted, so that the accuracy of data transmission is ensured. After the data to be sent is coded, bit codes are formed, then the data are packaged into frames, and the data in the frames sequentially comprise check bits, synchronous bits, the sequence of the data packets in the whole information sequence and data bits.
As shown in fig. 5, the remote face recognition module 5 provided by the present invention specifically includes:
(1) coherent loss using a twin neural network: the device consists of 6 convolutional layers, 4 pooling layers and 1 full-connection layer. Splitting a 5 × 5 convolution kernel into two layers of 3 × 3 convolution kernels, increasing the depth of a network, and taking P-RELU as an activation function without increasing the calculated amount:
Figure BDA0003339447190000101
(2) the whole network is trained by using CASIA-Webface:
firstly organizing training data, then carrying out feature learning on the network, learning to discrete Features by using the Microsoft max loss + center loss, and finally carrying out feature extraction or classification through a trained model.
In order to make the training result better, the types with too few samples are removed firstly, only 10010 people with the most images in 10575 are selected, and image cleaning is performed at the same time, so that the identity of the people in the test set is ensured not to be repeated. The data set is balanced by randomly selecting the same number of sheets in each class, and the image is subjected to horizontally-reversed data enhancement. And (4) carrying out face key point detection and face normalization on the balanced data set, and training the network after the size is unified by 112 x 96.
(3) The feature comparison is carried out by adopting a method for calculating cosine similarity, the larger the cosine value is, the more similar the two human faces are, and the cosine formula is as follows:
Figure BDA0003339447190000111
if vectorization is performed on both sides b and c, the cosine values for calculating two vector angles can be written as:
Figure BDA0003339447190000112
and a feature extraction stage, namely extracting depth feature vectors from the test image through a trained network, and splicing and representing feature graphs of the original graph and a horizontal turnover graph of the original graph into representation of a modified graph.
As shown in fig. 6, the feature extraction of the remote face recognition module 5 provided by the present invention mainly includes the following steps:
obtaining a plurality of local areas according to the segmented face images;
comparing one pixel point in each region with a pixel point in a neighborhood, if the gray value of the pixel point in the neighborhood is larger than that of the central pixel point, marking the position of the pixel point in the neighborhood as 1, and otherwise marking as 0, thereby obtaining the LBP value of the central pixel point;
calculating histograms of all neighborhoods, namely calculating the frequency of each number, and then normalizing the histograms;
and connecting the calculated histograms of each neighborhood into a feature vector, wherein the feature vector is a vector for describing LBP texture features of the whole image.
Describing and matching the human face: the LBP texture feature vector obtained by feature extraction can represent a face image, and an SVM classifier in Open CV can be used for classification, so that the purpose of face recognition is achieved.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (10)

1. An artificial intelligence recognition device based on big data, characterized in that, the artificial intelligence recognition device based on big data includes:
the three-dimensional laser scanning module acquires face information data through the micro-dimensional scanner, wherein the face information data usually comprises redundant information of hair and shoulders which are irrelevant to face recognition besides a face; in order to reduce the operation amount and improve the recognition rate, the three-dimensional face point cloud is preprocessed at the same time; the pretreatment comprises the following steps: face cutting and posture correction;
the three-dimensional image preprocessing module is used for analyzing the two aspects of extracting the characteristics of the visual two-dimensional multimedia target image and generating the change characteristics of the three-dimensional multimedia visual target image to extract the characteristics of the three-dimensional multimedia visual target image; constructing a sparse representation algorithm-based model, limiting the minimum value of a target function by utilizing gradient projection, optimizing the gradient direction to obtain a sparsely represented identification visual image, and realizing artificial intelligent identification of the three-dimensional multimedia visual image;
the face model training module is used for providing the algorithm in OpenCV2.4.9 based on the OpenCV technology, only the faces in the database are classified according to different people and then read in respectively, and an OpenCV internal function is called to obtain a trained model;
the interactive recognition module is used for inputting face information data and recognizing the face information data;
the remote face recognition module is used for transmitting data to the host computer through a wireless channel by using the NI USRP equipment; the host receives the data of the slave from the connected NI USRP, performs data processing and face recognition in a LabVIEW development environment, and finally feeds back the recognized identity information to the slave.
2. The artificial intelligence identification device based on big data as claimed in claim 1, wherein the three-dimensional laser scanning module comprises:
the human face cutting unit is used for determining the position of a nose tip quickly according to the fact that the nose tip has the maximum value on a human face under the geometric constraint of the human face, then drawing a sphere by taking the nose tip as a center and taking a human face area contained in the sphere as an effective area of the human face;
the gesture correction unit is used for analyzing PCA (principal component analysis) by taking the cut face point cloud as a principal component, taking the feature vector corresponding to the maximum feature value as a Y axis of a new coordinate system, taking the feature vector corresponding to the minimum feature value as a Z axis, and establishing a right-hand coordinate system which is called a posture coordinate system PCS (personal computing system); taking the nose tip point as the origin of the PCS, converting the point cloud of the cut human face into the PCS, and finishing the correction of the human face posture; all the faces are converted to the frontal pose by establishing a PCS of the face point cloud, which is then converted to the same resolution.
3. The artificial intelligence recognition apparatus based on big data according to claim 1, wherein the three-dimensional image preprocessing module comprises:
the two-dimensional multimedia target image feature extraction unit is used for finding a two-dimensional coordinate point by utilizing a visual target three-dimensional contour image and setting a multimedia geometric projection model capable of showing a perspective conversion model in a real environment;
and the three-dimensional multimedia visual target image unit is used for extracting the characteristics of the three-dimensional multimedia visual dynamic image through the optimization conversion of a nonlinear algorithm.
4. The artificial intelligence based recognition device of claim 1, wherein the remote face recognition module employs a coherent loss of a twin neural network: the device consists of 6 convolution layers, 4 pooling layers and 1 full-connection layer; the 5 x 5 convolution kernel is split into two layers of 3 x 3 convolution kernels.
5. A big data based artificial intelligence recognition method for operating the big data based artificial intelligence recognition device according to any one of claims 1 to 4, wherein the big data based artificial intelligence recognition method comprises the following steps:
the method comprises the steps that a micro-dimensional optical scanner obtains face information data, wherein the face information data comprise redundant information such as hairs and shoulders which are irrelevant to face recognition besides face parts; simultaneously, preprocessing the three-dimensional face point cloud; the pretreatment comprises the following steps: face cutting and posture correction;
extracting three-dimensional multimedia visual target image features by analyzing the two aspects of visual two-dimensional multimedia target image feature extraction and three-dimensional multimedia visual target image change feature generation; constructing a sparse representation algorithm-based model, limiting the minimum value of a target function by utilizing gradient projection, optimizing the gradient direction to obtain a sparsely represented identification visual image, and realizing artificial intelligent identification of the three-dimensional multimedia visual image;
based on the OpenCV technology, the algorithm is provided in OpenCV2.4.9, and trained models can be obtained only by classifying human faces in a database according to different people, respectively reading the human faces and calling OpenCV internal functions;
inputting face information data for recognition;
the NI USRP equipment is used for sending data to the host through a wireless channel; the host receives the data of the slave from the connected NI USRP, performs data processing and face recognition in a LabVIEW development environment, and finally feeds back the recognized identity information to the slave.
6. The artificial intelligence recognition method based on big data according to claim 5, wherein the PCA dimension reduction of the artificial intelligence recognition method based on big data comprises the following steps:
(1) removing a classification label of the data, and taking the d-dimensional data after removal as a sample;
(2) calculating a mean vector of d dimensions, the mean of each dimension vector of all data;
(3) calculating a dispersion matrix or covariance matrix of all data;
(4) calculating a feature value e1, e 2.., ed and a corresponding feature vector lambda1, lambda 2.., lambda d;
(5) sorting the eigenvectors in a descending order according to the magnitude of the eigenvalues, selecting the first k largest eigenvectors to form a matrix W with dimensions d x k, wherein each column represents one eigenvector;
(6) transforming the sample data into a new subspace by using the eigenvector matrix W of d × K:
y=WT×x
where x is a vector in d x 1 dimensions representing one sample and y is a vector in K x 1 dimensions in the new subspace.
7. The artificial intelligence recognition method based on big data according to claim 5, wherein the wireless channel of the remote face recognition of the artificial intelligence recognition method based on big data is realized by a wireless communication system, and the information transmission is completed by asynchronous communication; the data is transmitted in blocks, and check bits are added when the data is transmitted; after the data to be sent is coded, bit codes are formed, then the data are packaged into frames, and the data sequence in the frames is check, synchronous bits, the sequence of the data packet in the whole information sequence and data bits.
8. The artificial intelligence recognition method based on big data according to claim 5, wherein the remote face recognition module of the artificial intelligence recognition method based on big data specifically comprises:
(1) coherent loss using a twin neural network: the device consists of 6 convolution layers, 4 pooling layers and 1 full-connection layer; splitting a 5 × 5 convolution kernel into two layers of 3 × 3 convolution kernels, increasing the depth of a network, and taking P-RELU as an activation function without increasing the calculated amount:
Figure FDA0003339447180000041
(2) the whole network is trained by using CASIA-Webface:
firstly organizing training data, then performing feature learning on a network, learning to discrete Features by using the Microsoft max loss + center loss, and finally performing feature extraction or classification through a trained model;
in order to make the training result better, firstly removing the types with too few samples, only selecting 10010 people with the most images in 10575, and simultaneously cleaning the images to ensure that the identities of the people in the test set are not repeated; the data set is balanced by randomly selecting the same number of sheets in each class, and the data of the image is enhanced by horizontal turning; carrying out face key point detection and face normalization on the balanced data set, and then training a network after unifying the size 112 x 96;
(3) the feature comparison is carried out by adopting a method for calculating cosine similarity, the larger the cosine value is, the more similar the two human faces are, and the cosine formula is as follows:
Figure FDA0003339447180000042
if vectorization is performed on both sides b and c, the cosine values for calculating two vector angles can be written as:
Figure FDA0003339447180000043
and in the characteristic extraction stage, extracting depth characteristic vectors from the test image through a trained network, and splicing and representing the original image and the characteristic graphs of the horizontal turning graph of the original image into the representation of the changed graph.
9. The artificial intelligence recognition method based on big data according to claim 5, wherein the feature extraction of the remote face recognition module of the artificial intelligence recognition method based on big data comprises the steps of:
obtaining a plurality of local areas according to the segmented face images;
comparing one pixel point in each region with a pixel point in a neighborhood, if the gray value of the pixel point in the neighborhood is larger than that of the central pixel point, marking the position of the pixel point in the neighborhood as 1, and otherwise, marking the position of the pixel point in the neighborhood as 0, thereby obtaining the LBP value of the central pixel point;
calculating histograms of all neighborhoods, namely calculating the frequency of each number, and then normalizing the histograms;
and connecting the calculated histograms of each neighborhood into a feature vector, wherein the feature vector is a vector for describing LBP texture features of the whole image.
10. An intelligent processing terminal, characterized in that the intelligent processing terminal is equipped with the artificial intelligence recognition device based on big data of any claim 1-4; the intelligent processing terminal comprises: computer, panel, cell-phone.
CN202111303955.0A 2021-11-05 2021-11-05 Artificial intelligence recognition device based on big data Pending CN114038035A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111303955.0A CN114038035A (en) 2021-11-05 2021-11-05 Artificial intelligence recognition device based on big data

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111303955.0A CN114038035A (en) 2021-11-05 2021-11-05 Artificial intelligence recognition device based on big data

Publications (1)

Publication Number Publication Date
CN114038035A true CN114038035A (en) 2022-02-11

Family

ID=80142890

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111303955.0A Pending CN114038035A (en) 2021-11-05 2021-11-05 Artificial intelligence recognition device based on big data

Country Status (1)

Country Link
CN (1) CN114038035A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115830762A (en) * 2023-01-17 2023-03-21 四川三思德科技有限公司 Safety community access control platform, control method and control terminal

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115830762A (en) * 2023-01-17 2023-03-21 四川三思德科技有限公司 Safety community access control platform, control method and control terminal

Similar Documents

Publication Publication Date Title
CN106682598B (en) Multi-pose face feature point detection method based on cascade regression
JP5121506B2 (en) Image processing apparatus, image processing method, program, and storage medium
US7912253B2 (en) Object recognition method and apparatus therefor
CN112784763B (en) Expression recognition method and system based on local and overall feature adaptive fusion
Sonkusare et al. A review on hand gesture recognition system
Yuan et al. Facial expression feature extraction using hybrid PCA and LBP
CN111339990A (en) Face recognition system and method based on dynamic update of face features
Kadhim et al. A Real-Time American Sign Language Recognition System using Convolutional Neural Network for Real Datasets.
CN109919077B (en) Gesture recognition method, device, medium and computing equipment
CN111126240A (en) Three-channel feature fusion face recognition method
CN111444488A (en) Identity authentication method based on dynamic gesture
CN110942110A (en) Feature extraction method and device of three-dimensional model
Angadi et al. Face recognition through symbolic modeling of face graphs and texture
CN110188646B (en) Human ear identification method based on fusion of gradient direction histogram and local binary pattern
CN114359917A (en) Handwritten Chinese character detection and recognition and font evaluation method
CN110610131A (en) Method and device for detecting face motion unit, electronic equipment and storage medium
CN114038035A (en) Artificial intelligence recognition device based on big data
CN103942572A (en) Method and device for extracting facial expression features based on bidirectional compressed data space dimension reduction
CN111881261A (en) Internet of things multipoint response interactive intelligent robot system
Muthukumar et al. Vision based hand gesture recognition for Indian sign languages using local binary patterns with support vector machine classifier
CN111898454A (en) Weight binarization neural network and transfer learning human eye state detection method and device
Srininvas et al. A framework to recognize the sign language system for deaf and dumb using mining techniques
Manolova et al. Facial expression classification using supervised descent method combined with PCA and SVM
Jaiswal et al. Brief description of image based 3D face recognition methods
CN113688864A (en) Human-object interaction relation classification method based on split attention

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20220211

WD01 Invention patent application deemed withdrawn after publication