CN109359544B - Portrait retrieval method and device - Google Patents

Portrait retrieval method and device Download PDF

Info

Publication number
CN109359544B
CN109359544B CN201811091048.2A CN201811091048A CN109359544B CN 109359544 B CN109359544 B CN 109359544B CN 201811091048 A CN201811091048 A CN 201811091048A CN 109359544 B CN109359544 B CN 109359544B
Authority
CN
China
Prior art keywords
result
pedestrian
preset
probability
identification
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811091048.2A
Other languages
Chinese (zh)
Other versions
CN109359544A (en
Inventor
姜黎
张仁辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Fiberhome Digtal Technology Co Ltd
Original Assignee
Wuhan Fiberhome Digtal Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Fiberhome Digtal Technology Co Ltd filed Critical Wuhan Fiberhome Digtal Technology Co Ltd
Priority to CN201811091048.2A priority Critical patent/CN109359544B/en
Publication of CN109359544A publication Critical patent/CN109359544A/en
Application granted granted Critical
Publication of CN109359544B publication Critical patent/CN109359544B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/23Recognition of whole body movements, e.g. for sport training
    • G06V40/25Recognition of walking or running movements, e.g. gait recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/049Temporal neural networks, e.g. delay elements, oscillating neurons or pulsed inputs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items

Abstract

The invention provides a portrait retrieval method and a portrait retrieval device, wherein the portrait retrieval method comprises the following steps: obtaining a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian; acquiring a gait sequence network model, inputting a gait feature sequence into the gait sequence network model, and acquiring an identification result and an identification probability for a pedestrian; judging whether the recognition probability is greater than a preset threshold value or not; if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result; and if not, judging that the obtained identification result is incorrect, and returning to execute to obtain the video to be detected. By applying the embodiment of the invention, the efficiency and the accuracy of portrait retrieval are improved.

Description

Portrait retrieval method and device
Technical Field
The invention relates to the field of data retrieval, in particular to a portrait retrieval method and a portrait retrieval device.
Background
With the development of internet technology, various video data are growing explosively, and various portrait retrieval methods are applied to quickly retrieve relevant information of a person from massive video data.
At present, a portrait retrieval method mainly adopts a manual identification method or a face identification technology to retrieve the portrait in a video to obtain a retrieval result. However, these methods have the problems of low efficiency or low accuracy, and it is still difficult to meet the actual requirements of users. For example, the size of the video to be retrieved may be several hundred T (terabytes), and it may take one or two months to identify by manual means, which is a huge and time-consuming task; although the processing speed of the face recognition technology is high, the technology is easily interfered by scenes and has strict requirements on the size of the face, the scenes in the video usually change greatly, and the size of the face of the same person in the video may change, so that the accuracy of searching the portrait by applying the technology in the video is not high.
Therefore, it is necessary to design a new portrait retrieval method to overcome the above problems.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a portrait retrieval method and a portrait retrieval device so as to improve the efficiency and accuracy of portrait retrieval.
The invention is realized by the following steps:
in a first aspect, the present invention provides a method for retrieving a portrait, the method comprising:
obtaining a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian;
acquiring a gait sequence network model, inputting the gait feature sequence into the gait sequence network model, and acquiring an identification result and an identification probability for the pedestrian;
judging whether the recognition probability is greater than a preset threshold value or not;
if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result;
and if not, judging that the obtained identification result is incorrect, and returning to execute the step of obtaining the video to be detected.
Optionally, the motion detection is performed on the pedestrian motion video to obtain a gait feature sequence of the pedestrian, including:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from each image frame containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
Optionally, the gait sequence network model is a target neural network model, and obtaining the gait sequence network model includes:
and training a preset initial neural network model by using a training sample set to obtain the target neural network model.
Optionally, the identification result includes an identity identification result and a motion identification result, the preset threshold includes a preset first threshold and a preset second threshold, and the identification probability includes an identity identification probability and a motion identification probability; judging whether the recognition probability is greater than a preset threshold value, if so, judging that the obtained recognition result is correct, and taking the recognition result as a retrieval result, wherein the steps of:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
Optionally, if the recognition probability is not greater than the preset threshold, determining that the obtained recognition result is incorrect, and returning to the step of executing the video capturing of the pedestrian, including:
and when the identification probability is not greater than a preset first threshold or the motion recognition probability is not greater than a preset second threshold, judging that the obtained recognition result is incorrect, and returning to the step of acquiring the video of the pedestrian.
Optionally, the initial neural network model is an LSTM temporal recurrent neural network model.
Optionally, when there are a plurality of search results, the method further includes:
and according to the identification probability of each retrieval result, performing ascending/descending arrangement on each retrieval result.
In a second aspect, the present invention provides a portrait retrieval apparatus, the apparatus comprising:
the first obtaining module is used for obtaining a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian;
the second obtaining module is used for obtaining a gait sequence network model, inputting the gait feature sequence into the gait sequence network model and obtaining an identification result and an identification probability aiming at the pedestrian;
the judging module is used for judging whether the recognition probability is greater than a preset threshold value or not; if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result; and if not, judging that the obtained identification result is incorrect, and returning to execute to obtain the video to be detected.
Optionally, the motion detection of the pedestrian motion video by the first obtaining module to obtain a gait feature sequence of the pedestrian includes:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from each image frame containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
Optionally, the gait sequence network model is a target neural network model, and the second obtaining module obtains the gait sequence network model, specifically:
and training a preset initial neural network model by using a training sample set to obtain the target neural network model.
Optionally, the identification result includes an identity identification result and a motion identification result, the preset threshold includes a preset first threshold and a preset second threshold, and the identification probability includes an identity identification probability and a motion identification probability; the judging module judges whether the recognition probability is greater than a preset threshold value, if so, the obtained recognition result is judged to be correct, and the recognition result is taken as a retrieval result, specifically:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
Optionally, when the recognition probability is not greater than the preset threshold, the determining module determines that the obtained recognition result is incorrect, and returns to execute to obtain the video to be detected, specifically:
and when the identity recognition probability is not greater than a preset first threshold or the motion recognition probability is not greater than a preset second threshold, judging that the obtained recognition result is incorrect, and returning to execute to obtain the video to be detected.
Optionally, the initial neural network model is an LSTM temporal recurrent neural network model.
Optionally, the apparatus further comprises:
and the sorting module is used for performing ascending/descending sorting on each retrieval result according to the identification probability of each retrieval result when a plurality of retrieval results exist.
The invention has the following beneficial effects: by applying the embodiment of the invention, the video to be detected is obtained firstly; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian; further, a gait sequence network model is obtained, a gait feature sequence is input into the gait sequence network model, and a recognition result and recognition probability for the pedestrian are obtained; judging whether the recognition probability is greater than a preset threshold value or not; if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result; and if not, judging that the obtained identification result is incorrect, and returning to execute to obtain the video to be detected.
Therefore, by applying the embodiment of the invention, the gait feature sequence is input into the gait sequence network model to obtain the identification result and the identification probability for the pedestrian, compared with the existing manual identification mode, the retrieval efficiency is improved, and when the identification probability is greater than the preset threshold value, the identification result is taken as the retrieval result; and when the recognition probability is not greater than the preset threshold value, returning to the step of obtaining the video to be detected so as to retrieve the pedestrian again, so that the accuracy of the retrieval result is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic flow chart of a human image retrieval method according to an embodiment of the present invention;
fig. 2 is a schematic structural diagram of a portrait retrieval apparatus according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the portrait retrieval method provided by the present invention can be applied to electronic devices, wherein in a specific application, the electronic devices can be computers, personal computers, tablets, mobile phones, and the like, which is reasonable.
Referring to fig. 1, an embodiment of the present invention provides a portrait retrieval method, including the following steps:
s101, obtaining a video to be detected; processing the video to obtain a gait feature sequence of the pedestrian;
the video to be detected can be a video acquired by a video acquisition device in real time, can also be a video pre-stored in an execution main body (such as an electronic device) of the invention, and can also be a video provided by a third-party device. The video acquisition equipment can be a camera, a video recorder and the like, and the invention does not limit the specific model of the video acquisition equipment, for example, the video acquisition equipment can be a monocular camera or a binocular camera. The video acquisition equipment can be fixedly arranged at each touch position; and the device can also be installed on a moving object, such as an unmanned aerial vehicle and an automobile.
The video acquisition equipment can acquire data to obtain a video and can send the video to the electronic equipment, so that the electronic equipment can acquire the video acquired by the video acquisition equipment and can process the video to obtain a gait feature sequence of the pedestrian.
Specifically, the method for processing the video to obtain the gait feature sequence of the pedestrian may be:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from the image frames containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
The video is composed of continuous image frames, can be regarded as a continuous image frame sequence, and each image frame containing the pedestrian can be detected from the video by utilizing a motion detection algorithm, and the gait feature of the pedestrian can be extracted. The specific motion detection algorithm may be one or a combination of a background image difference method, an inter-frame difference method, an optical flow method, and the like.
The gait characteristics are used for reflecting characteristics of the pedestrian during movement and can comprise characteristics of step length, stride length, step frequency and the like. Gait features can be extracted from image frames containing pedestrians, and the extracted gait features are combined to obtain a gait feature sequence. The gait characteristic sequence has uniqueness, and can uniquely distinguish the pedestrian and the motion state thereof. The states of motion of the pedestrian include loitering, jogging, normal walking, running, falling, and the like.
In addition, in order to eliminate noise in the video and improve the identification accuracy, before the video is subjected to motion detection to obtain a gait feature sequence of a pedestrian, the method may further include:
and carrying out filtering processing on the video.
Correspondingly, the step S101 of performing motion detection on the video to obtain a gait feature sequence of the pedestrian may be:
and carrying out motion detection on the filtered video to obtain a gait feature sequence of the pedestrian.
Because the video collected by the video collecting device may have noise interference, the video is filtered before the video is subjected to motion detection, so that unnecessary noise interference in the video can be reduced, and the definition of the video is improved.
The embodiment of the present invention does not limit the implementation manner of the filtering process, and for example, one or a combination of filtering algorithms such as median filtering, linear filtering, kalman filtering, and the like may be used to perform the filtering process on the video.
Therefore, by applying the embodiment of the invention, after the video is filtered, the noise in the video can be removed, and the filtered video is subjected to motion detection, which is beneficial to improving the accuracy of the gait feature sequence.
In still another implementation, to increase the speed of feature extraction, after detecting each image frame containing a pedestrian from a video, the method may further include:
performing image skeletonization processing on each image frame containing pedestrians to obtain a skeletonized image sequence;
correspondingly, the gait features of the pedestrian are extracted from the image frame containing the pedestrian, and the method specifically comprises the following steps:
and extracting gait features of the pedestrian from the skeletonized image sequence.
The image skeletonization processing can realize the refinement of the image frame, namely, some unimportant points are removed from the original image, so that the skeleton of the object in the image can be obtained. The removed points do not affect the overall shape of the object, and the skeleton can be understood as the central axis of the object, for example, a rectangular skeleton is the central axis in the long direction; the square skeleton is its central point; the skeleton of the circle is the center of the circle, the skeleton of the straight line is the skeleton itself, and the skeleton of the isolated point is the skeleton itself. The obtained image skeleton is equivalent to the main structure and shape information of the salient object, and the extraction of the gait features of the pedestrians can be realized according to the information.
By applying the embodiment of the invention, the redundant information is removed, so that the characteristic extraction speed is improved.
S102, acquiring a gait sequence network model, inputting the gait feature sequence into the gait sequence network model, and obtaining an identification result and an identification probability for the pedestrian;
the gait sequence network model can be one or a combination of a neural network model, a Support Vector Machine (SVM) model, a genetic network model and other Machine learning models. The gait sequence network model is a machine learning model which is trained to be convergent by using a training sample set in advance, so that after a gait feature sequence is obtained, a recognition result and recognition probability can be output. The recognition probability is used to evaluate the probability that the correct result was recognized.
Specifically, the gait sequence network model may be a target neural network model, and the acquiring of the gait sequence network model includes:
and training a preset initial neural network model by using the training sample set to obtain a target neural network model.
The training sample set may be a sample set used for training an initial neural network model, each sample includes a gait feature sequence and a corresponding recognition result, the gait feature sequence in the training sample set is derived from a video acquired by a video acquisition device, or a video pre-stored by an electronic device, or a video provided by another third-party device, and the recognition result in the training sample set may be labeled by an expert in advance, or may be learned by another machine learning model in advance.
All parameters in the initial neural network model are initial default parameters, after training, a target neural network model formed by the trained and mature model parameters can be obtained, and the model parameters determine the identification accuracy of the target neural network model. The initial neural network model is an LSTM (Long Short-Term Memory, time recursive neural network model), the LSTM is a modified recurrent neural network, Long-Term information can be remembered, the problem of Long-Term dependence is solved, and the method has a good learning effect on data with large information processing quantity.
Of course, in other implementations, the initial neural network model may also be a convolutional neural network model, a cyclic neural network model, or the like.
S103, judging whether the recognition probability is larger than a preset threshold value or not; if yes, executing S104; if not, executing S105;
s104, judging that the obtained identification result is correct, and taking the identification result as a retrieval result;
and S105, judging that the obtained identification result is incorrect, and returning to the step of obtaining the video to be detected.
The identification result may include an identity identification result and a motion identification result, and correspondingly, the preset threshold may include a preset first threshold and a preset second threshold, and the identification probability includes an identity identification probability and a motion identification probability; judging whether the recognition probability is greater than a preset threshold value, if so, judging that the obtained recognition result is correct, and taking the recognition result as a retrieval result, wherein the judgment result can be:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
The preset first threshold and the preset second threshold may be preset according to requirements, and may be the same or different, for example, they may be 0.75 and 0.65, respectively, that is, when the identification probability is greater than 0.75 and the motion recognition probability is greater than 0.65, it is determined that the obtained recognition result is correct, and the identification result and the motion recognition result are used as the retrieval result.
Of course, in other implementation manners, when the identification probability is greater than the preset first threshold or the motion recognition probability is greater than the preset second threshold, it may be determined that the identification is successful or the motion recognition is correct, and the identification result or the motion recognition result is used as the search result.
For example, the preset first threshold and the preset second threshold are 0.75 and 0.65, respectively, when the identity recognition probability is greater than 0.75, the identity recognition is determined to be successful, and the identity recognition result is taken as the retrieval result; and when the motion recognition probability is larger than 0.65, judging that the motion recognition is correct, and taking the motion recognition result as a retrieval result.
Or, in another implementation manner, the purpose of the portrait recognition is only to recognize the identity of the pedestrian, the recognition result may only include the identity recognition result, and when the identity recognition probability is greater than a preset threshold, it is determined that the obtained recognition result is correct, and the identity recognition result is taken as the retrieval result.
Or, in another implementation manner, the purpose of the portrait recognition is only to recognize the motion of the pedestrian, and the identity of the pedestrian is not required to be concerned, the recognition result may only include the motion recognition result, and when the motion recognition probability is greater than a preset threshold, it is determined that the obtained recognition result is correct, and the identity recognition result is used as the search result.
The preset threshold may be set in advance, and may be 0.6, 0.65, 0.7, 0.75, and the like.
The identity recognition result can be used for uniquely identifying the identity of the pedestrian, and the specific content of the identity recognition result is not limited, for example, the identity recognition result can be a combination of information such as the name of the pedestrian, an identity card number, an address and the like, and can also only comprise the identity card number and the name. The motion recognition result may include one of loitering, jogging, normal walking, running, falling, and the like.
In addition, if the recognition probability is not greater than the preset threshold, it is determined that the obtained recognition result is incorrect, and the step of executing the video acquisition of the pedestrian is returned, and the steps may be:
and when the identification probability is not greater than a preset first threshold or the motion identification probability is not greater than a preset second threshold, judging that the obtained identification result is incorrect, and returning to the step of obtaining the video to be detected.
Or, in other implementation manners, when the identity recognition probability is not greater than the preset first threshold and the motion recognition probability is not greater than the preset second threshold, it may be determined that the obtained recognition result is incorrect, and the step of obtaining the video to be detected is returned to be executed.
For example, it may be 0.75 and 0.65, respectively, that is, when the identification probability is not more than 0.75 and the motion recognition probability is not more than 0.65, it is determined that the obtained recognition result is incorrect.
Therefore, by applying the embodiment of the invention, the gait feature sequence is input into the gait sequence network model to obtain the identification result and the identification probability for the pedestrian, compared with the existing manual identification mode, the retrieval efficiency is improved, and when the identification probability is greater than the preset threshold value, the identification result is taken as the retrieval result; and when the recognition probability is not greater than the preset threshold value, returning to execute to obtain the video to be detected so as to retrieve the pedestrian again, so that the accuracy of the retrieval result is improved.
In addition, in order to improve user experience, after the obtained recognition result is judged to be incorrect, prompt information can be given, wherein the prompt information is used for prompting the user that the obtained recognition result is incorrect, or the prompt information is also used for prompting the user whether to accept the recognition result, and if the user chooses to accept, the electronic equipment can take the recognition result as a retrieval result; if the user does not select to accept, the electronic device may return to perform the step of obtaining the video to be detected.
By applying the embodiment of the invention, the user can independently select whether to accept the identification result, so that the retrieval result is obtained according to the selection of the user, and the user experience is improved.
When the retrieval result is multiple, the method further comprises the following steps:
and according to the identification probability of each retrieval result, performing ascending/descending arrangement on each retrieval result.
When the video comprises a plurality of pedestrians, aiming at each pedestrian, an identification result and an identification probability can be obtained, each identification result with the identification probability larger than a preset threshold value can be used as a retrieval result, so that the retrieval results are multiple, and the identification probability of the retrieval result is that: the recognition probability of the recognition result as the search result.
Or, when only one pedestrian is included in the video, the pedestrian may be in different motion states under different conditions, so that the number of the recognition results may be multiple, and each recognition result with the recognition probability greater than the preset threshold may be used as the search result, so that the number of the search results is multiple.
By applying the embodiment of the invention, the retrieval results can be arranged in an ascending/descending order, thereby facilitating subsequent checking and analysis.
Corresponding to the above method embodiment, the embodiment of the present invention further provides a portrait retrieval apparatus.
Referring to fig. 2, fig. 2 is a schematic structural diagram of a portrait retrieval apparatus according to an embodiment of the present invention, the apparatus includes:
a first obtaining module 201, configured to obtain a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian;
a second obtaining module 202, configured to obtain a gait sequence network model, input the gait feature sequence into the gait sequence network model, and obtain an identification result and an identification probability for the pedestrian;
a judging module 203, configured to judge whether the recognition probability is greater than a preset threshold; if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result; and if not, judging that the obtained identification result is incorrect, and returning to execute to obtain the video to be detected.
Therefore, by applying the embodiment of the invention, the gait feature sequence is input into the gait sequence network model to obtain the identification result and the identification probability for the pedestrian, compared with the existing manual identification mode, the retrieval efficiency is improved, and when the identification probability is greater than the preset threshold value, the identification result is taken as the retrieval result; and when the recognition probability is not greater than the preset threshold value, returning to execute to obtain the video to be detected so as to retrieve the pedestrian again, so that the accuracy of the retrieval result is improved.
Optionally, the motion detection of the pedestrian motion video by the first obtaining module to obtain a gait feature sequence of the pedestrian includes:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from each image frame containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
Optionally, the gait sequence network model is a target neural network model, and the second obtaining module obtains the gait sequence network model, specifically:
and training a preset initial neural network model by using a training sample set to obtain the target neural network model.
Optionally, the identification result includes an identity identification result and a motion identification result, the preset threshold includes a preset first threshold and a preset second threshold, and the identification probability includes an identity identification probability and a motion identification probability; the judging module judges whether the recognition probability is greater than a preset threshold value, if so, the obtained recognition result is judged to be correct, and the recognition result is taken as a retrieval result, specifically:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
Optionally, when the recognition probability is not greater than the preset threshold, the determining module determines that the obtained recognition result is incorrect, and returns to execute to obtain the video to be detected, specifically:
and when the identity recognition probability is not greater than a preset first threshold or the motion recognition probability is not greater than a preset second threshold, judging that the obtained recognition result is incorrect, and returning to execute to obtain the video to be detected.
Optionally, the initial neural network model is an LSTM temporal recurrent neural network model.
Optionally, the apparatus further comprises:
and the sorting module is used for performing ascending/descending sorting on each retrieval result according to the identification probability of each retrieval result when a plurality of retrieval results exist.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (9)

1. A method for retrieving a portrait, the method comprising:
obtaining a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian;
acquiring a gait sequence network model, inputting the gait feature sequence into the gait sequence network model, and acquiring an identification result and an identification probability for the pedestrian;
judging whether the recognition probability is greater than a preset threshold value or not;
if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result;
if not, judging that the obtained identification result is incorrect, and returning to the step of obtaining the video to be detected;
the identification result comprises an identity identification result and a motion identification result, the preset threshold comprises a preset first threshold and a preset second threshold, and the identification probability comprises an identity identification probability and a motion identification probability; judging whether the recognition probability is greater than a preset threshold value, if so, judging that the obtained recognition result is correct, and taking the recognition result as a retrieval result, wherein the steps of:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
2. The method according to claim 1, wherein the motion detection of the pedestrian motion video to obtain a gait feature sequence of a pedestrian comprises:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from each image frame containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
3. The method of claim 1, wherein the gait sequence network model is a target neural network model, and obtaining the gait sequence network model comprises:
and training a preset initial neural network model by using a training sample set to obtain the target neural network model.
4. The method according to claim 1, wherein if the recognition probability is not greater than the preset threshold, determining that the obtained recognition result is incorrect, and returning to the step of collecting the video of the pedestrian, the method comprises:
and when the identification probability is not greater than a preset first threshold or the motion recognition probability is not greater than a preset second threshold, judging that the obtained recognition result is incorrect, and returning to the step of acquiring the video of the pedestrian.
5. The method of claim 3, wherein the initial neural network model is an LSTM temporal recurrent neural network model.
6. The method according to claim 1, wherein when there are a plurality of search results, the method further comprises:
and according to the identification probability of each retrieval result, performing ascending/descending arrangement on each retrieval result.
7. A portrait retrieval apparatus, characterized in that the apparatus comprises:
the first obtaining module is used for obtaining a video to be detected; carrying out motion detection on the video to obtain a gait feature sequence of the pedestrian;
the second obtaining module is used for obtaining a gait sequence network model, inputting the gait feature sequence into the gait sequence network model and obtaining an identification result and an identification probability aiming at the pedestrian;
the judging module is used for judging whether the recognition probability is greater than a preset threshold value or not; if so, judging that the obtained identification result is correct, and taking the identification result as a retrieval result; if not, judging that the obtained identification result is incorrect, and returning to execute to obtain the video to be detected;
the identification result comprises an identity identification result and a motion identification result, the preset threshold comprises a preset first threshold and a preset second threshold, and the identification probability comprises an identity identification probability and a motion identification probability; judging whether the recognition probability is greater than a preset threshold value, if so, judging that the obtained recognition result is correct, and taking the recognition result as a retrieval result, wherein the steps of:
and when the identity recognition probability is greater than a preset first threshold and the motion recognition probability is greater than a preset second threshold, judging that the obtained recognition result is correct, and taking the identity recognition result and the motion recognition result as a retrieval result.
8. The apparatus according to claim 7, wherein the first obtaining module performs motion detection on the pedestrian motion video to obtain a gait feature sequence of a pedestrian, specifically:
detecting each image frame containing the pedestrian from the video by using a preset motion detection algorithm, and extracting gait features of the pedestrian from each image frame containing the pedestrian; and combining the extracted gait features to obtain a gait feature sequence of the pedestrian.
9. The apparatus according to claim 7, wherein the gait sequence network model is a target neural network model, and the second obtaining module obtains the gait sequence network model by:
and training a preset initial neural network model by using a training sample set to obtain the target neural network model.
CN201811091048.2A 2018-09-19 2018-09-19 Portrait retrieval method and device Active CN109359544B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811091048.2A CN109359544B (en) 2018-09-19 2018-09-19 Portrait retrieval method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811091048.2A CN109359544B (en) 2018-09-19 2018-09-19 Portrait retrieval method and device

Publications (2)

Publication Number Publication Date
CN109359544A CN109359544A (en) 2019-02-19
CN109359544B true CN109359544B (en) 2022-01-21

Family

ID=65351305

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811091048.2A Active CN109359544B (en) 2018-09-19 2018-09-19 Portrait retrieval method and device

Country Status (1)

Country Link
CN (1) CN109359544B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113821689A (en) * 2021-09-22 2021-12-21 沈春华 Pedestrian retrieval method and device based on video sequence and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105354548A (en) * 2015-10-30 2016-02-24 武汉大学 Surveillance video pedestrian re-recognition method based on ImageNet retrieval
CN107590452A (en) * 2017-09-04 2018-01-16 武汉神目信息技术有限公司 A kind of personal identification method and device based on gait and face fusion
CN108520216A (en) * 2018-03-28 2018-09-11 电子科技大学 A kind of personal identification method based on gait image

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2589979A1 (en) * 2011-11-03 2013-05-08 Thales Nederland B.V. System for characterizing motion of an individual, notably a human individual

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105354548A (en) * 2015-10-30 2016-02-24 武汉大学 Surveillance video pedestrian re-recognition method based on ImageNet retrieval
CN107590452A (en) * 2017-09-04 2018-01-16 武汉神目信息技术有限公司 A kind of personal identification method and device based on gait and face fusion
CN108520216A (en) * 2018-03-28 2018-09-11 电子科技大学 A kind of personal identification method based on gait image

Also Published As

Publication number Publication date
CN109359544A (en) 2019-02-19

Similar Documents

Publication Publication Date Title
CN110751022B (en) Urban pet activity track monitoring method based on image recognition and related equipment
CN110070029B (en) Gait recognition method and device
CN110807385A (en) Target detection method and device, electronic equipment and storage medium
CN113095124A (en) Face living body detection method and device and electronic equipment
CN110826476A (en) Image detection method and device for identifying target object, electronic equipment and storage medium
CN109993102B (en) Similar face retrieval method, device and storage medium
CN108009466B (en) Pedestrian detection method and device
JP2011059810A (en) Image recognition system
CN104615986A (en) Method for utilizing multiple detectors to conduct pedestrian detection on video images of scene change
CN111008576B (en) Pedestrian detection and model training method, device and readable storage medium
CN111091101B (en) High-precision pedestrian detection method, system and device based on one-step method
CN115713715B (en) Human behavior recognition method and recognition system based on deep learning
CN110991397B (en) Travel direction determining method and related equipment
CN110610123A (en) Multi-target vehicle detection method and device, electronic equipment and storage medium
CN110991278A (en) Human body action recognition method and device in video of computer vision system
CN113516113A (en) Image content identification method, device, equipment and storage medium
CN112668438A (en) Infrared video time sequence behavior positioning method, device, equipment and storage medium
CN113688804B (en) Multi-angle video-based action identification method and related equipment
CN109359544B (en) Portrait retrieval method and device
CN116758590B (en) Palm feature processing method, device, equipment and medium for identity authentication
CN112613496A (en) Pedestrian re-identification method and device, electronic equipment and storage medium
CN116597361A (en) Image recognition tracking method, device and equipment of cleaning machine and readable storage medium
CN115393755A (en) Visual target tracking method, device, equipment and storage medium
CN114445787A (en) Non-motor vehicle weight recognition method and related equipment
WO2023197887A1 (en) Intelligent control method for starting washing machine and device thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant