CN110059577A - Pedestrian's attribute information extracting method and device - Google Patents

Pedestrian's attribute information extracting method and device Download PDF

Info

Publication number
CN110059577A
CN110059577A CN201910232030.8A CN201910232030A CN110059577A CN 110059577 A CN110059577 A CN 110059577A CN 201910232030 A CN201910232030 A CN 201910232030A CN 110059577 A CN110059577 A CN 110059577A
Authority
CN
China
Prior art keywords
pedestrian
feature
attributive character
preliminary
sample
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910232030.8A
Other languages
Chinese (zh)
Other versions
CN110059577B (en
Inventor
石娟峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Megvii Technology Co Ltd
Beijing Maigewei Technology Co Ltd
Original Assignee
Beijing Maigewei Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Maigewei Technology Co Ltd filed Critical Beijing Maigewei Technology Co Ltd
Priority to CN201910232030.8A priority Critical patent/CN110059577B/en
Publication of CN110059577A publication Critical patent/CN110059577A/en
Application granted granted Critical
Publication of CN110059577B publication Critical patent/CN110059577B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides pedestrian's attribute information extracting method and devices.This method is executed using convolutional neural networks.Convolutional neural networks include preliminary Feature Selection Model, pedestrian divides Feature Selection Model, pedestrian's attributive character extracts model and full articulamentum.This method comprises: preliminary characteristic extraction step, pedestrian image is input in preliminary Feature Selection Model, the preliminary feature of pedestrian image is obtained;Divide characteristic extraction step, preliminary feature is input to pedestrian and is divided in Feature Selection Model, extracts pedestrian and divide feature;Preliminary feature is input to pedestrian's attributive character and extracted in model, extracts pedestrian's attributive character by attributive character extraction step;Pedestrian is divided feature and merged with pedestrian's attributive character, obtains fusion feature by Fusion Features step;Fusion feature is input in full articulamentum by attribute information prediction steps, obtains pedestrian's attribute information of prediction.The combination for dividing feature and pedestrian's attributive character by pedestrian, improves the accuracy rate of pedestrian's attribute information.

Description

Pedestrian's attribute information extracting method and device
Technical field
This invention relates generally to field of artificial intelligence, relate more specifically to pedestrian's attribute information extracting method and dress It sets, electronic equipment and computer readable storage medium.
Background technique
In many applications of video structural, the analysis of pedestrian is most important, and the identification especially for people exists The various fields such as security protection, video frequency searching play central role.
Pedestrian's attribute is important a part in video structural, and the accuracy rate of pedestrian's attribute is right in practical application scene There is vital effect in promoting security protection working efficiency etc..
Summary of the invention
In order to solve the above-mentioned problems in the prior art, the embodiment of the invention provides pedestrian's attribute information extraction sides Method and device, electronic equipment and computer readable storage medium.
In a first aspect, the embodiment of the present invention provides a kind of pedestrian's attribute information extracting method, wherein this method utilizes convolution Neural network executes, wherein convolutional neural networks include preliminary Feature Selection Model, pedestrian segmentation Feature Selection Model, pedestrian Attributive character extracts model and full articulamentum, this method comprises: preliminary characteristic extraction step, is input to preliminary spy for pedestrian image Sign is extracted in model, and the preliminary feature of pedestrian image is obtained;Divide characteristic extraction step, preliminary feature is input to pedestrian's segmentation In Feature Selection Model, extracts pedestrian and divide feature;Preliminary feature is input to pedestrian's attributive character by attributive character extraction step It extracts in model, extracts pedestrian's attributive character;Pedestrian is divided feature and melted with pedestrian's attributive character by Fusion Features step It closes, obtains fusion feature;Fusion feature is input in full articulamentum by attribute information prediction steps, and the pedestrian for obtaining prediction belongs to Property information.
In one example, it includes multilayer convolution and multilayer deconvolution neural network that pedestrian, which divides Feature Selection Model,.
In one example, Fusion Features step include: by pedestrian divide feature and pedestrian's attributive character execute bit-wise addition and/ Or step-by-step multiplication operation, obtain fusion feature.
In one example, Fusion Features step includes: and divides feature according to pedestrian to obtain pedestrian's division position information;Utilize row Part pedestrian's attributive character in people's division position information mask pedestrian's attributive character, the pedestrian's attribute for obtaining area-of-interest are special Sign, to obtain fusion feature.
In one example, it includes that pedestrian divides semantic information that pedestrian, which divides feature, and pedestrian's attributive character includes pedestrian's attribute language Adopted information.Also, Fusion Features step includes: the semanteme divided between semantic information and pedestrian's attribute semantemes information according to pedestrian Relevance obtains fusion feature.
In one example, convolutional neural networks are through the following steps that being trained and obtaining: preliminary feature extraction training Sample pedestrian image is input in preliminary Feature Selection Model by step, obtains the preliminary feature of sample of sample pedestrian image;Point Feature extraction training step is cut, the preliminary feature of sample is input to pedestrian and is divided in Feature Selection Model, sample pedestrian point is extracted Cut feature;Attributive character extracts training step, and the preliminary feature of sample is input to pedestrian's attributive character and is extracted in model, sample is extracted The humanized feature of current row;Sample pedestrian is divided feature and merged with sample pedestrian's attributive character by Fusion Features training step, Obtain samples fusion feature;Attribute information predicts training step, and samples fusion feature is input in full articulamentum, is predicted Sample pedestrian's attribute information;Loss function calculates step, divides feature, sample pedestrian attributive character and sample according to sample pedestrian The humanized information of current row calculates the whole loss function of convolutional neural networks;Loss function feedback step, by whole loss function Feed back convolutional neural networks;Parameter tuning step adjusts the parameter of convolutional neural networks, Zhi Daojuan according to whole loss function Product neural network convergence.
In one example, it includes: to divide feature calculation segmentation loss function according to sample pedestrian that loss function, which calculates step,;Root According to sample pedestrian's attributive character computation attribute characteristic loss function;According to sample pedestrian's attribute information computation attribute information loss letter Number;Summation is weighted to segmentation loss function, attributive character loss function and attribute information loss function, obtains whole loss Function.
Second aspect, the embodiment of the present invention provide a kind of pedestrian's attribute information extraction element, wherein the device utilizes convolution Neural fusion, wherein convolutional neural networks include preliminary Feature Selection Model, pedestrian segmentation Feature Selection Model, pedestrian Attributive character extracts model and full articulamentum, the device include: preliminary characteristic extracting module, is configured to input pedestrian image Into preliminary Feature Selection Model, the preliminary feature of pedestrian image is obtained;Divide characteristic extracting module, is configured to preliminary spy Sign is input to pedestrian and divides in Feature Selection Model, extracts pedestrian and divides feature;Attributive character extraction module, being configured to will be first Step feature is input to pedestrian's attributive character and extracts in model, extracts pedestrian's attributive character;Fusion Features module is configured to go People's segmentation feature is merged with pedestrian's attributive character, obtains fusion feature;Attribute information prediction module is configured to merge Feature is input in full articulamentum, obtains pedestrian's attribute information of prediction.
The third aspect, the embodiment of the present invention provide a kind of electronic equipment, and electronic equipment includes: memory, refer to for storing It enables;And processor, the instruction execution above method for calling memory to store.
Fourth aspect, the embodiment of the present invention provide a kind of computer readable storage medium, and computer readable storage medium is deposited Computer executable instructions are contained, computer executable instructions when executed by the processor, execute the above method.
It pedestrian's attribute information extracting method provided in an embodiment of the present invention and device, electronic equipment and computer-readable deposits Storage media greatly improves the accuracy rate of pedestrian's attribute information by the combination of pedestrian's segmentation information and pedestrian's attribute information.
Detailed description of the invention
The following detailed description is read with reference to the accompanying drawings, above-mentioned and other purposes, the feature of embodiment of the present invention It will become prone to understand with advantage.In the accompanying drawings, several implementations of the invention are shown by way of example rather than limitation Mode, in which:
Fig. 1 shows the flow chart of pedestrian's attribute information extracting method of embodiment according to the present invention;
Fig. 2 shows the block diagrams of pedestrian's attribute information extraction element of embodiment according to the present invention;
Fig. 3 shows the block diagram of the electronic equipment of embodiment according to the present invention.
In the accompanying drawings, identical or corresponding label indicates identical or corresponding part.
Specific embodiment
The principle and spirit of the invention are described below with reference to several illustrative embodiments.It should be appreciated that providing this A little embodiments are used for the purpose of making those skilled in the art can better understand that realizing the present invention in turn, and be not with any Mode limits the scope of the invention.
As shown in Figure 1, an embodiment of the invention provides a kind of pedestrian's attribute information extracting method 100.Method 100 are executed using convolutional neural networks.Convolutional neural networks are current image recognition, common a kind of network in video analysis, It is made of several convolution units (i.e. convolution kernel), each convolution unit extracts different features.Convolutional neural networks can wrap Include preliminary Feature Selection Model, pedestrian divides Feature Selection Model, pedestrian's attributive character extracts model and full articulamentum.Method 100 include step S101-S105.
Step S101 is preliminary characteristic extraction step, and pedestrian image is input in preliminary Feature Selection Model, is gone The preliminary feature of people's image.
It is, for example, possible to use classical network structures as preliminary Feature Selection Model.Only as an example, classical network Structure such as GoogleNet, VGG, ResNet etc..In some embodiments, it for an image, is first entered into just It walks in Feature Selection Model, the parameter trained basic model initialization of the preliminary Feature Selection Model.
In some embodiments, preliminary Feature Selection Model may include one or more layers convolutional neural networks.
Step S102 is segmentation characteristic extraction step, and preliminary feature is input to pedestrian and is divided in Feature Selection Model, is mentioned Pedestrian is taken to divide feature.
In some embodiments, extract pedestrian divide feature may include by the image information of one or more pedestrians from It is extracted in background image, the image of a part of pedestrian and other pedestrians is split or is partitioned into the partial zones of pedestrian The image information in domain.The head of the regional area of pedestrian such as pedestrian, the upper part of the body, the lower part of the body etc..
Step S103 is attributive character extraction step, and preliminary feature is input to pedestrian's attributive character and is extracted in model, is mentioned Take pedestrian's attributive character.
In some embodiments, pedestrian's attributive character may include the attributive character of all pedestrians in image, part row The attributive character of people or the attributive character of specific one pedestrian.Only as an example, pedestrian's attributive character may include: property Not, the age, clothing style, clothing length, clothing color, whether be branded as, whether knapsack, the style of packet, hair length, whether Cycle etc..Above-mentioned pedestrian's attributive character is not exhaustive, can also include other any pedestrian's attributive character.
Although showing step S103 in Fig. 1 to execute after step s 102, it is however noted that, step S102 and The sequence of S103 is without being limited thereto.Alternatively, step S102 can be executed after step s 103.As another reality Mode is applied, step S102 and step S103 may be performed simultaneously.The present invention is in this regard with no restrictions.
Step S104 is Fusion Features step, and pedestrian is divided feature and is merged with pedestrian's attributive character, is merged Feature.
In some embodiments, fusion feature may include pedestrian local attribute feature, and pedestrian local attribute feature can To be to be also possible to the local features relative to a pedestrian relative to single pedestrian's feature in multiple pedestrians.
Step S105 is attribute information prediction steps, and fusion feature is input in full articulamentum, obtains the pedestrian of prediction Attribute information.
Pedestrian's attribute information extracting method 100 that embodiments of the present invention provide is by by pedestrian's segmentation information and pedestrian Attribute information combines, and improves the precision of pedestrian's attribute information using relationship between the two.
As an embodiment of the invention, it may include multilayer convolution and multilayer that pedestrian, which divides Feature Selection Model, Deconvolution neural network.
As an embodiment of the invention, Fusion Features step S104 may include: that pedestrian is divided feature and row Humanized feature executes bit-wise addition and/or step-by-step multiplication operation, obtains fusion feature.
As an embodiment of the invention, Fusion Features step S104 may include: to divide feature according to pedestrian to obtain Obtain pedestrian's division position information;Using part pedestrian's attributive character in pedestrian's division position information mask pedestrian's attributive character, Pedestrian's attributive character of area-of-interest is obtained, to obtain fusion feature.
For example, if the relevant attribute of analysis cap, so that it may the cap region that pedestrian is divided is analyzed, and incite somebody to action Other regional coverages such as upper body lower part of the body are lived;If the relevant attribute of analysis bag, so that it may which the packet region that pedestrian is divided carries out Analysis obtains local message, other regional coverages is lived, the interference in other regions is avoided.
By that in conjunction with pedestrian's attributive character, next processing step can be made to close using pedestrian's segmentation feature as mask The interested part in pedestrian's attributive character is infused, and other uninterested part overlaids are fallen, avoids the dry of other parts It disturbs, improves the accuracy of attribute information.
As an embodiment of the invention, it may include that pedestrian divides semantic information that pedestrian, which divides feature, and pedestrian belongs to Property feature may include pedestrian's attribute semantemes information.Fusion Features step S104 may include: to divide semantic information according to pedestrian Semantic relevance between pedestrian's attribute semantemes information obtains fusion feature.
Only as an example, pedestrian divides the semanteme that semantic information may include each physical feeling of pedestrian, such as head, Pedestrian's attribute semantemes information may include the semantic information whether pedestrian is branded as, by the semantic relevance on head and cap, Segmentation feature can be merged with attributive character.
It is associated by the way that pedestrian is divided feature and pedestrian's attributive character semantically, finds cross-correlation with can be convenient The pedestrian of connection divides feature and pedestrian's attributive character, to obtain fusion feature using both features.
As an embodiment of the invention, convolutional neural networks can be through the following steps that being trained and obtaining : sample pedestrian image is input in preliminary Feature Selection Model by preliminary feature extraction training step, obtains sample pedestrian figure The preliminary feature of the sample of picture;Divide feature extraction training step, the preliminary feature of sample is input to pedestrian and divides feature extraction mould In type, extracts sample pedestrian and divide feature;Attributive character extracts training step, and it is special that the preliminary feature of sample is input to pedestrian's attribute Sign is extracted in model, and sample pedestrian attributive character is extracted;Sample pedestrian is divided feature and sample row by Fusion Features training step Humanized feature is merged, and samples fusion feature is obtained;Attribute information predicts training step, and samples fusion feature is input to In full articulamentum, sample pedestrian's attribute information of prediction is obtained;Loss function calculate step, according to sample pedestrian divide feature, Sample pedestrian attributive character and sample pedestrian's attribute information calculate the whole loss function of convolutional neural networks;Loss function feedback Step, by whole loss function feedback to convolutional neural networks;Parameter tuning step adjusts convolution mind according to whole loss function Parameter through network, until convolutional neural networks are restrained.
In some embodiments, loss function can be used to estimate the predicted value of model and the inconsistent journey of true value Degree.It can be a non-negative real-valued function.
For example, the pedestrian that can be obtained according to current convolutional neural networks is divided feature, pedestrian's attributive character and pedestrian and is belonged to Property information predicted value divide feature with true pedestrian respectively, pedestrian's attributive character and pedestrian's attribute information are compared, obtain To whole loss function.
In some embodiments, loss function may include hinge loss, mutual entropy loss, Squared Error Loss, figure penalties Etc..
In some embodiments, it may include: to divide feature calculation point according to sample pedestrian that loss function, which calculates step, Cut loss function;According to sample pedestrian's attributive character computation attribute characteristic loss function;It is calculated according to sample pedestrian's attribute information Attribute information loss function;Segmentation loss function, attributive character loss function and attribute information loss function are weighted and are asked With acquisition whole loss function.
In some embodiments, weighted sum can be to each individual event loss function multiplied by respective weight ratio phase again Add, obtains population value loss function as a whole.It is alternatively possible to be added after weighting again divided by the loss function for participating in weighting Item number obtains weighted average loss function as a whole.It is of course also possible to use other any weighting schemes, the present invention exist In this respect with no restrictions.
In some embodiments, according to whole loss function, the parameter in convolutional neural networks is adjusted, it can be with So that convolutional neural networks tend to restrain.For example, can determine the convolutional Neural when the value of loss function is lower than a certain threshold value Network convergence.
In some embodiments, it is described above by whole loss function adjust convolutional neural networks parameter come pair Network is trained can realize by adjusting the weight of each individual event loss function.
As shown in Fig. 2, an embodiment of the invention provides a kind of pedestrian's attribute information extraction element 200.Device 200 are realized using convolutional neural networks.Wherein, convolutional neural networks include preliminary Feature Selection Model, pedestrian divide feature mention Modulus type, pedestrian's attributive character extract model and full articulamentum.Device 200 includes module 201-205.
Preliminary characteristic extracting module 201 may be configured to be input to pedestrian image in preliminary Feature Selection Model, obtain Obtain the preliminary feature of pedestrian image.
Segmentation characteristic extracting module 202 may be configured to for preliminary feature to be input to pedestrian's segmentation Feature Selection Model In, it extracts pedestrian and divides feature.
Attributive character extraction module 203 may be configured to for preliminary feature to be input to pedestrian's attributive character extraction model In, extract pedestrian's attributive character.
Fusion Features module 204 may be configured to merge pedestrian's segmentation feature with pedestrian's attributive character, obtain Fusion feature.
Attribute information prediction module 205 may be configured to be input to fusion feature in full articulamentum, obtain prediction Pedestrian's attribute information.
As an embodiment of the invention, it may include multilayer convolution and multilayer that pedestrian, which divides Feature Selection Model, Deconvolution neural network.
As an embodiment of the invention, Fusion Features module 204 can also be configured to: pedestrian is divided feature Bit-wise addition and/or step-by-step multiplication operation are executed with pedestrian's attributive character, obtains fusion feature.
As an embodiment of the invention, Fusion Features module 204 can also be configured to: be divided according to pedestrian special Sign obtains pedestrian's division position information;It is special using part pedestrian's attribute in pedestrian's division position information mask pedestrian's attributive character Sign, obtains pedestrian's attributive character of area-of-interest, to obtain fusion feature.
As an embodiment of the invention, it may include that pedestrian divides semantic information that pedestrian, which divides feature, and pedestrian belongs to Property feature includes pedestrian's attribute semantemes information.Also, Fusion Features module 204 can also be configured to: divide language according to pedestrian Semantic relevance between adopted information and pedestrian's attribute semantemes information obtains fusion feature.
As an embodiment of the invention, convolutional neural networks can be to be obtained and being trained with lower module : preliminary feature extraction training module is configured to for sample pedestrian image being input in preliminary Feature Selection Model, obtains sample The preliminary feature of the sample of this pedestrian image;Divide feature extraction training module, is configured to the preliminary feature of sample being input to row People is divided in Feature Selection Model, extracts sample pedestrian and divides feature;Attributive character extracts training module, is configured to sample Preliminary feature is input to pedestrian's attributive character and extracts in model, extracts sample pedestrian attributive character;Fusion Features training module, matches It sets and is merged for sample pedestrian to be divided feature with sample pedestrian's attributive character, obtain samples fusion feature;Attribute information It predicts training module, is configured to for samples fusion feature being input in full articulamentum, obtain sample pedestrian's attribute letter of prediction Breath;Loss function computing module is configured to divide feature, sample pedestrian attributive character and sample pedestrian category according to sample pedestrian Property information calculate convolutional neural networks whole loss function;Loss function feedback module is configured to whole loss function Feed back convolutional neural networks;Parameter adjustment module is configured to adjust the ginseng of convolutional neural networks according to whole loss function Number, until convolutional neural networks are restrained.
As an embodiment of the invention, loss function computing module can also be configured to: according to sample pedestrian Divide feature calculation and divides loss function;According to sample pedestrian's attributive character computation attribute characteristic loss function;According to sample row Humanized information computation attribute information loss function;To segmentation loss function, attributive character loss function and attribute information loss Function is weighted summation, obtains whole loss function.
The function that modules in device are realized is corresponding with the step in method as described above, specific implementation The description for method and step above is referred to technical effect, details are not described herein.
As shown in figure 3, an embodiment of the invention provides a kind of electronic equipment 300.Wherein, the electronic equipment 300 include memory 301, processor 302, input/output (Input/Output, I/O) interface 303.Wherein, memory 301, For storing instruction.Processor 302, pedestrian's attribute of the instruction execution embodiment of the present invention for calling memory 301 to store Information extracting method.Wherein, processor 302 is connect with memory 301, I/O interface 303 respectively, such as can pass through bus system And/or bindiny mechanism's (not shown) of other forms is attached.Memory 301 can be used for storing program and data, including this Pedestrian's attribute information extraction procedure involved in inventive embodiments, processor 302 are stored in the program of memory 301 by operation Thereby executing the various function application and data processing of electronic equipment 300.
Processor 302 can use digital signal processor (Digital Signal in the embodiment of the present invention Processing, DSP), field programmable gate array (Field-Programmable Gate Array, FPGA), programmable patrol At least one of volume array (Programmable Logic Array, PLA) example, in hardware realizes, the processor 302 It can be central processing unit (Central Processing Unit, CPU) or there is data-handling capacity and/or instruction The combination of one or more of the processing unit of other forms of executive capability.
Memory 301 in the embodiment of the present invention may include one or more computer program products, the computer Program product may include various forms of computer readable storage mediums, such as volatile memory and/or non-volatile deposit Reservoir.The volatile memory for example may include random access memory (Random Access Memory, RAM) and/ Or cache memory (Cache) etc..The nonvolatile memory for example may include read-only memory (Read-Only Memory, ROM), flash memory (Flash Memory), hard disk (Hard Disk Drive, HDD) or solid state hard disk (Solid-State Drive, SSD) etc..
In the embodiment of the present invention, I/O interface 303 can be used for receiving input instruction (such as number or character information, and Generate key signals input related with the user setting of electronic equipment 300 and function control etc.), it can also be output to the outside various Information (for example, image or sound etc.).In the embodiment of the present invention I/O interface 303 may include physical keyboard, function button (such as Volume control button, switch key etc.), mouse, operating stick, trace ball, microphone, one in loudspeaker and touch panel etc. It is a or multiple.
An embodiment of the invention provides a kind of computer readable storage medium, the computer readable storage medium Computer executable instructions are stored with, computer executable instructions when executed by the processor, execute any side described above Method.
Although description operation in a particular order in the accompanying drawings should not be construed as requiring specific shown in Sequence or serial order operate to execute these operations, or shown in requirement execution whole to obtain desired result.? In specific environment, multitask and parallel processing be may be advantageous.
Methods and apparatus of the present invention can be completed using standard programming technology, using rule-based logic or its His logic realizes various method and steps.It should also be noted that herein and the terms used in the claims " device " " module " is intended to include using the realization of a line or multirow software code and/or hardware realization and/or for receiving input Equipment.
One or more combined individually or with other equipment can be used in any step, operation or program described herein A hardware or software module are executed or are realized.In one embodiment, software module use includes comprising computer program The computer program product of the computer-readable medium of code is realized, can be executed by computer processor any for executing Or whole described step, operation or programs.
For the purpose of example and description, the preceding description that the present invention is implemented is had been presented for.Preceding description is not poor Also not the really wanting of act property limits the invention to exact form disclosed, according to the above instruction there is likely to be various modifications and Modification, or various changes and modifications may be obtained from the practice of the present invention.Select and describe these embodiments and be in order to Illustrate the principle of the present invention and its practical application, so that those skilled in the art can be to be suitable for the special-purpose conceived Come in a variety of embodiments with various modifications and utilize the present invention.

Claims (10)

1. a kind of pedestrian's attribute information extracting method, wherein the method is executed using convolutional neural networks, wherein the volume Product neural network includes preliminary Feature Selection Model, pedestrian divides Feature Selection Model, pedestrian's attributive character extracts model and complete Articulamentum, which comprises
Pedestrian image is input in the preliminary Feature Selection Model, obtains the pedestrian image by preliminary characteristic extraction step Preliminary feature;
Divide characteristic extraction step, the preliminary feature is input to the pedestrian and is divided in Feature Selection Model, pedestrian is extracted Divide feature;
The preliminary feature is input to pedestrian's attributive character and extracted in model, extracts pedestrian by attributive character extraction step Attributive character;
The pedestrian is divided feature and merged with pedestrian's attributive character, obtains fusion feature by Fusion Features step;
The fusion feature is input in the full articulamentum by attribute information prediction steps, obtains pedestrian's attribute letter of prediction Breath.
2. according to the method described in claim 1, wherein, it includes multilayer convolution and more that the pedestrian, which divides Feature Selection Model, Layer deconvolution neural network.
3. according to the method described in claim 1, wherein, the Fusion Features step includes:
The pedestrian is divided into feature and pedestrian's attributive character executes bit-wise addition and/or step-by-step multiplication operation, obtains institute State fusion feature.
4. according to the method described in claim 1, wherein, the Fusion Features step includes:
Divide feature according to the pedestrian and obtains pedestrian's division position information;
Using part pedestrian's attributive character in pedestrian's attributive character described in pedestrian's division position information mask, obtain feeling emerging Pedestrian's attributive character in interesting region, to obtain the fusion feature.
5. according to the method described in claim 1, wherein, it includes that pedestrian divides semantic information that the pedestrian, which divides feature, described Pedestrian's attributive character includes pedestrian's attribute semantemes information, also,
The Fusion Features step includes:
Divide the semantic relevance between semantic information and pedestrian's attribute semantemes information according to the pedestrian, melts described in acquisition Close feature.
6. according to the method described in claim 1, wherein, the convolutional neural networks are through the following steps that being trained and obtaining :
Preliminary feature extraction training step, sample pedestrian image is input in the preliminary Feature Selection Model, described in acquisition The preliminary feature of the sample of sample pedestrian image;
Divide feature extraction training step, the preliminary feature of the sample be input to the pedestrian and is divided in Feature Selection Model, It extracts sample pedestrian and divides feature;
Attributive character extracts training step, and the preliminary feature of the sample is input to pedestrian's attributive character and is extracted in model, Extract sample pedestrian attributive character;
The sample pedestrian is divided feature and merged with the sample pedestrian attributive character, obtained by Fusion Features training step Obtain samples fusion feature;
Attribute information predicts training step, and the samples fusion feature is input in the full articulamentum, the sample of prediction is obtained The humanized information of current row;
Loss function calculates step, divides feature, the sample pedestrian attributive character and the sample according to the sample pedestrian Pedestrian's attribute information calculates the whole loss function of the convolutional neural networks;
Loss function feedback step, by whole loss function feedback to the convolutional neural networks;
Parameter tuning step adjusts the parameter of the convolutional neural networks according to the whole loss function, until the convolution Neural network convergence.
7. according to the method described in claim 6, wherein, the loss function calculates step and includes:
Divide feature calculation according to the sample pedestrian and divides loss function;
According to the sample pedestrian attributive character computation attribute characteristic loss function;
According to the sample pedestrian attribute information computation attribute information loss function;
The segmentation loss function, the attributive character loss function and the attribute information loss function are weighted and are asked With obtain the whole loss function.
8. a kind of pedestrian's attribute information extraction element, wherein described device is realized using convolutional neural networks, wherein the volume Product neural network includes preliminary Feature Selection Model, pedestrian divides Feature Selection Model, pedestrian's attributive character extracts model and complete Articulamentum, described device include:
Preliminary characteristic extracting module is configured to be input to pedestrian image in the preliminary Feature Selection Model, described in acquisition The preliminary feature of pedestrian image;
Divide characteristic extracting module, be configured to be input to the preliminary feature in pedestrian's segmentation Feature Selection Model, It extracts pedestrian and divides feature;
Attributive character extraction module is configured to be input to the preliminary feature in pedestrian's attributive character extraction model, Extract pedestrian's attributive character;
Fusion Features module is configured to merge pedestrian segmentation feature with pedestrian's attributive character, be melted Close feature;
Attribute information prediction module is configured to for the fusion feature being input in the full articulamentum, obtains the row of prediction Humanized information.
9. a kind of electronic equipment, the electronic equipment include:
Memory, for storing instruction;And
Processor, for calling the described instruction of the memory storage to execute such as side of any of claims 1-7 Method.
10. a kind of computer readable storage medium, the computer-readable recording medium storage has computer executable instructions, institute It states computer executable instructions when executed by the processor, executes such as method of any of claims 1-7.
CN201910232030.8A 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device Active CN110059577B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910232030.8A CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910232030.8A CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Publications (2)

Publication Number Publication Date
CN110059577A true CN110059577A (en) 2019-07-26
CN110059577B CN110059577B (en) 2022-02-18

Family

ID=67316332

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910232030.8A Active CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Country Status (1)

Country Link
CN (1) CN110059577B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110969087A (en) * 2019-10-31 2020-04-07 浙江省北大信息技术高等研究院 Gait recognition method and system
CN111046950A (en) * 2019-12-11 2020-04-21 北京迈格威科技有限公司 Image processing method and device, storage medium and electronic device
CN112232173A (en) * 2020-10-12 2021-01-15 济南博观智能科技有限公司 Pedestrian attribute identification method, deep learning model, equipment and medium
WO2021174941A1 (en) * 2020-03-03 2021-09-10 平安科技(深圳)有限公司 Physical attribute recognition method, system, computer device, and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107784282A (en) * 2017-10-24 2018-03-09 北京旷视科技有限公司 The recognition methods of object properties, apparatus and system
CN108229267A (en) * 2016-12-29 2018-06-29 北京市商汤科技开发有限公司 Object properties detection, neural metwork training, method for detecting area and device
CN108875536A (en) * 2018-02-06 2018-11-23 北京迈格威科技有限公司 Pedestrian's analysis method, device, system and storage medium
CN108921054A (en) * 2018-06-15 2018-11-30 华中科技大学 A kind of more attribute recognition approaches of pedestrian based on semantic segmentation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229267A (en) * 2016-12-29 2018-06-29 北京市商汤科技开发有限公司 Object properties detection, neural metwork training, method for detecting area and device
CN107784282A (en) * 2017-10-24 2018-03-09 北京旷视科技有限公司 The recognition methods of object properties, apparatus and system
CN108875536A (en) * 2018-02-06 2018-11-23 北京迈格威科技有限公司 Pedestrian's analysis method, device, system and storage medium
CN108921054A (en) * 2018-06-15 2018-11-30 华中科技大学 A kind of more attribute recognition approaches of pedestrian based on semantic segmentation

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110969087A (en) * 2019-10-31 2020-04-07 浙江省北大信息技术高等研究院 Gait recognition method and system
CN110969087B (en) * 2019-10-31 2023-11-21 杭州未名信科科技有限公司 Gait recognition method and system
CN111046950A (en) * 2019-12-11 2020-04-21 北京迈格威科技有限公司 Image processing method and device, storage medium and electronic device
CN111046950B (en) * 2019-12-11 2023-09-22 北京迈格威科技有限公司 Image processing method and device, storage medium and electronic device
WO2021174941A1 (en) * 2020-03-03 2021-09-10 平安科技(深圳)有限公司 Physical attribute recognition method, system, computer device, and storage medium
CN112232173A (en) * 2020-10-12 2021-01-15 济南博观智能科技有限公司 Pedestrian attribute identification method, deep learning model, equipment and medium
CN112232173B (en) * 2020-10-12 2023-04-07 济南博观智能科技有限公司 Pedestrian attribute identification method, deep learning model, equipment and medium

Also Published As

Publication number Publication date
CN110059577B (en) 2022-02-18

Similar Documents

Publication Publication Date Title
Ozcan et al. Transfer learning-based convolutional neural networks with heuristic optimization for hand gesture recognition
Jung et al. Discriminative feature learning for unsupervised video summarization
CN110059577A (en) Pedestrian's attribute information extracting method and device
US10958748B2 (en) Resource push method and apparatus
Fauvel et al. XEM: An explainable-by-design ensemble method for multivariate time series classification
CN103530540B (en) User identity attribute detection method based on man-machine interaction behavior characteristics
US11741711B2 (en) Video classification method and server
Yan et al. Hierarchical multi-scale attention networks for action recognition
CN110263733B (en) Image processing method, nomination evaluation method and related device
Kempe et al. Detecting tactical patterns in basketball: comparison of merge self-organising maps and dynamic controlled neural networks
CN110599492A (en) Training method and device for image segmentation model, electronic equipment and storage medium
CN106529511A (en) Image structuring method and device
Liu et al. Hybrid design for sports data visualization using AI and big data analytics
JP7292657B2 (en) DATA PROCESSING METHOD, DATA PROCESSING DEVICE, COMPUTER PROGRAM AND ELECTRONIC DEVICE
Hii et al. Multigap: Multi-pooled inception network with text augmentation for aesthetic prediction of photographs
CN110147882A (en) Training method, crowd's method of diffusion, device and the equipment of neural network model
Zhang et al. Video event classification based on two-stage neural network
CN111400615B (en) Resource recommendation method, device, equipment and storage medium
Yudistira et al. Correlation net: Spatiotemporal multimodal deep learning for action recognition
CN113742580B (en) Recall method and device for target type data, electronic equipment and storage medium
Werling et al. On-the-job learning with bayesian decision theory
Amirgholipour et al. PDANet: Pyramid density-aware attention based network for accurate crowd counting
Han et al. Robust human action recognition using global spatial-temporal attention for human skeleton data
Sun et al. Triplet attention multiple spacetime-semantic graph convolutional network for skeleton-based action recognition
Peng et al. Motion boundary emphasised optical flow method for human action recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant