CN110059577B - Pedestrian attribute information extraction method and device - Google Patents

Pedestrian attribute information extraction method and device Download PDF

Info

Publication number
CN110059577B
CN110059577B CN201910232030.8A CN201910232030A CN110059577B CN 110059577 B CN110059577 B CN 110059577B CN 201910232030 A CN201910232030 A CN 201910232030A CN 110059577 B CN110059577 B CN 110059577B
Authority
CN
China
Prior art keywords
pedestrian
features
attribute
segmentation
feature extraction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910232030.8A
Other languages
Chinese (zh)
Other versions
CN110059577A (en
Inventor
石娟峰
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Megvii Technology Co Ltd
Original Assignee
Beijing Megvii Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Megvii Technology Co Ltd filed Critical Beijing Megvii Technology Co Ltd
Priority to CN201910232030.8A priority Critical patent/CN110059577B/en
Publication of CN110059577A publication Critical patent/CN110059577A/en
Application granted granted Critical
Publication of CN110059577B publication Critical patent/CN110059577B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/41Higher-level, semantic clustering, classification or understanding of video scenes, e.g. detection, labelling or Markovian modelling of sport events or news items
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Software Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Biophysics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a pedestrian attribute information extraction method and device. The method is performed using a convolutional neural network. The convolutional neural network comprises a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model and a full connection layer. The method comprises the following steps: a preliminary feature extraction step, namely inputting the pedestrian image into a preliminary feature extraction model to obtain a preliminary feature of the pedestrian image; a segmentation feature extraction step, namely inputting the preliminary features into a pedestrian segmentation feature extraction model to extract pedestrian segmentation features; an attribute feature extraction step, namely inputting the preliminary features into a pedestrian attribute feature extraction model to extract pedestrian attribute features; a feature fusion step, namely fusing the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features; and an attribute information prediction step, wherein the fusion characteristics are input into the full connection layer to obtain the predicted pedestrian attribute information. The accuracy of the pedestrian attribute information is improved through the combination of the pedestrian segmentation features and the pedestrian attribute features.

Description

Pedestrian attribute information extraction method and device
Technical Field
The present invention relates generally to the field of artificial intelligence technology, and more particularly, to a method and apparatus for extracting pedestrian attribute information, an electronic device, and a computer-readable storage medium.
Background
In many applications of video structuring, analysis of pedestrians is crucial, and especially, the pedestrian identification method plays a core role in many fields such as security and video retrieval for human identification.
The pedestrian attribute is an important part in video structuring, and the accuracy of the pedestrian attribute plays a vital role in improving the working efficiency of security protection and the like in a real application scene.
Disclosure of Invention
In order to solve the above problems in the prior art, embodiments of the present invention provide a method and an apparatus for extracting pedestrian attribute information, an electronic device, and a computer-readable storage medium.
In a first aspect, an embodiment of the present invention provides a method for extracting pedestrian attribute information, where the method is performed using a convolutional neural network, where the convolutional neural network includes a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model, and a full connection layer, and the method includes: a preliminary feature extraction step, namely inputting the pedestrian image into a preliminary feature extraction model to obtain a preliminary feature of the pedestrian image; a segmentation feature extraction step, namely inputting the preliminary features into a pedestrian segmentation feature extraction model to extract pedestrian segmentation features; an attribute feature extraction step, namely inputting the preliminary features into a pedestrian attribute feature extraction model to extract pedestrian attribute features; a feature fusion step, namely fusing the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features; and an attribute information prediction step, wherein the fusion characteristics are input into the full connection layer to obtain the predicted pedestrian attribute information.
In one example, the pedestrian segmentation feature extraction model includes a multi-layer convolution and a multi-layer deconvolution neural network.
In one example, the feature fusion step includes: and performing bitwise addition and/or bitwise multiplication operation on the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features.
In one example, the feature fusion step includes: acquiring pedestrian segmentation position information according to the pedestrian segmentation features; and utilizing the pedestrian segmentation position information mask to mask partial pedestrian attribute features in the pedestrian attribute features to obtain pedestrian attribute features of the region of interest, thereby obtaining fusion features.
In one example, the pedestrian segmentation feature includes pedestrian segmentation semantic information and the pedestrian attribute feature includes pedestrian attribute semantic information. And, the feature fusion step includes: and obtaining fusion characteristics according to semantic relevance between the pedestrian segmentation semantic information and the pedestrian attribute semantic information.
In one example, the convolutional neural network is obtained by training through the following steps: a preliminary feature extraction training step, namely inputting the pedestrian sample image into a preliminary feature extraction model to obtain a preliminary sample feature of the pedestrian sample image; a segmentation feature extraction training step, namely inputting the preliminary features of the sample into a pedestrian segmentation feature extraction model and extracting the pedestrian segmentation features of the sample; an attribute feature extraction training step, wherein the preliminary features of the sample are input into a pedestrian attribute feature extraction model, and the pedestrian attribute features of the sample are extracted; a characteristic fusion training step, namely fusing the pedestrian segmentation characteristics of the sample and the pedestrian attribute characteristics of the sample to obtain fusion characteristics of the sample; an attribute information prediction training step, wherein the sample fusion characteristics are input into a full connection layer to obtain predicted pedestrian attribute information of the sample; a loss function calculation step, namely calculating the integral loss function of the convolutional neural network according to the sample pedestrian segmentation characteristic, the sample pedestrian attribute characteristic and the sample pedestrian attribute information; a loss function feedback step, in which the whole loss function is fed back to the convolutional neural network; and a parameter adjusting step, adjusting the parameters of the convolutional neural network according to the overall loss function until the convolutional neural network is converged.
In one example, the step of calculating the loss function comprises: calculating a segmentation loss function according to the pedestrian segmentation characteristics of the sample; calculating an attribute feature loss function according to the pedestrian attribute features of the sample; calculating an attribute information loss function according to the pedestrian attribute information of the sample; and carrying out weighted summation on the segmentation loss function, the attribute characteristic loss function and the attribute information loss function to obtain an integral loss function.
In a second aspect, an embodiment of the present invention provides a device for extracting pedestrian attribute information, where the device is implemented by using a convolutional neural network, where the convolutional neural network includes a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model, and a full connection layer, and the device includes: the preliminary feature extraction module is configured for inputting the pedestrian image into the preliminary feature extraction model to obtain a preliminary feature of the pedestrian image; the segmentation feature extraction module is configured to input the preliminary features into a pedestrian segmentation feature extraction model and extract pedestrian segmentation features; the attribute feature extraction module is configured for inputting the preliminary features into a pedestrian attribute feature extraction model and extracting pedestrian attribute features; the feature fusion module is configured to fuse the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features; and the attribute information prediction module is configured to input the fusion features into the full-connection layer to obtain predicted pedestrian attribute information.
In a third aspect, an embodiment of the present invention provides an electronic device, where the electronic device includes: a memory to store instructions; and the processor is used for calling the instructions stored in the memory to execute the method.
In a fourth aspect, embodiments of the present invention provide a computer-readable storage medium storing computer-executable instructions that, when executed by a processor, perform the above-described method.
The pedestrian attribute information extraction method and device, the electronic device and the computer readable storage medium provided by the embodiment of the invention greatly improve the accuracy of the pedestrian attribute information by combining the pedestrian segmentation information and the pedestrian attribute information.
Drawings
The above and other objects, features and advantages of embodiments of the present invention will become readily apparent from the following detailed description read in conjunction with the accompanying drawings. Several embodiments of the invention are illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which:
fig. 1 shows a flowchart of a pedestrian attribute information extraction method according to an embodiment of the present invention;
fig. 2 shows a block diagram of a pedestrian attribute information extraction device according to an embodiment of the present invention;
fig. 3 shows a block diagram of an electronic device according to an embodiment of the invention.
In the drawings, the same or corresponding reference numerals indicate the same or corresponding parts.
Detailed Description
The principles and spirit of the present invention will be described with reference to a number of exemplary embodiments. It is understood that these embodiments are given solely for the purpose of enabling those skilled in the art to better understand and to practice the invention, and are not intended to limit the scope of the invention in any way.
As shown in fig. 1, one embodiment of the present invention provides a pedestrian attribute information extraction method 100. The method 100 is performed using a convolutional neural network. The convolutional neural network is a network commonly used in image recognition and video analysis at present, and is composed of a plurality of convolution units (namely convolution kernels), and each convolution unit extracts different features. The convolutional neural network may include a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model, and a full connection layer. The method 100 includes steps S101-S105.
Step S101 is a preliminary feature extraction step of inputting a pedestrian image into a preliminary feature extraction model to obtain a preliminary feature of the pedestrian image.
For example, a classical network structure may be used as a preliminary feature extraction model. By way of example only, classical network structures such as GoogleNet, VGG, ResNet, and the like. In some embodiments, an image is first input into a preliminary feature extraction model whose parameters are initialized with a base model that has been trained.
In some embodiments, the preliminary feature extraction model may include one or more layers of convolutional neural networks.
Step S102 is a segmentation feature extraction step of inputting the preliminary features into a pedestrian segmentation feature extraction model to extract pedestrian segmentation features.
In some embodiments, extracting the pedestrian segmentation features may include extracting image information of one or more pedestrians from a background image, segmenting a portion of the pedestrians from images of other pedestrians, or segmenting image information of a local region of the pedestrians. A partial region of a pedestrian such as the head, upper body, lower body, etc. of the pedestrian.
Step S103 is an attribute feature extraction step of inputting the preliminary features into a pedestrian attribute feature extraction model to extract pedestrian attribute features.
In some embodiments, the pedestrian attribute features may include attribute features of all pedestrians, attribute features of a portion of pedestrians, or attribute features of a particular one of the pedestrians in the image. For example only, the pedestrian attribute features may include: gender, age, style of clothing, length of clothing, color of clothing, whether to wear a hat, whether to wear a backpack, style of bag, length of hair, whether to ride a bike, and the like. The above pedestrian attribute features are not exhaustive and may include any other pedestrian attribute feature.
Although step S103 is shown in fig. 1 as being performed after step S102, it should be noted that the order of steps S102 and S103 is not limited thereto. As another embodiment, step S102 may be performed after step S103. As still another embodiment, step S102 and step S103 may be performed simultaneously. The invention is not limited in this respect.
Step S104 is a feature fusion step of fusing the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features.
In some embodiments, the fused feature may include a pedestrian local attribute feature, which may be a pedestrian feature relative to a single pedestrian of a plurality of pedestrians, or a local region feature relative to a pedestrian.
Step S105 is an attribute information prediction step of inputting the fusion feature into the all-connected layer to obtain predicted pedestrian attribute information.
The pedestrian attribute information extraction method 100 provided by the embodiment of the invention combines the pedestrian segmentation information and the pedestrian attribute information, and improves the accuracy of the pedestrian attribute information by using the relationship between the two.
As one embodiment of the present invention, the pedestrian segmentation feature extraction model may include a multilayer convolution and a multilayer deconvolution neural network.
As an embodiment of the present invention, the feature fusing step S104 may include: and performing bitwise addition and/or bitwise multiplication operation on the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features.
As an embodiment of the present invention, the feature fusing step S104 may include: acquiring pedestrian segmentation position information according to the pedestrian segmentation features; and utilizing the pedestrian segmentation position information mask to mask partial pedestrian attribute features in the pedestrian attribute features to obtain pedestrian attribute features of the region of interest, thereby obtaining fusion features.
For example, if the attributes related to the hat are analyzed, the hat area obtained by dividing the pedestrian can be analyzed, and other areas such as the upper body and the lower body can be covered; if the related attributes of the packet are analyzed, the packet area obtained by dividing the pedestrian can be analyzed to obtain local information, and other areas are covered to avoid the interference of other areas.
By combining the pedestrian segmentation feature as a mask with the pedestrian attribute feature, the following processing steps can focus on the interested part in the pedestrian attribute feature, and cover other uninteresting parts, so that the interference of other parts is avoided, and the accuracy of the attribute information is improved.
As one embodiment of the present invention, the pedestrian segmentation feature may include pedestrian segmentation semantic information, and the pedestrian attribute feature may include pedestrian attribute semantic information. The feature fusion step S104 may include: and obtaining fusion characteristics according to semantic relevance between the pedestrian segmentation semantic information and the pedestrian attribute semantic information.
For example only, the pedestrian segmentation semantic information may include semantics of various body parts of the pedestrian, such as a head, the pedestrian attribute semantic information may include semantic information of whether the pedestrian wears a hat, and the segmentation features and the attribute features may be fused by semantic association of the head and the hat.
By semantically associating the pedestrian segmentation features and the pedestrian attribute features, the pedestrian segmentation features and the pedestrian attribute features which are mutually associated can be conveniently found so as to obtain the fusion features by utilizing the two features.
As an embodiment of the present invention, the convolutional neural network may be obtained by training through the following steps: a preliminary feature extraction training step, namely inputting the pedestrian sample image into a preliminary feature extraction model to obtain a preliminary sample feature of the pedestrian sample image; a segmentation feature extraction training step, namely inputting the preliminary features of the sample into a pedestrian segmentation feature extraction model and extracting the pedestrian segmentation features of the sample; an attribute feature extraction training step, wherein the preliminary features of the sample are input into a pedestrian attribute feature extraction model, and the pedestrian attribute features of the sample are extracted; a characteristic fusion training step, namely fusing the pedestrian segmentation characteristics of the sample and the pedestrian attribute characteristics of the sample to obtain fusion characteristics of the sample; an attribute information prediction training step, wherein the sample fusion characteristics are input into a full connection layer to obtain predicted pedestrian attribute information of the sample; a loss function calculation step, namely calculating the integral loss function of the convolutional neural network according to the sample pedestrian segmentation characteristic, the sample pedestrian attribute characteristic and the sample pedestrian attribute information; a loss function feedback step, in which the whole loss function is fed back to the convolutional neural network; and a parameter adjusting step, adjusting the parameters of the convolutional neural network according to the overall loss function until the convolutional neural network is converged.
In some embodiments, a loss function may be used to measure how inconsistent the predicted values of the model are from the true values. Which may be a non-negative real-valued function.
For example, the predicted values of the pedestrian segmentation features, the pedestrian attribute features and the pedestrian attribute information obtained according to the current convolutional neural network are respectively compared with the real pedestrian segmentation features, the real pedestrian attribute features and the real pedestrian attribute information to obtain the overall loss function.
In some embodiments, the loss function may include hinge loss, mutual entropy loss, square loss, exponential loss, and the like.
In some embodiments, the loss function calculating step may include: calculating a segmentation loss function according to the pedestrian segmentation characteristics of the sample; calculating an attribute feature loss function according to the pedestrian attribute features of the sample; calculating an attribute information loss function according to the pedestrian attribute information of the sample; and carrying out weighted summation on the segmentation loss function, the attribute characteristic loss function and the attribute information loss function to obtain an integral loss function.
In some embodiments, the weighted summation may multiply each individual loss function by a respective weighted ratio and add to obtain an overall value as the overall loss function. Alternatively, the weighted average may be obtained as the overall loss function by weighting and adding the sum and dividing the sum by the number of terms of the loss function involved in the weighting. Of course, any other weighting may be used, and the invention is not limited in this respect.
In some embodiments, adjusting parameters in the convolutional neural network according to the overall loss function may cause the convolutional neural network to tend to converge. For example, it may be determined that the convolutional neural network converges when the value of the loss function is below a certain threshold.
In some embodiments, training the network by adjusting the parameters of the convolutional neural network through the global loss function as described above may be achieved by adjusting the weight of each of the individual loss functions.
As shown in fig. 2, one embodiment of the present invention provides a pedestrian attribute information extraction device 200. The apparatus 200 is implemented using a convolutional neural network. The convolutional neural network comprises a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model and a full connection layer. The apparatus 200 includes a module 201 and 205.
The preliminary feature extraction module 201 may be configured to input the pedestrian image into a preliminary feature extraction model, obtaining preliminary features of the pedestrian image.
The segmentation feature extraction module 202 may be configured to input the preliminary features into a pedestrian segmentation feature extraction model, extracting pedestrian segmentation features.
The attribute feature extraction module 203 may be configured to input the preliminary features into a pedestrian attribute feature extraction model, extracting pedestrian attribute features.
The feature fusion module 204 may be configured to fuse the pedestrian segmentation features with the pedestrian attribute features to obtain fusion features.
The attribute information prediction module 205 may be configured to input the fusion feature into the fully-connected layer to obtain predicted pedestrian attribute information.
As one embodiment of the present invention, the pedestrian segmentation feature extraction model may include a multilayer convolution and a multilayer deconvolution neural network.
As an embodiment of the present invention, the feature fusion module 204 may be further configured to: and performing bitwise addition and/or bitwise multiplication operation on the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features.
As an embodiment of the present invention, the feature fusion module 204 may be further configured to: acquiring pedestrian segmentation position information according to the pedestrian segmentation features; and utilizing the pedestrian segmentation position information mask to mask partial pedestrian attribute features in the pedestrian attribute features to obtain pedestrian attribute features of the region of interest, thereby obtaining fusion features.
As one embodiment of the present invention, the pedestrian segmentation feature may include pedestrian segmentation semantic information, and the pedestrian attribute feature includes pedestrian attribute semantic information. Moreover, the feature fusion module 204 may be further configured to: and obtaining fusion characteristics according to semantic relevance between the pedestrian segmentation semantic information and the pedestrian attribute semantic information.
As an embodiment of the present invention, the convolutional neural network may be obtained by training through the following modules: the preliminary feature extraction training module is configured for inputting the pedestrian sample image into the preliminary feature extraction model to obtain preliminary features of the pedestrian sample image; the segmentation feature extraction training module is configured for inputting the preliminary features of the sample into a pedestrian segmentation feature extraction model and extracting pedestrian segmentation features of the sample; the attribute feature extraction training module is configured for inputting the preliminary features of the sample into a pedestrian attribute feature extraction model and extracting pedestrian attribute features of the sample; the characteristic fusion training module is configured for fusing the sample pedestrian segmentation characteristics and the sample pedestrian attribute characteristics to obtain sample fusion characteristics; the attribute information prediction training module is configured to input the sample fusion characteristics into the full-connection layer to obtain predicted pedestrian attribute information of the sample; the loss function calculation module is configured for calculating an overall loss function of the convolutional neural network according to the sample pedestrian segmentation feature, the sample pedestrian attribute feature and the sample pedestrian attribute information; a loss function feedback module configured to feed back the overall loss function to the convolutional neural network; and the parameter adjusting module is configured for adjusting the parameters of the convolutional neural network according to the overall loss function until the convolutional neural network converges.
As an embodiment of the present invention, the loss function calculation module may be further configured to: calculating a segmentation loss function according to the pedestrian segmentation characteristics of the sample; calculating an attribute feature loss function according to the pedestrian attribute features of the sample; calculating an attribute information loss function according to the pedestrian attribute information of the sample; and carrying out weighted summation on the segmentation loss function, the attribute characteristic loss function and the attribute information loss function to obtain an integral loss function.
The functions implemented by the modules in the apparatus correspond to the steps in the method described above, and for concrete implementation and technical effects, please refer to the description of the method steps above, which is not described herein again.
As shown in fig. 3, one embodiment of the invention provides an electronic device 300. The electronic device 300 includes a memory 301, a processor 302, and an Input/Output (I/O) interface 303. The memory 301 is used for storing instructions. And the processor 302 is used for calling the instructions stored in the memory 301 to execute the pedestrian attribute information extraction method according to the embodiment of the invention. The processor 302 is connected to the memory 301 and the I/O interface 303, respectively, for example, via a bus system and/or other connection mechanism (not shown). The memory 301 may be used to store programs and data, including a pedestrian attribute information extraction program according to an embodiment of the present invention, and the processor 302 executes various functional applications and data processing of the electronic device 300 by executing the programs stored in the memory 301.
In an embodiment of the present invention, the processor 302 may be implemented in at least one hardware form of a Digital Signal Processor (DSP), a Field-Programmable Gate Array (FPGA), a Programmable Logic Array (PLA), and the processor 302 may be one or a combination of several Central Processing Units (CPUs) or other forms of Processing units with data Processing capability and/or instruction execution capability.
Memory 301 in embodiments of the present invention may comprise one or more computer program products that may include various forms of computer-readable storage media, such as volatile memory and/or non-volatile memory. The volatile Memory may include, for example, a Random Access Memory (RAM), a Cache Memory (Cache), and/or the like. The nonvolatile Memory may include, for example, a Read-Only Memory (ROM), a Flash Memory (Flash Memory), a Hard Disk Drive (HDD), a Solid-State Drive (SSD), or the like.
In the embodiment of the present invention, the I/O interface 303 may be used to receive input instructions (e.g., numeric or character information, and generate key signal inputs related to user settings and function control of the electronic device 300, etc.), and may also output various information (e.g., images or sounds, etc.) to the outside. The I/O interface 303 may comprise one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a mouse, a joystick, a trackball, a microphone, a speaker, a touch panel, and the like.
One embodiment of the present invention provides a computer-readable storage medium having stored thereon computer-executable instructions that, when executed by a processor, perform any of the methods described above.
Although operations are depicted in the drawings in a particular order, this should not be understood as requiring that such operations be performed in the particular order shown or in serial order, or that all illustrated operations be performed, to achieve desirable results. In certain environments, multitasking and parallel processing may be advantageous.
The methods and apparatus of the present invention can be accomplished with standard programming techniques with rule based logic or other logic to accomplish the various method steps. It should also be noted that the words "means" and "module," as used herein and in the claims, is intended to encompass implementations using one or more lines of software code, and/or hardware implementations, and/or equipment for receiving inputs.
Any of the steps, operations, or procedures described herein may be performed or implemented using one or more hardware or software modules, alone or in combination with other devices. In one embodiment, the software modules are implemented using a computer program product comprising a computer readable medium containing computer program code, which is executable by a computer processor for performing any or all of the described steps, operations, or procedures.
The foregoing description of the implementation of the invention has been presented for purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed, and modifications and variations are possible in light of the above teachings or may be acquired from practice of the invention. The embodiments were chosen and described in order to explain the principles of the invention and its practical application to enable one skilled in the art to utilize the invention in various embodiments and with various modifications as are suited to the particular use contemplated.

Claims (9)

1. A pedestrian attribute information extraction method, wherein the method is performed using a convolutional neural network, wherein the convolutional neural network includes a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model, and a full-connected layer, the method comprising:
a preliminary feature extraction step of inputting a pedestrian image into the preliminary feature extraction model to obtain a preliminary feature of the pedestrian image;
a segmentation feature extraction step, namely inputting the preliminary features into the pedestrian segmentation feature extraction model to extract pedestrian segmentation features;
an attribute feature extraction step, namely inputting the preliminary features into the pedestrian attribute feature extraction model to extract pedestrian attribute features;
a feature fusion step, namely fusing the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features;
an attribute information prediction step of inputting the fusion features into the full-link layer to obtain predicted pedestrian attribute information;
correspondingly, the feature fusion step comprises the following steps: acquiring pedestrian segmentation position information according to the pedestrian segmentation features; and masking partial pedestrian attribute features in the pedestrian attribute features by using the pedestrian segmentation position information to obtain pedestrian attribute features of the region of interest, so as to obtain the fusion features.
2. The method of claim 1, wherein the pedestrian segmentation feature extraction model comprises a multi-layer convolution and a multi-layer deconvolution neural network.
3. The method of claim 1, wherein the feature fusion step comprises:
and performing bitwise addition and/or bitwise multiplication operation on the pedestrian segmentation features and the pedestrian attribute features to obtain the fusion features.
4. The method of claim 1, wherein the pedestrian segmentation features comprise pedestrian segmentation semantic information, the pedestrian attribute features comprise pedestrian attribute semantic information, and,
the feature fusion step includes:
and acquiring the fusion feature according to the semantic relevance between the pedestrian segmentation semantic information and the pedestrian attribute semantic information.
5. The method of claim 1, wherein the convolutional neural network is obtained by training by:
a preliminary feature extraction training step, namely inputting a sample pedestrian image into the preliminary feature extraction model to obtain a sample preliminary feature of the sample pedestrian image;
a segmentation feature extraction training step, namely inputting the preliminary features of the sample into the pedestrian segmentation feature extraction model to extract pedestrian segmentation features of the sample;
an attribute feature extraction training step, namely inputting the preliminary features of the samples into the pedestrian attribute feature extraction model to extract pedestrian attribute features of the samples;
a characteristic fusion training step, namely fusing the sample pedestrian segmentation characteristic and the sample pedestrian attribute characteristic to obtain a sample fusion characteristic;
an attribute information prediction training step, namely inputting the sample fusion characteristics into the full-connection layer to obtain predicted pedestrian attribute information of the sample;
a loss function calculation step, calculating the integral loss function of the convolutional neural network according to the sample pedestrian segmentation feature, the sample pedestrian attribute feature and the sample pedestrian attribute information;
a loss function feedback step of feeding back the overall loss function to the convolutional neural network;
and a parameter adjusting step, adjusting the parameters of the convolutional neural network according to the overall loss function until the convolutional neural network is converged.
6. The method of claim 5, wherein the loss function calculating step comprises:
calculating a segmentation loss function according to the sample pedestrian segmentation features;
calculating an attribute feature loss function according to the sample pedestrian attribute features;
calculating an attribute information loss function according to the sample pedestrian attribute information;
and carrying out weighted summation on the segmentation loss function, the attribute characteristic loss function and the attribute information loss function to obtain the integral loss function.
7. A pedestrian attribute information extraction device, wherein the device is implemented using a convolutional neural network, wherein the convolutional neural network includes a preliminary feature extraction model, a pedestrian segmentation feature extraction model, a pedestrian attribute feature extraction model, and a full connection layer, the device comprising:
the preliminary feature extraction module is configured to input a pedestrian image into the preliminary feature extraction model to obtain a preliminary feature of the pedestrian image;
the segmentation feature extraction module is configured to input the preliminary features into the pedestrian segmentation feature extraction model and extract pedestrian segmentation features;
the attribute feature extraction module is configured to input the preliminary features into the pedestrian attribute feature extraction model and extract pedestrian attribute features;
the feature fusion module is configured to fuse the pedestrian segmentation features and the pedestrian attribute features to obtain fusion features;
the attribute information prediction module is configured to input the fusion features into the full-connection layer to obtain predicted pedestrian attribute information;
the feature fusion module is specifically used for obtaining pedestrian segmentation position information according to the pedestrian segmentation features; and masking partial pedestrian attribute features in the pedestrian attribute features by using the pedestrian segmentation position information to obtain pedestrian attribute features of the region of interest, so as to obtain the fusion features.
8. An electronic device, the electronic device comprising:
a memory to store instructions; and
a processor for invoking the instructions stored by the memory to perform the method of any of claims 1-6.
9. A computer-readable storage medium having stored thereon computer-executable instructions that, when executed by a processor, perform the method of any one of claims 1-6.
CN201910232030.8A 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device Active CN110059577B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910232030.8A CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910232030.8A CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Publications (2)

Publication Number Publication Date
CN110059577A CN110059577A (en) 2019-07-26
CN110059577B true CN110059577B (en) 2022-02-18

Family

ID=67316332

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910232030.8A Active CN110059577B (en) 2019-03-26 2019-03-26 Pedestrian attribute information extraction method and device

Country Status (1)

Country Link
CN (1) CN110059577B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110969087B (en) * 2019-10-31 2023-11-21 杭州未名信科科技有限公司 Gait recognition method and system
CN111046950B (en) * 2019-12-11 2023-09-22 北京迈格威科技有限公司 Image processing method and device, storage medium and electronic device
CN111414812A (en) * 2020-03-03 2020-07-14 平安科技(深圳)有限公司 Human body attribute identification method, system, computer device and storage medium
CN112232173B (en) * 2020-10-12 2023-04-07 济南博观智能科技有限公司 Pedestrian attribute identification method, deep learning model, equipment and medium

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108229267B (en) * 2016-12-29 2020-10-16 北京市商汤科技开发有限公司 Object attribute detection, neural network training and region detection method and device
CN107784282B (en) * 2017-10-24 2020-04-03 北京旷视科技有限公司 Object attribute identification method, device and system
CN108875536A (en) * 2018-02-06 2018-11-23 北京迈格威科技有限公司 Pedestrian's analysis method, device, system and storage medium
CN108921054B (en) * 2018-06-15 2021-08-03 华中科技大学 Pedestrian multi-attribute identification method based on semantic segmentation

Also Published As

Publication number Publication date
CN110059577A (en) 2019-07-26

Similar Documents

Publication Publication Date Title
CN110059577B (en) Pedestrian attribute information extraction method and device
Li et al. Spatio-temporal graph routing for skeleton-based action recognition
Liu et al. Multimodal video classification with stacked contractive autoencoders
Kao et al. Visual aesthetic quality assessment with a regression model
Rukhovich et al. Iterdet: iterative scheme for object detection in crowded environments
Ma et al. Knowledge adaptation with partiallyshared features for event detectionusing few exemplars
CN112784670A (en) Object detection based on pixel differences
JP7559063B2 (en) FACE PERSHING METHOD AND RELATED DEVICE
Hii et al. Multigap: Multi-pooled inception network with text augmentation for aesthetic prediction of photographs
Le et al. Global-local attention for emotion recognition
Tan et al. Fine-grained classification via hierarchical bilinear pooling with aggregated slack mask
CN112804558B (en) Video splitting method, device and equipment
Hu et al. Multimodal learning for image popularity prediction on social media
WO2021127841A1 (en) Property identification method and apparatus, storage medium, and electronic device
CN108171189A (en) Video coding method, video coding device and electronic equipment
Zhang et al. Bioinspired scene classification by deep active learning with remote sensing applications
Gao et al. A hierarchical recurrent approach to predict scene graphs from a visual‐attention‐oriented perspective
CN110096617A (en) Video classification methods, device, electronic equipment and computer readable storage medium
CN111783692A (en) Action recognition method and device, electronic equipment and storage medium
Kompella et al. A semi-supervised recurrent neural network for video salient object detection
Sun et al. Triplet attention multiple spacetime-semantic graph convolutional network for skeleton-based action recognition
Han et al. Panoramic human activity recognition
WO2021138911A1 (en) Image classification method and apparatus, storage medium, and electronic device
CN114332893A (en) Table structure identification method and device, computer equipment and storage medium
WO2022063076A1 (en) Adversarial example identification method and apparatus

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant