CN110322509A - Object localization method, system and computer equipment based on level Class Activation figure - Google Patents

Object localization method, system and computer equipment based on level Class Activation figure Download PDF

Info

Publication number
CN110322509A
CN110322509A CN201910559655.5A CN201910559655A CN110322509A CN 110322509 A CN110322509 A CN 110322509A CN 201910559655 A CN201910559655 A CN 201910559655A CN 110322509 A CN110322509 A CN 110322509A
Authority
CN
China
Prior art keywords
layer
convolutional layer
unit
level class
convolutional
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910559655.5A
Other languages
Chinese (zh)
Other versions
CN110322509B (en
Inventor
李鸿健
程卓
曾祥燕
段小林
汪美琦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN201910559655.5A priority Critical patent/CN110322509B/en
Publication of CN110322509A publication Critical patent/CN110322509A/en
Application granted granted Critical
Publication of CN110322509B publication Critical patent/CN110322509B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20016Hierarchical, coarse-to-fine, multiscale or multiresolution image processing; Pyramid transform
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to deep learning field and object detecting areas, the invention discloses object localization method, system and computer equipments based on level Class Activation figure, this method constructs a hierarchal model, traditional full articulamentum is substituted using the average pond layer of the overall situation or pyramid pond layer behind convolutional layer, loses image structure information to avoid in full articulamentum.Multiple convolutional layers of this method in low layer acquire corresponding characteristic information, to obtain level Class Activation figure.Level Class Activation figure of the invention is acquired in multiple convolutional layers in low layer not only from the last layer acquisition characteristics figure, to reduce the loss of low layer pictures information, improves framing ability.

Description

Object localization method, system and computer equipment based on level Class Activation figure
Technical field
The present invention relates to deep learning field and object detecting areas, specifically realize that object is examined using depth learning technology Target positioning under surveying;Specially a kind of object localization method based on level Class Activation figure.
Background technique
In recent years, with the quick emergence of depth learning technology, the research of object detection is had been achieved under image domains Critically important progress.Wherein most popular object detection algorithms can be divided into two types: (1) two-step method first passes through CNN net Network structure generates a series of sparse candidate frames, then carries out classification processing to these candidate frames.(2) one-step method is similar to SSD Thought carries out intensive sampling to image using the length-width ratio of different scale in image different location, extracts feature using CNN, directly Tap into capable classification.The wherein target positioning under object detection is mainly, our targets of interest from image or video It is separated with background.Weakly supervised method and strong measure of supervision are divided into for the method for target location tasks.
Weakly supervised object localization method is different with strong measure of supervision, and Weakly supervised method only needs the mark of image level, is not required to Very important person is the location information of label target, size information, such as bounding box.The workload of the mankind can be reduced in this way, while Reduce calculation amount.Due to the data set of bounding box be it is a small number of, more data sets only have the mark of image level, so phase For strong measure of supervision, Weakly supervised method applicability is wider.
Many researchers are all selected using Weakly supervised method at present, but during the experiment, they have found using convolution mind Convolutional layer through network (CNN) can directly position target, however but will lead to targeting capability after full articulamentum and lack It loses.In order to enhance the ability of target positioning, many people propose the network structure using full convolutional layer, such as the network in network , (NIN) and the proposition of full convolutional network (FCN) quantity of parameter is minimized to avoid using full articulamentum, while can also protect Hold high-performance.
For the network structure used before, many persons select the extraction feature map in top convolutional layer, carry out Pondization operation, with retaining space characteristic.In the adaptive convolutional layer of such as oquab and global maximum pondization substitution convolutional network Full coupling part enhances stationkeeping ability of the convolutional network in Weakly supervised.But this method can only position a point, and used Evaluation method be author oneself definition, cannot be general.The 3*3 convolution in one layer of 1024 channel such as Zhou on this basis Layer and global average pondization replace full articulamentum.Stationkeeping ability has greatly improved compared with the method for oquab etc., but can only determine A part of target is arrived in position.This method only extracts characteristic information from the top of convolutional network, excessive for wisp background Situation, the opposite information that can lack low layer.Zhiqiang etc. is substituted on the basis of the methods of zhou with spatial pyramid pondization The average pond of the overall situation, further increases stationkeeping ability.Comprehensive above method has a problem that they are all from convolutional layer Top extract characteristic information, will cause bottom-up information missing, locating accuracy relative reduction.
Summary of the invention
Based on problem of the existing technology, the present invention is directed in target positioning because positioning energy caused by bottom-up information missing This disadvantage of power inaccuracy, has carried out some modifications to basic convolutional network.Respectively in convolutional layer 4-3, convolutional layer 4-4, convolution Layer 5-3, convolutional layer 5-4 etc. add the 3*3 convolutional layer in one layer of 1024 channel below, and the padding of a1 and a2 are set as 0, It is proposed a kind of novel hierarchical network structure.And level Class Activation figure is obtained from the characteristic pattern of different levels, it proposes a kind of new The Class Activation figure of type, to make up the missing of opposite low level information, to improve stationkeeping ability.
A kind of object localization method based on level Class Activation figure of the invention, the method includes being by image to be predicted Input picture inputs convolution hierarchical structure, and extracts the hierarchy characteristic of image to be predicted, generates the level class of image to be predicted Activation figure, retains the partial value in level Class Activation figure, and generate the bounding box that can predict object to be measured in image to be predicted; The target position after positioning according to the bounding box to export image to be predicted;The position of bounding box namely target, in target In positioning or object detection task, it should be clear to those skilled in the art that the position of ideal is with bounding box table Show.
Wherein, the map generalization of level Class Activation the following steps are included:
The convolution hierarchical structure of S1, building image to be predicted, including the convolutional layer 4-3 in VGG19 network structure, convolution Layer 4-4, convolutional layer 5-3, one layer of customized convolutional layer is added respectively behind convolutional layer 5-4;
The step-length and padding of customized convolutional layer are added in S2, setting S1;
S3, the corresponding customized convolutional layer of convolutional layer 4-3, convolutional layer 4-4 in S2 is overlapped according to channel To the first superimposed layer;The corresponding customized convolutional layer of convolutional layer 5-3, convolutional layer 5-4 is overlapped to obtain according to channel Second superimposed layer;
S4, the first superimposed layer and the second superimposed layer progress pond are respectively obtained into TAnAnd TBn
S5, TAnAnd TBnIt is sent in linear layer, acquires the score Sc of classification;
S6, it is obtained according to classification score Sc using softmax function and cross entropy loss function to convolutional network training WeightWith
S7, the notable figure I for finding out the first superimposed layer and the second superimposed layer respectivelyAAnd IB, put using linear two interpolation method Greatly to input picture it is consistent after, two notable figures are added, level Class Activation figure is obtained;Retain it greater than maximum activation value 20% part, for generating the bounding box of prediction.
Further, in the step S2, the customized convolutional layer step-length after convolutional layer 4-3 and convolutional layer 4-4 is set as 1, padding is set as 0;It is equal that customized convolutional layer step-length after convolutional layer 5-3, convolutional layer 5-4 is set as 1, padding It is set as 1.
Further, the calculation formula of the score of classification includes:
Wherein, PcIndicate that object to be measured is the probability of c class;ScIndicate that object to be measured is the score of c class;Indicate first The weight of c class in superimposed layer;Indicate the weight of c class in the second superimposed layer, n indicates the number of characteristic pattern.
Further, the calculation formula of the cell value of the notable figure of the first superimposed layer and the second superimposed layer is represented sequentially as:
Wherein, n indicates the number of characteristic pattern;Indicate the weight of c class in the first superimposed layer;Indicate that second is folded Add the weight of c class in layer;FAkn(x, y) indicates (x, y) unit of n-th characteristic pattern in the first superimposed layer;FBkn(x, y) table Show (x, y) unit of n-th characteristic pattern in the second superimposed layer.
Further, after causing input picture consistent using the amplification of linear two interpolation method the first superimposed layer and the second superimposed layer It is added, obtains level Class Activation figure I, the part greater than maximum activation value 20% for retaining I generates prediction block.Wherein obtain level The formula of Class Activation figure I are as follows:
I=IA+IB
The invention also provides a kind of object locating system based on level Class Activation figure, the system comprises:
Image collection module, for obtaining image to be predicted;
Hierarchy characteristic extraction module, for extracting the hierarchy characteristic in image to be predicted;
Level Class Activation figure constructs module, for hierarchy characteristic to be constructed level Class Activation figure;
Predicted boundary frame computing module, for predicting the side of object to be measured in image to be predicted according to level Class Activation figure Boundary's frame;
Wherein, the hierarchy characteristic extraction module includes VGG19 network structure, customized convolution layer unit, superimposed layer list Member and pond layer unit;The customized convolution layer unit is to the convolutional layer 4-3 of VGG19 network structure, convolutional layer 4-4, volume One layer of customized convolutional layer is added behind lamination 5-3, convolutional layer 5-4 respectively;The superposition layer unit is respectively by 4-3, convolutional layer The characteristic pattern of the corresponding customized convolutional layer output of 4-4 is overlapped according to channel and by convolutional layer 5-3, convolutional layer 5- The characteristic pattern of 4 corresponding customized convolutional layer outputs is overlapped according to channel;The pond layer unit will be for that will be superimposed Layer unit treated characteristic pattern carries out pond;
The level Class Activation figure building module includes classification score calculation unit, classification function computing unit, loss letter Counting unit, notable figure superpositing unit and activation computing unit;The classification score calculation unit is used for computing pool layer unit Export the classification score of result;The classification function computing unit and the loss function unit are used for convolutional neural networks Training;The notable figure superpositing unit is for the notable figure that superimposed layer exports to be overlapped;The activation computing unit is used for The part for calculating the maximum activation value of level Class Activation figure, and will be greater than maximum activation value 20% generates the bounding box of prediction.
Further, a kind of computer equipment, including memory, processor and storage on a memory and can handled The computer program run on device;For realizing the object localization method when processor executes described program.
Beneficial effects of the present invention:
1. the present invention can extract information from the convolutional layer of opposite bottom, the missing of opposite bottom-up information can be made up.
2. the present invention can test on multiple data sets, and its targeting capability is significant.
3. every input picture only needs to carry out a propagated forward in the present invention, reduce computational complexity, saves Time cost.
4. the present invention can be used for fine-grained classification, target tracking etc. task.
Detailed description of the invention
Fig. 1 is level Class Activation map generalization procedure chart in the present invention;
Fig. 2 is flow chart of the method for the present invention;
Fig. 3 is feature extraction figure of the invention;
Fig. 4 is the hierarchical structure figure of the invention based on Class Activation figure;
Fig. 5 is level Class Activation figure of the invention;
Specific embodiment
In order to make the objectives, technical solutions, and advantages of the present invention clearer, below in conjunction with attached drawing to of the invention real The technical solution applied in example is clearly and completely described, it is clear that described embodiment is only that present invention a part is implemented Example, instead of all the embodiments.
Embodiment 1
A kind of object localization method based on level Class Activation figure of the invention, the method includes image to be predicted is defeated Enter into convolution hierarchical structure, and extract the hierarchy characteristic of image to be predicted, generates the level Class Activation figure of image to be predicted; Retain the partial value in level Class Activation figure, and generates the bounding box that can predict object to be measured;
Wherein, as shown in Figure 1, the map generalization of level Class Activation the following steps are included:
The convolution hierarchical structure of S1, building image to be predicted, including the convolutional layer 4-3 in VGG19 network structure, convolution Layer 4-4, convolutional layer 5-3, one layer of customized convolutional layer is added respectively behind convolutional layer 5-4;
The step-length and padding of customized convolutional layer are added in S2, setting S1;
S3, the corresponding customized convolutional layer of convolutional layer 4-3, convolutional layer 4-4 in S2 is overlapped according to channel To the first superimposed layer;The corresponding customized convolutional layer of convolutional layer 5-3, convolutional layer 5-4 is overlapped to obtain according to channel Second superimposed layer;
S4, the first superimposed layer and the second superimposed layer progress pond are respectively obtained into TAnAnd TBn
S5, TAnAnd TBnIt is sent in linear layer, acquires the score Sc of classification;
S6, it is obtained according to classification score Sc using softmax function and cross entropy loss function to convolutional network training WeightWith
S7, the notable figure for finding out the first superimposed layer and the second superimposed layer respectively are amplified to using linear two interpolation method After input picture is consistent, two importance are added, level Class Activation figure is obtained;Retain it and is greater than the 20% of maximum activation value Part, for generating the bounding box of prediction.
Embodiment 2
The present embodiment provides another embodiment of the present invention, in the embodiment, inputs testing image into model, calculates damage Function is lost, until loss function is restrained, then trains model, otherwise parameters is updated using gradient descent algorithm and continues to input It is trained into model;After model training is good, testing image is inputted, convolutional layer 4-3, convolutional layer 4-4, convolutional layer are extracted The characteristic pattern of 5-3, convolutional layer 5-4 determine the notable figure I of classification according to formula (5)AAnd IB;And the two notable figures are carried out Superposition obtains level Class Activation figure;Retain and activate partial value in figure, in the present embodiment, selection is greater than the value of maximum activation value 20% Retained;It is used for generating the bounding box of prediction.
Wherein, a few class loss functions well known to those of ordinary skill in the art, such as cross entropy can be used in loss function Loss function, hinge loss function, figure penalties function etc..
Specifically,
As shown in figure 3, after inputting image to be predicted in the present invention, in the Part IV and convolutional layer of the convolutional layer of VGG Part V carries out feature extraction respectively;Maximum pond is carried out to the feature after extraction respectively or pyramid pond, formation obtain Level Class Activation figure, retain partial value in level Class Activation figure, it is after output category as a result, so that it is determined that testing image out Target positioning.
Since each part includes multiple convolutional layers in VGG network structure.Such as the 4th in VGG19 network structure Part includes conv4-1, conv4-2, conv4-3, conv4-4.In the present invention at preferred Part IV and Part V Reason.
As shown in figure 4, the present embodiment is basic network with VGG19net, in convolutional layer 4-3, convolutional layer 4-4, convolutional layer 5- 3, the 3*3 convolutional layer in one layer of 1024 channel is added behind convolutional layer 5-4 respectively, and be successively named as convolutional layer a1, convolutional layer a2, Convolutional layer b1, convolutional layer b2.Convolutional layer a1 is set, the step-length of convolutional layer a2 is 1, padding 0;Convolutional layer b1, convolution are set Layer b2 step-length be 1, padding 1.Convolutional layer a1 and a2, b1 and b2 are overlapped to the formula for respectively obtaining A, B by channel For (1)
The size of A and B is not identical, so to carry out pond respectively.Here we turn to example with the average pond of the overall situation, empty Between the pyramid pondization that is averaged it is similar.Global average pondization is carried out to A and utilizes formula (2)
Fkn(x, y) indicates (x, y) unit of n-th characteristic pattern in A, NAIt is the unit number of every characteristic pattern in A. B also carries out same pondization operation.
Then obtained TAnAnd TBnIt is transmitted in linear layer and softmax function, such as formula (3) and formula (4) institute Show.
WithIt is the weight of c class.ScIt is the score that image to be predicted is c class.Then, formula 5 illustrate for Notable figure I of (x, the y) unit for classification in AA.Also there is similar I for BB
As shown in figure 5, using linear two interpolation method IAAnd IBAfter being amplified to big size as input picture, IA And IBIt is added, level Class Activation figure is obtained, shown in formula (6).
I=IA+IB (6)
20% part greater than maximum activation value for finally retaining level Class Activation figure I, for generating the boundary of prediction Frame;Prediction block in Fig. 5 can be used for predicting the position of object to be measured in image to be predicted.
Embodiment 4
The present embodiment provides the related description of object locating system in the present invention;
The invention also provides a kind of object locating system based on level Class Activation figure, the system comprises:
Image collection module, for obtaining image to be predicted;
Hierarchy characteristic extraction module, for extracting the hierarchy characteristic in image to be predicted;
Level Class Activation figure constructs module, for hierarchy characteristic to be constructed level Class Activation figure;
Predicted boundary frame computing module, for predicting the side of object to be measured in image to be predicted according to level Class Activation figure Boundary's frame;
Wherein, the hierarchy characteristic extraction module includes VGG19 network structure, customized convolution layer unit, superimposed layer list Member and pond layer unit;The customized convolution layer unit is to the convolutional layer 4-3 of VGG19 network structure, convolutional layer 4-4, volume One layer of customized convolutional layer is added behind lamination 5-3, convolutional layer 5-4 respectively;The superposition layer unit is respectively by 4-3, convolutional layer The characteristic pattern of the corresponding customized convolutional layer output of 4-4 is overlapped according to channel and by convolutional layer 5-3, convolutional layer 5- The characteristic pattern of 4 corresponding customized convolutional layer outputs is overlapped according to channel;The pond layer unit will be for that will be superimposed Layer unit treated characteristic pattern carries out pond;
The level Class Activation figure building module includes classification score calculation unit, classification function computing unit, loss letter Counting unit, notable figure superpositing unit and activation computing unit;The classification score calculation unit is used for computing pool layer unit Export the classification score of result;The classification function computing unit and the loss function unit are used for convolutional neural networks Training;The notable figure superpositing unit is for the notable figure that superimposed layer exports to be overlapped;The activation computing unit is used for The part for calculating the maximum activation value of level Class Activation figure, and will be greater than maximum activation value 20% generates the bounding box of prediction.
Embodiment 5
The embodiment of the invention also provides a kind of computer equipment, including memory, processor and it is stored in memory Computer program that is upper and can running on a processor;The processor executes fixed for realizing the target when described program Position method.
Each technical characteristic of embodiment described above can be combined arbitrarily, in order to avoid repeating, mesh in the present invention The feature of mark localization method, system and computer equipment can be quoted mutually.
Those of ordinary skill in the art will appreciate that all or part of the steps in the various methods of above-described embodiment is can It is completed with instructing relevant hardware by program, which can be stored in a computer readable storage medium, storage Medium may include: ROM, RAM, disk or CD etc..
Embodiment provided above has carried out further detailed description, institute to the object, technical solutions and advantages of the present invention It should be understood that embodiment provided above is only the preferred embodiment of the present invention, be not intended to limit the invention, it is all Any modification, equivalent substitution, improvement and etc. made for the present invention, should be included in the present invention within the spirit and principles in the present invention Protection scope within.

Claims (6)

1. a kind of object localization method based on level Class Activation figure, the method includes image to be predicted is input to convolutional layer In level structure, and the hierarchy characteristic of image to be predicted is extracted, generates the level Class Activation figure of image to be predicted;Retain level class The partial value in figure is activated, and generates the bounding box that can predict object to be measured in image to be predicted;
It is characterized in that, the map generalization of level Class Activation the following steps are included:
The convolution hierarchical structure of S1, building image to be predicted, including convolutional layer 4-3, the convolutional layer 4- in VGG19 network structure 4, one layer of customized convolutional layer is added respectively behind convolutional layer 5-3, convolutional layer 5-4;
The step-length and padding of customized convolutional layer are added in S2, setting S1;
S3, by the characteristic pattern of the corresponding customized convolutional layer output of convolutional layer 4-3, convolutional layer 4-4 in S2 according to channel into Row superposition obtains the first superimposed layer;By the characteristic pattern of the corresponding customized convolutional layer output of convolutional layer 5-3, convolutional layer 5-4 It is overlapped to obtain the second superimposed layer according to channel;
S4, the first superimposed layer and the second superimposed layer are carried out to the TA that pondization successively obtains the output of pond layer respectivelynAnd TBn
S5, TAnAnd TBnIt is sent in linear layer, acquires the score Sc of classification;
S6, weight is obtained using softmax function and cross entropy loss function to convolutional network training according to classification score ScWith
S7, the notable figure for finding out the first superimposed layer and the second superimposed layer respectively are amplified to using linear two interpolation method to pre- After altimetric image is consistent, two notable figures are added, level Class Activation figure is obtained;Retain its 20% portion for being greater than maximum activation value Point, and be used for generating the bounding box of prediction.
2. a kind of object localization method based on level Class Activation figure according to claim 1, which is characterized in that the step In rapid S2, the customized convolutional layer step-length after convolutional layer 4-3 and convolutional layer 4-4 is set as 1, padding and is set as 0;Convolution Customized convolutional layer step-length after layer 5-3, convolutional layer 5-4 is set as 1, padding and is set as 1.
3. a kind of object localization method based on level Class Activation figure according to claim 1, which is characterized in that classification The calculation formula of score includes:
Wherein, PcIndicate that object to be measured is the probability of c classification;ScIndicate that object to be measured is the score of c classification;Indicate first Object to be measured is the weight of c in superimposed layer;Indicate the weight that object to be measured is c in the second superimposed layer, n indicates characteristic pattern Number.
4. a kind of object localization method based on level Class Activation figure according to claim 1, which is characterized in that first is folded The calculation formula of the notable figure cell value of layer and the second superimposed layer is added to be represented sequentially as:
Wherein, n indicates the number of characteristic pattern;Indicate the weight of c class in the first superimposed layer;Indicate the second superimposed layer The weight of middle c class;FAkn(x, y) indicates (x, y) unit of n-th characteristic pattern in the first superimposed layer;FBkn(x, y) is indicated (x, y) unit of n-th characteristic pattern in second superimposed layer.
5. a kind of object locating system based on level Class Activation figure, which is characterized in that the system comprises:
Image collection module, for obtaining image to be predicted;
Hierarchy characteristic extraction module, for extracting the hierarchy characteristic in image to be predicted;
Level Class Activation figure constructs module, for hierarchy characteristic to be constructed level Class Activation figure;
Predicted boundary frame computing module, for predicting the boundary of object to be measured in image to be predicted according to level Class Activation figure Frame;
Wherein, the hierarchy characteristic extraction module include VGG19 network structure, customized convolution layer unit, superposition layer unit with And pond layer unit;Convolutional layer 4-3, convolutional layer 4-4, convolutional layer of the customized convolution layer unit to VGG19 network structure One layer of customized convolutional layer is added behind 5-3, convolutional layer 5-4 respectively;The superposition layer unit is each by 4-3, convolutional layer 4-4 respectively The characteristic pattern of self-corresponding customized convolutional layer output be overlapped according to channel and by convolutional layer 5-3, convolutional layer 5-4 respectively The characteristic pattern of corresponding customized convolutional layer output is overlapped according to channel;The pond layer unit will be for that will be superimposed layer unit Treated, and characteristic pattern carries out pond;
The level Class Activation figure building module includes classification score calculation unit, classification function computing unit, loss function list Member, notable figure superpositing unit and activation computing unit;The classification score calculation unit is exported for computing pool layer unit As a result classification score;The classification function computing unit and the loss function unit are used for the instruction to convolutional neural networks Practice;The notable figure superpositing unit is for the notable figure that superimposed layer exports to be overlapped;The activation computing unit is based on The part for calculating the maximum activation value of level Class Activation figure, and will be greater than maximum activation value 20% generates the bounding box of prediction.
6. a kind of computer equipment, can run on a memory and on a processor including memory, processor and storage Computer program;It is characterized in that, for realizing as described in Claims 1 to 4 is any when the processor executes described program Method.
CN201910559655.5A 2019-06-26 2019-06-26 Target positioning method, system and computer equipment based on hierarchical class activation graph Active CN110322509B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910559655.5A CN110322509B (en) 2019-06-26 2019-06-26 Target positioning method, system and computer equipment based on hierarchical class activation graph

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910559655.5A CN110322509B (en) 2019-06-26 2019-06-26 Target positioning method, system and computer equipment based on hierarchical class activation graph

Publications (2)

Publication Number Publication Date
CN110322509A true CN110322509A (en) 2019-10-11
CN110322509B CN110322509B (en) 2021-11-12

Family

ID=68121186

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910559655.5A Active CN110322509B (en) 2019-06-26 2019-06-26 Target positioning method, system and computer equipment based on hierarchical class activation graph

Country Status (1)

Country Link
CN (1) CN110322509B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110910366A (en) * 2019-11-18 2020-03-24 湖北工业大学 Visualization method of brain nuclear magnetic resonance abnormal image based on 3D CAM
CN111026898A (en) * 2019-12-10 2020-04-17 云南大学 Weak supervision image emotion classification and positioning method based on cross space pooling strategy
CN111046939A (en) * 2019-12-06 2020-04-21 中国人民解放军战略支援部队信息工程大学 CNN (CNN) class activation graph generation method based on attention
CN111553462A (en) * 2020-04-08 2020-08-18 哈尔滨工程大学 Class activation mapping method
CN113569860A (en) * 2021-07-29 2021-10-29 北京地平线信息技术有限公司 Example segmentation method, training method of example segmentation network and device thereof

Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104866829A (en) * 2015-05-25 2015-08-26 苏州大学 Cross-age face verify method based on characteristic learning
US20160352733A1 (en) * 2015-05-27 2016-12-01 Rubicon Labs, Inc. Distributed and hierarchical device activation mechanisms
CN107145842A (en) * 2017-04-19 2017-09-08 西安电子科技大学 With reference to LBP characteristic patterns and the face identification method of convolutional neural networks
CN107239565A (en) * 2017-06-14 2017-10-10 电子科技大学 A kind of image search method based on salient region
WO2017191463A1 (en) * 2016-05-06 2017-11-09 Magic Pony Technology Limited Encoder pre-analyser
CN107563999A (en) * 2017-09-05 2018-01-09 华中科技大学 A kind of chip defect recognition methods based on convolutional neural networks
US20180130203A1 (en) * 2016-11-06 2018-05-10 International Business Machines Corporation Automated skin lesion segmentation using deep side layers
CN108399406A (en) * 2018-01-15 2018-08-14 中山大学 The method and system of Weakly supervised conspicuousness object detection based on deep learning
CN108399380A (en) * 2018-02-12 2018-08-14 北京工业大学 A kind of video actions detection method based on Three dimensional convolution and Faster RCNN
CN108509954A (en) * 2018-04-23 2018-09-07 合肥湛达智能科技有限公司 A kind of more car plate dynamic identifying methods of real-time traffic scene
CN108596058A (en) * 2018-04-11 2018-09-28 西安电子科技大学 Running disorder object distance measuring method based on computer vision
CN108647585A (en) * 2018-04-20 2018-10-12 浙江工商大学 A kind of traffic mark symbol detection method based on multiple dimensioned cycle attention network
CN108875812A (en) * 2018-06-01 2018-11-23 宁波工程学院 A kind of driving behavior classification method based on branch's convolutional neural networks
CN108898078A (en) * 2018-06-15 2018-11-27 上海理工大学 A kind of traffic sign real-time detection recognition methods of multiple dimensioned deconvolution neural network
CN108960184A (en) * 2018-07-20 2018-12-07 天津师范大学 A kind of recognition methods again of the pedestrian based on heterogeneous components deep neural network
CN108985317A (en) * 2018-05-25 2018-12-11 西安电子科技大学 A kind of image classification method based on separable convolution sum attention mechanism
CN109002752A (en) * 2018-01-08 2018-12-14 北京图示科技发展有限公司 A kind of complicated common scene rapid pedestrian detection method based on deep learning
CN109214505A (en) * 2018-08-29 2019-01-15 中山大学 A kind of full convolution object detection method of intensive connection convolutional neural networks
US20190065817A1 (en) * 2017-08-29 2019-02-28 Konica Minolta Laboratory U.S.A., Inc. Method and system for detection and classification of cells using convolutional neural networks
CN109583277A (en) * 2017-09-29 2019-04-05 大连恒锐科技股份有限公司 A kind of sex determination's method that is barefoot or wearing sock print based on CNN

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104866829A (en) * 2015-05-25 2015-08-26 苏州大学 Cross-age face verify method based on characteristic learning
US20160352733A1 (en) * 2015-05-27 2016-12-01 Rubicon Labs, Inc. Distributed and hierarchical device activation mechanisms
WO2017191463A1 (en) * 2016-05-06 2017-11-09 Magic Pony Technology Limited Encoder pre-analyser
US20180130203A1 (en) * 2016-11-06 2018-05-10 International Business Machines Corporation Automated skin lesion segmentation using deep side layers
CN107145842A (en) * 2017-04-19 2017-09-08 西安电子科技大学 With reference to LBP characteristic patterns and the face identification method of convolutional neural networks
CN107239565A (en) * 2017-06-14 2017-10-10 电子科技大学 A kind of image search method based on salient region
US20190065817A1 (en) * 2017-08-29 2019-02-28 Konica Minolta Laboratory U.S.A., Inc. Method and system for detection and classification of cells using convolutional neural networks
CN107563999A (en) * 2017-09-05 2018-01-09 华中科技大学 A kind of chip defect recognition methods based on convolutional neural networks
CN109583277A (en) * 2017-09-29 2019-04-05 大连恒锐科技股份有限公司 A kind of sex determination's method that is barefoot or wearing sock print based on CNN
CN109002752A (en) * 2018-01-08 2018-12-14 北京图示科技发展有限公司 A kind of complicated common scene rapid pedestrian detection method based on deep learning
CN108399406A (en) * 2018-01-15 2018-08-14 中山大学 The method and system of Weakly supervised conspicuousness object detection based on deep learning
CN108399380A (en) * 2018-02-12 2018-08-14 北京工业大学 A kind of video actions detection method based on Three dimensional convolution and Faster RCNN
CN108596058A (en) * 2018-04-11 2018-09-28 西安电子科技大学 Running disorder object distance measuring method based on computer vision
CN108647585A (en) * 2018-04-20 2018-10-12 浙江工商大学 A kind of traffic mark symbol detection method based on multiple dimensioned cycle attention network
CN108509954A (en) * 2018-04-23 2018-09-07 合肥湛达智能科技有限公司 A kind of more car plate dynamic identifying methods of real-time traffic scene
CN108985317A (en) * 2018-05-25 2018-12-11 西安电子科技大学 A kind of image classification method based on separable convolution sum attention mechanism
CN108875812A (en) * 2018-06-01 2018-11-23 宁波工程学院 A kind of driving behavior classification method based on branch's convolutional neural networks
CN108898078A (en) * 2018-06-15 2018-11-27 上海理工大学 A kind of traffic sign real-time detection recognition methods of multiple dimensioned deconvolution neural network
CN108960184A (en) * 2018-07-20 2018-12-07 天津师范大学 A kind of recognition methods again of the pedestrian based on heterogeneous components deep neural network
CN109214505A (en) * 2018-08-29 2019-01-15 中山大学 A kind of full convolution object detection method of intensive connection convolutional neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
CHEN, C. F.等: "Efficient Fusion of Sparse and Complementary Convolutions", 《ARXIV》 *
郑德炼: "图像检索中局部特征的几何信息建模", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110910366A (en) * 2019-11-18 2020-03-24 湖北工业大学 Visualization method of brain nuclear magnetic resonance abnormal image based on 3D CAM
CN110910366B (en) * 2019-11-18 2023-10-24 湖北工业大学 Visualization method of brain nuclear magnetic resonance abnormal image based on 3D CAM
CN111046939A (en) * 2019-12-06 2020-04-21 中国人民解放军战略支援部队信息工程大学 CNN (CNN) class activation graph generation method based on attention
CN111046939B (en) * 2019-12-06 2023-08-04 中国人民解放军战略支援部队信息工程大学 Attention-based CNN class activation graph generation method
CN111026898A (en) * 2019-12-10 2020-04-17 云南大学 Weak supervision image emotion classification and positioning method based on cross space pooling strategy
CN111553462A (en) * 2020-04-08 2020-08-18 哈尔滨工程大学 Class activation mapping method
CN113569860A (en) * 2021-07-29 2021-10-29 北京地平线信息技术有限公司 Example segmentation method, training method of example segmentation network and device thereof
CN113569860B (en) * 2021-07-29 2024-02-27 北京地平线信息技术有限公司 Instance segmentation method, training method of instance segmentation network and device thereof

Also Published As

Publication number Publication date
CN110322509B (en) 2021-11-12

Similar Documents

Publication Publication Date Title
CN110322509A (en) Object localization method, system and computer equipment based on level Class Activation figure
CN108364097B (en) Based on the typhoon cloud system prediction technique for generating confrontation network
Liu et al. Super-resolution-based change detection network with stacked attention module for images with different resolutions
CN110276316B (en) Human body key point detection method based on deep learning
CN108334847B (en) A kind of face identification method based on deep learning under real scene
CN109271888A (en) Personal identification method, device, electronic equipment based on gait
Ying et al. Multi-attention object detection model in remote sensing images based on multi-scale
CN109344821A (en) Small target detecting method based on Fusion Features and deep learning
CN110147743A (en) Real-time online pedestrian analysis and number system and method under a kind of complex scene
CN109918532A (en) Image search method, device, equipment and computer readable storage medium
CN109902677A (en) A kind of vehicle checking method based on deep learning
CN110163187A (en) Remote road traffic sign detection recognition methods based on F-RCNN
CN107463919A (en) A kind of method that human facial expression recognition is carried out based on depth 3D convolutional neural networks
CN107122736A (en) A kind of human body based on deep learning is towards Forecasting Methodology and device
CN105160400A (en) L21 norm based method for improving convolutional neural network generalization capability
CN110163836A (en) Based on deep learning for the excavator detection method under the inspection of high-altitude
Wang et al. BANet: Small and multi-object detection with a bidirectional attention network for traffic scenes
CN110472542A (en) A kind of infrared image pedestrian detection method and detection system based on deep learning
CN111160111B (en) Human body key point detection method based on deep learning
CN113111767A (en) Fall detection method based on deep learning 3D posture assessment
CN109711401A (en) A kind of Method for text detection in natural scene image based on Faster Rcnn
CN104298974A (en) Human body behavior recognition method based on depth video sequence
CN109034152A (en) License plate locating method and device based on LSTM-CNN built-up pattern
CN107563409A (en) A kind of description method based on area image feature concern network with arest neighbors sequence
CN110008861A (en) A kind of recognition methods again of the pedestrian based on global and local feature learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant