CN107194318A - The scene recognition method of target detection auxiliary - Google Patents

The scene recognition method of target detection auxiliary Download PDF

Info

Publication number
CN107194318A
CN107194318A CN201710270013.4A CN201710270013A CN107194318A CN 107194318 A CN107194318 A CN 107194318A CN 201710270013 A CN201710270013 A CN 201710270013A CN 107194318 A CN107194318 A CN 107194318A
Authority
CN
China
Prior art keywords
picture
identified
scene
training
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710270013.4A
Other languages
Chinese (zh)
Other versions
CN107194318B (en
Inventor
王蕴红
孙宇航
赵文婷
陈训逊
刘庆杰
王博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beihang University
Original Assignee
Beihang University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beihang University filed Critical Beihang University
Priority to CN201710270013.4A priority Critical patent/CN107194318B/en
Publication of CN107194318A publication Critical patent/CN107194318A/en
Application granted granted Critical
Publication of CN107194318B publication Critical patent/CN107194318B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/35Categorising the entire scene, e.g. birthday party or wedding scene
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of scene recognition method of target detection auxiliary.The scene recognition method of target detection auxiliary of the present invention includes:Picture to be identified is obtained, picture to be identified is sampled, the sample of predetermined number and default size is obtained, scene Recognition is carried out to each sample according to convolutional neural networks model, corresponding at least two scene of picture to be identified is obtained;Fisrt feature figure corresponding with picture to be identified is advised in the region for obtaining picture to be identified, according to region suggestion and picture to be identified, obtains the classification score of each target in picture to be identified;According to corresponding at least two scene of picture to be identified and the classification score of each target, the corresponding scene of picture to be identified is obtained.The present invention advises that the object detection method of network integration aids in scene recognition method by Fast R CNN networks and region so that the accuracy rate of scene recognition method is improved.

Description

The scene recognition method of target detection auxiliary
Technical field
The present invention relates to the scene recognition method of technical field of computer vision, more particularly to a kind of target detection auxiliary.
Background technology
Scene Recognition is a major issue of computer vision field, is by the computer automatic decision image or photo Which kind of specific scene belonged to, especially important work is played in terms of video monitoring, social network user Behavior mining With.
However, due to scene complexity in itself and illumination, block, the factor, existing scene Recognition such as dimensional variation The scene type that method can not still be distinguished well, such as a pictures have group, and computer is difficult to picture classification to arrive The scene such as market or station or rally parade, so lessens the accuracy rate to scene Recognition.
The content of the invention
The present invention provides a kind of scene recognition method of target detection auxiliary, to realize target detection auxiliary scene Recognition Method, the problem of to reduce in scene recognition method to scene error differentiating.
The present invention provides a kind of scene recognition method of target detection auxiliary, including:
Picture to be identified is obtained, the picture to be identified is sampled, the sample of predetermined number and default size is obtained, Scene Recognition is carried out to each sample according to convolutional neural networks model, the picture to be identified corresponding at least two is obtained Scene;
The region suggestion fisrt feature figure corresponding with the picture to be identified of the picture to be identified is obtained, according to described Region is advised and the picture to be identified, obtains the classification score of each target in the picture to be identified;Wherein, the region At least part of region for indicating to include at least one target is proposed to be used in, the region suggestion advises network to second for region Characteristic pattern progress handles what is obtained, and the fisrt feature figure and the second feature figure are to be treated to described in Fast R-CNN networks Identification picture carries out what process of convolution was obtained;
According to corresponding at least two scene of picture to be identified and the classification score of each target, described treat is obtained Recognize the corresponding scene of picture.
Alternatively, scene Recognition carried out to each sample according to convolutional network model described, obtains described to be identified Before corresponding at least two scene of picture, in addition to:
Training picture and the corresponding label of the training picture are obtained, the label is used to indicate the training picture pair The scene answered;
According to the convolutional neural networks model and the corresponding label of the training picture, each scene is obtained corresponding The corresponding network parameter of the convolutional neural networks model;
It is described that scene Recognition is carried out to each sample according to convolutional network model, obtain the picture to be identified corresponding At least two scenes, including:
According to the corresponding network parameter of the corresponding convolutional neural networks model of each scene, each sample is entered Row scene Recognition, obtains corresponding at least two scene of the picture to be identified.
Alternatively, it is described according to the convolutional neural networks model and the corresponding label of the training picture, obtain each institute The corresponding network parameter of the corresponding convolutional neural networks model of scene is stated, including:
Division Sampling, the training picture after being expanded are carried out to the training picture;
According to the first default training parameter, default processing is carried out to the training picture after the amplification, predetermined number is obtained Third feature figure, it is described it is default processing include convolution, Chi Hua, normalized;
Full connection processing is carried out repeatedly to the third feature figure of the predetermined number, the corresponding field of the training picture is obtained Scape probability;
According to the scene probability and the corresponding label of the training picture, the described first default training parameter is adjusted It is whole, obtain the corresponding network parameter of the corresponding convolutional neural networks model of the scene.
Alternatively, the region suggestion for obtaining the picture to be identified fisrt feature corresponding with the picture to be identified Figure, including:
Process of convolution is carried out to the picture to be identified by the Fast R-CNN networks, shared convolutional layer is obtained;
The second feature figure is extracted from the shared convolutional layer, advises that the corresponding network of network is joined according to the region Number, advises that network carries out region suggestion processing to the second feature figure by the region, obtains the region of each target area Score, according to the area score, obtains the region suggestion;
Using the convolutional layer of the shared default number of plies of convolutional layer superposition as peculiar convolutional layer, the fisrt feature is obtained Figure, the number for the convolutional layer that the peculiar convolutional layer is included is more than the number for the convolutional layer that the shared convolutional layer is included.
Alternatively, it is described according to region suggestion and the picture to be identified, obtain each in the picture to be identified The classification score of target, including:
Advised according to the region, zone marker processing is carried out to the fisrt feature figure, obtained at the zone marker Fisrt feature figure after reason;
Pond processing is carried out to the fisrt feature figure after zone marker processing by the Fast R-CNN networks, obtained Fisrt feature figure to after pondization processing;
Full connection processing is carried out to the fisrt feature figure after pondization processing;
According to the corresponding network parameter of the Fast R-CNN networks, point of each target in the picture to be identified is obtained Class score.
Alternatively, in the region suggestion for obtaining the picture to be identified and corresponding first spy of the picture to be identified Before levying figure, in addition to:
Training picture and the corresponding target area of the training picture are obtained, the target area is complete for indicating Position of the target in the training picture;
According to the Fast R-CNN networks, region suggestion network and the corresponding target area of the training picture Domain, obtains the corresponding network parameter of the Fast R-CNN networks and the corresponding net of region suggestion network of each target Network parameter.
Alternatively, it is described that network and the training picture pair are advised according to the Fast R-CNN networks, the region The target area answered, obtains the corresponding network parameter of the Fast R-CNN networks and region suggestion net of each target The corresponding network parameter of network, including:
Process of convolution is carried out to the training picture by the Fast R-CNN networks, shared convolutional layer is obtained;
The second feature figure is extracted from the shared convolutional layer, according to the second default training parameter, passes through the region It is recommended that network carries out region suggestion processing to the second feature figure, the area score is obtained, according to the area score, is obtained To region suggestion;
Using the convolutional layer of the shared default number of plies of convolutional layer superposition as peculiar convolutional layer, the fisrt feature is obtained Figure, the number for the convolutional layer that the peculiar convolutional layer is included is more than the number for the convolutional layer that the shared convolutional layer is included;
Advised according to the region, zone marker processing is carried out to the fisrt feature figure, obtained at the zone marker Fisrt feature figure after reason;
Fisrt feature figure after being gone out by the Fast R-CNN networks to the zone marker carries out pond processing, obtains Fisrt feature figure after pondization processing;
Full connection processing is carried out to the fisrt feature figure after pondization processing, according to the 3rd default training parameter, obtained The classification score of each target in the training picture;
The classification score of each target in region suggestion, the training picture and the target area, are adjusted Whole described second default training parameter and the 3rd default training parameter, obtain the corresponding network ginseng of the region suggestion network Number network parameter corresponding with the Fast R-CNN networks of each target.
The scene recognition method of target detection that the present invention is provided auxiliary, by according to convolutional neural networks model to respectively treating The sample that identification picture sampling is obtained carries out scene Recognition, obtains corresponding at least two scene of picture to be identified.Again by treating Recognize that fisrt feature figure corresponding with picture to be identified is advised in the region of picture, obtain the classification of each target in picture to be identified Score completes target detection process.Finally by corresponding at least two scene of picture to be identified and the classification score of each target, Obtain the corresponding scene of picture to be identified.The present invention advises the object detection method of network by Fast R-CNN networks and region Aid in scene recognition method so that the accuracy rate of scene recognition method is improved.Wherein, region advises network to Fast R-CNN nets Network provides the region for having target area and set up, and greatly reduces the time that Fast R-CNN networks carry out target detection, improves Fast R-CNN networks carry out the speed of target detection.
Brief description of the drawings
The flow chart one of the scene recognition method for the target detection auxiliary that Fig. 1 provides for the present invention;
The flowchart 2 of the scene recognition method for the target detection auxiliary that Fig. 2 provides for the present invention;
The flow chart 3 of the scene recognition method for the target detection auxiliary that Fig. 3 provides for the present invention;
The flow chart four of the scene recognition method for the target detection auxiliary that Fig. 4 provides for the present invention;
The training process of Alexnet models is shown in the scene recognition method for the target detection auxiliary that Fig. 5 provides for the present invention It is intended to;
The flow chart five of the scene recognition method for the target detection auxiliary that Fig. 6 provides for the present invention;
Fast R-CNN networks and region suggestion in the scene recognition method for the target detection auxiliary that Fig. 7 provides for the present invention The training process schematic diagram of network.
Embodiment
The flow chart one of the scene recognition method for the target detection auxiliary that Fig. 1 provides for the present invention, Fig. 2 provides for the present invention Target detection auxiliary scene recognition method flowchart 2, Fig. 3 for the present invention provide target detection auxiliary scene knowledge The flow chart 3 of other method, as shown in figure 1, the scene recognition method of the target detection auxiliary of the present embodiment includes:
Step 101, acquisition picture to be identified, sample to the picture to be identified, obtain predetermined number and default big Small sample, scene Recognition is carried out according to convolutional neural networks model to each sample, obtains the picture correspondence to be identified At least two scenes.
Specifically, the size of picture to be identified can be according to the network of specifically chosen convolutional neural networks model in the present embodiment Selected, the quantity of picture to be identified can be entered according in convolutional neural networks model training learning process to different quality data cases Row adjustment, the present embodiment is not specifically limited to this.And convolutional neural networks model can select Alexnet, VGG16, VGG19, The models such as ResNet, the present embodiment is not also limited this.
Further, method that can be to each picture to be identified by cutting out cutting sampling in the present embodiment, by one Picture to be identified expands many times, obtains the sample of predetermined number and default size, to prevent over-fitting, waits to know so as to improve every The accuracy of the scene Recognition of other picture.Predetermined number and default size are not limited in the present embodiment, figure to be identified is only needed Quantity after piece amplification disclosure satisfy that convolutional neural networks model is trained demand study.
For example, for convenience of explanation, the present embodiment is illustrated from the citing of Alexnet models.It can be treated respectively from one Identification picture upper left corresponding with the picture to be identified after this flip horizontal, upper right, lower-left, bottom right, centre position are cut Cut out sampling, then a picture to be identified has just acquired 10 samples, and this 10 sample standard deviations are input in Alexnet models Scene Recognition is carried out, the result of corresponding 10 scene Recognitions is obtained, the result of scene Recognition can be carried out by matrix here Represent, wherein the line number of matrix is a line, and columns is scene type, each matrix value is the decimal between 0 to 1, and each square Battle array value sum is 1.The scene knowledge averaged as the picture to be identified is added to the corresponding matrix of this 10 samples again Other probability matrix, takes the scene wherein corresponding to maximum probability value, the scene results of the final identification of picture as to be identified.Its Remaining picture to be identified carries out scene Recognition in the manner described above, obtains the result of corresponding scene Recognition.
Further, for the nondescript scene such as station and rally parade, according to convolutional neural networks to be identified The scene Recognition that picture is carried out will be unable to identification, and the result of obtained scene Recognition may swim the picture recognition at station into rally OK, by the picture recognition for parade of gathering into station, picture so to be identified is likely to corresponding at least two scene.
Step 102, the region suggestion for obtaining the picture to be identified fisrt feature figure corresponding with the picture to be identified, According to region suggestion and the picture to be identified, the classification score of each target in the picture to be identified is obtained.
Wherein, the region is proposed to be used at least part of region for indicating to include at least one target, and the region is built Discuss and what is obtained is handled to the progress of second feature figure for region suggestion network, the fisrt feature figure and the second feature figure are fast Fast version base advises target identification method (the Fast Region-based Convolutional Network of network in region Method, Fast R-CNN) what process of convolution was obtained is carried out to the picture to be identified in network.
Specifically, target detection is carried out to picture to be identified in the present embodiment, can be by the target identification in picture to be identified Out.
On the one hand, as shown in Fig. 2 obtaining the region suggestion of the picture to be identified and the figure to be identified in the present embodiment The corresponding fisrt feature figure of piece, including:
Step 201, process of convolution carried out to the picture to be identified by the Fast R-CNN networks, obtain shared volume Lamination.
Step 202, from the shared convolutional layer second feature figure is extracted, advise that network is corresponding according to the region Network parameter, advises that network carries out region suggestion processing to the second feature figure by the region, obtains each target area Area score, according to the area score, obtain region suggestion.
Specifically, picture to be identified is input in Fast R-CNN networks in the present embodiment and carries out process of convolution, obtained Shared convolutional layer, extracts one layer of corresponding second feature figure from shared convolutional layer, is input in region suggestion network and carries out area Domain suggestion processing, just can obtain area score.Wherein, region suggestion is processed as carrying out slide window processing to picture to be identified and carried out Multiple convolution or full connection processing, area score are to express the region with the presence or absence of target and the probability size of target.
Step 203, the convolutional layer that the shared convolutional layer is superimposed to the default number of plies obtain described the as peculiar convolutional layer One characteristic pattern, the number for the convolutional layer that the peculiar convolutional layer is included is more than for the convolutional layer that the shared convolutional layer is included Number.
Specifically, in order to not lose other information, Fast R-CNN networks also may proceed on the basis of shared convolutional layer Carry out process of convolution so that information is more complete, you can the convolutional layer that shared convolutional layer is superimposed to the default number of plies is used as peculiar volume Lamination, obtains fisrt feature figure.The default number of plies is not specifically limited in the present embodiment.For example, extracting the 5th from shared convolutional layer The corresponding second feature figure of layer sends region suggestion network to, using the 9th layer of corresponding fisrt feature figure of shared convolutional layer as Peculiar convolutional layer.
On the other hand, as shown in figure 3, according to region suggestion and the picture to be identified described in the present embodiment, obtaining The classification score of each target in the picture to be identified, including:
Step 301, according to the region advise, to the fisrt feature figure carry out zone marker processing, obtain the area Fisrt feature figure after field mark processing.
Specifically, the higher some regions of area score can be selected in the present embodiment, by these regions correspondence to be identified The particular location of picture is transferred to Fast R-CNN networks as region suggestion, and Fast R-CNN networks just can according to region suggestion Fisrt feature figure is subjected to zone marker processing, such Fast R-CNN networks need to only carry out the classification of each target to marked region Score process, the process for score of then classifying for aimless region without progress, so as to save to picture to be identified The target detection time.
Step 302, by the Fast R-CNN networks to the zone marker processing after fisrt feature figure carry out pond Change is handled, and obtains the fisrt feature figure after pondization processing.
Further, because the size of the fisrt feature figure after zone marker processing is inconsistent, then need by pond Change and handle that its is size normalised, obtain the fisrt feature figure after pondization processing.
Step 303, full connection processing is carried out to the fisrt feature figure after pondization processing.
Specifically, comprehensive figure is much opened because the fisrt feature figure after pondization processing has, is handled by full connection, can Its dimension is reduced, is easy to the operation of classification score.
Step 304, according to the corresponding network parameter of the Fast R-CNN networks, obtain each in the picture to be identified The classification score of target.
Specifically, score of classifying in the present embodiment can not only illustrate whether there is target in picture to be identified, moreover it is possible to illustrate There is the probability size of various targets in picture to be identified.And the corresponding network parameter of Fast R-CNN networks is Fast R-CNN nets The most optimized parameter that network training is obtained, accurate can must come out the target detection in picture to be identified.So by obtaining Whether the classification score for stating each target in picture to be identified just can determine whether out certain target in picture to be identified.
Herein it should be noted that the order of step 101 can be before step 102, the order of step 102 can be in step Before 101, the order of step 101 and step 102 can be carried out simultaneously.It is suitable to the priority of step 101 and step 102 in the present embodiment Sequence is not limited.
Step 103, the classification score according to corresponding at least two scene of picture to be identified and each target, are obtained To the corresponding scene of the picture to be identified.
Specifically, step 101 can come out corresponding at least two scene Recognition of picture to be identified, so become and reduce field The scope of scape selection, such as may be by station and the scene of the corresponding picture to be identified of the more scene of the such number of rally parade. Step 102 can detect the classification score of each target in picture to be identified, such as, and the scene for parade of gathering is most The presence of banner is had, banner can be regard as target.So by corresponding at least two scene of picture to be identified and each target Classification score combines, to identify the corresponding scene of picture to be identified.If for example, having banner in picture to be identified, being somebody's turn to do The corresponding scene of picture to be identified is just rally parade.If not having banner in picture to be identified, the picture to be identified is corresponding Scene is just station.
The present embodiment provide target detection auxiliary scene recognition method, by according to convolutional neural networks model to each The sample that picture sampling to be identified is obtained carries out scene Recognition, obtains corresponding at least two scene of picture to be identified.Pass through again Fisrt feature figure corresponding with picture to be identified is advised in the region of picture to be identified, obtains point of each target in picture to be identified Class score completes target detection process.Finally by classifying for corresponding at least two scene of picture to be identified and each target Point, obtain the corresponding scene of picture to be identified.Advise the target of network in the present embodiment by Fast R-CNN networks and region Detection method aids in scene recognition method so that the accuracy rate of scene recognition method is improved.Wherein, region advises network to Fast R-CNN networks provide have target area region set up, greatly reduce Fast R-CNN networks carry out target detection when Between, improve the speed that Fast R-CNN networks carry out target detection.
The flow chart four of the scene recognition method for the target detection auxiliary that Fig. 4 provides for the present invention, as shown in figure 4, in institute State and scene Recognition is carried out to each sample according to convolutional network model, obtain the picture to be identified corresponding at least two Before scape, the method for the present embodiment also includes:
Step 401, acquisition training picture and the corresponding label of the training picture, the label are used to indicate the instruction Practice the corresponding scene of picture.
Specifically, train picture can be by being obtained in database in the present embodiment, also can be by obtaining manually, the present embodiment This is not limited, need to only ensure the quantity of the corresponding training picture of each scene can reach a thousand sheets.If manually The training picture of acquisition, can carry out sample expansion flip horizontal and by way of cutting sampling to training figure.And if convolution god The size for training picture is required through network model, by way of cutting sampling picture can will be trained to become unification It can be made whether to need to cut out sentencing for sampling process according to the concrete model of convolutional neural networks model in size, the present embodiment It is disconnected.Meanwhile, also need to obtain the corresponding label of training picture in the present embodiment, its label is the corresponding scene class of training picture Not.
Step 402, according to the convolutional neural networks model and the corresponding label of the training picture, obtain each field The corresponding network parameter of the corresponding convolutional neural networks model of scape.
Specifically, training picture label corresponding with its is input in convolutional neural networks model, convolutional neural networks Model just can be trained the process of study.Because training picture label corresponding with training picture is all known, it can pass through The default training parameter of dynamic adjustment first so that convolutional neural networks model can learn to various types of other scene, and then will Different scenes make a distinction.Specific method is as follows:
Step 4021, to it is described training picture carry out Division Sampling, the training picture after being expanded.
Specifically, that a kind of training picture of scene type is become into several times uniform sizes by the process of Division Sampling is big Small, Same Scene training picture, has expanded the quantity of training picture, simple to operate, and is convolutional neural networks model training Study provides more references.
Step 4022, according to the first default training parameter, default processing is carried out to the training picture after the amplification, is obtained The third feature figure of predetermined number, the default processing includes convolution, Chi Hua, normalized.
Specifically, the random parameter for setting convolutional neural networks model is the first default training parameter, to the instruction after amplification Practice picture and carry out convolution, Chi Hua, after normalized, just can obtain predetermined number, can clearly assertiveness training picture for field The third feature figure of scape.
Step 4023, the third feature figure to the predetermined number carry out repeatedly full connection and handled, and obtain the training figure The corresponding scene probability of piece.
Specifically, comprehensive figure is much opened because the third feature figure after default processing has, by multiple full junction Reason, just can arrive the corresponding scene probability of training picture.Wherein, repeatedly full connection processing can reduce the dimension of third feature figure, no Multi information can be lost, classifying quality is effectively ensured.
Step 4024, according to the scene probability and the corresponding label of the training picture, to the described first default training Parameter is adjusted, and obtains the corresponding network parameter of the corresponding convolutional neural networks model of the scene.
Specifically, whether consistent by contrasting scene probability label corresponding with training picture, dynamic adjustment first is preset Training parameter so that scene probability label corresponding with training picture is consistent, just can make the first default training parameter now For the corresponding network parameter of the corresponding convolutional neural networks model of scene.
Step 403, according to the corresponding network parameter of the corresponding convolutional neural networks model of each scene, to each institute State sample and carry out scene Recognition, obtain corresponding at least two scene of the picture to be identified.
Specifically, by convolutional Neural net that the corresponding network parameter of the corresponding convolutional neural networks model of scene is set up Network model can preferably identify scene sample corresponding to picture to be identified carries out scene Recognition, just can be treated Recognize corresponding at least two scene of picture.
The training process of Alexnet models is shown in the scene recognition method for the target detection auxiliary that Fig. 5 provides for the present invention It is intended to, in a specific embodiment, it is necessary to first learn to Alexnet model trainings by taking Alexnet models as an example, then enters Row carries out scene Recognition to picture to be identified, as shown in Figure 5.
First, the training picture of the 256*256 sizes of triple channel is input in Alexnet models, Alexnet model energy The label of enough detection training pictures, and the picture of the 256*256 sizes of triple channel is obtained after Division Sampling, by convolution meter Calculate and activate processing, wherein activation processing has been sampled activation primitive, such as amendment linear unit (Rectified linear unit, RELU, ReLU).Wherein, ReLU mathematic(al) representation is:(0, x), wherein x represents input signal to f (x)=max, and f (x) is represented Output signal.When input signal is less than 0, output signal is 0;If input signal is equal to or more than 0, output signal is equal to defeated Enter signal.Understand, the receipts of stochastic gradient descent (Stochastic gradient descent, SGD) method are obtained using ReLU Holding back speed can be more many soon than other functions (the activation primitive sigmoid/tanh in such as conventional method).Because ReLU is linear , ReLU only needs to a threshold value and can be obtained by activation value, and a lot of complicated computings are calculated without spending.Therefore, it is possible to excellent Change convolution process, so as to obtain the characteristic pattern of 96 55*55 sizes, then carry out pond and normalized, obtain 96 27*27 The characteristic pattern of size, again passes by convolutional calculation and activates processing, obtain the characteristic pattern of 256 27*27 sizes, and pond is carried out again And normalized, the characteristic pattern of 256 13*13 sizes is obtained, pond and normalized are carried out again, 384 13*13 are obtained The characteristic pattern of size, passes through convolutional calculation and activates processing, obtain the characteristic pattern of 256 13*13 sizes, pond Hua Chu is carried out again again Reason, obtains the characteristic pattern of 256 6*6 sizes, then carries out three full connections and handles line activating processing of going forward side by side, just can be trained The result of picture correspondence scene, that is, train the corresponding scene probability of picture.Further according to the label of training picture, adjust above-mentioned whole Process, that is, change the first default training parameter of Alexnet models, until the corresponding scene probability of training picture is schemed with training The label of piece is consistent, and institute's Alexnet model trainings converge to certain interval, and penalty values are controlled within the acceptable range, Now deconditioning.It regard the first default training parameter now as the corresponding network parameter of the corresponding Alexnet models of scene. So, the overall process of Alexnet model trainings study is just completed.
Then, when carrying out scene Recognition to one picture to be identified of Alexnet mode inputs, because above-mentioned training learns Process the corresponding network parameter of the corresponding Alexnet models of the scene of Alexnet models is determined, just can be to be identified by this The corresponding scene Recognition of picture comes out.
The flow chart five of the scene recognition method for the target detection auxiliary that Fig. 6 provides for the present invention, as shown in fig. 6, in institute Before stating the region suggestion fisrt feature figure corresponding with the picture to be identified for obtaining the picture to be identified, the present embodiment Method also includes:
Step 601, acquisition training picture and the corresponding target area of the training picture, the target area is used to refer to Show position of the complete target in the training picture.
Specifically, train picture can be by being obtained in database in the present embodiment, also can be by obtaining manually, the present embodiment This is not limited.Because the size of the input picture of Fast R-CNN networks is not limited, then without being carried out to training picture Split sampling processing.Meanwhile, also need to obtain the corresponding target area of training picture in the present embodiment, its target area is used to refer to Show physical location of the complete target in training picture.
Step 602, according to the Fast R-CNN networks, the region advise network and the training picture it is corresponding Target area, obtains the corresponding network parameter of the Fast R-CNN networks and region suggestion network pair of each target The network parameter answered.
Specifically, training picture target area corresponding with its is input in Fast R-CNN networks, Fast R-CNN Network and region suggestion network just can be trained the process of study.Due to training picture target area corresponding with training picture All it is known, the default training parameter of dynamic adjustment second can be passed through so that region suggestion network can learn to various targets Whether exist in training picture, obtain the region suggestion of target presence, send region suggestion to Fast R-CNN networks, can Reduce Fast R-CNN networks carry out target detection it is real-time between, improve efficiency.Meanwhile, it can also be preset by dynamic adjustment the 3rd Training parameter so that Fast R-CNN networks can learn all kinds of targets into various training pictures, and then will train in picture All kinds of targets make a distinction.Specific method is as follows:
Step 6021, by the Fast R-CNN networks to it is described training picture carry out process of convolution, obtain shared volume Lamination.
Specifically, training picture is input in Fast R-CNN networks in the present embodiment and carries out process of convolution, be total to Enjoy convolutional layer.
Step 6022, from the shared convolutional layer second feature figure is extracted, according to the second default training parameter, passed through The region suggestion network carries out region suggestion processing to the second feature figure, the area score is obtained, according to the area Domain score, obtains the region suggestion.
Specifically, one layer of corresponding second feature figure is extracted from shared convolutional layer, is input in region suggestion network The suggestion processing of row region, just can obtain area score.Wherein, region suggestion is processed as going forward side by side to training picture progress slide window processing Row multiple convolution or full connection processing.Area score is to express the region with the presence or absence of target and the probability size of target.
Step 6023, the convolutional layer that the shared convolutional layer is superimposed into the default number of plies obtain described as peculiar convolutional layer Fisrt feature figure, the number for the convolutional layer that the peculiar convolutional layer is included is more than for the convolutional layer that the shared convolutional layer is included Number.
Specifically, in order to not lose other information, Fast R-CNN networks also may proceed on the basis of shared convolutional layer Carry out process of convolution so that information is more complete, you can the convolutional layer that shared convolutional layer is superimposed to the default number of plies is used as peculiar volume Lamination, obtains fisrt feature figure.The default number of plies is not specifically limited in the present embodiment.For example, extracting the 5th from shared convolutional layer The corresponding second feature figure of layer sends region suggestion network to, using the 9th layer of corresponding fisrt feature figure of shared convolutional layer as Peculiar convolutional layer.
Step 6024, according to the region advise, to the fisrt feature figure carry out zone marker processing, obtain the area Fisrt feature figure after field mark processing.
Specifically, the higher some regions of area score can be selected in the present embodiment, these regions correspondence is schemed in training The particular location of piece is transferred to Fast R-CNN networks as region suggestion, and Fast R-CNN networks just can be by according to region suggestion Fisrt feature figure carries out zone marker processing so that Fast R-CNN networks only need to carry out classifying for each target to marked region Divide process, for aimless region then without carrying out the process of classification score, so as to save the target to training picture Detection time.
Step 6025, the zone marker is gone out by the Fast R-CNN networks after fisrt feature figure carry out pond Processing, obtains the fisrt feature figure after pondization processing.
Specifically, because the size of the fisrt feature figure after zone marker processing is inconsistent, then need by pond Processing is size normalised by its, obtains the fisrt feature figure after pondization processing.
Step 6026, full connection processing is carried out to the fisrt feature figure after pondization processing, according to the 3rd default training Parameter, obtains the classification score of each target in the training picture.
Specifically, comprehensive figure is much opened because the fisrt feature figure after pondization processing has, is handled by full connection, can Its dimension is reduced, is easy to the operation of classification score.
Step 6027, according to the region advise, it is described training picture in each target classification score and the mesh Region is marked, the described second default training parameter and the 3rd default training parameter is adjusted, obtains the region suggestion network pair The network parameter and the corresponding network parameter of the Fast R-CNN networks of each target answered.
Specifically, because target area is known, the 3rd default training parameter just can be dynamically adjusted according to target area, So that target detection is according to accurate, so as to preset training parameter as the corresponding net of Fast R-CNN networks of each target using the 3rd Network parameter.Meanwhile, can also be further according to region suggestion and target area, the default training parameter of dynamic adjustment second so that region is obtained Divide corresponding region suggestion more accurate, become the time that Fast R-CNN networks carry out target detection to training picture that reduces, So as to which the second default training parameter is advised into the corresponding network parameter of network as region.
Fast R-CNN networks and RPN networks in the scene recognition method for the target detection auxiliary that Fig. 7 provides for the present invention Training process schematic diagram, it is necessary to first advise network to Fast R-CNN networks and region in a specific embodiment (Region Proposal Networks, RPN) is trained study to training picture together, then carries out the field of picture to be identified Scape identification process, as shown in Figure 7.
First, the training picture of arbitrary size is input in Fast R-CNN networks, obtains the corresponding mesh of training picture Mark region.Process of convolution is carried out to training picture by Fast R-CNN networks, shared convolutional layer is obtained.RPN networks are from shared Convolutional layer extracts second feature figure, slide window processing is carried out to second feature figure, then carry out process of convolution twice or full connection processing Afterwards, i.e., the pixel in each sliding window forms 9 different size of regions, is compared respectively with target area, according to Two default training parameters, obtain corresponding area score, 300 regions for choosing highest scoring are advised as region, are sent to In Fast R-CNN networks.Shared convolutional layer is superimposed to the convolutional layer of some numbers of plies again as peculiar convolutional layer, the is just can obtain One characteristic pattern.It can be advised in the present embodiment according to region, zone marker processing is carried out to fisrt feature figure, obtained at zone marker Fisrt feature after reason, then fisrt feature figure after being gone out by Fast R-CNN networks to zone marker carry out pond processing, obtain Fisrt feature figure to after pondization processing.Full connection processing is carried out to the fisrt feature figure after pondization processing again, it is pre- according to the 3rd If training parameter, the classification score of each target in the training picture is obtained.Now, the 3rd default training parameter and mesh is judged Whether consistent mark region.If inconsistent, the default training parameter of dynamic adjustment the 3rd, while also according to region suggestion and target area Domain, the default training parameter of adjustment second, until the 3rd default training parameter is consistent with target area, and Fast R-CNN networks with RPN network trainings converge to certain interval, and penalty values are controlled within the acceptable range, now deconditioning.By now Second default training parameter presets training parameter as Fast as the corresponding network parameter of RPN networks, and by the 3rd now The corresponding network parameter of R-CNN networks.So, the overall process of Fast R-CNN networks and the study of RPN network trainings is just completed.
Then, when carrying out target detection to one picture to be identified of Fast R-CNN network inputs, due to above-mentioned training The process of study determines the corresponding network parameter of the corresponding Fast R-CNN networks of the target of Fast R-CNN networks, RPN nets The corresponding network parameter of the corresponding RPN networks of target of network also determines, just can be by the corresponding target identification of this picture to be identified Out, and the region suggestions of RPN networks can also be saved also save the detection times of Fast R-CNN networks.
One of ordinary skill in the art will appreciate that:Realizing all or part of step of above-mentioned each method embodiment can lead to The related hardware of programmed instruction is crossed to complete.Foregoing program can be stored in a computer read/write memory medium.The journey Sequence upon execution, performs the step of including above-mentioned each method embodiment;And foregoing storage medium includes:ROM, RAM, magnetic disc or Person's CD etc. is various can be with the medium of store program codes.
Finally it should be noted that:Various embodiments above is merely illustrative of the technical solution of the present invention, rather than its limitations;To the greatest extent The present invention is described in detail with reference to foregoing embodiments for pipe, it will be understood by those within the art that:Its according to The technical scheme described in foregoing embodiments can so be modified, or which part or all technical characteristic are entered Row equivalent substitution;And these modifications or replacement, the essence of appropriate technical solution is departed from various embodiments of the present invention technology The scope of scheme.

Claims (7)

1. a kind of scene recognition method of target detection auxiliary, it is characterised in that including:
Picture to be identified is obtained, the picture to be identified is sampled, the sample of predetermined number and default size is obtained, according to Convolutional neural networks model carries out scene Recognition to each sample, obtains the picture to be identified corresponding at least two Scape;
The region suggestion fisrt feature figure corresponding with the picture to be identified of the picture to be identified is obtained, according to the region It is recommended that and the picture to be identified, obtain the classification score of each target in the picture to be identified;Wherein, the region suggestion For at least part of region for indicating to include at least one target, the region suggestion advises network to second feature for region Figure progress handles what is obtained, and the fisrt feature figure and the second feature figure are to described to be identified in Fast R-CNN networks Picture carries out what process of convolution was obtained;
According to corresponding at least two scene of picture to be identified and the classification score of each target, obtain described to be identified The corresponding scene of picture.
2. according to the method described in claim 1, it is characterised in that each sample is entered according to convolutional network model described Row scene Recognition, before obtaining corresponding at least two scene of the picture to be identified, in addition to:
Training picture and the corresponding label of the training picture are obtained, the label is used to indicate that the training picture is corresponding Scene;
According to the convolutional neural networks model and the corresponding label of the training picture, each scene is obtained corresponding described The corresponding network parameter of convolutional neural networks model;
It is described that scene Recognition is carried out to each sample according to convolutional network model, obtain the picture to be identified corresponding at least Two scenes, including:
According to the corresponding network parameter of the corresponding convolutional neural networks model of each scene, field is carried out to each sample Scape is recognized, obtains corresponding at least two scene of the picture to be identified.
3. method according to claim 2, it is characterised in that described according to the convolutional neural networks model and the instruction Practice the corresponding label of picture, obtain the corresponding network parameter of the corresponding convolutional neural networks model of each scene, including:
Division Sampling, the training picture after being expanded are carried out to the training picture;
According to the first default training parameter, default processing is carried out to the training picture after the amplification, the of predetermined number is obtained Three characteristic patterns, the default processing includes convolution, Chi Hua, normalized;
Full connection processing is carried out repeatedly to the third feature figure of the predetermined number, the corresponding scene of the training picture is obtained general Rate;
According to the scene probability and the corresponding label of the training picture, the described first default training parameter is adjusted, Obtain the corresponding network parameter of the corresponding convolutional neural networks model of the scene.
4. according to the method described in claim 1, it is characterised in that the region suggestion for obtaining the picture to be identified and institute The corresponding fisrt feature figure of picture to be identified is stated, including:
Process of convolution is carried out to the picture to be identified by the Fast R-CNN networks, shared convolutional layer is obtained;
The second feature figure is extracted from the shared convolutional layer, the corresponding network parameter of network is advised according to the region, led to Cross the region suggestion network and region suggestion processing carried out to the second feature figure, obtain the area score of each target area, According to the area score, the region suggestion is obtained;
Using the convolutional layer of the shared default number of plies of convolutional layer superposition as peculiar convolutional layer, the fisrt feature figure, institute are obtained The number for stating the convolutional layer that peculiar convolutional layer is included is more than the number for the convolutional layer that the shared convolutional layer is included.
5. method according to claim 4, it is characterised in that described according to region suggestion and the figure to be identified Piece, obtains the classification score of each target in the picture to be identified, including:
Advised according to the region, zone marker processing is carried out to the fisrt feature figure, obtained after the zone marker processing Fisrt feature figure;
Pond processing is carried out to the fisrt feature figure after zone marker processing by the Fast R-CNN networks, pond is obtained Fisrt feature figure after change processing;
Full connection processing is carried out to the fisrt feature figure after pondization processing;
According to the corresponding network parameter of the Fast R-CNN networks, classifying for each target in the picture to be identified is obtained Point.
6. method according to claim 5, it is characterised in that in the region suggestion for obtaining the picture to be identified and Before the corresponding fisrt feature figure of the picture to be identified, in addition to:
Training picture and the corresponding target area of the training picture are obtained, the target area is used to indicate complete target Position in the training picture;
According to the Fast R-CNN networks, region suggestion network and the corresponding target area of the training picture, obtain Take the corresponding network parameter of the Fast R-CNN networks and the corresponding network ginseng of region suggestion network of each target Number.
7. method according to claim 6, it is characterised in that described according to the Fast R-CNN networks, the region It is recommended that network and the corresponding target area of the training picture, obtain the Fast R-CNN networks correspondence of each target Network parameter and the corresponding network parameter of region suggestion network, including:
Process of convolution is carried out to the training picture by the Fast R-CNN networks, shared convolutional layer is obtained;
The second feature figure is extracted from the shared convolutional layer, according to the second default training parameter, is advised by the region Network carries out region suggestion processing to the second feature figure, obtains the area score, according to the area score, obtains institute State region suggestion;
Using the convolutional layer of the shared default number of plies of convolutional layer superposition as peculiar convolutional layer, the fisrt feature figure, institute are obtained The number for stating the convolutional layer that peculiar convolutional layer is included is more than the number for the convolutional layer that the shared convolutional layer is included;
Advised according to the region, zone marker processing is carried out to the fisrt feature figure, obtained after the zone marker processing Fisrt feature figure;
Fisrt feature figure after being gone out by the Fast R-CNN networks to the zone marker carries out pond processing, obtains pond Fisrt feature figure after processing;
Full connection processing is carried out to the fisrt feature figure after pondization processing, according to the 3rd default training parameter, obtains described Train the classification score of each target in picture;
The classification score of each target in region suggestion, the training picture and the target area, adjust institute State the second default training parameter and the 3rd default training parameter, obtain the corresponding network parameter of the region suggestion network and The corresponding network parameter of the Fast R-CNN networks of each target.
CN201710270013.4A 2017-04-24 2017-04-24 Target detection assisted scene identification method Active CN107194318B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710270013.4A CN107194318B (en) 2017-04-24 2017-04-24 Target detection assisted scene identification method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710270013.4A CN107194318B (en) 2017-04-24 2017-04-24 Target detection assisted scene identification method

Publications (2)

Publication Number Publication Date
CN107194318A true CN107194318A (en) 2017-09-22
CN107194318B CN107194318B (en) 2020-06-12

Family

ID=59872812

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710270013.4A Active CN107194318B (en) 2017-04-24 2017-04-24 Target detection assisted scene identification method

Country Status (1)

Country Link
CN (1) CN107194318B (en)

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107563357A (en) * 2017-09-29 2018-01-09 北京奇虎科技有限公司 Live dress ornament based on scene cut, which is dressed up, recommends method, apparatus and computing device
CN107610146A (en) * 2017-09-29 2018-01-19 北京奇虎科技有限公司 Image scene segmentation method, apparatus, computing device and computer-readable storage medium
CN107622498A (en) * 2017-09-29 2018-01-23 北京奇虎科技有限公司 Image penetration management method, apparatus and computing device based on scene cut
CN107730514A (en) * 2017-09-29 2018-02-23 北京奇虎科技有限公司 Scene cut network training method, device, computing device and storage medium
CN107808138A (en) * 2017-10-31 2018-03-16 电子科技大学 A kind of communication signal recognition method based on FasterR CNN
CN107832795A (en) * 2017-11-14 2018-03-23 深圳码隆科技有限公司 Item identification method, system and electronic equipment
CN107844977A (en) * 2017-10-09 2018-03-27 中国银联股份有限公司 A kind of method of payment and device
CN108681752A (en) * 2018-05-28 2018-10-19 电子科技大学 A kind of image scene mask method based on deep learning
CN108734162A (en) * 2018-04-12 2018-11-02 上海扩博智能技术有限公司 Target identification method, system, equipment and storage medium in commodity image
CN108765033A (en) * 2018-06-08 2018-11-06 Oppo广东移动通信有限公司 Transmitting advertisement information method and apparatus, storage medium, electronic equipment
CN108764235A (en) * 2018-05-23 2018-11-06 中国民用航空总局第二研究所 Neural network model, object detection method, equipment and medium
CN108960209A (en) * 2018-08-09 2018-12-07 腾讯科技(深圳)有限公司 Personal identification method, device and computer readable storage medium
CN109086742A (en) * 2018-08-27 2018-12-25 Oppo广东移动通信有限公司 scene recognition method, scene recognition device and mobile terminal
CN109325491A (en) * 2018-08-16 2019-02-12 腾讯科技(深圳)有限公司 Identification code recognition methods, device, computer equipment and storage medium
CN109614876A (en) * 2018-11-16 2019-04-12 北京市商汤科技开发有限公司 Critical point detection method and device, electronic equipment and storage medium
CN109688351A (en) * 2017-10-13 2019-04-26 华为技术有限公司 A kind of image-signal processing method, device and equipment
CN109727268A (en) * 2018-12-29 2019-05-07 西安天和防务技术股份有限公司 Method for tracking target, device, computer equipment and storage medium
CN109784131A (en) * 2017-11-15 2019-05-21 深圳光启合众科技有限公司 Method for checking object, device, storage medium and processor
CN109981695A (en) * 2017-12-27 2019-07-05 广东欧珀移动通信有限公司 Content delivery method, device and equipment
CN110012210A (en) * 2018-01-05 2019-07-12 广东欧珀移动通信有限公司 Photographic method, device, storage medium and electronic equipment
CN110390262A (en) * 2019-06-14 2019-10-29 平安科技(深圳)有限公司 Video analysis method, apparatus, server and storage medium
CN111062441A (en) * 2019-12-18 2020-04-24 武汉大学 Scene classification method and device based on self-supervision mechanism and regional suggestion network
CN111104942A (en) * 2019-12-09 2020-05-05 熵智科技(深圳)有限公司 Template matching network training method, template matching network recognition method and template matching network recognition device
CN111178367A (en) * 2018-11-09 2020-05-19 财团法人资讯工业策进会 Feature determination device and method for adapting to multiple object sizes
CN111383246A (en) * 2018-12-29 2020-07-07 杭州海康威视数字技术股份有限公司 Scroll detection method, device and equipment
CN113569734A (en) * 2021-07-28 2021-10-29 山东力聚机器人科技股份有限公司 Image identification and classification method and device based on feature recalibration
CN112633064B (en) * 2020-11-19 2023-12-15 深圳银星智能集团股份有限公司 Scene recognition method and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103778443A (en) * 2014-02-20 2014-05-07 公安部第三研究所 Method for achieving scene analysis description based on theme model method and field rule library
CN106504233A (en) * 2016-10-18 2017-03-15 国网山东省电力公司电力科学研究院 Image electric power widget recognition methodss and system are patrolled and examined based on the unmanned plane of Faster R CNN
US20170098162A1 (en) * 2015-10-06 2017-04-06 Evolv Technologies, Inc. Framework for Augmented Machine Decision Making

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103778443A (en) * 2014-02-20 2014-05-07 公安部第三研究所 Method for achieving scene analysis description based on theme model method and field rule library
US20170098162A1 (en) * 2015-10-06 2017-04-06 Evolv Technologies, Inc. Framework for Augmented Machine Decision Making
CN106504233A (en) * 2016-10-18 2017-03-15 国网山东省电力公司电力科学研究院 Image electric power widget recognition methodss and system are patrolled and examined based on the unmanned plane of Faster R CNN

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
SHAOQING REN ETC.: ""Faster R-CNN:Towards Real-Time Object Detection with Region Proposal Networks"", 《ARXIV》 *

Cited By (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107610146B (en) * 2017-09-29 2021-02-23 北京奇虎科技有限公司 Image scene segmentation method and device, electronic equipment and computer storage medium
CN107610146A (en) * 2017-09-29 2018-01-19 北京奇虎科技有限公司 Image scene segmentation method, apparatus, computing device and computer-readable storage medium
CN107622498A (en) * 2017-09-29 2018-01-23 北京奇虎科技有限公司 Image penetration management method, apparatus and computing device based on scene cut
CN107730514A (en) * 2017-09-29 2018-02-23 北京奇虎科技有限公司 Scene cut network training method, device, computing device and storage medium
CN107563357A (en) * 2017-09-29 2018-01-09 北京奇虎科技有限公司 Live dress ornament based on scene cut, which is dressed up, recommends method, apparatus and computing device
CN107730514B (en) * 2017-09-29 2021-02-12 北京奇宝科技有限公司 Scene segmentation network training method and device, computing equipment and storage medium
CN107563357B (en) * 2017-09-29 2021-06-04 北京奇虎科技有限公司 Live-broadcast clothing dressing recommendation method and device based on scene segmentation and computing equipment
CN107622498B (en) * 2017-09-29 2021-06-04 北京奇虎科技有限公司 Image crossing processing method and device based on scene segmentation and computing equipment
CN107844977A (en) * 2017-10-09 2018-03-27 中国银联股份有限公司 A kind of method of payment and device
CN109688351A (en) * 2017-10-13 2019-04-26 华为技术有限公司 A kind of image-signal processing method, device and equipment
US11430209B2 (en) 2017-10-13 2022-08-30 Huawei Technologies Co., Ltd. Image signal processing method, apparatus, and device
CN107808138B (en) * 2017-10-31 2021-03-30 电子科技大学 Communication signal identification method based on FasterR-CNN
CN107808138A (en) * 2017-10-31 2018-03-16 电子科技大学 A kind of communication signal recognition method based on FasterR CNN
CN107832795A (en) * 2017-11-14 2018-03-23 深圳码隆科技有限公司 Item identification method, system and electronic equipment
WO2019096180A1 (en) * 2017-11-14 2019-05-23 深圳码隆科技有限公司 Object recognition method and system, and electronic device
CN109784131B (en) * 2017-11-15 2023-08-22 深圳光启合众科技有限公司 Object detection method, device, storage medium and processor
CN109784131A (en) * 2017-11-15 2019-05-21 深圳光启合众科技有限公司 Method for checking object, device, storage medium and processor
CN109981695A (en) * 2017-12-27 2019-07-05 广东欧珀移动通信有限公司 Content delivery method, device and equipment
CN110012210B (en) * 2018-01-05 2020-09-22 Oppo广东移动通信有限公司 Photographing method and device, storage medium and electronic equipment
US11503205B2 (en) 2018-01-05 2022-11-15 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Photographing method and device, and related electronic apparatus
CN110012210A (en) * 2018-01-05 2019-07-12 广东欧珀移动通信有限公司 Photographic method, device, storage medium and electronic equipment
CN108734162A (en) * 2018-04-12 2018-11-02 上海扩博智能技术有限公司 Target identification method, system, equipment and storage medium in commodity image
CN108734162B (en) * 2018-04-12 2021-02-09 上海扩博智能技术有限公司 Method, system, equipment and storage medium for identifying target in commodity image
CN108764235B (en) * 2018-05-23 2021-06-29 中国民用航空总局第二研究所 Target detection method, apparatus and medium
CN108764235A (en) * 2018-05-23 2018-11-06 中国民用航空总局第二研究所 Neural network model, object detection method, equipment and medium
CN108681752B (en) * 2018-05-28 2023-08-15 电子科技大学 Image scene labeling method based on deep learning
CN108681752A (en) * 2018-05-28 2018-10-19 电子科技大学 A kind of image scene mask method based on deep learning
CN108765033B (en) * 2018-06-08 2021-01-12 Oppo广东移动通信有限公司 Advertisement information pushing method and device, storage medium and electronic equipment
CN108765033A (en) * 2018-06-08 2018-11-06 Oppo广东移动通信有限公司 Transmitting advertisement information method and apparatus, storage medium, electronic equipment
CN108960209B (en) * 2018-08-09 2023-07-21 腾讯科技(深圳)有限公司 Identity recognition method, identity recognition device and computer readable storage medium
CN108960209A (en) * 2018-08-09 2018-12-07 腾讯科技(深圳)有限公司 Personal identification method, device and computer readable storage medium
US11494577B2 (en) 2018-08-16 2022-11-08 Tencent Technology (Shenzhen) Company Limited Method, apparatus, and storage medium for identifying identification code
CN109325491A (en) * 2018-08-16 2019-02-12 腾讯科技(深圳)有限公司 Identification code recognition methods, device, computer equipment and storage medium
CN109086742A (en) * 2018-08-27 2018-12-25 Oppo广东移动通信有限公司 scene recognition method, scene recognition device and mobile terminal
CN111178367A (en) * 2018-11-09 2020-05-19 财团法人资讯工业策进会 Feature determination device and method for adapting to multiple object sizes
CN111178367B (en) * 2018-11-09 2023-02-24 财团法人资讯工业策进会 Feature determination device and method for adapting to multiple object sizes
CN109614876B (en) * 2018-11-16 2021-07-27 北京市商汤科技开发有限公司 Key point detection method and device, electronic equipment and storage medium
CN109614876A (en) * 2018-11-16 2019-04-12 北京市商汤科技开发有限公司 Critical point detection method and device, electronic equipment and storage medium
CN111383246A (en) * 2018-12-29 2020-07-07 杭州海康威视数字技术股份有限公司 Scroll detection method, device and equipment
CN109727268A (en) * 2018-12-29 2019-05-07 西安天和防务技术股份有限公司 Method for tracking target, device, computer equipment and storage medium
CN111383246B (en) * 2018-12-29 2023-11-07 杭州海康威视数字技术股份有限公司 Scroll detection method, device and equipment
WO2020248386A1 (en) * 2019-06-14 2020-12-17 平安科技(深圳)有限公司 Video analysis method and apparatus, computer device and storage medium
CN110390262B (en) * 2019-06-14 2023-06-30 平安科技(深圳)有限公司 Video analysis method, device, server and storage medium
CN110390262A (en) * 2019-06-14 2019-10-29 平安科技(深圳)有限公司 Video analysis method, apparatus, server and storage medium
CN111104942A (en) * 2019-12-09 2020-05-05 熵智科技(深圳)有限公司 Template matching network training method, template matching network recognition method and template matching network recognition device
CN111104942B (en) * 2019-12-09 2023-11-03 熵智科技(深圳)有限公司 Template matching network training method, recognition method and device
CN111062441A (en) * 2019-12-18 2020-04-24 武汉大学 Scene classification method and device based on self-supervision mechanism and regional suggestion network
CN112633064B (en) * 2020-11-19 2023-12-15 深圳银星智能集团股份有限公司 Scene recognition method and electronic equipment
CN113569734A (en) * 2021-07-28 2021-10-29 山东力聚机器人科技股份有限公司 Image identification and classification method and device based on feature recalibration

Also Published As

Publication number Publication date
CN107194318B (en) 2020-06-12

Similar Documents

Publication Publication Date Title
CN107194318A (en) The scene recognition method of target detection auxiliary
CN109583322B (en) Face recognition deep network training method and system
CN104573688B (en) Mobile platform tobacco laser code intelligent identification Method and device based on deep learning
CN103971102B (en) Static gesture recognition method based on finger contour and decision-making trees
CN106203395B (en) Face attribute recognition method based on multitask deep learning
CN104217214B (en) RGB D personage's Activity recognition methods based on configurable convolutional neural networks
CN108510194A (en) Air control model training method, Risk Identification Method, device, equipment and medium
CN106980858A (en) The language text detection of a kind of language text detection with alignment system and the application system and localization method
CN105590099B (en) A kind of more people's Activity recognition methods based on improvement convolutional neural networks
CN108171112A (en) Vehicle identification and tracking based on convolutional neural networks
CN107844784A (en) Face identification method, device, computer equipment and readable storage medium storing program for executing
CN106446930A (en) Deep convolutional neural network-based robot working scene identification method
CN107016405A (en) A kind of insect image classification method based on classification prediction convolutional neural networks
CN106897738A (en) A kind of pedestrian detection method based on semi-supervised learning
CN107180226A (en) A kind of dynamic gesture identification method based on combination neural net
CN108073888A (en) A kind of teaching auxiliary and the teaching auxiliary system using this method
Blok et al. The effect of data augmentation and network simplification on the image‐based detection of broccoli heads with Mask R‐CNN
CN105144239A (en) Image processing device, program, and image processing method
CN109558902A (en) A kind of fast target detection method
CN106204779A (en) The check class attendance method learnt based on plurality of human faces data collection strategy and the degree of depth
CN107480575A (en) The training method of model, across age face identification method and corresponding device
CN110135231A (en) Animal face recognition methods, device, computer equipment and storage medium
CN106022273A (en) Handwritten form identification system of BP neural network based on dynamic sample selection strategy
CN107633242A (en) Training method, device, equipment and the storage medium of network model
CN109410211A (en) The dividing method and device of target object in a kind of image

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant