CN109685780A - A kind of Retail commodity recognition methods based on convolutional neural networks - Google Patents

A kind of Retail commodity recognition methods based on convolutional neural networks Download PDF

Info

Publication number
CN109685780A
CN109685780A CN201811541070.2A CN201811541070A CN109685780A CN 109685780 A CN109685780 A CN 109685780A CN 201811541070 A CN201811541070 A CN 201811541070A CN 109685780 A CN109685780 A CN 109685780A
Authority
CN
China
Prior art keywords
training
target
classification
neural networks
convolutional neural
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811541070.2A
Other languages
Chinese (zh)
Other versions
CN109685780B (en
Inventor
王敏
方仁渊
范晓烨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hohai University HHU
Original Assignee
Hohai University HHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hohai University HHU filed Critical Hohai University HHU
Priority to CN201811541070.2A priority Critical patent/CN109685780B/en
Publication of CN109685780A publication Critical patent/CN109685780A/en
Application granted granted Critical
Publication of CN109685780B publication Critical patent/CN109685780B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

The present invention discloses a kind of Retail commodity recognition methods based on convolutional neural networks, first using the yolo3 object detector of data set one customization of training of a general coarseness, then image to be detected is inputted, obtain a series of primary semantic objects, then combining primary semantic object according to series of rules is high-level semantics object, and the similarity comparison between the attribute of target and the attribute of each high-level semantics object that need to detect finally by judgement obtains required target.The invention enables the detectors obtained based on general coarseness data set training can also be used for completing fine grit classification task under certain condition;The data that target category is directly acquired compared to conventional method are trained, the threshold that the present invention can greatly reduce data acquisition cost and use in a production environment.

Description

A kind of Retail commodity recognition methods based on convolutional neural networks
Technical field
The invention belongs to target identification technologies, and in particular to a kind of Retail commodity identification side based on convolutional neural networks Method.
Background technique
In retail trade, traditional retail mode is to be known by either scans' bar code or two dimensional code to Retail commodity Not.In recent years, the application with deep learning method in every field deepens continuously, and artificial intelligence technology has been widely applied It is exactly the appearance of unmanned supermarket or convenience store in the embodiment of retail trade in daily life.Since there is no receive for it Human cost is greatly saved compared to traditional supermarket in silver-colored member, while compared to artificial cashier, sky shared by unmanned version Between it is smaller, so that storekeeper be allowed to arrange more checkout aisles in same space, directly enhance cash register efficiency.And unmanned receipts One primary clustering of silver-colored platform, commercial detector then rely on target detection technique.
As existing, basic and challenging problem long-term in a computer vision field, target detection Hot spot always is in the research of recent decades.The target of target detection is in given picture, from given classification (example Such as the mankind, automobile, bicycle, dog and cat) in tell and whether there is, there are several object instances.If target exists, return The spatial position of each object instance and size.Target detection is in artificial intelligence and information technology, including robot vision, disappears Take grade electronic product, security fields, automatic Pilot, human-computer interaction, the image retrieval of Context-dependent, intelligent video surveillance, There is very extensive application with numerous areas such as augmented realities.And in target detection technique, it is based on depth convolutional network (DCNN) object detector possesses relatively good effect.Object detector based on DCNN, which generally relies on, largely to be needed to detect The image data of target is trained, due to the limitation of data collection capability, it is necessary to be relied on and be carried out to target context information Explicit code.
In physical world, visual object is appeared in specific environment, and is usually coexisted with other related objects.In the heart There is very conclusive evidence to show that context plays a crucial role in human object's identification on pharmacological research.People are Through recognizing, especially when the barment tag of target because of too small target sizes, block or poor picture quality and show When insufficient, context modeling appropriate facilitates object detection and recognition.
At present the state-of-the-art technology of object detection field can not explicitly utilize any contextual information in the case where Detect target.Because DCNN learns layering from multiple abstraction hierarchies, it is considered that the utilization that DCNN can be implicit is up and down Literary information, but implicit context brings a problem: detector extremely relies on training set to the recognition capability of specific objective. We can not identify new target using existing training result, even if it possesses characteristic attribute similar with former target.Cause It is still valuable that this finds explicit contextual information in the detector based on DCNN.
Summary of the invention
Goal of the invention: it is an object of the invention to solve the deficiencies in the prior art, provides a kind of based on convolution mind Retail commodity recognition methods through network carries out model training based on coarseness data set disclosed on network, to complete needle To specific objective, fine-grained target detection.
Technical solution: a kind of Retail commodity recognition methods based on convolutional neural networks of the invention, first with general Coarseness data set trains a yolo3 object detector, then input picture obtains a series of class of coarseness targets thereto Then other and location information combines rudimentary semantic object according to the inclusion relation of predefined and obtains a series of high-level semantics pair As;The fine granularity target for needing to detect finally is chosen in these objects;
Specifically includes the following steps:
Step 1: according to the following suitable classification of Rules Filtering granularity from the category set of public data collection: 1. categories (such as box, bottle etc.) related to the appearance of targeted retailer product or packaging external appearance;2. the category and targeted retailer product Packing content is related;3. item name cannot include any brand message;Then by figure corresponding to these classifications chosen Piece forms training set and (suggests that classification based training concentrates each classification picture number at 1000 or more, classification number is 30 or more and figure Piece size is not less than 512*512 pixel);
Step 2: the training set training yolo3 object detector chosen using step 1 obtains one accurately generally Object detector;
Step 3: by object detector obtained in image to be detected input step 2, obtaining a series of rudimentary semantic objects Classification and their position;
Step 4: choosing suitable μab, determine inclusion relation:
For rudimentary semantic object A, B, if A ∩ B=A, orAnd ThenWherein S (A) refers to the area of the bounding box of rudimentary semantic object A, μaAnd μbFor some object bounding box simultaneously It is determined when not exclusively falling in the bounding box of another object between both with the presence or absence of inclusion relation;
Step 5: by rudimentary semanteme object set S={ t obtained in step 31,t2,…tnBy mapping f:Turn Change initial high-level semantics object set S '={ T1,T2…Tn};Wherein Ti={ ti};
Step 6: according to following compatible rule merging object:
For high-level semantics object Ta,Tb, T if it existsa,Tb,ti,tj, ti∈TaAnd tj∈Tb,Or cti≠ ctj, andThen merge Ta, TbFor their union, i.e. Tn+1=Ta∪Tb, from deletion T in S 'a, TbAnd add Tn+1;Wherein, high-level semantics object refers to the set of rudimentary semantic object;
The step is repeated, until being previously mentioned regular object pair in high-level semantics object set there is no satisfaction, i.e., Until S ' does not finally obtain S "={ T in variationo,To+1,…Tp};
Step 7: for the object that need to be detected, manually provide it includes rudimentary semantic object type, obtain in step 6 High-level semantics object in retrieve, if the category set manually provided is the subset of the category set of some high-level semantics object, Then it is exactly the target for needing to detect.
Further, in the step 1 public data collection refer to from network public data concentration filter out categorized data set With target detection data set, wherein categorized data set only needs target category information contained in image, and target detection data set is then Need two attributes of the classification of target and bounding box in image.
Further, the detailed step of the step 2 are as follows:
(2.1) it training characteristics extractor: is averaged pond on the feature extractor top of yolov3 plus an overall situation first Layer, two linear layers and one softmax layers, and then obtain a classifier;Then the classifier is trained, reads instruction Whether white silk integrates image and does center cutting as RGB image, inputs network after normalized, correct using cross entropy damage for exporting Function is lost to be measured;Continuous training terminates until accuracy to 90% or more;
(2.2) modify on the convolutional neural networks constructed: display removes the last overall situation and is averaged pond layer, two A linear layer and one softmax layers;Then the Feature Mapping for therefrom extracting 3 scales, passes through multiple convolutional layers and up-sampling Layer is connected, and finally exports the tensor of a 52*52* (c+5b), and wherein c is the classification number in data set, and b is that each unit is pre- The bounding box quantity (generally taking 2) of survey, and then constitute complete yolo3 network structure, the wherein each convolution of characteristic extraction part The initial weight of core uses gained target detection data set training in step (2.1) using weight obtained in pre-training step New network, the object detector finally used.
The utility model has the advantages that the present invention by the basis of existing object detector to the contextual information of complex target into The explicit modeling of row, to realize the detection to fine granularity target;So that being obtained using disclosed coarseness data set training network Detector can also complete the fine grit classification task under certain condition.
In conclusion the data that the present invention directly acquires target category are trained, data acquisition cost is greatly reduced, Also use threshold in a production environment is reduced.
Detailed description of the invention
The network structure that Fig. 1 is used when being training characteristics extractor;
The network structure that Fig. 2 is used when being training objective detector;
Fig. 3 is the sub-structure schematic diagram of each template in Fig. 1 and Fig. 2;
Fig. 4 is the identification process figure in embodiment 1.
Specific embodiment
Technical solution of the present invention is described in detail below, but protection scope of the present invention is not limited to the implementation Example.
As shown in Figure 1 to Figure 3, a kind of Retail commodity recognition methods based on convolutional neural networks of the invention, including under State step:
Step 1: the collection of data set, the data set being had disclosed from network (such as OpenIamge, ImageNet etc.) In filter out categorized data set and target detection data set, wherein categorized data set only need target category contained in image believe Breath, target detection data set then need two attributes of the classification of target and bounding box in image.It should be selected in the selection of classification (such as relative to " Coke bottle ", " Sprite bottle " etc., " beverage bottle " or " bottle " is obviously more to the lower classification of semantization degree Good selection) to obtain better sensing capability.This embodies a concentrated reflection of the WordNet ID of classification because to the greatest extent in ImageNet data May close to whole WordNet tree middle part (all categories of the data set and its between relationship be WordNet a son Figure) and as Open Image then can directly choose the classification of second level classification.
Step 2: training characteristics extractor is averaged pond on the feature extractor top of yolov3 plus an overall situation first Layer, two linear layers and one softmax layers obtain a classifier (structure of network is shown in Fig. 1 at this time), to the classifier into Row training, reading training set image are RGB image, and depth is 8, and zoomed image makes its shorter edge length 416, then does center It cuts, obtains 416 × 416 × 3 tensor, the value by value in [0,255] normalizes to [- 0.5,0.5] and inputs network again, right Correctly whether cross entropy loss function is used to be measured in output.Training method uses stochastic gradient descent (SGD), study speed Rate takes 0.01.Continuous training terminates until top1 accuracy to 90% or more.
Step 3: training objective detector is modified on the convolutional neural networks constructed in step 2: 1. remove it is last The overall situation be averaged pond layer, two linear layers and one softmax layers;2. the Feature Mapping of 3 scales is therefrom extracted, by more A convolutional layer is connected with up-sampling layer, finally exports one 52 × 52 × tensor of (c+5b), and wherein c is the class in data set Not Shuo, b is the bounding box quantity (generally taking 2) of each unit prediction, as above and constitutes complete yolo3 network structure, specifically Structure is shown in Fig. 2, and wherein the initial weight of each convolution kernel of characteristic extraction part is used using weight obtained in pre-training step The new network of the training of target detection data set obtained in step 1, the object detector that can be used.
Step 4: inclusion relation modeling
Defining rudimentary semantic object is (c, B), and wherein c is the classification of rudimentary semantic object, and B is rudimentary semantic object bounds Frame, definition bounding box are (xlt,ylt,xrb,yrb) wherein xlt,ylt,xrb,yrbThe upper left of respectively rudimentary semantic object bounds frame, The x of lower right coordinate, y-coordinate value, and xrb≥xlt, ylt≥yrb
Define the calculating A ∩ B and A ∪ B between frame calculating: for rudimentary semantic object A=(ca,(xlta,ylta,xrba, yrba)) and B=(cb,(xltb,yltb,xrbb,yrbb)):
A ∪ B=(min (xlta,xltb),max(xrba,xrbb),min(ylta,yltb),max(yrba,yrbb))
If max (xlta,xltb)≤min(xrba,xrbb) and max (ylta,yltb)≤min(yrba,yrbb):
A ∩ B=(max (xlta,xltb),min(xrba,xrbb),max(ylta,yltb),min(yrba,yrbb))
Otherwise A ∩ B is not present
Define area S (B)=(x of bounding boxrb-xlt)(yrb-ylt)
Define inclusion relation: for Primary objectives A, B:
If A ∩ B=A,
IfAndThen
Step 5: building high-level semantics object
It include the set of rudimentary semantic object from former steps available one, wherein each rudimentary semantic object includes one A classification and a bounding box.
For set S={ t1,t2,…tn, construct S '={ T1,T2…Tn, wherein S is the rudimentary semanteme that former steps obtain The set of object, tkFor rudimentary semantic object, Tk={ tk}。
Circular test S ', if it exists Ta,Tb,ti,tj, ti∈TaAnd tj∈Tb,Or cti≠ctj, andThen Tn+1=Ti∪Tj, the lasting set pair checked up to not met the requirements in S ', finally Obtain S "={ To,To+1,…Tp}
Step 6: object of interest extracts
For the high level goal that needs detect, its rudimentary target category for being included is manually specified, obtains Tt={ c1, c2,…cn}.T is retrieved in S "s={ ts1,ts2,…tsn, ifThen TsFor the target that need to be detected. Its bounding box isWherein ctsiFor tsiClassification, BtsiFor tsiBounding box.
Embodiment 1:
The present embodiment by taking certain commodity as an example, using the Retail commodity recognition methods of the invention based on convolutional neural networks into Row detection identification, as shown in figure 4, explicitly being modeled by carrying out contextual information to recognition detection target, thus realization pair The detection of fine granularity target;So that being completed under certain condition using the detector that disclosed coarseness data set training network obtains Fine grit classification task, finally can precisely quickly finish the identification of commodity.

Claims (5)

1. a kind of Retail commodity recognition methods based on convolutional neural networks, it is characterised in that: first with general coarse grain degree A yolo3 object detector is trained according to collecting, then input picture obtains classification and the position of a series of coarseness targets thereto Then information combines rudimentary semantic object according to the inclusion relation of predefined and obtains a series of high-level semantics objects;Finally exist The fine granularity target for needing to detect is chosen in these objects;
Specifically includes the following steps:
Step 1: screening the suitable classification of granularity according to the rule of correspondence from the category set of public data collection, then will choose Classification corresponding to picture form training set:
Step 2: the training set training yolo3 object detector chosen using step 1 obtains an accurate general objectives Detector;
Step 3: by object detector obtained in image to be detected input step 2, obtaining a series of class of rudimentary semantic objects Not and their position;
Step 4: choosing suitable μab, determine inclusion relation:
For rudimentary semantic object A, B, if A ∩ B=A, orAndThen Wherein S (A) refers to the area of the bounding box of rudimentary semantic object A, μaAnd μbBe respectively used to some object bounding box not fully It is determined when falling in the bounding box of another object between both with the presence or absence of inclusion relation;
Step 5: by rudimentary semanteme object set S={ t obtained in step 31,t2,…tnBy mapping f:Conversion is just Beginning high-level semantics object set S '={ T1,T2…Tn};Wherein Ti={ ti};
Step 6: according to following compatible rule merging object:
For high-level semantics object Ta,Tb, T if it existsa,Tb,ti,tj, ti∈TαAnd tj∈Tb,Or cti≠ctj, andThen merge Ta, TbFor their union, i.e. Tn+1=Ta∪Tb, from deletion T in S 'a,TbAnd add Add Tn+1;Wherein, high-level semantics object refers to the set of rudimentary semantic object;
Repeat the step, until in high-level semantics object set there is no meet be previously mentioned rule object pair, i.e., until S ' does not finally obtain S "={ T in variationo,To+1,…Tp};
Step 7: for the object that need to be detected, manually provide it includes rudimentary semantic object type, the height obtained in step 6 It is retrieved in grade semantic object, if the category set manually provided is the subset of the category set of some high-level semantics object, It is exactly the target for needing to detect.
2. the Retail commodity recognition methods according to claim 1 based on convolutional neural networks, it is characterised in that: the step Public data collection, which refers to from network public data concentration, in rapid 1 filters out categorized data set and target detection data set, wherein dividing Class data set only needs target category information contained in image, and target detection data set then needs the classification of target and boundary in image Two attributes of frame.
3. the Retail commodity recognition methods according to claim 1 based on convolutional neural networks, it is characterised in that: the step Rapid 1 classification based training concentrates each classification picture number at 1000 or more, classification number 30 or more and picture size be not less than 512*512 pixel.
4. the Retail commodity recognition methods according to claim 1 based on convolutional neural networks, it is characterised in that: the step Rapid 2 detailed step are as follows:
(2.1) it training characteristics extractor: is averaged pond layer, two on the feature extractor top of yolov3 plus an overall situation first A linear layer and one softmax layers, and then obtain a classifier;Then the classifier is trained, reads training set Whether image is that RGB image does center cutting, inputs network after normalized, correct using intersection entropy loss letter for exporting Number is measured;Continuous training terminates until accuracy to 90% or more;
(2.2) modify on the convolutional neural networks constructed: display removes the last overall situation and is averaged pond layer, two lines Property layer and one softmax layers;Then the Feature Mapping for therefrom extracting 3 scales passes through multiple convolutional layers and up-sampling layer phase Connection finally exports the tensor of a 52*52* (c+5b), and wherein c is the classification number in data set, and b is the prediction of each unit Bounding box quantity, and then complete yolo3 network structure is constituted, wherein the initial weight of each convolution kernel of characteristic extraction part makes The weight obtained in pre-training step is obtained most using the new network of gained target detection data set training in step (2.1) The object detector used eventually.
5. the Retail commodity recognition methods according to claim 1 based on convolutional neural networks, it is characterised in that: step 1 In screening rule are as follows:
(a) category is related to the appearance of targeted retailer product or packaging external appearance;
(b) category is related to the packing content of targeted retailer product;
(c) item name cannot include any brand message.
CN201811541070.2A 2018-12-17 2018-12-17 Retail commodity identification method based on convolutional neural network Active CN109685780B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811541070.2A CN109685780B (en) 2018-12-17 2018-12-17 Retail commodity identification method based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811541070.2A CN109685780B (en) 2018-12-17 2018-12-17 Retail commodity identification method based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN109685780A true CN109685780A (en) 2019-04-26
CN109685780B CN109685780B (en) 2021-05-11

Family

ID=66186116

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811541070.2A Active CN109685780B (en) 2018-12-17 2018-12-17 Retail commodity identification method based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN109685780B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110119811A (en) * 2019-05-15 2019-08-13 电科瑞达(成都)科技有限公司 A kind of convolution kernel method of cutting out based on entropy significance criteria model
CN110287369A (en) * 2019-06-25 2019-09-27 中科软科技股份有限公司 A kind of semantic-based video retrieval method and system
CN110414559A (en) * 2019-06-26 2019-11-05 武汉大学 The construction method and commodity recognition method of intelligence retail cabinet commodity target detection Unified frame
CN110751195A (en) * 2019-10-12 2020-02-04 西南交通大学 Fine-grained image classification method based on improved YOLOv3
CN110909660A (en) * 2019-11-19 2020-03-24 佛山市南海区广工大数控装备协同创新研究院 Plastic bottle detection and positioning method based on target detection
CN110992422A (en) * 2019-11-04 2020-04-10 浙江工业大学 Medicine box posture estimation method based on 3D vision
CN112241755A (en) * 2019-07-17 2021-01-19 东芝泰格有限公司 Article specifying device and storage medium
CN112929380A (en) * 2021-02-22 2021-06-08 中国科学院信息工程研究所 Trojan horse communication detection method and system combining meta-learning and spatiotemporal feature fusion
CN113344108A (en) * 2021-06-25 2021-09-03 视比特(长沙)机器人科技有限公司 Commodity identification and attitude estimation method and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170109625A1 (en) * 2015-10-14 2017-04-20 Microsoft Technology Licensing, Llc System for training networks for semantic segmentation
CN107133569A (en) * 2017-04-06 2017-09-05 同济大学 The many granularity mask methods of monitor video based on extensive Multi-label learning
US20180260793A1 (en) * 2016-04-06 2018-09-13 American International Group, Inc. Automatic assessment of damage and repair costs in vehicles
CN108596101A (en) * 2018-04-25 2018-09-28 上海交通大学 A kind of remote sensing images multi-target detection method based on convolutional neural networks
CN108776807A (en) * 2018-05-18 2018-11-09 复旦大学 It is a kind of based on can the double branch neural networks of skip floor image thickness grain-size classification method
CN109002845A (en) * 2018-06-29 2018-12-14 西安交通大学 Fine granularity image classification method based on depth convolutional neural networks
CN109002834A (en) * 2018-06-15 2018-12-14 东南大学 Fine granularity image classification method based on multi-modal characterization

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170109625A1 (en) * 2015-10-14 2017-04-20 Microsoft Technology Licensing, Llc System for training networks for semantic segmentation
US20180260793A1 (en) * 2016-04-06 2018-09-13 American International Group, Inc. Automatic assessment of damage and repair costs in vehicles
CN107133569A (en) * 2017-04-06 2017-09-05 同济大学 The many granularity mask methods of monitor video based on extensive Multi-label learning
CN108596101A (en) * 2018-04-25 2018-09-28 上海交通大学 A kind of remote sensing images multi-target detection method based on convolutional neural networks
CN108776807A (en) * 2018-05-18 2018-11-09 复旦大学 It is a kind of based on can the double branch neural networks of skip floor image thickness grain-size classification method
CN109002834A (en) * 2018-06-15 2018-12-14 东南大学 Fine granularity image classification method based on multi-modal characterization
CN109002845A (en) * 2018-06-29 2018-12-14 西安交通大学 Fine granularity image classification method based on depth convolutional neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JINGSONG LI.ET: "Supermarket Commodity Identification Using Convolutional Neural Networks", 《2016 2ND INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND INTERNET OF THINGS (CCIOT)》 *
陈鹤森: "基于深度学习的细粒度图像识别研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110119811A (en) * 2019-05-15 2019-08-13 电科瑞达(成都)科技有限公司 A kind of convolution kernel method of cutting out based on entropy significance criteria model
CN110119811B (en) * 2019-05-15 2021-07-27 电科瑞达(成都)科技有限公司 Convolution kernel cutting method based on entropy importance criterion model
CN110287369A (en) * 2019-06-25 2019-09-27 中科软科技股份有限公司 A kind of semantic-based video retrieval method and system
CN110287369B (en) * 2019-06-25 2022-02-22 中科软科技股份有限公司 Semantic-based video retrieval method and system
CN110414559A (en) * 2019-06-26 2019-11-05 武汉大学 The construction method and commodity recognition method of intelligence retail cabinet commodity target detection Unified frame
CN112241755A (en) * 2019-07-17 2021-01-19 东芝泰格有限公司 Article specifying device and storage medium
JP2021018470A (en) * 2019-07-17 2021-02-15 東芝テック株式会社 Article specification device and program
CN110751195A (en) * 2019-10-12 2020-02-04 西南交通大学 Fine-grained image classification method based on improved YOLOv3
CN110751195B (en) * 2019-10-12 2023-02-07 西南交通大学 Fine-grained image classification method based on improved YOLOv3
CN110992422A (en) * 2019-11-04 2020-04-10 浙江工业大学 Medicine box posture estimation method based on 3D vision
CN110992422B (en) * 2019-11-04 2023-11-07 浙江工业大学 Medicine box posture estimation method based on 3D vision
CN110909660A (en) * 2019-11-19 2020-03-24 佛山市南海区广工大数控装备协同创新研究院 Plastic bottle detection and positioning method based on target detection
CN112929380A (en) * 2021-02-22 2021-06-08 中国科学院信息工程研究所 Trojan horse communication detection method and system combining meta-learning and spatiotemporal feature fusion
CN113344108A (en) * 2021-06-25 2021-09-03 视比特(长沙)机器人科技有限公司 Commodity identification and attitude estimation method and device

Also Published As

Publication number Publication date
CN109685780B (en) 2021-05-11

Similar Documents

Publication Publication Date Title
CN109685780A (en) A kind of Retail commodity recognition methods based on convolutional neural networks
US11151427B2 (en) Method and apparatus for checkout based on image identification technique of convolutional neural network
Hsiao et al. Occlusion reasoning for object detectionunder arbitrary viewpoint
Yuan et al. Gated CNN: Integrating multi-scale feature layers for object detection
CN105844263B (en) The schematic diagram of the video object of shared predicable
CN103729848B (en) High-spectrum remote sensing small target detecting method based on spectrum saliency
CN109829429A (en) Security protection sensitive articles detection method under monitoring scene based on YOLOv3
CN110188807A (en) Tunnel pedestrian target detection method based on cascade super-resolution network and improvement Faster R-CNN
CN109829467A (en) Image labeling method, electronic device and non-transient computer-readable storage medium
CN107273106A (en) Object information is translated and derivation information acquisition methods and device
CN108345912A (en) Commodity rapid settlement system based on RGBD information and deep learning
CN109559453A (en) Human-computer interaction device and its application for Automatic-settlement
CN110245663A (en) One kind knowing method for distinguishing for coil of strip information
CN110298297A (en) Flame identification method and device
CN110119726A (en) A kind of vehicle brand multi-angle recognition methods based on YOLOv3 model
CN111597870A (en) Human body attribute identification method based on attention mechanism and multi-task learning
CN101853397A (en) Bionic human face detection method based on human visual characteristics
CN104050460B (en) The pedestrian detection method of multiple features fusion
Nguwi et al. Automatic road sign recognition using neural networks
CN103971106A (en) Multi-view human facial image gender identification method and device
CN115272652A (en) Dense object image detection method based on multiple regression and adaptive focus loss
CN104063721A (en) Human behavior recognition method based on automatic semantic feature study and screening
CN109712324A (en) A kind of automatic vending machine image-recognizing method, good selling method and vending equipment
CN114937298A (en) Micro-expression recognition method based on feature decoupling
Sakthimohan et al. Detection and Recognition of Face Using Deep Learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant