CN105512676A - Food recognition method at intelligent terminal - Google Patents
Food recognition method at intelligent terminal Download PDFInfo
- Publication number
- CN105512676A CN105512676A CN201510862931.7A CN201510862931A CN105512676A CN 105512676 A CN105512676 A CN 105512676A CN 201510862931 A CN201510862931 A CN 201510862931A CN 105512676 A CN105512676 A CN 105512676A
- Authority
- CN
- China
- Prior art keywords
- layer
- training
- network
- intelligent terminal
- stage
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/285—Selection of pattern recognition techniques, e.g. of classifiers in a multi-classifier system
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/29—Graphical models, e.g. Bayesian networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/086—Learning methods using evolutionary algorithms, e.g. genetic algorithms or genetic programming
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Physiology (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a food recognition method at an intelligent terminal, and the method is based on a convolution neural network method, and comprises a training process and an automatic classification process. The training process comprises the following steps: firstly building a training sample set; secondly constructing a network structure model according to a classic AlexNet network; thirdly carrying out network training through an open-source Caffe network framework, and continuously adjusting initial conditions so as to obtain an optimal network structure model and parameter configuration. The automatic classification process comprises the steps: enabling a to-be-recognized image photographed by a user to serve as an input image of the network; configuring the network according to parameters of a trained network structure in a computer, and classifying input images; and finally displaying the former 10 optimal classification results to the user. The method can achieve the automatic recognition of the type of food at the intelligent terminal, is high in speed, is small in storage capacity, is high in accuracy, and is good in user experience.
Description
Technical field
The present invention relates to artificial intelligence field, the food recognition methods particularly on a kind of intelligent terminal.
Background technology
Images Classification is a key areas of computer vision, has obtained significant progress and development at present in this field, and wherein automated graphics classification can provide convenient for our life, improves our quality of life.Multinomial patent of invention has been had in automated graphics classification, but these patents are all much adopt traditional feature extracting method, namely by extracting as SIFT feature, DenseSIFT feature, HOG feature etc., build dictionary by such as k-means clustering method again to encode to feature, carry out training by support vector machine (SVM) method again and obtain multiple SVM, finally utilize these SVM to carry out classification to test picture and obtain result, as the image classification method that patent CN104077597A extracts, use image Segmentation Technology to after each Object Segmentation in image, and extract the feature of this object in segmentation result, and carry out encoding and training, obtain the sorter of this object, then this sorter is utilized to carry out sort operation to test pattern.But this traditional extraction characteristics of image method be decide the final recognition accuracy of method by feature, but good feature needs good priori and design experiences, be difficult to design the optimum feature with discrimination in real system exploitation.
Convolutional neural networks is developed recently, and has been widely used in a kind of degree of depth learning method of many Images Classification problems, because convolutional neural networks can avoid characteristic extraction procedure complicated in traditional recognition method.In patent CN103544506A, propose a kind of image classification method based on convolutional neural networks and device, the method, by receiving the image pattern of multiple classifications of input, calculates the neural network weight that the image of each classification is corresponding; Adopt the corresponding neural network weight of hierarchy to multiple classification image to distribute, every one deck forms corresponding learning database; The category of test image sample data of input is processed and obtains corresponding one-dimensional characteristic and describe and carry out feedforward with the neural network weight in described learning database and learn, thus judge described category of test whether in learnt classification image.
Along with extensively popularizing of intelligent terminal, it has become the important tool that people obtain data.In addition, increasing people starts to pay close attention to own health problem, people wish there is a system software quick and precisely recording food, classic method lacks the food recognizer for this specific demand, thus needs to design a kind of food image categorizing system that may operate on intelligent terminal.Due to the restriction of intelligent terminal hardware condition, so the application on intelligent terminal must have real-time, need high accuracy of identification simultaneously.
Summary of the invention
In order to overcome the above-mentioned shortcoming of prior art with not enough, the object of the present invention is to provide the food recognition methods on a kind of intelligent terminal, the task of food identification can be completed on mobile terminals rapidly.
Object of the present invention is achieved through the following technical solutions:
A food recognition methods on intelligent terminal, comprises the following steps:
S1 carries out training process on computers:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
S1.2 builds convolutional neural networks model:
Described convolutional neural networks model comprises eight stages, is respectively convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model;
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
Sample picture chosen by the food of S1.6 for each classification in the training picture of each classification, and puts into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture.
Also following steps are carried out after step S2.2:
S2.3 carries out ascending sort to the classification results of step S2.2, the classification results of several optimum before showing on a user interface, and the result of each classification uses the sample picture chosen in S1.6 to show.
Convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage include the convolutional layer, active coating, abstraction, layer and the normalization layer that contain and connect successively; Described convolutional layer is all adopt Gaussian filter, and active coating uses ReLU function, and abstraction, layer adopts maximal value abstracting method, and normalization layer uses LRN method.
Described convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.
Described active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.
Described abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, and the mode of extraction is not overlapping extraction, and extracting unit size is P*P, and the interval s of adjacent extraction unit is taken as 2.
Described normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant,
represent the activation unit at i-th feature layer place of (x, y) position;
for right
the activation unit exported after normalization; I is positive integer.
Described full articulamentum the 6th stage and full articulamentum the 7th stage all comprise full articulamentum, active coating and abandon layer; Described full articulamentum the 8th stage comprises full articulamentum, output layer and loss layer.
Describedly abandon layer in the training process, network concealed node layer, in the mode of random chance, selects fractional weight to calculate, and the weighted value of unchecked hidden layer node not participation network upgrade, but its weighted value retains;
Described full articulamentum input value is launched into an one-dimensional vector X
c, then proceed as follows:
Y
c=w*X
c+b
Wherein, X
cinput vector, Y
cbe output vector, w is weighted value, and b is bias;
Described output layer selects accuracy layer prediction accuracy; Described accuracy layer obtains final degree of accuracy according to contrasting the prediction class label of output vector and its true class label and add up;
The described loss layer Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
Training convolutional neural networks model described in step S1.3, is specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying.
Compared with prior art, the present invention has the following advantages and beneficial effect:
(1) the present invention adopts the method training data based on convolutional neural networks, overcomes the feature extraction designs that tradition is complicated, and has automated characterization learning ability.
(2) the present invention adopts the Caffe network frame of increasing income, and this framework has the advantages such as modularization, speed is fast, and person easy to use is optimized network structure model.
(3) training process of the present invention is owing to carrying out off-line training on the server.This off-line training pattern, can utilize powerful server resource to save the training time.And the network profile after training is deposited on intelligent terminal, intelligent terminal recognizer is called network profile and is realized food identification fast.
Accompanying drawing explanation
Fig. 1 be embodiments of the invention intelligent terminal on food recognition methods carry out the process flow diagram of training process on computers.
Fig. 2 is the overall construction drawing of the convolutional neural networks of embodiments of the invention.
Fig. 3 is the process flow diagram of the identifying of food recognition methods on intelligent terminal on the intelligent terminal of embodiments of the invention.
Embodiment
Below in conjunction with embodiment, the present invention is described in further detail, but embodiments of the present invention are not limited thereto.
Embodiment
Food recognition methods on a kind of intelligent terminal of the present embodiment, comprises the following steps:
S1 carries out training process on computers, as shown in Figure 1:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
The present embodiment downloads Food-101 food database and the class label (link http://www.vision.ee.ethz.ch/datasets/food-101/) in Eidgenoess Tech Hochschule's computer vision laboratory from network.Food-101 food database downloads most popular front 101 kinds of foods from website foodspotting.com, and often kind of food comprises 1000 pictures, then picture is divided into training set and test set.The packet mode that the present embodiment provides according to Eidgenoess Tech Hochschule's computer vision laboratory, every group food uses 750 to train picture and 250 test pictures, and final training set has 75750 pictures, and test set has 25250 pictures;
S1.2 builds convolutional neural networks model, as shown in Figure 2:
Described convolutional neural networks model comprises 8 stages, is respectively convolutional layer the 1st stage, convolutional layer the 2nd stage, convolutional layer the 3rd stage, convolutional layer the 4th stage, convolutional layer the 5th stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
Described convolutional layer 1 stage comprises the 1st convolutional layer, the 1st active coating, the 1st abstraction, layer and the 1st normalization layer that connect successively; Described convolutional layer the 2nd stage comprises the 2nd convolutional layer, the 2nd active coating, the 2nd abstraction, layer and the 2nd normalization layer that connect successively; Described convolutional layer the 3rd stage comprises the 3rd convolutional layer, the 3rd active coating, the 3rd abstraction, layer and the 3rd normalization layer that connect successively; Described convolutional layer the 4th stage comprises the 4th convolutional layer, the 4th active coating, the 4th abstraction, layer and the 4th normalization layer that connect successively; Described convolutional layer the 5th stage is containing the 5th convolutional layer, the 5th active coating, the 5th abstraction, layer and the 5th normalization layer that connect successively.Wherein the 1st convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.Gaussian filter size in the present embodiment is set as 11*11, corresponding convolution kernel is 11*11 size, standard deviation sigma is 1, and always have 96 wave filters, and the step-length between each wave filter is 4 pixel sizes, the bias (bias) of this layer is set to 0 simultaneously, and therefore the output of the 1st convolutional layer has 96 feature layer.Because input size of data is 227*227*3, then the output through the 1st convolutional layer is 55*55*96.
Described 1st active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.Adopt this function that network can be made to have certain openness, can training speed be accelerated simultaneously.Concrete list of references (Rectifiedlinearunitsimproverestrictedboltzmannmachines.V .NairandG.E.Hinton.2010.);
Described 1st abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, extraction mode is not overlapping extraction, extracting unit size is P*P (P>1, in the present invention, the P of abstraction, layer is 3), the interval s of adjacent extraction unit is taken as 2.To being input as 55*55*96 after the maximal value extraction of s=2, the output of the 1st abstraction, layer is 27*27*96, is still 96 feature layer.
Described 1st normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant,
represent the activation unit at i-th feature layer place of (x, y) position;
for right
the activation unit exported after normalization; I is positive integer.(in normalization layer, all adopt LRN method in the present embodiment, and set k=2, n=5, α=0.0001, β=0.75).Thus be have 27*27*96 output in the convolutional layer first stage.
In the present embodiment, convolutional layer 2 employs the Gaussian filter of 256 5*5 sizes, be provided with every input on one side at this layer and all can increase by 2 pixel values, i.e. pad=2, bias value is 1 simultaneously, step value is set to 1, so the output of convolutional layer 2 is 27*27*256, altogether has 256 feature layer; The output of final convolutional layer subordinate phase is 13*13*256.
In the present embodiment, the 3rd convolutional layer employs the Gaussian filter of 384 3*3 sizes, pad=1, and step-length is also 1, bias=0, altogether has 384 feature layer; The output of final convolutional layer phase III is 13*13*384.
In the present embodiment, the optimum configurations of the 4th convolutional layer and the optimum configurations of the 3rd convolutional layer are identical.So be 13*13*384 in the output that convolutional layer the 4th stage is final.
In the present embodiment, the 5th convolutional layer employs the Gaussian filter of 256 3*3 sizes, pad=1, and step-length is 1, bias=1, always has 256 feature layer; Therefore the output that lamination the 4th stage is final is 6*6*256.
Articulamentum the 6th stage comprises the 6th full articulamentum, the 6th active coating and the 6th loss (Dropout) layer of connecting successively.
Input value is launched into an one-dimensional vector X by articulamentum 6
c, then proceed as follows:
Y
c=w*X
c+b
Wherein, X
cinput vector, Y
cbe output vector, w is weighted value, and b is bias.Full articulamentum 6 employs 4096 Gaussian filters in the present embodiment, and exporting so final is one 4096 vector tieed up.
Dropout layer the weight of subnetwork hidden layer node can be used to calculate according to certain probability when model training, in addition the weight of a part of network concealed node layer be then retain but and be used for calculate, effectively can prevent over-fitting like this, arranging probable value in the present embodiment is 0.5, concrete list of references (Improvingneuralnetworksbypreventingco-adaptationoffeatur edetectors.G.E.Hinton, N.Srivastava, A.Krizhevsky, I.Sutskever, andR.R.Salakhutdinov.2012).
Full articulamentum the 7th stage contains the 7th full articulamentum, the 7th active coating and the 7th loss layer that connect successively; In the present embodiment, each layer parameter in this stage arranges and arranges the same with each layer parameter in full articulamentum 6 stage.Thus the final output in this stage remains the vector of one 4096 dimension.
Full articulamentum the 8th stage comprises full articulamentum 8, output layer and loss (Loss) layer.The output of this layer corresponds to the class label of food picture; In the present embodiment, output layer selects accuracy layer to carry out prediction accuracy, and Loss layer selects the Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model, be specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying; In the present embodiment, the initial learn rate of Setup Experiments is 0.01,0.008,0.006 respectively, and maximum training iterations is then 450,000 times respectively, 400,000 times and 350,000 times, final select initial learn rate be 0.008 and maximum frequency of training be 450,000 times.
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
3 sample pictures chosen by the food of S1.6 for each classification in the training picture of each classification, and put into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal, as shown in Figure 3:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture; And ascending sort is carried out to classification results;
The classification results that S2.3 obtains according to S2.2, shows front 10 optimum classification results on a user interface, and the result of each classification uses 3 the sample pictures chosen in S1.6 to show.
Above-described embodiment is the present invention's preferably embodiment; but embodiments of the present invention are not limited by the examples; change, the modification done under other any does not deviate from Spirit Essence of the present invention and principle, substitute, combine, simplify; all should be the substitute mode of equivalence, be included within protection scope of the present invention.
Claims (10)
1. the food recognition methods on intelligent terminal, is characterized in that, comprise the following steps:
S1 carries out training process on computers:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
S1.2 builds convolutional neural networks model:
Described convolutional neural networks model comprises eight stages, is respectively convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model;
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
Sample picture chosen by the food of S1.6 for each classification in the training picture of each classification, and puts into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture.
2. the food recognition methods on intelligent terminal according to claim 1, is characterized in that, also carries out following steps after step S2.2:
S2.3 carries out ascending sort to the classification results of step S2.2, the classification results of several optimum before showing on a user interface, and the result of each classification uses the sample picture chosen in S1.6 to show.
3. the food recognition methods on intelligent terminal according to claim 1, it is characterized in that, convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage include the convolutional layer, active coating, abstraction, layer and the normalization layer that contain and connect successively; Described convolutional layer is all adopt Gaussian filter, and active coating uses ReLU function, and abstraction, layer adopts maximal value abstracting method, and normalization layer uses LRN method.
4. the food recognition methods on described intelligent terminal according to claim 3, it is characterized in that, described convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.
5. the food recognition methods on described intelligent terminal according to claim 3, is characterized in that, described active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.
6. the food recognition methods on described intelligent terminal according to claim 3, it is characterized in that, described abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, extraction mode is not overlapping extraction, extracting unit size is P*P, and the interval s of adjacent extraction unit is taken as 2.
7. the food recognition methods on described intelligent terminal according to claim 3, is characterized in that, described normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant,
represent the activation unit at i-th feature layer place of (x, y) position;
for right
the activation unit exported after normalization; I is positive integer.
8. the food recognition methods on described intelligent terminal according to claim 1, it is characterized in that, described full articulamentum the 6th stage and full articulamentum the 7th stage all comprise full articulamentum, active coating and abandon layer; Described full articulamentum the 8th stage comprises full articulamentum, output layer and loss layer.
9. the food recognition methods on intelligent terminal according to claim 8, it is characterized in that, describedly abandon layer in the training process, network concealed node layer is in the mode of random chance, fractional weight is selected to calculate, and the not participation network renewal of the weighted value of unchecked hidden layer node, but its weighted value retains;
Described full articulamentum input value is launched into an one-dimensional vector X
c, then proceed as follows:
Y
c=w*X
c+b
Wherein, X
cinput vector, Y
cbe output vector, w is weighted value, and b is bias;
Described output layer selects accuracy layer prediction accuracy; Described accuracy layer obtains final degree of accuracy according to contrasting the prediction class label of output vector and its true class label and add up;
The described loss layer Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
10. the food recognition methods on described intelligent terminal according to claim 1, it is characterized in that, training convolutional neural networks model described in step S1.3, is specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510862931.7A CN105512676A (en) | 2015-11-30 | 2015-11-30 | Food recognition method at intelligent terminal |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510862931.7A CN105512676A (en) | 2015-11-30 | 2015-11-30 | Food recognition method at intelligent terminal |
Publications (1)
Publication Number | Publication Date |
---|---|
CN105512676A true CN105512676A (en) | 2016-04-20 |
Family
ID=55720642
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510862931.7A Pending CN105512676A (en) | 2015-11-30 | 2015-11-30 | Food recognition method at intelligent terminal |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN105512676A (en) |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106203330A (en) * | 2016-07-08 | 2016-12-07 | 西安理工大学 | A kind of vehicle classification method based on convolutional neural networks |
CN106228177A (en) * | 2016-06-30 | 2016-12-14 | 浙江大学 | Daily life subject image recognition methods based on convolutional neural networks |
CN106295707A (en) * | 2016-08-17 | 2017-01-04 | 北京小米移动软件有限公司 | Image-recognizing method and device |
CN106529564A (en) * | 2016-09-26 | 2017-03-22 | 浙江工业大学 | Food image automatic classification method based on convolutional neural networks |
CN106845527A (en) * | 2016-12-29 | 2017-06-13 | 南京江南博睿高新技术研究院有限公司 | A kind of vegetable recognition methods |
CN106874913A (en) * | 2016-12-29 | 2017-06-20 | 南京江南博睿高新技术研究院有限公司 | A kind of vegetable detection method |
CN106980817A (en) * | 2017-02-27 | 2017-07-25 | 南京邮电大学 | A kind of terrified video frequency identifying method based on Caffe frameworks |
CN107024073A (en) * | 2017-04-26 | 2017-08-08 | 中国石油大学(华东) | Multi-sensor intelligent controlling method for refrigerator and intelligent refrigerator based on deep learning |
CN107067043A (en) * | 2017-05-25 | 2017-08-18 | 哈尔滨工业大学 | A kind of diseases and pests of agronomic crop detection method |
CN107480773A (en) * | 2017-08-09 | 2017-12-15 | 北京小米移动软件有限公司 | The method, apparatus and storage medium of training convolutional neural networks model |
CN107742128A (en) * | 2017-10-20 | 2018-02-27 | 百度在线网络技术(北京)有限公司 | Method and apparatus for output information |
WO2018040105A1 (en) * | 2016-09-05 | 2018-03-08 | 合肥华凌股份有限公司 | System and method for food recognition, food model training method, refrigerator and server |
CN107798277A (en) * | 2016-09-05 | 2018-03-13 | 合肥美的智能科技有限公司 | Food materials identifying system and method, food materials model training method, refrigerator and server |
CN107832804A (en) * | 2017-10-30 | 2018-03-23 | 上海寒武纪信息科技有限公司 | A kind of information processing method and Related product |
CN107886108A (en) * | 2017-10-11 | 2018-04-06 | 中国农业大学 | Fruit and vegetable classification method and device based on AlexNet network models |
CN108009525A (en) * | 2017-12-25 | 2018-05-08 | 北京航空航天大学 | A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks |
CN108171318A (en) * | 2017-11-30 | 2018-06-15 | 河南大学 | One kind is based on the convolutional neural networks integrated approach of simulated annealing-Gaussian function |
CN108229979A (en) * | 2018-01-24 | 2018-06-29 | 亳州职业技术学院 | Based on wechat small routine and the American Ginseng of depth learning technology tasting assistant |
WO2018121690A1 (en) * | 2016-12-29 | 2018-07-05 | 北京市商汤科技开发有限公司 | Object attribute detection method and device, neural network training method and device, and regional detection method and device |
CN108280474A (en) * | 2018-01-19 | 2018-07-13 | 广州市派客朴食信息科技有限责任公司 | A kind of food recognition methods based on neural network |
CN108507270A (en) * | 2017-02-24 | 2018-09-07 | 九阳股份有限公司 | A kind of food materials of refrigerator determine method and food materials determining device |
CN108647734A (en) * | 2018-05-15 | 2018-10-12 | 上海达显智能科技有限公司 | A kind of food image big data acquisition method, acquisition system and food recognition methods |
CN108846419A (en) * | 2018-05-25 | 2018-11-20 | 平安科技(深圳)有限公司 | Single page high load image-recognizing method, device, computer equipment and storage medium |
CN108875959A (en) * | 2018-05-24 | 2018-11-23 | 四川斐讯信息技术有限公司 | A kind of intelligence sees the method and system of object knowledge name |
CN109214409A (en) * | 2018-07-10 | 2019-01-15 | 上海斐讯数据通信技术有限公司 | A kind of vegetable recognition methods and system |
CN109214272A (en) * | 2018-07-17 | 2019-01-15 | 北京陌上花科技有限公司 | A kind of image-recognizing method and device |
WO2019011249A1 (en) * | 2017-07-14 | 2019-01-17 | 腾讯科技(深圳)有限公司 | Method, apparatus, and device for determining pose of object in image, and storage medium |
WO2020057455A1 (en) * | 2018-09-17 | 2020-03-26 | 鲁班嫡系机器人(深圳)有限公司 | Food material processing robot control method, apparatus, system, storage medium and device |
CN111797719A (en) * | 2020-06-17 | 2020-10-20 | 武汉大学 | Food component identification method |
US11990137B2 (en) | 2018-09-13 | 2024-05-21 | Shanghai Cambricon Information Technology Co., Ltd. | Image retouching method and terminal device |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103544506A (en) * | 2013-10-12 | 2014-01-29 | Tcl集团股份有限公司 | Method and device for classifying images on basis of convolutional neural network |
CN104090871A (en) * | 2014-07-18 | 2014-10-08 | 百度在线网络技术(北京)有限公司 | Picture translation method and system |
CN104346622A (en) * | 2013-07-31 | 2015-02-11 | 富士通株式会社 | Convolutional neural network classifier, and classifying method and training method thereof |
CN104636757A (en) * | 2015-02-06 | 2015-05-20 | 中国石油大学(华东) | Deep learning-based food image identifying method |
-
2015
- 2015-11-30 CN CN201510862931.7A patent/CN105512676A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104346622A (en) * | 2013-07-31 | 2015-02-11 | 富士通株式会社 | Convolutional neural network classifier, and classifying method and training method thereof |
CN103544506A (en) * | 2013-10-12 | 2014-01-29 | Tcl集团股份有限公司 | Method and device for classifying images on basis of convolutional neural network |
CN104090871A (en) * | 2014-07-18 | 2014-10-08 | 百度在线网络技术(北京)有限公司 | Picture translation method and system |
CN104636757A (en) * | 2015-02-06 | 2015-05-20 | 中国石油大学(华东) | Deep learning-based food image identifying method |
Non-Patent Citations (2)
Title |
---|
KRIZHEVSKY A ETC,: ""ImageNet classification with deep convolutional neural networks"", 《INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING SYSTEMS.CURRAN ASSOCIATES INC》 * |
王振 等,: ""基于卷积神经网络的图像识别算法设计与实现"", 《图形图像》 * |
Cited By (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106228177A (en) * | 2016-06-30 | 2016-12-14 | 浙江大学 | Daily life subject image recognition methods based on convolutional neural networks |
CN106203330A (en) * | 2016-07-08 | 2016-12-07 | 西安理工大学 | A kind of vehicle classification method based on convolutional neural networks |
CN106295707A (en) * | 2016-08-17 | 2017-01-04 | 北京小米移动软件有限公司 | Image-recognizing method and device |
CN106295707B (en) * | 2016-08-17 | 2019-07-02 | 北京小米移动软件有限公司 | Image-recognizing method and device |
CN107798277A (en) * | 2016-09-05 | 2018-03-13 | 合肥美的智能科技有限公司 | Food materials identifying system and method, food materials model training method, refrigerator and server |
WO2018040105A1 (en) * | 2016-09-05 | 2018-03-08 | 合肥华凌股份有限公司 | System and method for food recognition, food model training method, refrigerator and server |
CN106529564A (en) * | 2016-09-26 | 2017-03-22 | 浙江工业大学 | Food image automatic classification method based on convolutional neural networks |
CN106529564B (en) * | 2016-09-26 | 2019-05-31 | 浙江工业大学 | A kind of food image automatic classification method based on convolutional neural networks |
CN106874913A (en) * | 2016-12-29 | 2017-06-20 | 南京江南博睿高新技术研究院有限公司 | A kind of vegetable detection method |
CN106845527A (en) * | 2016-12-29 | 2017-06-13 | 南京江南博睿高新技术研究院有限公司 | A kind of vegetable recognition methods |
WO2018121690A1 (en) * | 2016-12-29 | 2018-07-05 | 北京市商汤科技开发有限公司 | Object attribute detection method and device, neural network training method and device, and regional detection method and device |
CN108507270B (en) * | 2017-02-24 | 2020-07-10 | 九阳股份有限公司 | Food material determining method and food material determining device for refrigerator |
CN108507270A (en) * | 2017-02-24 | 2018-09-07 | 九阳股份有限公司 | A kind of food materials of refrigerator determine method and food materials determining device |
CN106980817A (en) * | 2017-02-27 | 2017-07-25 | 南京邮电大学 | A kind of terrified video frequency identifying method based on Caffe frameworks |
CN107024073A (en) * | 2017-04-26 | 2017-08-08 | 中国石油大学(华东) | Multi-sensor intelligent controlling method for refrigerator and intelligent refrigerator based on deep learning |
CN107067043B (en) * | 2017-05-25 | 2020-07-24 | 哈尔滨工业大学 | Crop disease and insect pest detection method |
CN107067043A (en) * | 2017-05-25 | 2017-08-18 | 哈尔滨工业大学 | A kind of diseases and pests of agronomic crop detection method |
US11107232B2 (en) | 2017-07-14 | 2021-08-31 | Tencent Technology (Shenzhen) Company Limited | Method and apparatus for determining object posture in image, device, and storage medium |
WO2019011249A1 (en) * | 2017-07-14 | 2019-01-17 | 腾讯科技(深圳)有限公司 | Method, apparatus, and device for determining pose of object in image, and storage medium |
CN107480773B (en) * | 2017-08-09 | 2020-11-13 | 北京小米移动软件有限公司 | Method and device for training convolutional neural network model and storage medium |
CN107480773A (en) * | 2017-08-09 | 2017-12-15 | 北京小米移动软件有限公司 | The method, apparatus and storage medium of training convolutional neural networks model |
CN107886108A (en) * | 2017-10-11 | 2018-04-06 | 中国农业大学 | Fruit and vegetable classification method and device based on AlexNet network models |
CN107742128A (en) * | 2017-10-20 | 2018-02-27 | 百度在线网络技术(北京)有限公司 | Method and apparatus for output information |
US11922132B2 (en) | 2017-10-30 | 2024-03-05 | Shanghai Cambricon Information Technology Co., Ltd. | Information processing method and terminal device |
CN107832804A (en) * | 2017-10-30 | 2018-03-23 | 上海寒武纪信息科技有限公司 | A kind of information processing method and Related product |
CN108171318A (en) * | 2017-11-30 | 2018-06-15 | 河南大学 | One kind is based on the convolutional neural networks integrated approach of simulated annealing-Gaussian function |
CN108009525A (en) * | 2017-12-25 | 2018-05-08 | 北京航空航天大学 | A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks |
CN108009525B (en) * | 2017-12-25 | 2018-10-12 | 北京航空航天大学 | A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks |
CN108280474A (en) * | 2018-01-19 | 2018-07-13 | 广州市派客朴食信息科技有限责任公司 | A kind of food recognition methods based on neural network |
CN108229979A (en) * | 2018-01-24 | 2018-06-29 | 亳州职业技术学院 | Based on wechat small routine and the American Ginseng of depth learning technology tasting assistant |
CN108647734A (en) * | 2018-05-15 | 2018-10-12 | 上海达显智能科技有限公司 | A kind of food image big data acquisition method, acquisition system and food recognition methods |
CN108647734B (en) * | 2018-05-15 | 2022-03-11 | 上海达显智能科技有限公司 | Food image big data acquisition method, acquisition system and food identification method |
CN108875959A (en) * | 2018-05-24 | 2018-11-23 | 四川斐讯信息技术有限公司 | A kind of intelligence sees the method and system of object knowledge name |
CN108846419A (en) * | 2018-05-25 | 2018-11-20 | 平安科技(深圳)有限公司 | Single page high load image-recognizing method, device, computer equipment and storage medium |
CN109214409A (en) * | 2018-07-10 | 2019-01-15 | 上海斐讯数据通信技术有限公司 | A kind of vegetable recognition methods and system |
CN109214272A (en) * | 2018-07-17 | 2019-01-15 | 北京陌上花科技有限公司 | A kind of image-recognizing method and device |
US11990137B2 (en) | 2018-09-13 | 2024-05-21 | Shanghai Cambricon Information Technology Co., Ltd. | Image retouching method and terminal device |
US11996105B2 (en) | 2018-09-13 | 2024-05-28 | Shanghai Cambricon Information Technology Co., Ltd. | Information processing method and terminal device |
WO2020057455A1 (en) * | 2018-09-17 | 2020-03-26 | 鲁班嫡系机器人(深圳)有限公司 | Food material processing robot control method, apparatus, system, storage medium and device |
CN111797719A (en) * | 2020-06-17 | 2020-10-20 | 武汉大学 | Food component identification method |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105512676A (en) | Food recognition method at intelligent terminal | |
CN108875674B (en) | Driver behavior identification method based on multi-column fusion convolutional neural network | |
CN108388927B (en) | Small sample polarization SAR terrain classification method based on deep convolution twin network | |
Hossain et al. | Improving consumer satisfaction in smart cities using edge computing and caching: A case study of date fruits classification | |
CN102314614B (en) | Image semantics classification method based on class-shared multiple kernel learning (MKL) | |
CN110598598A (en) | Double-current convolution neural network human behavior identification method based on finite sample set | |
CN105320961A (en) | Handwriting numeral recognition method based on convolutional neural network and support vector machine | |
CN106372648A (en) | Multi-feature-fusion-convolutional-neural-network-based plankton image classification method | |
CN109344884A (en) | The method and device of media information classification method, training picture classification model | |
Zhang et al. | Unsupervised difference representation learning for detecting multiple types of changes in multitemporal remote sensing images | |
CN104182772A (en) | Gesture recognition method based on deep learning | |
CN104537647A (en) | Target detection method and device | |
CN112883839A (en) | Remote sensing image interpretation method based on adaptive sample set construction and deep learning | |
CN108537119A (en) | A kind of small sample video frequency identifying method | |
WO2022062419A1 (en) | Target re-identification method and system based on non-supervised pyramid similarity learning | |
CN107169106A (en) | Video retrieval method, device, storage medium and processor | |
CN108764084A (en) | Video classification methods based on spatial domain sorter network and the time domain network integration | |
CN109492596A (en) | A kind of pedestrian detection method and system based on K-means cluster and region recommendation network | |
CN113569895A (en) | Image processing model training method, processing method, device, equipment and medium | |
CN106778910A (en) | Deep learning system and method based on local training | |
CN109241392A (en) | Recognition methods, device, system and the storage medium of target word | |
CN110503082A (en) | A kind of model training method and relevant apparatus based on deep learning | |
CN109978074A (en) | Image aesthetic feeling and emotion joint classification method and system based on depth multi-task learning | |
CN113109782A (en) | Novel classification method directly applied to radar radiation source amplitude sequence | |
CN113822130A (en) | Model training method, scene recognition method, computing device, and medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20160420 |