CN105512676A - Food recognition method at intelligent terminal - Google Patents

Food recognition method at intelligent terminal Download PDF

Info

Publication number
CN105512676A
CN105512676A CN201510862931.7A CN201510862931A CN105512676A CN 105512676 A CN105512676 A CN 105512676A CN 201510862931 A CN201510862931 A CN 201510862931A CN 105512676 A CN105512676 A CN 105512676A
Authority
CN
China
Prior art keywords
layer
training
network
intelligent terminal
stage
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510862931.7A
Other languages
Chinese (zh)
Inventor
郭礼华
罗才
廖启俊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201510862931.7A priority Critical patent/CN105512676A/en
Publication of CN105512676A publication Critical patent/CN105512676A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/285Selection of pattern recognition techniques, e.g. of classifiers in a multi-classifier system
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/29Graphical models, e.g. Bayesian networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/086Learning methods using evolutionary algorithms, e.g. genetic algorithms or genetic programming

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Physiology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a food recognition method at an intelligent terminal, and the method is based on a convolution neural network method, and comprises a training process and an automatic classification process. The training process comprises the following steps: firstly building a training sample set; secondly constructing a network structure model according to a classic AlexNet network; thirdly carrying out network training through an open-source Caffe network framework, and continuously adjusting initial conditions so as to obtain an optimal network structure model and parameter configuration. The automatic classification process comprises the steps: enabling a to-be-recognized image photographed by a user to serve as an input image of the network; configuring the network according to parameters of a trained network structure in a computer, and classifying input images; and finally displaying the former 10 optimal classification results to the user. The method can achieve the automatic recognition of the type of food at the intelligent terminal, is high in speed, is small in storage capacity, is high in accuracy, and is good in user experience.

Description

Food recognition methods on a kind of intelligent terminal
Technical field
The present invention relates to artificial intelligence field, the food recognition methods particularly on a kind of intelligent terminal.
Background technology
Images Classification is a key areas of computer vision, has obtained significant progress and development at present in this field, and wherein automated graphics classification can provide convenient for our life, improves our quality of life.Multinomial patent of invention has been had in automated graphics classification, but these patents are all much adopt traditional feature extracting method, namely by extracting as SIFT feature, DenseSIFT feature, HOG feature etc., build dictionary by such as k-means clustering method again to encode to feature, carry out training by support vector machine (SVM) method again and obtain multiple SVM, finally utilize these SVM to carry out classification to test picture and obtain result, as the image classification method that patent CN104077597A extracts, use image Segmentation Technology to after each Object Segmentation in image, and extract the feature of this object in segmentation result, and carry out encoding and training, obtain the sorter of this object, then this sorter is utilized to carry out sort operation to test pattern.But this traditional extraction characteristics of image method be decide the final recognition accuracy of method by feature, but good feature needs good priori and design experiences, be difficult to design the optimum feature with discrimination in real system exploitation.
Convolutional neural networks is developed recently, and has been widely used in a kind of degree of depth learning method of many Images Classification problems, because convolutional neural networks can avoid characteristic extraction procedure complicated in traditional recognition method.In patent CN103544506A, propose a kind of image classification method based on convolutional neural networks and device, the method, by receiving the image pattern of multiple classifications of input, calculates the neural network weight that the image of each classification is corresponding; Adopt the corresponding neural network weight of hierarchy to multiple classification image to distribute, every one deck forms corresponding learning database; The category of test image sample data of input is processed and obtains corresponding one-dimensional characteristic and describe and carry out feedforward with the neural network weight in described learning database and learn, thus judge described category of test whether in learnt classification image.
Along with extensively popularizing of intelligent terminal, it has become the important tool that people obtain data.In addition, increasing people starts to pay close attention to own health problem, people wish there is a system software quick and precisely recording food, classic method lacks the food recognizer for this specific demand, thus needs to design a kind of food image categorizing system that may operate on intelligent terminal.Due to the restriction of intelligent terminal hardware condition, so the application on intelligent terminal must have real-time, need high accuracy of identification simultaneously.
Summary of the invention
In order to overcome the above-mentioned shortcoming of prior art with not enough, the object of the present invention is to provide the food recognition methods on a kind of intelligent terminal, the task of food identification can be completed on mobile terminals rapidly.
Object of the present invention is achieved through the following technical solutions:
A food recognition methods on intelligent terminal, comprises the following steps:
S1 carries out training process on computers:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
S1.2 builds convolutional neural networks model:
Described convolutional neural networks model comprises eight stages, is respectively convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model;
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
Sample picture chosen by the food of S1.6 for each classification in the training picture of each classification, and puts into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture.
Also following steps are carried out after step S2.2:
S2.3 carries out ascending sort to the classification results of step S2.2, the classification results of several optimum before showing on a user interface, and the result of each classification uses the sample picture chosen in S1.6 to show.
Convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage include the convolutional layer, active coating, abstraction, layer and the normalization layer that contain and connect successively; Described convolutional layer is all adopt Gaussian filter, and active coating uses ReLU function, and abstraction, layer adopts maximal value abstracting method, and normalization layer uses LRN method.
Described convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
G ( x , y , σ ) = 1 2 πσ 2 e - ( x 2 + y 2 ) / 2 σ 2
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.
Described active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.
Described abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, and the mode of extraction is not overlapping extraction, and extracting unit size is P*P, and the interval s of adjacent extraction unit is taken as 2.
Described normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
b x , y i = u x , y i ( k + α Σ j = max ( 0 , i - n / 2 ) min ( N - 1 , i + n / 2 ( u x , y j ) 2 ) β
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant, represent the activation unit at i-th feature layer place of (x, y) position; for right the activation unit exported after normalization; I is positive integer.
Described full articulamentum the 6th stage and full articulamentum the 7th stage all comprise full articulamentum, active coating and abandon layer; Described full articulamentum the 8th stage comprises full articulamentum, output layer and loss layer.
Describedly abandon layer in the training process, network concealed node layer, in the mode of random chance, selects fractional weight to calculate, and the weighted value of unchecked hidden layer node not participation network upgrade, but its weighted value retains;
Described full articulamentum input value is launched into an one-dimensional vector X c, then proceed as follows:
Y c=w*X c+b
Wherein, X cinput vector, Y cbe output vector, w is weighted value, and b is bias;
Described output layer selects accuracy layer prediction accuracy; Described accuracy layer obtains final degree of accuracy according to contrasting the prediction class label of output vector and its true class label and add up;
The described loss layer Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
Training convolutional neural networks model described in step S1.3, is specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying.
Compared with prior art, the present invention has the following advantages and beneficial effect:
(1) the present invention adopts the method training data based on convolutional neural networks, overcomes the feature extraction designs that tradition is complicated, and has automated characterization learning ability.
(2) the present invention adopts the Caffe network frame of increasing income, and this framework has the advantages such as modularization, speed is fast, and person easy to use is optimized network structure model.
(3) training process of the present invention is owing to carrying out off-line training on the server.This off-line training pattern, can utilize powerful server resource to save the training time.And the network profile after training is deposited on intelligent terminal, intelligent terminal recognizer is called network profile and is realized food identification fast.
Accompanying drawing explanation
Fig. 1 be embodiments of the invention intelligent terminal on food recognition methods carry out the process flow diagram of training process on computers.
Fig. 2 is the overall construction drawing of the convolutional neural networks of embodiments of the invention.
Fig. 3 is the process flow diagram of the identifying of food recognition methods on intelligent terminal on the intelligent terminal of embodiments of the invention.
Embodiment
Below in conjunction with embodiment, the present invention is described in further detail, but embodiments of the present invention are not limited thereto.
Embodiment
Food recognition methods on a kind of intelligent terminal of the present embodiment, comprises the following steps:
S1 carries out training process on computers, as shown in Figure 1:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
The present embodiment downloads Food-101 food database and the class label (link http://www.vision.ee.ethz.ch/datasets/food-101/) in Eidgenoess Tech Hochschule's computer vision laboratory from network.Food-101 food database downloads most popular front 101 kinds of foods from website foodspotting.com, and often kind of food comprises 1000 pictures, then picture is divided into training set and test set.The packet mode that the present embodiment provides according to Eidgenoess Tech Hochschule's computer vision laboratory, every group food uses 750 to train picture and 250 test pictures, and final training set has 75750 pictures, and test set has 25250 pictures;
S1.2 builds convolutional neural networks model, as shown in Figure 2:
Described convolutional neural networks model comprises 8 stages, is respectively convolutional layer the 1st stage, convolutional layer the 2nd stage, convolutional layer the 3rd stage, convolutional layer the 4th stage, convolutional layer the 5th stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
Described convolutional layer 1 stage comprises the 1st convolutional layer, the 1st active coating, the 1st abstraction, layer and the 1st normalization layer that connect successively; Described convolutional layer the 2nd stage comprises the 2nd convolutional layer, the 2nd active coating, the 2nd abstraction, layer and the 2nd normalization layer that connect successively; Described convolutional layer the 3rd stage comprises the 3rd convolutional layer, the 3rd active coating, the 3rd abstraction, layer and the 3rd normalization layer that connect successively; Described convolutional layer the 4th stage comprises the 4th convolutional layer, the 4th active coating, the 4th abstraction, layer and the 4th normalization layer that connect successively; Described convolutional layer the 5th stage is containing the 5th convolutional layer, the 5th active coating, the 5th abstraction, layer and the 5th normalization layer that connect successively.Wherein the 1st convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
G ( x , y , σ ) = 1 2 πσ 2 e - ( x 2 + y 2 ) / 2 σ 2
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.Gaussian filter size in the present embodiment is set as 11*11, corresponding convolution kernel is 11*11 size, standard deviation sigma is 1, and always have 96 wave filters, and the step-length between each wave filter is 4 pixel sizes, the bias (bias) of this layer is set to 0 simultaneously, and therefore the output of the 1st convolutional layer has 96 feature layer.Because input size of data is 227*227*3, then the output through the 1st convolutional layer is 55*55*96.
Described 1st active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.Adopt this function that network can be made to have certain openness, can training speed be accelerated simultaneously.Concrete list of references (Rectifiedlinearunitsimproverestrictedboltzmannmachines.V .NairandG.E.Hinton.2010.);
Described 1st abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, extraction mode is not overlapping extraction, extracting unit size is P*P (P>1, in the present invention, the P of abstraction, layer is 3), the interval s of adjacent extraction unit is taken as 2.To being input as 55*55*96 after the maximal value extraction of s=2, the output of the 1st abstraction, layer is 27*27*96, is still 96 feature layer.
Described 1st normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
b x , y i = u x , y i ( k + α Σ j = max ( 0 , i - n / 2 ) min ( N - 1 , i + n / 2 ( u x , y j ) 2 ) β
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant, represent the activation unit at i-th feature layer place of (x, y) position; for right the activation unit exported after normalization; I is positive integer.(in normalization layer, all adopt LRN method in the present embodiment, and set k=2, n=5, α=0.0001, β=0.75).Thus be have 27*27*96 output in the convolutional layer first stage.
In the present embodiment, convolutional layer 2 employs the Gaussian filter of 256 5*5 sizes, be provided with every input on one side at this layer and all can increase by 2 pixel values, i.e. pad=2, bias value is 1 simultaneously, step value is set to 1, so the output of convolutional layer 2 is 27*27*256, altogether has 256 feature layer; The output of final convolutional layer subordinate phase is 13*13*256.
In the present embodiment, the 3rd convolutional layer employs the Gaussian filter of 384 3*3 sizes, pad=1, and step-length is also 1, bias=0, altogether has 384 feature layer; The output of final convolutional layer phase III is 13*13*384.
In the present embodiment, the optimum configurations of the 4th convolutional layer and the optimum configurations of the 3rd convolutional layer are identical.So be 13*13*384 in the output that convolutional layer the 4th stage is final.
In the present embodiment, the 5th convolutional layer employs the Gaussian filter of 256 3*3 sizes, pad=1, and step-length is 1, bias=1, always has 256 feature layer; Therefore the output that lamination the 4th stage is final is 6*6*256.
Articulamentum the 6th stage comprises the 6th full articulamentum, the 6th active coating and the 6th loss (Dropout) layer of connecting successively.
Input value is launched into an one-dimensional vector X by articulamentum 6 c, then proceed as follows:
Y c=w*X c+b
Wherein, X cinput vector, Y cbe output vector, w is weighted value, and b is bias.Full articulamentum 6 employs 4096 Gaussian filters in the present embodiment, and exporting so final is one 4096 vector tieed up.
Dropout layer the weight of subnetwork hidden layer node can be used to calculate according to certain probability when model training, in addition the weight of a part of network concealed node layer be then retain but and be used for calculate, effectively can prevent over-fitting like this, arranging probable value in the present embodiment is 0.5, concrete list of references (Improvingneuralnetworksbypreventingco-adaptationoffeatur edetectors.G.E.Hinton, N.Srivastava, A.Krizhevsky, I.Sutskever, andR.R.Salakhutdinov.2012).
Full articulamentum the 7th stage contains the 7th full articulamentum, the 7th active coating and the 7th loss layer that connect successively; In the present embodiment, each layer parameter in this stage arranges and arranges the same with each layer parameter in full articulamentum 6 stage.Thus the final output in this stage remains the vector of one 4096 dimension.
Full articulamentum the 8th stage comprises full articulamentum 8, output layer and loss (Loss) layer.The output of this layer corresponds to the class label of food picture; In the present embodiment, output layer selects accuracy layer to carry out prediction accuracy, and Loss layer selects the Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model, be specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying; In the present embodiment, the initial learn rate of Setup Experiments is 0.01,0.008,0.006 respectively, and maximum training iterations is then 450,000 times respectively, 400,000 times and 350,000 times, final select initial learn rate be 0.008 and maximum frequency of training be 450,000 times.
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
3 sample pictures chosen by the food of S1.6 for each classification in the training picture of each classification, and put into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal, as shown in Figure 3:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture; And ascending sort is carried out to classification results;
The classification results that S2.3 obtains according to S2.2, shows front 10 optimum classification results on a user interface, and the result of each classification uses 3 the sample pictures chosen in S1.6 to show.
Above-described embodiment is the present invention's preferably embodiment; but embodiments of the present invention are not limited by the examples; change, the modification done under other any does not deviate from Spirit Essence of the present invention and principle, substitute, combine, simplify; all should be the substitute mode of equivalence, be included within protection scope of the present invention.

Claims (10)

1. the food recognition methods on intelligent terminal, is characterized in that, comprise the following steps:
S1 carries out training process on computers:
S1.1 gathers the food picture for training, and obtains training picture; According to the classification of food, add class label to respectively the training picture of each classification;
S1.2 builds convolutional neural networks model:
Described convolutional neural networks model comprises eight stages, is respectively convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage, full articulamentum the 6th stage, full articulamentum the 7th stage, full articulamentum the 8th stage;
S1.3 training convolutional neural networks model: adopt the Caffe network frame of increasing income to carry out training convolutional neural networks model;
Training data is sent into Caffe network in step S1.3 in batches and is trained by S1.4, and can obtain current training sample by training and concentrate best classify accuracy, network parameter is now the network parameter of final optimum network structure model;
The configuration file of the network parameter of the final optimum network structure model that S1.4 obtains by S1.5 puts into the installation kit of intelligent terminal software;
Sample picture chosen by the food of S1.6 for each classification in the training picture of each classification, and puts into the installation kit of intelligent terminal software;
The automatic classification process of S2 on intelligent terminal:
S2.2 is using picture to be identified as input picture, and the convolutional neural networks of the network structure model parameter configuration test obtained according to S1.4, classifies to input picture.
2. the food recognition methods on intelligent terminal according to claim 1, is characterized in that, also carries out following steps after step S2.2:
S2.3 carries out ascending sort to the classification results of step S2.2, the classification results of several optimum before showing on a user interface, and the result of each classification uses the sample picture chosen in S1.6 to show.
3. the food recognition methods on intelligent terminal according to claim 1, it is characterized in that, convolutional layer first stage, convolutional layer subordinate phase, convolutional layer phase III, convolutional layer fourth stage, convolutional layer five-stage include the convolutional layer, active coating, abstraction, layer and the normalization layer that contain and connect successively; Described convolutional layer is all adopt Gaussian filter, and active coating uses ReLU function, and abstraction, layer adopts maximal value abstracting method, and normalization layer uses LRN method.
4. the food recognition methods on described intelligent terminal according to claim 3, it is characterized in that, described convolutional layer is Gaussian convolution layer, for realizing the convolution operation of received image signal and Gaussian filter; The function of Gaussian filter is defined as follows:
G ( x , y , σ ) = 1 2 πσ 2 e - ( x 2 + y 2 ) / 2 σ 2
X, y are the coordinate of Gaussian filter function in two-dimensional image plane, and σ is that Gaussian filter functional standard is poor, control the smoothness of Gaussian filter, be worth larger, its Gaussian filter function is more level and smooth, and G (x, y, σ) then Gaussian filter function respective value.
5. the food recognition methods on described intelligent terminal according to claim 3, is characterized in that, described active coating adopts the output of ReLU number to convolutional layer to process, and its function expression is:
f(IN)=max(0,IN)
Wherein IN is input signal, and f (IN) is output signal.
6. the food recognition methods on described intelligent terminal according to claim 3, it is characterized in that, described abstraction, layer adopts the output of maximal value abstracting method to the convolutional layer after active coating process to sample, extraction mode is not overlapping extraction, extracting unit size is P*P, and the interval s of adjacent extraction unit is taken as 2.
7. the food recognition methods on described intelligent terminal according to claim 3, is characterized in that, described normalization layer adopts the method for LRN the output of abstraction, layer to be carried out to the normalization of regional area, and its function expression adopted is:
b x , y i = u x , y i ( k + α Σ j = m a x ( 0 , i - n / 2 ) min ( N - 1 , i + n / 2 ( u x , y j ) 2 ) β
Wherein, N is the sum of feature layer in this layer, and n is the feature layer number of neighbour, and k, α, β are constant, represent the activation unit at i-th feature layer place of (x, y) position; for right the activation unit exported after normalization; I is positive integer.
8. the food recognition methods on described intelligent terminal according to claim 1, it is characterized in that, described full articulamentum the 6th stage and full articulamentum the 7th stage all comprise full articulamentum, active coating and abandon layer; Described full articulamentum the 8th stage comprises full articulamentum, output layer and loss layer.
9. the food recognition methods on intelligent terminal according to claim 8, it is characterized in that, describedly abandon layer in the training process, network concealed node layer is in the mode of random chance, fractional weight is selected to calculate, and the not participation network renewal of the weighted value of unchecked hidden layer node, but its weighted value retains;
Described full articulamentum input value is launched into an one-dimensional vector X c, then proceed as follows:
Y c=w*X c+b
Wherein, X cinput vector, Y cbe output vector, w is weighted value, and b is bias;
Described output layer selects accuracy layer prediction accuracy; Described accuracy layer obtains final degree of accuracy according to contrasting the prediction class label of output vector and its true class label and add up;
The described loss layer Softmax function error of calculation, and by error back propagation to seek the optimum solution of each network layer weight and bias.
10. the food recognition methods on described intelligent terminal according to claim 1, it is characterized in that, training convolutional neural networks model described in step S1.3, is specially:
A, training set changed into the document format data that Caffe specifies;
The average of b, calculation training picture
C, define grid parameter and training parameter;
D, training network: iteration is carried out to the network model that S1.2 builds by adopting the method for stochastic gradient descent, every iteration once just detects a subgradient, with the optimum solution of the weight and bias of seeking each network layer, finally obtain the optimum convolutional neural networks structural model of this training;
E, test network: the network structure model utilizing Step d to train is tested the food picture in test set, obtain the accuracy of classifying.
CN201510862931.7A 2015-11-30 2015-11-30 Food recognition method at intelligent terminal Pending CN105512676A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510862931.7A CN105512676A (en) 2015-11-30 2015-11-30 Food recognition method at intelligent terminal

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510862931.7A CN105512676A (en) 2015-11-30 2015-11-30 Food recognition method at intelligent terminal

Publications (1)

Publication Number Publication Date
CN105512676A true CN105512676A (en) 2016-04-20

Family

ID=55720642

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510862931.7A Pending CN105512676A (en) 2015-11-30 2015-11-30 Food recognition method at intelligent terminal

Country Status (1)

Country Link
CN (1) CN105512676A (en)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106203330A (en) * 2016-07-08 2016-12-07 西安理工大学 A kind of vehicle classification method based on convolutional neural networks
CN106228177A (en) * 2016-06-30 2016-12-14 浙江大学 Daily life subject image recognition methods based on convolutional neural networks
CN106295707A (en) * 2016-08-17 2017-01-04 北京小米移动软件有限公司 Image-recognizing method and device
CN106529564A (en) * 2016-09-26 2017-03-22 浙江工业大学 Food image automatic classification method based on convolutional neural networks
CN106845527A (en) * 2016-12-29 2017-06-13 南京江南博睿高新技术研究院有限公司 A kind of vegetable recognition methods
CN106874913A (en) * 2016-12-29 2017-06-20 南京江南博睿高新技术研究院有限公司 A kind of vegetable detection method
CN106980817A (en) * 2017-02-27 2017-07-25 南京邮电大学 A kind of terrified video frequency identifying method based on Caffe frameworks
CN107024073A (en) * 2017-04-26 2017-08-08 中国石油大学(华东) Multi-sensor intelligent controlling method for refrigerator and intelligent refrigerator based on deep learning
CN107067043A (en) * 2017-05-25 2017-08-18 哈尔滨工业大学 A kind of diseases and pests of agronomic crop detection method
CN107480773A (en) * 2017-08-09 2017-12-15 北京小米移动软件有限公司 The method, apparatus and storage medium of training convolutional neural networks model
CN107742128A (en) * 2017-10-20 2018-02-27 百度在线网络技术(北京)有限公司 Method and apparatus for output information
WO2018040105A1 (en) * 2016-09-05 2018-03-08 合肥华凌股份有限公司 System and method for food recognition, food model training method, refrigerator and server
CN107798277A (en) * 2016-09-05 2018-03-13 合肥美的智能科技有限公司 Food materials identifying system and method, food materials model training method, refrigerator and server
CN107832804A (en) * 2017-10-30 2018-03-23 上海寒武纪信息科技有限公司 A kind of information processing method and Related product
CN107886108A (en) * 2017-10-11 2018-04-06 中国农业大学 Fruit and vegetable classification method and device based on AlexNet network models
CN108009525A (en) * 2017-12-25 2018-05-08 北京航空航天大学 A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks
CN108171318A (en) * 2017-11-30 2018-06-15 河南大学 One kind is based on the convolutional neural networks integrated approach of simulated annealing-Gaussian function
CN108229979A (en) * 2018-01-24 2018-06-29 亳州职业技术学院 Based on wechat small routine and the American Ginseng of depth learning technology tasting assistant
WO2018121690A1 (en) * 2016-12-29 2018-07-05 北京市商汤科技开发有限公司 Object attribute detection method and device, neural network training method and device, and regional detection method and device
CN108280474A (en) * 2018-01-19 2018-07-13 广州市派客朴食信息科技有限责任公司 A kind of food recognition methods based on neural network
CN108507270A (en) * 2017-02-24 2018-09-07 九阳股份有限公司 A kind of food materials of refrigerator determine method and food materials determining device
CN108647734A (en) * 2018-05-15 2018-10-12 上海达显智能科技有限公司 A kind of food image big data acquisition method, acquisition system and food recognition methods
CN108846419A (en) * 2018-05-25 2018-11-20 平安科技(深圳)有限公司 Single page high load image-recognizing method, device, computer equipment and storage medium
CN108875959A (en) * 2018-05-24 2018-11-23 四川斐讯信息技术有限公司 A kind of intelligence sees the method and system of object knowledge name
CN109214409A (en) * 2018-07-10 2019-01-15 上海斐讯数据通信技术有限公司 A kind of vegetable recognition methods and system
CN109214272A (en) * 2018-07-17 2019-01-15 北京陌上花科技有限公司 A kind of image-recognizing method and device
WO2019011249A1 (en) * 2017-07-14 2019-01-17 腾讯科技(深圳)有限公司 Method, apparatus, and device for determining pose of object in image, and storage medium
WO2020057455A1 (en) * 2018-09-17 2020-03-26 鲁班嫡系机器人(深圳)有限公司 Food material processing robot control method, apparatus, system, storage medium and device
CN111797719A (en) * 2020-06-17 2020-10-20 武汉大学 Food component identification method
US11990137B2 (en) 2018-09-13 2024-05-21 Shanghai Cambricon Information Technology Co., Ltd. Image retouching method and terminal device

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103544506A (en) * 2013-10-12 2014-01-29 Tcl集团股份有限公司 Method and device for classifying images on basis of convolutional neural network
CN104090871A (en) * 2014-07-18 2014-10-08 百度在线网络技术(北京)有限公司 Picture translation method and system
CN104346622A (en) * 2013-07-31 2015-02-11 富士通株式会社 Convolutional neural network classifier, and classifying method and training method thereof
CN104636757A (en) * 2015-02-06 2015-05-20 中国石油大学(华东) Deep learning-based food image identifying method

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104346622A (en) * 2013-07-31 2015-02-11 富士通株式会社 Convolutional neural network classifier, and classifying method and training method thereof
CN103544506A (en) * 2013-10-12 2014-01-29 Tcl集团股份有限公司 Method and device for classifying images on basis of convolutional neural network
CN104090871A (en) * 2014-07-18 2014-10-08 百度在线网络技术(北京)有限公司 Picture translation method and system
CN104636757A (en) * 2015-02-06 2015-05-20 中国石油大学(华东) Deep learning-based food image identifying method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KRIZHEVSKY A ETC,: ""ImageNet classification with deep convolutional neural networks"", 《INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING SYSTEMS.CURRAN ASSOCIATES INC》 *
王振 等,: ""基于卷积神经网络的图像识别算法设计与实现"", 《图形图像》 *

Cited By (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106228177A (en) * 2016-06-30 2016-12-14 浙江大学 Daily life subject image recognition methods based on convolutional neural networks
CN106203330A (en) * 2016-07-08 2016-12-07 西安理工大学 A kind of vehicle classification method based on convolutional neural networks
CN106295707A (en) * 2016-08-17 2017-01-04 北京小米移动软件有限公司 Image-recognizing method and device
CN106295707B (en) * 2016-08-17 2019-07-02 北京小米移动软件有限公司 Image-recognizing method and device
CN107798277A (en) * 2016-09-05 2018-03-13 合肥美的智能科技有限公司 Food materials identifying system and method, food materials model training method, refrigerator and server
WO2018040105A1 (en) * 2016-09-05 2018-03-08 合肥华凌股份有限公司 System and method for food recognition, food model training method, refrigerator and server
CN106529564A (en) * 2016-09-26 2017-03-22 浙江工业大学 Food image automatic classification method based on convolutional neural networks
CN106529564B (en) * 2016-09-26 2019-05-31 浙江工业大学 A kind of food image automatic classification method based on convolutional neural networks
CN106874913A (en) * 2016-12-29 2017-06-20 南京江南博睿高新技术研究院有限公司 A kind of vegetable detection method
CN106845527A (en) * 2016-12-29 2017-06-13 南京江南博睿高新技术研究院有限公司 A kind of vegetable recognition methods
WO2018121690A1 (en) * 2016-12-29 2018-07-05 北京市商汤科技开发有限公司 Object attribute detection method and device, neural network training method and device, and regional detection method and device
CN108507270B (en) * 2017-02-24 2020-07-10 九阳股份有限公司 Food material determining method and food material determining device for refrigerator
CN108507270A (en) * 2017-02-24 2018-09-07 九阳股份有限公司 A kind of food materials of refrigerator determine method and food materials determining device
CN106980817A (en) * 2017-02-27 2017-07-25 南京邮电大学 A kind of terrified video frequency identifying method based on Caffe frameworks
CN107024073A (en) * 2017-04-26 2017-08-08 中国石油大学(华东) Multi-sensor intelligent controlling method for refrigerator and intelligent refrigerator based on deep learning
CN107067043B (en) * 2017-05-25 2020-07-24 哈尔滨工业大学 Crop disease and insect pest detection method
CN107067043A (en) * 2017-05-25 2017-08-18 哈尔滨工业大学 A kind of diseases and pests of agronomic crop detection method
US11107232B2 (en) 2017-07-14 2021-08-31 Tencent Technology (Shenzhen) Company Limited Method and apparatus for determining object posture in image, device, and storage medium
WO2019011249A1 (en) * 2017-07-14 2019-01-17 腾讯科技(深圳)有限公司 Method, apparatus, and device for determining pose of object in image, and storage medium
CN107480773B (en) * 2017-08-09 2020-11-13 北京小米移动软件有限公司 Method and device for training convolutional neural network model and storage medium
CN107480773A (en) * 2017-08-09 2017-12-15 北京小米移动软件有限公司 The method, apparatus and storage medium of training convolutional neural networks model
CN107886108A (en) * 2017-10-11 2018-04-06 中国农业大学 Fruit and vegetable classification method and device based on AlexNet network models
CN107742128A (en) * 2017-10-20 2018-02-27 百度在线网络技术(北京)有限公司 Method and apparatus for output information
US11922132B2 (en) 2017-10-30 2024-03-05 Shanghai Cambricon Information Technology Co., Ltd. Information processing method and terminal device
CN107832804A (en) * 2017-10-30 2018-03-23 上海寒武纪信息科技有限公司 A kind of information processing method and Related product
CN108171318A (en) * 2017-11-30 2018-06-15 河南大学 One kind is based on the convolutional neural networks integrated approach of simulated annealing-Gaussian function
CN108009525A (en) * 2017-12-25 2018-05-08 北京航空航天大学 A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks
CN108009525B (en) * 2017-12-25 2018-10-12 北京航空航天大学 A kind of specific objective recognition methods over the ground of the unmanned plane based on convolutional neural networks
CN108280474A (en) * 2018-01-19 2018-07-13 广州市派客朴食信息科技有限责任公司 A kind of food recognition methods based on neural network
CN108229979A (en) * 2018-01-24 2018-06-29 亳州职业技术学院 Based on wechat small routine and the American Ginseng of depth learning technology tasting assistant
CN108647734A (en) * 2018-05-15 2018-10-12 上海达显智能科技有限公司 A kind of food image big data acquisition method, acquisition system and food recognition methods
CN108647734B (en) * 2018-05-15 2022-03-11 上海达显智能科技有限公司 Food image big data acquisition method, acquisition system and food identification method
CN108875959A (en) * 2018-05-24 2018-11-23 四川斐讯信息技术有限公司 A kind of intelligence sees the method and system of object knowledge name
CN108846419A (en) * 2018-05-25 2018-11-20 平安科技(深圳)有限公司 Single page high load image-recognizing method, device, computer equipment and storage medium
CN109214409A (en) * 2018-07-10 2019-01-15 上海斐讯数据通信技术有限公司 A kind of vegetable recognition methods and system
CN109214272A (en) * 2018-07-17 2019-01-15 北京陌上花科技有限公司 A kind of image-recognizing method and device
US11990137B2 (en) 2018-09-13 2024-05-21 Shanghai Cambricon Information Technology Co., Ltd. Image retouching method and terminal device
US11996105B2 (en) 2018-09-13 2024-05-28 Shanghai Cambricon Information Technology Co., Ltd. Information processing method and terminal device
WO2020057455A1 (en) * 2018-09-17 2020-03-26 鲁班嫡系机器人(深圳)有限公司 Food material processing robot control method, apparatus, system, storage medium and device
CN111797719A (en) * 2020-06-17 2020-10-20 武汉大学 Food component identification method

Similar Documents

Publication Publication Date Title
CN105512676A (en) Food recognition method at intelligent terminal
CN108875674B (en) Driver behavior identification method based on multi-column fusion convolutional neural network
CN108388927B (en) Small sample polarization SAR terrain classification method based on deep convolution twin network
Hossain et al. Improving consumer satisfaction in smart cities using edge computing and caching: A case study of date fruits classification
CN102314614B (en) Image semantics classification method based on class-shared multiple kernel learning (MKL)
CN110598598A (en) Double-current convolution neural network human behavior identification method based on finite sample set
CN105320961A (en) Handwriting numeral recognition method based on convolutional neural network and support vector machine
CN106372648A (en) Multi-feature-fusion-convolutional-neural-network-based plankton image classification method
CN109344884A (en) The method and device of media information classification method, training picture classification model
Zhang et al. Unsupervised difference representation learning for detecting multiple types of changes in multitemporal remote sensing images
CN104182772A (en) Gesture recognition method based on deep learning
CN104537647A (en) Target detection method and device
CN112883839A (en) Remote sensing image interpretation method based on adaptive sample set construction and deep learning
CN108537119A (en) A kind of small sample video frequency identifying method
WO2022062419A1 (en) Target re-identification method and system based on non-supervised pyramid similarity learning
CN107169106A (en) Video retrieval method, device, storage medium and processor
CN108764084A (en) Video classification methods based on spatial domain sorter network and the time domain network integration
CN109492596A (en) A kind of pedestrian detection method and system based on K-means cluster and region recommendation network
CN113569895A (en) Image processing model training method, processing method, device, equipment and medium
CN106778910A (en) Deep learning system and method based on local training
CN109241392A (en) Recognition methods, device, system and the storage medium of target word
CN110503082A (en) A kind of model training method and relevant apparatus based on deep learning
CN109978074A (en) Image aesthetic feeling and emotion joint classification method and system based on depth multi-task learning
CN113109782A (en) Novel classification method directly applied to radar radiation source amplitude sequence
CN113822130A (en) Model training method, scene recognition method, computing device, and medium

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160420