CN107016415B - A kind of color image Color Semantic classification method based on full convolutional network - Google Patents

A kind of color image Color Semantic classification method based on full convolutional network Download PDF

Info

Publication number
CN107016415B
CN107016415B CN201710236763.XA CN201710236763A CN107016415B CN 107016415 B CN107016415 B CN 107016415B CN 201710236763 A CN201710236763 A CN 201710236763A CN 107016415 B CN107016415 B CN 107016415B
Authority
CN
China
Prior art keywords
color
characteristic pattern
color image
image
pixel
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201710236763.XA
Other languages
Chinese (zh)
Other versions
CN107016415A (en
Inventor
张骏
熊高敏
高隽
张旭东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN201710236763.XA priority Critical patent/CN107016415B/en
Publication of CN107016415A publication Critical patent/CN107016415A/en
Application granted granted Critical
Publication of CN107016415B publication Critical patent/CN107016415B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Image Analysis (AREA)

Abstract

The color image Color Semantic classification method based on full convolutional network that the invention discloses a kind of, comprising: the 1 full convolutional network of building;2 obtain the color image data collection with Pixel-level mark;3 are trained full convolutional network using color image data collection, obtain the characteristic model that the classification of Pixel-level Color Semantic can be carried out to arbitrary dimension color image;4 are classified using the Color Semantic that characteristic model carries out Pixel-level to any color image, assess the nicety of grading of characteristic model;5 optimize processing to network class result using the method for full condition of contact random field, obtain the colors label of each pixel in image, according to the mapping relations of class label and color space, by the label converting Color Semantic classification results to corresponding color space display pixel grade of colors.The present invention is able to achieve the Color Semantic classification of color image pixel grade, effectively improves the precision that color image Color Semantic is classified under environment complicated and changeable.

Description

A kind of color image Color Semantic classification method based on full convolutional network
Technical field
The invention belongs in computer/machine vision, image procossing and analysis field, it is specifically a kind of based on full volume The color image Color Semantic classification method of product network.
Background technique
In computer vision, it is the important channel of human perception image information that color, which is a kind of important attribute of image,. By assigning image color class label, it can further apply to image retrieval, image labeling, colour blindness auxiliary, visual pursuit, language Say the fields such as human-computer interaction.Therefore, good Color Semantic classification results facilitate further image procossing and image analysis.
Existing image color semantic classification method includes: the method based on statistical model and the side based on deep learning Method.
Method based on statistical model is mainly based upon chromatic stimulus, such as by the tone of perceived color, brightness and satisfies Color Semantic classification is carried out with these three color appearance attributes are spent.Class inherited of the different color on these three attributes is less than same Kind color difference in the class on these three attributes causes the decision boundaries of Color Semantic classification to be difficult to determine, and in different fields There are minor changes under scape, cause classifying quality poor.
In the method based on statistical model, representative method is the hidden semantic analysis (Probabilistic of probability Latent SemanticAnalysis, PLSA).This method is on Lab color space by the word finder and group of composition model document It is mapped at each color pixel collection of color image, regards colors as hiding main body variable;Use natural scene Image training PLSA model obtains the colors hidden in pictures, to realize that Color Semantic is classified.Natural scene image The quality of training set determines the accuracy of model, therefore under the complex scenes such as bloom, shade, the classifying quality of the model is simultaneously It is undesirable.
Method based on deep learning mainly utilizes convolutional neural networks (Convolutional Neural Network, CNN) carry out Color Semantic classification.This method learns colors from great amount of images sample, it is intended to predict small The colors of image block.This method training convolutional neural networks include two stages: first stage, are carried out certainly using image block Supervised training, color histogram is as its supervision message, to obtain Color Semantic disaggregated model;Second stage, image block Label is changed to inherit from dad image and obtain, and by iterating, filters out noise sample, further increases Color Semantic classification mould The precision of type.However under complex scene, the variation of intensity of illumination, light source etc. will lead to pixel rgb value in a channel or It changes on multiple channels, the color histogram obtained at this time cannot really react the COLOR COMPOSITION THROUGH DISTRIBUTION of image block, by it It is ineffective that the Color Semantic disaggregated model come is trained as supervision message;In second stage, partial color semantic classification Successful image block is given up to fall because of inconsistent with the colors of dad image by mistake, to reduce available Amount of training data.
Cheng et al. is improved in the method for convolutional neural networks, constructs the network structure of PCN-CNN, the net Full articulamentum is changed to convolutional layer, the quantity for reducing pond layer, addition one on the basis of VGG16 network structure by network structure Warp lamination realizes the color image pixel grade Color Semantic classification based on convolutional neural networks.Method compared with PLSA is in picture It improves a lot in plain grade Color Semantic nicety of grading, but the model uses less pond layer, to the biggish coloured silk of resolution ratio It is low that chromatic graph picture carries out precision when Color Semantic classification.
The Color Semantic that above-mentioned Color Semantic classification method realizes color image is classified, however point under complex scene Class precision is not high.
Summary of the invention
The present invention is to solve shortcoming in the prior art, provides a kind of color image color based on full convolutional network Semantic classification method, it is intended to the color attribute semantic classification for solving the problems, such as color image pixel grade, by constructing full convolutional Neural Network obtains the good color attribute semantic classification network characterization model of the nicety of grading under complex scene, to improve complexity The precision that color image Color Semantic is classified under changeable environment.
In order to solve the above technical problems, the technical solution adopted by the present invention is that:
A kind of the characteristics of Color Semantic classification method of color image based on full convolutional network of the present invention is according to as follows Step carries out:
Step 1, the full convolutional network of building, and Pixel-level Color Semantic is carried out to the color image I (x, y, k) of arbitrary dimension Classification;The full convolutional network is made of convolutional layer, pond layer, warp lamination;
Step 1.1, the input data of the full convolutional network are color image I (x, y, k), and x, y, k respectively indicates described Height, width and the port number of color image I;
Defining operation number is t, and initializes t=1;
Definition pond number is v, and initializes v=1;
Initialize δ=0;
The color image I is sent into convolutional layer L by step 1.2c(ks,nt) the t times convolution operation is carried out, it obtains described complete The characteristic pattern F of t-th of convolutional layer of convolutional networkc(x,y,nt);Ks indicates the size of convolution kernel, ntIndicate t-th of convolutional layer Convolution kernel number;
Step 1.3, to the characteristic pattern F of t-th of convolutional layerc(x,y,nt) the activation operation of t sublinear is carried out, obtain feature Scheme Fr(x,y,nt);
Step 1.4 after enabling t+1 be assigned to t, judges whether t≤δ+2 is true, if so, then repeat step 1.2 and step 1.3, otherwise, after enabling t-1 be assigned to t, execute step 1.5;
The t sublinear is activated the characteristic pattern F after operation by step 1.5r(x,y,nt) it is sent into pond layer Lp(kp,s) After carrying out non-overlapping sampling, the characteristic pattern F in v stage is obtainedv(x,y,nv), wherein the range that kp, s respectively indicate sampling is big Small and sliding step, nvIndicate the quantity of the characteristic pattern in v stage;
Step 1.6 enables δ+2 be assigned to δ, judges whether δ=4 are true, if so, then follow the steps 1.7;Otherwise, by v+1 After being assigned to v, return step 1.4;
Step 1.7 after enabling t+1 be assigned to t, judges whether t≤δ+3 is true, if so, then repeat step 1.2 and step Step 1.7 is returned again to after 1.3, otherwise, t-1 is enabled to be assigned to t, after v+1 is assigned to v, jumps to step again after executing step 1.5 1.8;
Step 1.8 enables δ+3 be assigned to δ, judges whether δ=13 are true, if so, then follow the steps 1.9;Otherwise, t+ is enabled After 1 is assigned to t, after repeating step 1.2 and step 1.3, return step 1.7;
Step 1.9, the characteristic pattern F to the v stagev(x,y,nv) convolution operation is carried out, obtained result carries out deconvolution behaviour Make, to obtain characteristic pattern Fv′(x,y,nv);
To the characteristic pattern F in v-1 stagev-1(x,y,nv-1) carry out convolution operation after with characteristic pattern Fv′(x,y,nv) be added, Deconvolution is carried out again to operate to obtain characteristic pattern F 'v-1(x,y,nv-1);
To the characteristic pattern F in v-2 stagev-2(x,y,nv-2) carry out convolution operation after with characteristic pattern F 'v-1(x,y,nv-1) phase Add, then carries out deconvolution and operate to obtain characteristic pattern F 'v-2(x,y,nv-2);
Step 1.10, according to the size of the color image I (x, y, k), to characteristic pattern Fv′(x,y,nv) cut, obtain coloured silk Chromatic graph is denoted as C (x, y, p as the corresponding colors probability of I (x, y, k)i), wherein piThe probability of i-th of colors is represented, I ∈ (0,1 ..., N), N are the quantity of colors;
Step 2 obtains the color image data collection with Pixel-level mark, is denoted as D={ Dtr,Dval,Dte};DtrIndicate instruction Practice data set;DvalIndicate validation data set;DteIndicate test data set;The training dataset Dtr, validation data set DvalWith Test data set DteComprising color image and mark image;And it is described mark image in each pixel value V ∈ (0, 1,...,N);
The training dataset DtrWith validation data set DvalIn all color images tri- channels RGB mean value, be denoted as M={ mr,mg,mb};
Step 3 is trained the full convolutional network;
Step 3.1, to the training dataset DtrIn all color images add σ1Gaussian noise, for changing colour The contrast and rgb value of image;σ is added to all color images again2Gaussian noise for changing brightness, to obtain colour Changing image;
Step 3.2, with the training dataset DtrIn all color images and its corresponding color transformation image as institute The input data of full convolutional network training stage is stated, with the training dataset DtrIn all mark images as the input number According to label;And color image and its color transformation image use identical colors label;
With the validation data set DvalIn test number of all color images as the full convolutional network training stage According to the validation data set DvalIn all labels of the mark images as the test data;
Step 3.3 is trained the full convolutional network using stochastic gradient descent algorithm, obtains the layer of Color Semantic Secondary characteristic model;
Step 4, with the test data set DteIn all color images and the mean value M difference as the character modules The input data of type, and by the calculating of the characteristic model, obtain the colors probability of every width color image;
Step 5 optimizes the colors probability of every width color image using the method for full condition of contact random field Processing, obtains the test data set DteThe colors of each pixel of middle color image, then by the color of each pixel It is color class switching into color space, to obtain the Color Semantic classification results of Pixel-level.
Compared with the prior art, the beneficial effects of the present invention are:
1, the present invention supervises full convolutional network using label data and is learnt, to obtain by constructing full convolutional network Must have the Color Semantic characteristic of division model of robustness, solve based in statistical model use color appearance attribute carry out color Coloured silk perception, the problem that classification boundaries are complicated and nicety of grading is not high.
2, the full convolutional network constructed by the present invention reduces network size without the network structure of full articulamentum, thus The huge parameter training in traditional convolutional neural networks to full articulamentum is avoided, training speed is greatly improved.Reasonably Network structure can handle the smaller and biggish image of resolution ratio;Color Semantic classification results are extended to Pixel-level, solve volume The problem of product neural network can not carry out Pixel-level classification, and Pixel-level Color Semantic nicety of grading greatly improved.
3, the present invention slightly changes color, the brightness of training data, for simulating in the full convolutional network training stage Complex scene increases the type and quantity of training data to change caused by color image, limited in training dataset In the case of, as far as possible the environmental factors such as illumination simulation condition, shooting visual angle on Color Semantic classification caused by influence, solve because For poor fitting problem caused by training data deficiency during training network.
4, the present invention is classified using the Color Semantic that full convolutional network carries out Pixel-level to color image, utilizes full connection strap Local color semantic association between the method combination pixel of part random field, to the Color Semantic classification results of full convolutional network It optimizes, corrects for the pixel of classification error on color area profile, keep classifying edge more smooth, improve color language The accuracy of justice classification.
Detailed description of the invention
Fig. 1 is the flow chart that the method for the present invention carries out Color Semantic classification to color image;
Fig. 2 is the structural schematic diagram of full convolutional network used herein;
Fig. 3 a is the schematic diagram of the part subset in the color image test set that the present invention uses;
Authentic signature schematic diagram corresponding to the part subset in color image test set that Fig. 3 b uses for the present invention;
Fig. 3 c is the partial color semantic classification result schematic diagram that the present invention is obtained using full convolutional network;
Fig. 3 d is that the partial color semantic classification result obtained after the present invention is optimized using full connection random field is illustrated Figure.
Specific embodiment
As shown in Figure 1, in the present embodiment, a kind of Color Semantic classification method of the color image based on full convolutional network, It is to carry out in accordance with the following steps:
Step 1, the full convolutional network of building, carry out Pixel-level color language for the color image I (x, y, k) to arbitrary dimension Justice classification;Full convolutional network is made of convolutional layer, pond layer, warp lamination,
As shown in Fig. 2, in the present embodiment, full convolutional network includes the convolution pondization operation in five stages: the first and second rank Section respectively includes two convolutional layers, a pond layer;Third and fourth, five stages respectively include three convolutional layers, a pond layer.Full volume Product network shares 13 convolutional layers, five pond layers, three warp laminations.The arbitrary dimension size of color image refers to: sending The size for entering the color image of network needs not be a fixed size;The color image that full convolutional network can be handled is most Large scale is determined by the size of the video memory of used graphics processor.
Step 1.1, the input data of full convolutional network are color image I (x, y, k), and x, y, k respectively indicates color image I Height, width and port number;
Defining operation number is t, and initializes t=1;
Definition pond number is v, and initializes v=1;
Initialize δ=0;
In this experiment test, the height of color image, width are respectively x=128, y=64, but not limited to this it takes Value;Port number k=3 indicates red R, the green G, tri- channels blue B of color image.
Color image I is sent into convolutional layer L by step 1.2c(ks,nt) and the t times convolution operation is carried out to it, it is rolled up entirely The characteristic pattern F of t-th of convolutional layer of product networkc(x,y,nt);Ks indicates the size of convolution kernel, ntIndicate the volume of t-th of convolutional layer Product core number;
Characteristic pattern F after convolutional layerc(x,y,nt) size are as follows: x=(x+2-ks)+1, y=(y+2-ks)+1. In this experiment test, size that the convolution kernel size ks=3 of all convolutional layers, i.e. convolution operation do not change characteristic pattern;t When≤2, convolution kernel number nt=64;When 2 < t≤4, convolution kernel number nt=128;When 4 < t≤7, convolution kernel number nt= 256;When 7 < t≤13, convolution kernel number nt=512;
Step 1.3, to the characteristic pattern F of t-th of convolutional layerc(x,y,nt) the activation operation of t sublinear is carried out, obtain feature Scheme Fr(x,y,nt);
Linear activation operate used in activation primitive be ReLU, the activation primitive is by characteristic pattern Fc(x,y,nt) in negative value Element value sets 0, keeps positive value element value constant;Characteristic pattern F will not be changedc(x,y,nt) size.
Step 1.4 after enabling t+1 be assigned to t, judges whether t≤δ+2 is true, if so, then repeat step 1.2 and step 1.3, otherwise, after enabling t-1 be assigned to t, execute step 1.5;
T sublinear is activated the characteristic pattern F after operation by step 1.5r(x,y,nt) it is sent into pond layer Lp(kp, s) is carried out After non-overlapping sampling, the characteristic pattern F in v stage is obtainedv(x,y,nv), wherein kp, s respectively indicate sampling range size and Sliding step, nvIndicate the quantity of the characteristic pattern in v stage;
In this experiment test, pond layer carries out the range size kp=2 of non-overlapping sampling, sliding step s=2;Chi Hua Layer takes characteristic pattern F of the maximum value pond after the sampling of pond layerv(x,y,nv) size are as follows:The characteristic pattern quantity in each stage and the convolution kernel number in corresponding stage are equal: when v=1, nv =64;When v=2, nv=128;When v=3, nv=256;When 4≤v≤5, nv=512;
Step 1.6 enables δ+2 be assigned to δ, judges whether δ=4 are true, if so, then follow the steps 1.7;Otherwise, by v+1 After being assigned to v, return step 1.4;
Step 1.7 after enabling t+1 be assigned to t, judges whether t≤δ+3 is true, if so, then repeat step 1.2 and step Step 1.7 is returned again to after 1.3, otherwise, t-1 is enabled to be assigned to t, after v+1 is assigned to v, jumps to step again after executing step 1.5 1.8;
Step 1.8 enables δ+3 be assigned to δ, judges whether δ=13 are true, if so, then follow the steps 1.9;Otherwise, t+ is enabled After 1 is assigned to t, after repeating step 1.2 and step 1.3, return step 1.7;
Step 1.9, the characteristic pattern F to the v stagev(x,y,nv) convolution operation is carried out, obtained result carries out deconvolution behaviour Make, to obtain characteristic pattern Fv′(x,y,nv);
To the characteristic pattern F in v-1 stagev-1(x,y,nv-1) carry out convolution operation after with characteristic pattern Fv′(x,y,nv) be added, Deconvolution is carried out again to operate to obtain characteristic pattern F 'v-1(x,y,nv-1);
To the characteristic pattern F in v-2 stagev-2(x,y,nv-2) carry out convolution operation after with characteristic pattern F 'v-1(x,y,nv-1) phase Add, then carries out deconvolution and operate to obtain characteristic pattern F 'v-2(x,y,nv-2);
Deconvolution operation is the inverse process of convolution operation, the size of gained characteristic pattern after deconvolution operates are as follows: x =(x-1) × ks+pad, y=(y-1) × ks+pad, in this experiment test, the deconvolution core in v stage and v-1 stage Size ks=2, pad=4;The deconvolution core size k in v-2 stages=8, pad=16.The convolution kernel number n of convolution operationt= 12, the convolution kernel number n of deconvolution operationv=12.Before two width characteristic patterns are added, need to cut out the big characteristic pattern of size It cuts, keeps the size of two width characteristic patterns consistent.
Step 1.10, to characteristic pattern F 'v-2(x,y,nv-2) cut, keep it in the same size with color image I (x, y, k), The corresponding colors probability of color image I (x, y, k) is obtained, C (x, y, p are denoted asi), wherein piRepresent i-th of colors Probability, i ∈ (0,1 ..., N), N be colors quantity;
In this experiment test, the mode of cutting is random cropping, the quantity N=11 of colors, this 11 kinds of colors point Not are as follows: black, blue, brown, grey, green, orange, pink, purple, red, white, yellow.It is specific fixed Justice refers to " Basic colorterms:Theiruniversality and evolution ", and the book is in 1991 by University of California Publishing house publishes.
Step 2 obtains the color image data collection with Pixel-level mark, is denoted as D={ Dtr,Dval,Dte};DtrIndicate instruction Practice data set;DvalIndicate validation data set;DteIndicate test data set;Training dataset Dtr, validation data set DvalAnd test Data set DteComprising color image and mark image;And mark the value V ∈ (0,1 ..., N) of each pixel in image;
Training dataset DtrWith validation data set DvalIn all color images tri- channels RGB mean value, be denoted as M= {mr,mg,mb};
Training dataset picture number is 10913, and validation data set picture number is 1500, test data set picture number It is 1800.In this experiment test, mark image is to assign colors label V, V ∈ to pixel each in color image (0,1 ..., 11), respectively indicates the color of the pixel are as follows: dark cyan, black, blue, brown, grey, green, orange,pink,purple,red,white,yellow;Wherein dark cyan indicates that the pixel is background.Data set Mean value M={ 93.53614,96.15632,102.91466 }.It as shown in Figure 3a, is the portion in color image test set of the present invention Divide color image;As shown in Figure 3b, the corresponding Pixel-level mark figure of concentrated part color image is tested for color image of the present invention Picture.The region that the present invention carries out Color Semantic classification to color image does not include background, hair, skin;These regions are unified to assign Give colors label 0.
Step 3 is trained full convolutional network;
Step 3.1, to training dataset DtrIn all color images add σ1Gaussian noise to change color image Contrast and rgb value;σ is added to all color images again2Gaussian noise for changing brightness, to obtain color transformation figure Picture;In the present embodiment, σ1∈ [0,0.04], σ2=0.2;
Step 3.2, with training dataset DtrIn all color images and its color transformation image instructed as full convolutional network Practice the input data in stage, with training dataset DtrIn all labels of the mark images as input data;And color image and Its color transformation image uses identical label;
With validation data set DvalIn test data of all color images as the full convolutional network training stage, with verifying Data set DvalIn all labels of the mark images as test data;
Step 3.3, using stochastic gradient descent algorithm (Stochastic Gradient Descent, SGD) to full convolution Network is trained, and obtains the level characteristics model of Color Semantic;
During network repetitive exercise, each random one sub-picture of selection of stochastic gradient descent algorithm It practises, Lai Gengxin model parameter.Specific method can refer to " the Stochastic GradientDescent of LeonBottou Tricks ", this article are published in the 421-436 pages of " Neural networks:Tricks ofthe trade " in 2012.
Step 4, with test data set DteIn all color images and mean value M input number of the difference as characteristic model According to, and by the calculating of characteristic model, obtain the colors probability of every width color image;
Colors of the color label as the pixel corresponding to maximum probability value in assumed appearance coloured silk class probability are such as schemed Shown in 3c, for the Pixel-level colors for the color image for using full convolutional network to obtain, the value V ∈ of each pixel (0, 1 ..., 11), indicate the color of the pixel are as follows: dark cyan, black, blue, brown, grey, green, orange, pink,purple,red,white,yellow.Compared with Fig. 3 c authentic signature shown in Fig. 3 d, the Pixel-level color of color image Color classification and authentic signature are substantially uniform, but there are classification errors on some zonules.
Step 5 optimizes the colors probability of every width color image using the method for full condition of contact random field Processing, obtains test data set DteThe colors of each pixel of middle color image, then by the color class of each pixel It is not transformed into color space, to obtain the Color Semantic classification results of Pixel-level.
The pass of each pixel He other each pixels is described in full condition of contact random field using binary potential function System encourages similar pixel to distribute identical label, that is, is classified as same color.Specific method can refer to paper " Efficient Inference in Fully Connected CRFs with Gaussian Edge Potentials ", this article is in 2011 Year is published in international conference " the Advances in neural informationprocessing systems " phase the 4th of volume 23 Page.It as shown in Figure 3d, is the Pixel-level color of the color image obtained after using full condition of contact Random Fields Method to optimize Classification, the value of each pixel be V ∈ (0,1 ..., 11), indicate the color of the pixel are as follows: dark cyan, black, blue,brown,grey,green,orange,pink,purple,red,white,yellow.Final Pixel-level color language Adopted classification results need colors being transformed into RGB color space.Fig. 3 d is compared with Fig. 3 c, corrects for the mistake of some zonules Accidentally Color Semantic classification, profile of the classification boundaries closer to color image;It is compared with the authentic signature of Fig. 3 b, the color of the same area Coloured silk is unified, and Color Semantic classification boundaries are smooth, clear-cut.Colors and RGB value there are one-to-one transforming relationship, In this experiment test, the relationship of colors label, rgb value and color is as shown in table 1.
Table 1
Table 2
Color image Color Semantic classification method Ours Cheng et al. PLSA
PNS 90.1 74.3 63.1
It as shown in table 2, is color image Color Semantic classification method of the invention with " pixel mark average " (Pixel Annotation Score, PNS) is module, is carried out with current existing color image Color Semantic classification method Quantify the analytical table of comparison.PNS value is bigger, shows that Color Semantic classification results precision is higher.In table 2, Ours indicates this hair Bright color image Color Semantic classification method;Cheng et al. indicates to use improved convolutional neural networks method;PLSA is indicated The hidden semantic analysis of probability.3 kinds of methods are tested on the color image test set that the present invention uses, it will thus be seen that When being tested using small in resolution image, the method for method of the invention than Cheng et al. has in precision Significantly promoted;It is compared with PLSA method, then there is absolute predominance.

Claims (1)

1. a kind of Color Semantic classification method of the color image based on full convolutional network, it is characterized in that in accordance with the following steps into Row:
Step 1, the full convolutional network of building, and Pixel-level Color Semantic point is carried out to the color image I (x, y, k) of arbitrary dimension Class;The full convolutional network is made of convolutional layer, pond layer, warp lamination;
Step 1.1, the input data of the full convolutional network are color image I (x, y, k), and x, y, k respectively indicates the colour Height, width and the port number of image I;
Defining operation number is t, and initializes t=1;
Definition pond number is v, and initializes v=1;
Initialize δ=0;
The color image I is sent into convolutional layer L by step 1.2c(ks,nt) the t times convolution operation is carried out, obtain the full convolution The characteristic pattern F of t-th of convolutional layer of networkc(x,y,nt);Ks indicates the size of convolution kernel, ntIndicate the convolution of t-th of convolutional layer Core number;
Step 1.3, to the characteristic pattern F of t-th of convolutional layerc(x,y,nt) the activation operation of t sublinear is carried out, obtain characteristic pattern Fr (x,y,nt);
Step 1.4 after enabling t+1 be assigned to t, judges whether t≤δ+2 is true, if so, step 1.2 and step 1.3 are then repeated, Otherwise, after enabling t-1 be assigned to t, step 1.5 is executed;
The t sublinear is activated the characteristic pattern F after operation by step 1.5r(x,y,nt) it is sent into pond layer Lp(kp, s) is carried out After non-overlapping sampling, the characteristic pattern F in v stage is obtainedv(x′,y′,nv), wherein kp, s respectively indicate the range size of sampling And sliding step, nvIndicate the quantity of the characteristic pattern in v stage;The height of the characteristic pattern in x ' expression v stage, y ' expression v The width of the characteristic pattern in stage, and
Step 1.6 enables δ+2 be assigned to δ, judges whether δ=4 are true, if so, then follow the steps 1.7;Otherwise, by v+1 assignment After v, return step 1.4;
Step 1.7 after enabling t+1 be assigned to t, judges whether t≤δ+3 is true, if so, after then repeating step 1.2 and step 1.3 Step 1.7 is returned again to, otherwise, t-1 is enabled to be assigned to t, after v+1 is assigned to v, jumps to step 1.8 again after executing step 1.5;
Step 1.8 enables δ+3 be assigned to δ, judges whether δ=13 are true, if so, then follow the steps 1.9;Otherwise, t+1 is enabled to assign It is worth to after t, after repeating step 1.2 and step 1.3, return step 1.7;
Step 1.9, the characteristic pattern F to the v stagev(x′,y′,nv) convolution operation is carried out, obtained result carries out deconvolution behaviour Make, to obtain characteristic pattern Fv′(x″,y″,nv);X " indicates characteristic pattern Fv' height, y " indicate characteristic pattern Fv' width, and X "=(x-1) × ks+ pad, y "=(y-1) × ks+pad;
To the characteristic pattern F in v-1 stagev-1(x′,y′,nv-1) carry out convolution operation after with characteristic pattern Fv′(x″,y″,nv) be added, Deconvolution is carried out again to operate to obtain characteristic pattern F 'v-1(x″,y″,nv-1);
To the characteristic pattern F in v-2 stagev-2(x′,y′,nv-2) carry out convolution operation after with characteristic pattern F 'v-1(x″,y″,nv-1) phase Add, then carries out deconvolution and operate to obtain characteristic pattern F 'v-2(x″,y″,nv-2);
Step 1.10, according to the size of the color image I (x, y, k), to characteristic pattern Fv′(x″,y″,nv) cut, obtain colour The corresponding colors probability of image I (x, y, k), is denoted as C (x, y, pi), wherein piRepresent the probability of i-th of colors, i ∈ (0,1 ..., N), N are the quantity of colors;
Step 2 obtains the color image data collection with Pixel-level mark, is denoted as D={ Dtr,Dval,Dte};DtrIndicate training number According to collection;DvalIndicate validation data set;DteIndicate test data set;The training dataset Dtr, validation data set DvalAnd test Data set DteComprising color image and mark image;And it is described mark image in each pixel value V ∈ (0,1 ..., N);
The training dataset DtrWith validation data set DvalIn all color images tri- channels RGB mean value, be denoted as M= {mr,mg,mb};
Step 3 is trained the full convolutional network;
Step 3.1, to the training dataset DtrIn all color images add σ1Gaussian noise, for changing color image Contrast and rgb value;σ is added to all color images again2Gaussian noise for changing brightness, to obtain color transformation Image;
Step 3.2, with the training dataset DtrIn all color images and its corresponding color transformation image as described complete The input data of convolutional network training stage, with the training dataset DtrIn all mark images as the input data Label;And color image and its color transformation image use identical colors label;
With the validation data set DvalIn test data of all color images as the full convolutional network training stage, with The validation data set DvalIn all labels of the mark images as the test data;
Step 3.3 is trained the full convolutional network using stochastic gradient descent algorithm, and the level for obtaining Color Semantic is special Levy model;
Step 4, with the test data set DteIn all color images and the mean value M difference as the characteristic model Input data, and by the calculating of the characteristic model, obtain the colors probability of every width color image;
Step 5 optimizes processing to the colors probability of every width color image using the method for full condition of contact random field, Obtain the test data set DteThe colors of each pixel of middle color image, then by the colors of each pixel It is transformed into color space, to obtain the Color Semantic classification results of Pixel-level.
CN201710236763.XA 2017-04-12 2017-04-12 A kind of color image Color Semantic classification method based on full convolutional network Expired - Fee Related CN107016415B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710236763.XA CN107016415B (en) 2017-04-12 2017-04-12 A kind of color image Color Semantic classification method based on full convolutional network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710236763.XA CN107016415B (en) 2017-04-12 2017-04-12 A kind of color image Color Semantic classification method based on full convolutional network

Publications (2)

Publication Number Publication Date
CN107016415A CN107016415A (en) 2017-08-04
CN107016415B true CN107016415B (en) 2019-07-19

Family

ID=59446649

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710236763.XA Expired - Fee Related CN107016415B (en) 2017-04-12 2017-04-12 A kind of color image Color Semantic classification method based on full convolutional network

Country Status (1)

Country Link
CN (1) CN107016415B (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107527351B (en) * 2017-08-31 2020-12-29 华南农业大学 Lactating sow image segmentation method fusing FCN and threshold segmentation
CN107992850B (en) * 2017-12-20 2020-01-14 大连理工大学 Outdoor scene three-dimensional color point cloud classification method
US11636665B2 (en) 2018-01-15 2023-04-25 Shenzhen Corerain Technologies Co., Ltd. Streaming image semantic segmentation method, logical integrated circuit system and electronic device
CN108830870B (en) * 2018-05-21 2021-12-28 千寻位置网络有限公司 Satellite image high-precision farmland boundary extraction method based on multi-scale structure learning
CN109063723B (en) * 2018-06-11 2020-04-28 清华大学 Weak supervision image semantic segmentation method based on common features of iteratively mined objects
CN109102457B (en) * 2018-06-12 2023-01-17 杭州米绘科技有限公司 Intelligent color changing system and method based on convolutional neural network
CN109379572B (en) 2018-12-04 2020-03-06 北京达佳互联信息技术有限公司 Image conversion method, image conversion device, electronic equipment and storage medium
CN110263151B (en) * 2019-05-06 2023-02-07 广东工业大学 Latent semantic learning method for multi-view multi-label data
CN110602527B (en) 2019-09-12 2022-04-08 北京小米移动软件有限公司 Video processing method, device and storage medium
CN111179272B (en) * 2019-12-10 2024-01-05 中国科学院深圳先进技术研究院 Rapid semantic segmentation method for road scene
CN111784709B (en) * 2020-07-07 2023-02-17 北京字节跳动网络技术有限公司 Image processing method, image processing device, electronic equipment and computer readable medium
CN112070760B (en) * 2020-09-17 2022-11-08 安徽大学 Bone mass detection method based on convolutional neural network
CN112348792A (en) * 2020-11-04 2021-02-09 广东工业大学 X-ray chest radiography image classification method based on small sample learning and self-supervision learning

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105426462A (en) * 2015-11-13 2016-03-23 深圳码隆科技有限公司 Image searching method and device based on image element
CN106407903A (en) * 2016-08-31 2017-02-15 四川瞳知科技有限公司 Multiple dimensioned convolution neural network-based real time human body abnormal behavior identification method
CN106447658A (en) * 2016-09-26 2017-02-22 西北工业大学 Significant target detection method based on FCN (fully convolutional network) and CNN (convolutional neural network)
CN106504064A (en) * 2016-10-25 2017-03-15 清华大学 Clothes classification based on depth convolutional neural networks recommends method and system with collocation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105426462A (en) * 2015-11-13 2016-03-23 深圳码隆科技有限公司 Image searching method and device based on image element
CN106407903A (en) * 2016-08-31 2017-02-15 四川瞳知科技有限公司 Multiple dimensioned convolution neural network-based real time human body abnormal behavior identification method
CN106447658A (en) * 2016-09-26 2017-02-22 西北工业大学 Significant target detection method based on FCN (fully convolutional network) and CNN (convolutional neural network)
CN106504064A (en) * 2016-10-25 2017-03-15 清华大学 Clothes classification based on depth convolutional neural networks recommends method and system with collocation

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Efficient convolutional neural networks for pixelwise classification on heterogeneous hardware systems;Fabian Tshopp 等;《IEEE International Symposium on Biomedical Imaging》;20161231;第1228-1228页
基于并行深度卷积神经网络的图像美感分类;王伟凝 等;《自动化学报》;20160630;第42卷(第6期);第904-914页
行人检测技术研究;张旭东;《中国优秀硕士学位论文全文数据库 信息科技辑》;20111215;第2011年卷(第12期);第I138-1074页

Also Published As

Publication number Publication date
CN107016415A (en) 2017-08-04

Similar Documents

Publication Publication Date Title
CN107016415B (en) A kind of color image Color Semantic classification method based on full convolutional network
Gijsenij et al. Color constancy using natural image statistics and scene semantics
US10535141B2 (en) Differentiable jaccard loss approximation for training an artificial neural network
US20200210773A1 (en) Neural network for image multi-label identification, related method, medium and device
Li et al. Aesthetic visual quality assessment of paintings
Narihira et al. Learning lightness from human judgement on relative reflectance
Falomir et al. Categorizing paintings in art styles based on qualitative color descriptors, quantitative global features and machine learning (QArt-Learn)
CN108921092B (en) Melanoma classification method based on convolution neural network model secondary integration
Rosenthal et al. Color statistics of objects, and color tuning of object cortex in macaque monkey
Ma et al. Shadow detection of moving objects based on multisource information in internet of things
CN108564589A (en) A kind of plant leaf blade dividing method based on the full convolutional neural networks of improvement
Yu Online teaching quality evaluation based on emotion recognition and improved AprioriTid algorithm
Zhang et al. Inkthetics: a comprehensive computational model for aesthetic evaluation of Chinese ink paintings
CN110807485B (en) Method for fusing two-classification semantic segmentation maps into multi-classification semantic map based on high-resolution remote sensing image
CN110889718B (en) Scheme screening method, scheme screening device, medium and electronic equipment
Niu et al. CF‐based optimisation for saliency detection
Murray et al. Toward automatic and flexible concept transfer
CN109978074A (en) Image aesthetic feeling and emotion joint classification method and system based on depth multi-task learning
Shete et al. TasselGAN: An application of the generative adversarial model for creating field-based maize tassel data
Pérez et al. Interactive feature space extension for multidimensional data projection
Saraswat et al. Colour normalisation of histopathological images
CN102184403B (en) Optimization-based intrinsic image extraction method
Velastegui et al. The importance of color spaces for image classification using artificial neural networks: a review
CN116778223A (en) Multi-category target domain data set labeling method and system based on zero labeling
Li et al. Neural network’s selection of color in UI design of social software

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20190719