CN113627528B - Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning - Google Patents
Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning Download PDFInfo
- Publication number
- CN113627528B CN113627528B CN202110917411.7A CN202110917411A CN113627528B CN 113627528 B CN113627528 B CN 113627528B CN 202110917411 A CN202110917411 A CN 202110917411A CN 113627528 B CN113627528 B CN 113627528B
- Authority
- CN
- China
- Prior art keywords
- chinese painting
- traditional chinese
- sample
- artist
- visual
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000010422 painting Methods 0.000 title claims abstract description 119
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000013135 deep learning Methods 0.000 title claims abstract description 11
- 230000000007 visual effect Effects 0.000 claims abstract description 51
- 238000012545 processing Methods 0.000 claims abstract description 42
- 230000000694 effects Effects 0.000 claims abstract description 35
- 238000012549 training Methods 0.000 claims abstract description 19
- 238000000605 extraction Methods 0.000 claims abstract description 17
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 13
- 230000008447 perception Effects 0.000 claims abstract description 9
- 230000010365 information processing Effects 0.000 claims abstract description 5
- 230000006870 function Effects 0.000 claims description 24
- 238000001914 filtration Methods 0.000 claims description 18
- 230000009466 transformation Effects 0.000 claims description 13
- 238000011176 pooling Methods 0.000 claims description 12
- 230000004913 activation Effects 0.000 claims description 9
- 230000008569 process Effects 0.000 claims description 7
- 230000035945 sensitivity Effects 0.000 claims description 6
- 238000010586 diagram Methods 0.000 claims description 4
- 238000007781 pre-processing Methods 0.000 claims description 4
- 230000008859 change Effects 0.000 claims description 3
- 238000009877 rendering Methods 0.000 claims description 3
- 230000007246 mechanism Effects 0.000 abstract description 2
- 238000010606 normalization Methods 0.000 description 6
- 238000012360 testing method Methods 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012706 support-vector machine Methods 0.000 description 2
- 235000003140 Panax quinquefolius Nutrition 0.000 description 1
- 240000005373 Panax quinquefolius Species 0.000 description 1
- 241000282376 Panthera tigris Species 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000003340 mental effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000000611 regression analysis Methods 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
- 230000005477 standard model Effects 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/047—Probabilistic or stochastic networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/10—Image enhancement or restoration using non-spatial domain filtering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20024—Filtering details
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20048—Transform domain processing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Health & Medical Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- General Health & Medical Sciences (AREA)
- Biophysics (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an automatic identification method of painters belonging to traditional Chinese painting based on human eye visual deep learning, which relates to application of deep learning based on visual processing mechanism in the art field, after visual front-end saliency processing and extraction are carried out on each traditional Chinese painting sample in a sample data set to obtain a visual saliency map, visual back-end perception information processing is carried out based on pixel value relation of each pixel point and neighborhood pixel points thereof to obtain an ordered style effect map, taking an ordered style effect graph obtained by processing each traditional Chinese painting sample as input and a corresponding artist as output, performing model training based on a convolutional neural network by using a sample data set to obtain an artist identification model, then using the artist identification model to identify and obtain an artist to which the traditional Chinese painting to be identified belongs, compared with the existing various identification methods, the method has the advantages that the accuracy of automatic identification is improved, and the judgment result of the observer is highly consistent.
Description
Technical Field
The invention relates to application of deep learning based on a visual processing mechanism in the field of art, in particular to an automatic recognition method for a traditional Chinese painting artist based on human eye visual deep learning.
Background
Along with the improvement of living standard, people pursue mental life more and more, wherein the appreciation requirement of Chinese painting is gradually improved. A large number of digital chinese drawings appear in networks or digital museums, but how to efficiently utilize and manage these drawings is a problem to be solved. Therefore, the method has great practical value in the research of automatically classifying and identifying the artistic style of the Chinese painting and the authors thereof. In the field of Chinese painting feature extraction and classification, related researches exist. Li et al designed the paintings of painters of related algorithm classification Shen Zhou, tiger, zhang Daqian, etc., but the paintings information could not be fully described, and its omission of other important information resulted in a non-ideal classification result. Jiang et al extract color features and texture features from the chinese painting of class 2, and classify the chinese painting by using a support vector machine (support vector machine, SVM). Chen Junjie and the like are used for extracting the characteristics of the mountain water and flower and bird Chinese painting in terms of color, shape and the like, and classifying by adopting SVM. Sun et al propose feature selection by using a Monte Carlo convex hull model and classify for Chinese painting authors. Liu Xiaowei, and the like, distinguish different drawings through 4 features such as image color board entropy, redundancy, order degree, complexity, and the like, but the algorithm only extracts the features related to colors, and does not consider the influence of other features on classification. Wang Zheng et al propose supervised heterogeneous sparse feature selection algorithms to perform feature screening and classify authors of Chinese paintings. Fusion algorithm is proposed by Sheng and the like, the classification results of the whole feature and the local feature are integrated by information entropy, and the texture feature is extracted by using a 3-layer wavelet transform method for Chinese painting classification. The existing literature adopts a traditional feature extraction method to extract the features of the Chinese painting, and the obtained features are single and one-sided. In addition, land et al use neural networks to extract Chinese painting features, tenenbaum et al propose feature recognition methods of embedded machine learning, li Yuzhi et al use Convolutional Neural Networks (CNNs) to extract Chinese painting visual features, an improved embedded learning algorithm is adopted to classify Chinese painting artist works, zhang Jia et al adopt manual evaluation and regression analysis strategies.
At present, the traditional Chinese painting still has the following problems of feature extraction and automatic identification: 1) The traditional Chinese painting sign extraction methods are all based on various mathematical modeling algorithms in the image pixel domain, and human eye perception characteristics and subjective judgment experience of an observer are not considered, so that the automatic recognition accuracy of various methods is not high, and a large gap exists between the automatic recognition accuracy and the judgment result of the observer; 2) The requirements on the quality of the traditional Chinese painting image are high, and any noise pollution and artificial intervention can reduce or disable the performance of the existing method.
Disclosure of Invention
Aiming at the problems and the technical requirements, the inventor provides an automatic artist identification method for traditional Chinese painting based on deep learning of human eyes, and the technical scheme of the invention is as follows:
an automatic identification method for a painting belonging to a traditional Chinese painting based on human eye vision deep learning, which comprises the following steps:
Constructing a sample data set, wherein the sample data set comprises a plurality of Chinese painting samples belonging to a plurality of different painters;
Performing visual front-end saliency processing and extraction on each traditional Chinese painting sample in the sample data set to obtain a visual saliency map;
performing visual back-end perception information processing on the visual saliency map based on the pixel value relation of each pixel point and the neighborhood pixel points to obtain an ordered style effect map;
taking an ordered style effect graph obtained by processing each traditional Chinese painting sample as input and a corresponding artist as output, and performing model training based on a convolutional neural network by using a sample data set to obtain an artist identification model;
And extracting an ordered style effect diagram of the traditional Chinese painting to be identified, inputting an artist identification model, and outputting an artist to which the identified traditional Chinese painting to be identified belongs.
The further technical scheme is that the traditional Chinese painting sample in the sample data set comprises an original traditional Chinese painting and a lossy traditional Chinese painting obtained by simulation after image pretreatment of the original traditional Chinese painting, wherein the image pretreatment comprises at least one of brightness enhancement, color rendering, contrast enhancement, gaussian pollution, random brightness change, sharpness treatment, image rotation and image overturning.
The further technical scheme is that the visual saliency processing and extraction are carried out on each traditional Chinese painting sample in the sample data set to obtain a visual saliency map, and the method comprises the following steps of for each traditional Chinese painting sample:
performing threshold filtering processing based on vision on the traditional Chinese painting sample;
Carrying out logarithmic transformation of frequency domain light intensity on the traditional Chinese painting sample after finishing threshold filtering treatment;
and performing feature extraction on the converted traditional Chinese painting sample by using an SDSP extraction algorithm to obtain a visual saliency map.
The further technical scheme is that the threshold value filtering processing based on vision is carried out on the traditional Chinese painting sample, and the method comprises the following steps:
According to the visual contrast sensitivity threshold CSF (1), the visual fovea function tau (i, j) and the visual threshold delta, carrying out threshold filtering processing on a pixel value y (i, j) of a pixel point at any coordinate (i, j) in a traditional Chinese painting sample to obtain a pixel value z (i, j) after the threshold filtering processing, wherein the pixel value z (i, j) is as follows:
Wherein the visual contrast sensitivity threshold is:
CSF(1)=2.6[0.0192+λ·2-1·r·v·tan(0.5°)]exp{-[λ·2-1·r·v·tan(0.5°)]1.1};
visual foveal function
Wherein r is a resolution parameter, v is a viewing distance parameter, lambda is a first adjustment parameter, d t is a second adjustment parameter, d (i, j) is a distance between a pixel point of coordinates (i, j) in the traditional Chinese painting sample and a center of a sample image, and d 0 is a distance from the center of the sample image of the traditional Chinese painting sample to an image edge.
The further technical scheme is that the logarithmic transformation of the frequency domain light intensity is carried out on the traditional Chinese painting sample after the threshold filtering treatment, and the method comprises the following steps:
The result of logarithmic transformation of the frequency domain light intensity is that the pixel value z (i, j) of the pixel point at any coordinate (i, j) in the traditional Chinese painting sample after finishing the threshold filtering process is:
Wherein, the function F (-) represents performing frequency domain positive transformation processing, F -1 (-) represents performing frequency domain inverse transformation processing, and G (omega, theta j) is a log-Gabor filter.
The further technical scheme is that the method for processing the visual back end perception information of the visual saliency map to obtain the ordered style effect map based on the pixel value relation of each pixel point and the neighborhood pixel points thereof comprises the following steps:
Calculating mutual information between a pixel value t (i, j) of a pixel point at the coordinate (i, j) and a pixel value t k (i, j) of a pixel point nearby the pixel point;
based on mutual information between the pixel value t (i, j) and the pixel value of each neighborhood pixel point, the pixel value Q (i, j) of the pixel point at the coordinate (i, j) in the ordered style effect map is obtained.
The further technical scheme is that mutual information between a pixel value t (i, j) and a pixel value t k (i, j) of a neighboring pixel point is as follows:
I[t(i,j),tk(i,j)]=logp[t(i,j)/tk(i.j)]-logp[t(i,j)];
Wherein p [ t (i, j) ] is the statistical probability of the pixel value t (i, j), p [ t (i, j)/t k (i, j) ] is the statistical conditional probability of the pixel value t (i, j) and the neighborhood pixel value t k (i, j), and log represents the logarithmic operation.
The further technical scheme is that the pixel value Q (i, j) of the pixel point at the coordinate (i, j) in the ordered style effect graph is as follows:
wherein T (i, j) is a neighborhood pixel value set formed by pixel values of all neighborhood pixel points of the pixel point at the coordinate (i, j), and C l is a normalization coefficient corresponding to the neighborhood pixel value T l (i, j) and has Ω (I, j) is a gaussian weight function, mxn is a window size of T (I, j), I [ T (I, j), T l (I, j) ] is mutual information between the pixel value T (I, j) and the neighborhood pixel value T l (I, j), and T k (I, j) represents a pixel value of any one of the neighborhood pixel points in T (I, j).
The further technical scheme is that when the artist identification model is obtained through training:
According to And carrying out normalization processing on the ordered style effect graph obtained by processing each pair of traditional Chinese painting samples, taking the ordered style effect graph obtained by normalizing each pair of traditional Chinese painting samples as input and the corresponding painter as output, and carrying out model training based on a convolutional neural network by utilizing a sample data set to obtain a painter identification model, wherein Q (i, j) is a pixel value in the ordered style effect graph.
The convolutional neural network for training to obtain the artist identification model comprises five convolutional modules and three full-connection layers, wherein each convolutional module comprises a convolutional layer, an activation function and a pooling layer, the activation function in each convolutional module is a ReLU, the pooling function of each pooling layer is the maximum pooling, and the activation function of each full-connection layer is a ReLU;
outputting the painters to which the Chinese painting to be identified obtained by identification belongs, comprising:
And outputting the prediction probability of the traditional Chinese painting to be identified belonging to each candidate artist, and taking the candidate artist with the highest prediction probability as the artist to which the traditional Chinese painting to be identified belongs, wherein the candidate artist is each artist to which the traditional Chinese painting sample in the sample data set belongs.
The beneficial technical effects of the invention are as follows:
The application discloses an automatic identification method of a traditional Chinese painting belonging to a painting based on deep learning of human eyes, which fully considers the perception characteristics of human eyes and subjective judgment experience of an observer. And the method still keeps higher recognition accuracy under various noise pollution and human intervention conditions.
Drawings
Fig. 1 is a schematic flow chart of an automatic identification method of painters to which the traditional Chinese painting disclosed by the application belongs.
Fig. 2 is a visual saliency map obtained by extracting a traditional Chinese painting sample after visual front saliency processing.
Fig. 3 is an ordered style effect diagram obtained by processing the visual back-end perception information of the visual saliency map shown in fig. 2.
Detailed Description
The following describes the embodiments of the present invention further with reference to the drawings.
The application discloses an automatic identification method of painters to which traditional Chinese painting belongs based on deep learning of human eyes, please refer to a flow chart shown in fig. 1, the method comprises the following steps:
step 1, a sample data set is constructed, wherein the sample data set comprises a plurality of Chinese painting samples belonging to a plurality of different painters. The traditional Chinese painting sample in the sample data set comprises an original traditional Chinese painting and a damaged traditional Chinese painting obtained by simulating the original traditional Chinese painting after image pretreatment, and the influence of factors such as the transition of a pseudo-age, technical restoration, ornamental mode and the like on the traditional Chinese painting can be simulated through the image pretreatment. Image preprocessing includes at least one of brightness enhancement, color rendering, contrast enhancement, gaussian stain, random brightness change, sharpness processing, image rotation, image flipping, image rotation processing including rotation at any angle, typically including rotation at 90 ° or 180 ° or 270 ° clockwise/counterclockwise, flipping including left-right and/or up-down flipping. The image preprocessing actually adopted comprises one or more of the processing means, for example, the original country can be rotated by 90 degrees anticlockwise and turned up and down to obtain the lossy traditional Chinese painting, and other combination modes can be used for pushing the lossy traditional Chinese painting.
The original traditional Chinese painting has more sources, for example, in an actual measurement experiment, the original traditional Chinese painting is obtained from relevant departments of all levels of China museums, the Yachang art network and the Flickr network download centralized approach and is subjected to relevant processing, so that 5 painters are obtained, each painter comprises 600 original traditional Chinese paintings and 600 damaged traditional Chinese paintings, and a sample data set containing 3000 traditional Chinese painting samples in total is constructed.
And 2, performing visual front-end saliency processing and extraction on each traditional Chinese painting sample in the sample data set to obtain a visual saliency map. Comprising the following steps:
(1) The traditional Chinese painting sample is subjected to threshold filtering processing based on vision, specifically, the pixel value y (i, j) of the pixel point at any coordinate (i, j) in the traditional Chinese painting sample is subjected to threshold filtering processing according to a visual contrast sensitivity threshold CSF (1), a visual fovea function tau (i, j) and a visual threshold delta, and the pixel value z (i, j) after the threshold filtering processing is obtained as follows:
Wherein the visual contrast sensitivity threshold is:
CSF(1)=2.6[0.0192+λ·2-1·r·v·tan(0.5°)]exp{-[λ·2-1·r·v·tan(0.5°)]1.1};
where r is a resolution parameter, v is a line-of-sight parameter, r=96 (pixels/inch), v=19.1 (inch), and pixels represents an image in pixels and inch represents a length in inches, according to subjective test conditions. Lambda is the first adjustment parameter. Preferably, λ=0.228.
Visual foveal functionWherein d t is a second adjustment parameter, d (i, j) is a distance between a pixel point of the coordinate (i, j) in the traditional Chinese painting sample and the center of the sample image, and d 0 is a distance from the center of the sample image of the traditional Chinese painting sample to the edge of the image. Preferably, d t =4.0 is taken.
Preferably, the visual threshold δ=6.0 is taken.
(2) Carrying out logarithmic transformation of frequency domain light intensity on the traditional Chinese painting sample after the threshold filtering treatment, and carrying out logarithmic transformation of the frequency domain light intensity on the pixel value z (i, j) of the pixel point at any coordinate (i, j) in the traditional Chinese painting sample after the threshold filtering treatment, wherein the result is as follows:
Wherein, the function F (-) represents performing frequency domain positive transformation processing, F -1 (-) represents performing frequency domain inverse transformation processing, and G (omega, theta j) is a log-Gabor filter.
(3) Performing feature extraction on the converted traditional Chinese painting sample by using an SDSP extraction algorithm to obtain a visual saliency map, and obtaining pixel values of pixel points at any coordinates (i, j) in the visual saliency mapSDSP (·) represents an SDSP extraction algorithm.
And step 3, performing visual back-end perception information processing on the visual saliency map based on the pixel value relation of each pixel point and the neighborhood pixel points thereof to obtain an ordered style effect map. Comprising the following steps:
(1) Mutual information between the pixel value t (i, j) and its neighborhood pixel point t k (i, j) is calculated. Specifically, the mutual information between the pixel value t (i, j) and the pixel value t k (i, j) of the neighboring pixel point is:
I[t(i,j),tk(i,j)]=logp[t(i,j)/tk(i.j)]-logp[t(i,j)]。
Wherein p [ t (i, j) ] is the statistical probability of the pixel value t (i, j), p [ t (i, j)/t k (i, j) ] is the statistical conditional probability of the pixel value t (i, j) and the neighborhood pixel value t k (i, j), and log represents the logarithmic operation.
(2) Based on mutual information between the pixel value t (i, j) and the pixel value of each neighborhood pixel point, the pixel value Q (i, j) of the pixel point at the coordinate (i, j) in the ordered style effect map is obtained. Specifically, the pixel value Q (i, j) of the pixel point at the coordinate (i, j) in the ordered style effect map is:
wherein T (i, j) is a neighborhood pixel value set formed by pixel values of all neighborhood pixel points of the pixel point at the coordinate (i, j), and C l is a normalization coefficient corresponding to the neighborhood pixel value T l (i, j) and has Ω (I, j) is a gaussian weight function, mxn is the window size of T (I, j), I [ T (I, j), T l (I, j) ] is the mutual information between the pixel value T (I, j) and the neighborhood pixel value T l (I, j), and T k (I, j) represents the pixel value of any one neighborhood pixel in T (I, j), i.e., T k (I, j) et (I, j).
As a preferred alternative to this, m×n=16×16.
For example, a visual saliency map obtained by performing visual front-end saliency processing and extraction on a traditional Chinese painting sample is shown in fig. 2, and an ordered style effect map obtained by continuing visual rear-end perception information processing is shown in fig. 3.
And 4, taking an ordered style effect graph obtained by processing each traditional Chinese painting sample as input and a corresponding artist as output, and performing model training based on a convolutional neural network by using a sample data set to obtain an artist identification model.
In practical application, the ordered style effect graph is not generally directly taken as input, but is firstly taken as followsAnd carrying out normalization processing on the ordered style effect graph obtained by processing each traditional Chinese painting sample, wherein Q (i, j) is a pixel value in the ordered style effect graph.
And then taking the normalized ordered style effect graph of each traditional Chinese painting sample as input and the corresponding painter as output, and performing model training based on a convolutional neural network by using a sample data set to obtain a painter identification model. During training, a training set and a testing set are firstly randomly divided into a sample data set according to the proportion of 8:2, then model training is carried out by using the training set, model testing is carried out by using the testing set, and finally an artist identification model is obtained.
The convolutional neural network used for training to obtain the artist identification model is modified on the basis of the VGG16 standard model, the convolutional neural network comprises five convolutional modules and three full-connection layers, each convolutional module comprises a convolutional layer, an activation function and a pooling layer, the activation function in each convolutional module is a ReLU, and the pooling function of each pooling layer is the maximum pooling. In the application, the first two convolution modules comprise two convolution layers, and the last three convolution modules comprise three convolution layers. The activation function of each fully connected layer is ReLU. The network parameters of each layer in one experiment were as follows:
As an optimization, training parameters were set as follows: the optimizer selects a random gradient descent SGD function, a loss function is selected as Cross entropy-Entropy, the iteration number is 30, the number of input images in each batch is 16, the learning rate is 1e-5, and the momentum is 0.9.
And 5, extracting an ordered style effect diagram of the traditional Chinese painting to be identified, inputting an artist identification model, and outputting an artist to which the identified traditional Chinese painting to be identified belongs.
The traditional Chinese painting to be identified can be an original traditional Chinese painting or a damaged traditional Chinese painting, the method for extracting the ordered style effect picture of the traditional Chinese painting to be identified is similar to the model training process, and if the ordered style effect picture after normalization processing is used as input during model training, the ordered style effect picture and the ordered style effect picture are also required to be input after normalization processing. The candidate painters with the largest prediction probability are used as the painters to which the traditional Chinese painting to be identified belongs, and the candidate painters are all the painters to which the traditional Chinese painting samples in the sample data set belong.
Under the condition of no noise pollution and artificial interference, the traditional Chinese painting drawing method respectively utilizes the traditional methods to automatically identify and count the accuracy, and under the condition of noise pollution and artificial interference, the traditional Chinese painting drawing method automatically identifies and counts the accuracy, and compared with the statistics result, the method provided by the application has the following steps:
the experimental results show that the method provided by the application still has higher automatic identification accuracy even under various environmental conditions of noise pollution and artificial interference.
The above is only a preferred embodiment of the present application, and the present application is not limited to the above examples. It is to be understood that other modifications and variations which may be directly derived or contemplated by those skilled in the art without departing from the spirit and concepts of the present application are deemed to be included within the scope of the present application.
Claims (8)
1. An automatic identification method for painters belonging to traditional Chinese painting based on deep learning of human eyes is characterized by comprising the following steps:
constructing a sample data set, wherein the sample data set comprises a plurality of Chinese painting samples belonging to a plurality of different painters;
performing visual front-end saliency processing and extraction on each traditional Chinese painting sample in the sample data set to obtain a visual saliency map, wherein the visual saliency map comprises the following steps of: performing threshold filtering processing based on vision on the traditional Chinese painting sample, performing logarithmic transformation of frequency domain light intensity on the traditional Chinese painting sample after the threshold filtering processing, and performing feature extraction on the traditional Chinese painting sample after transformation by using an SDSP extraction algorithm to obtain the visual saliency map;
Performing visual back-end perception information processing on the visual saliency map based on the pixel value relation of each pixel point and the neighborhood pixel points thereof to obtain an ordered style effect map, wherein the method comprises the following steps: calculating coordinates Pixel value of pixel point atPixel value of adjacent pixel pointMutual information between them, based on pixel valuesMutual information between the pixel values of each neighborhood pixel point to obtain coordinatesPixel values of pixel points in the ordered style effect graph;
Taking an ordered style effect graph obtained by processing each traditional Chinese painting sample as input and a corresponding artist as output, and performing model training based on a convolutional neural network by using the sample data set to obtain an artist identification model;
And extracting an ordered style effect diagram of the traditional Chinese painting to be identified, inputting the artist identification model, and outputting the artist to which the identified traditional Chinese painting to be identified belongs.
2. The method of claim 1, wherein the chinese painting samples in the sample dataset include original chinese painting and lossy chinese painting simulated after image preprocessing of the original chinese painting, the image preprocessing including at least one of brightness enhancement, color rendering, contrast enhancement, gaussian pollution, random brightness change, sharpness processing, image rotation, image flipping.
3. The method of claim 1, wherein said performing a vision-based thresholding process on said chinese painting sample comprises:
Sensitivity threshold according to visual contrast Visual foveal functionAnd vision thresholdFor any coordinates in the Chinese painting samplePixel value of pixel point atPerforming threshold filtering processing to obtain pixel values after the threshold filtering processingThe method comprises the following steps:
;
Wherein the visual contrast sensitivity threshold is:
;
visual foveal function ;
Wherein, In order to be able to carry out the resolution parameters,As a function of the viewing distance parameter,For the first adjustment parameter, a first parameter is provided,For the second adjustment parameter, a second adjustment parameter,Is the coordinates in the Chinese painting sampleIs the distance between the pixel point of (c) and the center of the sample image,Is the distance from the center of the sample image of the traditional Chinese painting sample to the edge of the image.
4. The method according to claim 1, wherein the performing logarithmic transformation of the frequency domain light intensity on the chinese painting sample after the threshold filtering process includes:
for any coordinates in the traditional Chinese painting sample after finishing the threshold filtering process Pixel value of pixel point atThe result of logarithmic transformation of the frequency domain light intensity is:
;
Wherein the function is The representation is subjected to a frequency domain forward transform process,The representation is subjected to an inverse frequency domain transform process,Is a log-Gabor filter.
5. The method of claim 1, wherein the pixel values arePixel value of adjacent pixel pointMutual information between the two is:
;
Wherein, For pixel valuesIs used for the statistical probability of (a),For pixel valuesAnd neighborhood pixel valuesIs used to determine the statistical conditional probability of (1),Representing a logarithmic operation.
6. The method of claim 1, wherein the coordinatesPixel values of pixel points in the ordered style effect graphThe method comprises the following steps:
;
Wherein, Is the coordinatesA neighborhood pixel value set formed by pixel values of all neighborhood pixel points of the pixel point,Is a neighborhood pixel valueCorresponding normalized coefficient and has,As a function of the gaussian weight,Is thatIs defined by a window size of (a),For pixel valuesAnd neighborhood pixel valuesThe mutual information between the two pieces of information,Representation ofThe pixel value of any one of the neighborhood pixel points.
7. The method of claim 1, wherein, in training to obtain the artist identification model:
According to Normalizing the ordered style effect graph obtained by processing each traditional Chinese painting sample, taking the normalized ordered style effect graph of each traditional Chinese painting sample as input and the corresponding painter as output, performing model training based on a convolutional neural network by using the sample data set to obtain a painter identification model,Is the pixel value in the ordered style effect map.
8. The method of claim 1, wherein the convolutional neural network for training to obtain the artist recognition model comprises five convolutional modules and three fully connected layers, each convolutional module comprises a convolutional layer, an activation function and a pooling layer, the activation function in each convolutional module is a ReLU, the pooling function of each pooling layer is a maximum pooling, and the activation function of each fully connected layer is a ReLU;
The outputting the obtained painters to which the to-be-identified traditional Chinese painting belongs, including:
And outputting the prediction probability that the traditional Chinese painting to be identified belongs to each candidate artist, and taking the candidate artist with the largest prediction probability as the artist to which the traditional Chinese painting to be identified belongs, wherein the candidate artist is each artist to which the traditional Chinese painting sample in the sample data set belongs.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110917411.7A CN113627528B (en) | 2021-08-11 | 2021-08-11 | Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110917411.7A CN113627528B (en) | 2021-08-11 | 2021-08-11 | Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113627528A CN113627528A (en) | 2021-11-09 |
CN113627528B true CN113627528B (en) | 2024-09-06 |
Family
ID=78384282
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110917411.7A Active CN113627528B (en) | 2021-08-11 | 2021-08-11 | Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113627528B (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109308697A (en) * | 2018-09-18 | 2019-02-05 | 安徽工业大学 | A kind of leaf disease recognition method based on machine learning algorithm |
CN111325290A (en) * | 2020-03-20 | 2020-06-23 | 西安邮电大学 | Chinese painting image classification method based on multi-view fusion and multi-example learning |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
ITMI20131244A1 (en) * | 2013-07-24 | 2015-01-25 | Telecom Italia Spa | IDENTIFICATION OF KEYPOINTS |
CN109118459B (en) * | 2017-06-23 | 2022-07-19 | 南开大学 | Image salient object detection method and device |
CN112434590B (en) * | 2020-11-18 | 2023-11-24 | 中国人民解放军国防科技大学 | SAR image wind stripe recognition method based on wavelet transformation |
-
2021
- 2021-08-11 CN CN202110917411.7A patent/CN113627528B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109308697A (en) * | 2018-09-18 | 2019-02-05 | 安徽工业大学 | A kind of leaf disease recognition method based on machine learning algorithm |
CN111325290A (en) * | 2020-03-20 | 2020-06-23 | 西安邮电大学 | Chinese painting image classification method based on multi-view fusion and multi-example learning |
Also Published As
Publication number | Publication date |
---|---|
CN113627528A (en) | 2021-11-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108562589B (en) | Method for detecting surface defects of magnetic circuit material | |
CN111553929B (en) | Mobile phone screen defect segmentation method, device and equipment based on converged network | |
Yiyang | The design of glass crack detection system based on image preprocessing technology | |
CN108960245B (en) | Tire mold character detection and recognition method, device, equipment and storage medium | |
CN111582294B (en) | Method for constructing convolutional neural network model for surface defect detection and application thereof | |
CN106897673B (en) | Retinex algorithm and convolutional neural network-based pedestrian re-identification method | |
CN105354865B (en) | The automatic cloud detection method of optic of multispectral remote sensing satellite image and system | |
US9449253B2 (en) | Learning painting styles for painterly rendering | |
CN111222519B (en) | Construction method, method and device of hierarchical colored drawing manuscript line extraction model | |
US10157450B2 (en) | Enhancing the legibility of images using monochromatic light sources | |
US10726535B2 (en) | Automatically generating image datasets for use in image recognition and detection | |
CN108985337A (en) | A kind of product surface scratch detection method based on picture depth study | |
Pei et al. | Effects of image degradations to cnn-based image classification | |
CN114926407A (en) | Steel surface defect detection system based on deep learning | |
Mai et al. | Back propagation neural network dehazing | |
CN114332079A (en) | Plastic lunch box crack detection method, device and medium based on image processing | |
Kaur et al. | An edge detection technique with image segmentation using ant colony optimization: A review | |
CN109376719A (en) | A kind of camera light Photo-Response Non-Uniformity fingerprint extraction and comparison method based on assemblage characteristic expression | |
CN113627528B (en) | Automatic identification method for painters belonging to traditional Chinese painting based on human eye vision deep learning | |
CN109376782B (en) | Support vector machine cataract classification method and device based on eye image features | |
CN116363114A (en) | Ceramic tile surface quality detection method and device, electronic equipment and storage medium | |
Kim et al. | Classification of oil painting using machine learning with visualized depth information | |
CN114548250B (en) | Mobile phone appearance detection method and device based on data analysis | |
Liu et al. | A novel image segmentation algorithm based on visual saliency detection and integrated feature extraction | |
CA2997335C (en) | Automatically generating image datasets for use in image recognition and detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |