CN108510456A - The sketch of depth convolutional neural networks based on perception loss simplifies method - Google Patents

The sketch of depth convolutional neural networks based on perception loss simplifies method Download PDF

Info

Publication number
CN108510456A
CN108510456A CN201810259452.XA CN201810259452A CN108510456A CN 108510456 A CN108510456 A CN 108510456A CN 201810259452 A CN201810259452 A CN 201810259452A CN 108510456 A CN108510456 A CN 108510456A
Authority
CN
China
Prior art keywords
sketch
model
network
layer
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810259452.XA
Other languages
Chinese (zh)
Other versions
CN108510456B (en
Inventor
徐雪妙
谢敏珊
缪佩琦
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
South China University of Technology SCUT
Original Assignee
South China University of Technology SCUT
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by South China University of Technology SCUT filed Critical South China University of Technology SCUT
Priority to CN201810259452.XA priority Critical patent/CN108510456B/en
Publication of CN108510456A publication Critical patent/CN108510456A/en
Application granted granted Critical
Publication of CN108510456B publication Critical patent/CN108510456B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses the sketches of the depth convolutional neural networks based on perception loss to simplify method, including step:1) data acquisition;Use rough draft figure and corresponding label data;2) data processing;It converts the rough draft figure of image data set and corresponding line manuscript base to the trained required format of depth convolutional neural networks by pretreatment;3) model construction;According to training objective, constructs one and be suitable for the depth convolutional neural networks that sketch simplifies problem;4) loss function is defined;5) model training;The penalty values of network are calculated according to loss function, then calculate the gradient of each network layer parameter by backpropagation, and the parameter of each layer of network is updated by stochastic gradient descent method;6) model is verified;The model that training obtains is verified using validation data set, tests its Generalization Capability.Method proposed by the present invention so that sketch simplifies network more coarse, the indefinite sketch of main structure line that can handle miscellaneous line, and the influence for illumination has stronger robustness.

Description

The sketch of depth convolutional neural networks based on perception loss simplifies method
Technical field
The present invention relates to the technical fields of Computer Image Processing, refer in particular to a kind of depth convolution based on perception loss The sketch of neural network simplifies method.
Background technology
Skeletonizing is the first step of artistic creation.In this step, creator usually merely desires to expression concept and composition, and It is not to focus on portraying for details.After having created sketch, creator needs to describe clean line original text on the basis of sketch, this is One lengthy and tedious and labor intensive work.
In recent years, with the fast development of deep learning clean line original text was may learn using depth convolutional neural networks The feature of figure, to which the sketch more than miscellaneous line to be changed into the simplification line manuscript base for possessing line manuscript base feature.But its there are the problem of It is that current deep learning method directly weighs loss by calculating the Euclidean distance of generation figure and artificial marking line original text, and Actually sketch and line original text are not matched one by one in pixel, but consistent in visual perception, this allows for existing depth Degree learning method can only handle fairly simple less mixed and disorderly line original text, and larger for some difficulty, and main structure line is unknown True, the problems such as needing the line original text semantically understood that lines will occur and lose, obscure, distort, simplify scarce capacity.
Invention content
The shortcomings that it is an object of the invention to overcome the prior art and deficiency provide a kind of depth volume based on perception loss The sketch of product neural network simplifies method, and this method can well simplify the sketch of arbitrary size, before breaching The problems such as loss, fuzzy, lines distort, simplify scarce capacity in sketch simplification method.
In order to achieve the above object, the present invention uses following technical scheme:
The present invention is based on the sketches of the depth convolutional neural networks of perception loss to simplify method, includes the following steps:
1) data acquisition;
Using rough draft figure and corresponding label data, training depth convolutional neural networks need training data, the label Data are exactly the correspondence line manuscript base manually drawn on the basis of rough draft figure, are then divided into training dataset and verify data Collection;
2) data processing;
It converts the rough draft figure of image data set and corresponding line manuscript base to trained depth convolutional neural networks by pretreatment Required format;
3) model construction;
According to training objective and the input/output format of model, constructs one and be suitable for the depth that sketch simplifies problem Convolutional neural networks;
4) loss function is defined;
According to training objective and the framework of model, required loss function is defined;
5) model training;
The parameter of each layer network is initialized, continuous iteration inputs training sample, network is calculated according to loss function Penalty values, then the gradient of each network layer parameter is calculated by backpropagation, by stochastic gradient descent method to each layer network Parameter is updated;
6) model is verified;
The model trained using verify data set pair is verified, its Generalization Capability is tested.
As a preferred technical solution, in the step 2), data processing is specially:
2.1) image is changed into gray-scale map;
2.2) horizontal, vertical, diagonal turnover is carried out to the image that data are concentrated, corresponding line manuscript base also does same overturning Operation;
2.3) picture after overturning is given to add tone reversal, noise and shade;
2.4) by treated, picture is cut to be sized, and corresponding line manuscript base is also cropped to correspondingly sized;
2.5) image concentrated to data is transformed into from [0,255] in the range of [- 1,1].
As a preferred technical solution, in the step 2.4), treated, and picture is cut to 384 × 384 pixel sizes.
As a preferred technical solution, in the step 3), model construction is specially:
3.1) sketch simplified network model is constructed;
The sketch simplified network model is responsible for the sketch of input to be simplified to line original text, and input is an arbitrary size Sketch, output are the line manuscript bases after a same size size simplifies;Sketch simplified network model is by two concatenated encoders Model and decoder model are constituted;The encoder model is for extracting high-rise semantic information from the sketch of input simultaneously It is saved in the coding of a low-dimensional;The decoder model is for after being disassembled in the coding of low-dimensional and restoring and be simplified Result figure;
3.2) construction feature extracts network model;
The feature extraction network model is used to extract high-rise semantic information from the image of input and in this height Be compared under layer semantic information, the input of feature extraction network be result figure after the simplification that step 3.1) is handled and Corresponding artificial marking line manuscript base needs input picture being extended to 3 × 384 × 384 before inputting pre-training VGG16 networks, Under a high-layer semantic information, result figure and the corresponding artificial respective characteristic pattern of marking line manuscript base is taken to be compared.
As a preferred technical solution, in the step 3.1), the encoder model includes multiple cascade down-samplings Layer, the down-sampling layer are made of concatenated convolutional layer, batch regularization layer and nonlinear activation layer, are increasing receptive field While be gradually reduced the size of image, make Web-based Self-regulated Learning down-sampling, batch regularization using the convolutional layer with step-length Layer plays the work stablized with acceleration model training by the mean value and standard deviation of the input sample of the same batch of normalization With it is simple linear model that the addition of nonlinear activation layer, which prevents model degradation, improves the descriptive power of model;
The decoder model includes multiple cascade up-sampling layers, the up-sampling layer by concatenated warp lamination, batch Regularization layer and nonlinear activation layer composition are measured, plays the role of encoding enlarged drawing size from low-dimensional, with the anti-of step-length Convolutional layer makes Web-based Self-regulated Learning up-sample, same in the effect of batch regularization layer and nonlinear activation layer and encoder, together When we also introduce residual block to accelerate training speed.
As a preferred technical solution, in the step 3.2),
The feature extraction network model includes multiple cascade down-sampling layers, and the down-sampling layer is by concatenated convolution Layer and nonlinear activation function layer, pond layer composition, wherein convolutional layer step-length is 1, and convolution kernel size is 3 × 3, is extracted Corresponding characteristic pattern, it is simple linear model that the addition of nonlinear activation layer, which prevents model degradation, improves the description energy of model Power, the effect of pond layer is to reduce the size of characteristic pattern, to increase the receptive field of convolution kernel.
As a preferred technical solution, in the step 4), defining loss function is specially:
The computational methods of the function be by the result figure after sketch simplifies network reduction and its it is corresponding in advance Through the good line manuscript base of handmarking together input feature vector extraction network result figure and line are taken under a high-rise semantic information The corresponding characteristic pattern of manuscript base is compared, and defining loss function makes the characteristic pattern of result figure and the characteristic pattern of line manuscript base to the greatest extent may be used Approach to energy;Therefore, perception loss can be defined as by simplifying the loss function of sketch, and formula is as follows:
Wherein, LpecTo perceive loss function,And IyIt indicates to simplify result figure line corresponding with its that network generates respectively Manuscript base, Wi,j, Hi,jAnd Ci,jThe width of characteristic pattern, height, number of channels, φ are indicated respectivelyi,j(I) indicate characteristic pattern function, i and J indicates feature graph laplacian.
As a preferred technical solution, in the step 5), the specific method of model training is:
5.1) each layer parameter of initialization model
The initialization of each layer parameter is using commonly used method in depth convolutional neural networks, to feature extraction network Convolution layer parameter using the convolutional layer parameter value in the good VGG16 network models of ImageNet pre-training as initial value, grass Figure simplify network in convolutional layer, use mean value be 0, standard deviation use The Gaussian Profile being calculated carries out Initialization, wherein scale is result figure with respect to the multiple that artwork is amplified, faninIt is the picture number of input;
5.2) training network model
Stochastic inputs pass through the original image of step 2) processing, after the sketch simplification network generation by step 3.1) simplifies Result figure, result figure Jing Guo step 3.1) and the artificial marking line manuscript base of correspondence are inputted, by the feature extraction of step 3.2) Network obtains corresponding characteristic pattern, calculates corresponding perception penalty values and by step 4.1), then loses this perception Value can obtain the gradient of each layer parameter in step 3.1) sketch simplified network model by backpropagation, then pass through boarding steps The gradient that degree descent algorithm makes optimizes each layer parameter, you can realizes the training of a wheel network model;
5.3) step 5.2) is repeated until network reaches the set goal about the ability that sketch simplifies.
As a preferred technical solution, in the step 6), the specific method of model verification is:
It is concentrated at random from verify data and takes out some original images, after step 2) processing, be input to step 5) training Good network model allows the network model to go to simplify the sketch inputted, passes through result and the corresponding artificial marking line original text of output Figure is compared, to judge that the sketch of the trained network model simplifies ability.
Compared with prior art, the present invention having the following advantages that and advantageous effect:
1, the present invention proposes a simple depth convolutional neural networks and perception loss function (Perceptual Loss, with the good VGG feature extractions real-time performance of a pre-training), it solves the problems, such as that complicated sketch simplifies, reaches visual effect More preferably, the wider array of purpose of applicability,
2, the method that the present invention proposes an amplification data collection so that sketch simplifies shadow of the network for illumination and shade Ringing has stronger robustness, suitable for the sketch for scanning and arbitrarily shooting.
Description of the drawings
Fig. 1 is the method for the present invention flow diagram;
Fig. 2 (a), Fig. 2 (b) are sketch and corresponding artificial mark figure respectively;
Fig. 3 is the schematic diagram that sketch of the present invention simplifies network;
Fig. 4 is the schematic diagram that feature of present invention extracts network.
Specific implementation mode
Present invention will now be described in further detail with reference to the embodiments and the accompanying drawings, but embodiments of the present invention are unlimited In this.
Embodiment
As shown in Figure 1, the sketch of the depth convolutional neural networks based on perception loss described in the present embodiment simplifies method, Its specific practice is as follows:
1) sketch data set is obtained, its corresponding artificial mark is then obtained, i.e., is manually drawn out on the basis of sketch dry Net line original text, and it is divided into training dataset and validation data set.
The sketch of acquisition and corresponding artificial mark are as shown in Fig. 2 (a) and Fig. 2 (b):
2) image of image data set and label data are converted to trained depth convolutional neural networks institute by pretreatment The format needed, includes the following steps:
2.1) image is changed into gray-scale map;
2.2) horizontal, vertical, diagonal turnover is carried out to the image that data are concentrated, corresponding line manuscript base also does same overturning Operation;
2.3) picture after overturning is given to add tone reversal, noise and shade;
2.4) by treated, picture is cut to 384 × 384 pixel sizes, and corresponding line manuscript base is also cropped to corresponding big It is small;
2.5) image concentrated to data is transformed into from [0,255] in the range of [- 1,1].
3) network model is built, including feature extraction network, sketch simplify network, include the following steps:
3.1) construction sketch simplifies network.Sketch simplify network input be arbitrary size sketch (training when, we Inputted using fixed, be 1 × 384 × 384 after step 2) processing), it exports as after simplification identical with artwork size dimension Result figure.The network includes 3 concatenated structures (up-sampling layer, residual block and down-sampling layer).It is that a sketch simplifies below The specific example of network model, as shown in Figure 3;
3.2) construction feature extracts network.The image that the input of feature extraction network is 3 × 384 × 384 (need to be by sketch letter Change 1 × 384 × 384 results that network obtains to be extended), exporting a series of low-dimensional coding characteristic figures, (present invention uses It is the characteristic pattern of relu5-1, size is 512 × 24 × 24).The network includes multiple cascade down-sampling layers.Down-sampling layer by Concatenated convolutional layer and nonlinear activation function layer, pond layer composition, as shown in Figure 4.
4) loss function that sketch simplifies network is defined, is included the following steps:
4.1) loss function that sketch simplifies network is defined.Defining loss function makes the result figure of output as far as possible in vision Perceptually close to the line manuscript base manually marked, defining loss function with Perceptual Loss herein makes sketch simplify network It exports result figure to approach with the line manuscript base manually marked as far as possible in visual perception, formula is as follows:
Wherein, LpecTo perceive loss function,And IyIt indicates to simplify result figure line corresponding with its that network generates respectively Manuscript base, Wi,j,Hi,jAnd Ci,jThe width of characteristic pattern, height, number of channels, φ are indicated respectivelyi,j(I) indicate characteristic pattern function, i and J indicates feature graph laplacian;
5) training network model, includes the following steps:
5.1) each layer parameter of initialization model
The initialization of each layer parameter is using commonly used method in depth convolutional neural networks, to feature extraction network Convolution layer parameter using the convolutional layer parameter value in the good VGG16 network models of ImageNet pre-training as initial value, grass Figure simplifies the parameter of the convolutional layer in network, and it is 0 to be all made of mean value, and standard deviation is used (scale is result figure The multiple of opposite artwork amplification, faninIt is the picture number of input) Gaussian Profile that is calculated initialized;
5.2) training network model
Stochastic inputs pass through the original image of step 2) processing, and the feature extraction network by step 3.1) obtains accordingly Low-dimensional coding characteristic simplifies network in the sketch by step 3.2) and generates the result figure after simplifying, and counted by step 4.1) Corresponding perception penalty values are calculated, this perception penalty values then can be obtained step 3) sketch by backpropagation simplifies network The gradient of each layer parameter in model, then the gradient made by stochastic gradient descent algorithm optimize each layer parameter, The training of a wheel network model can be realized;
5.3) step 5.2) is repeated until network reaches the set goal about the ability that sketch simplifies.
6) model is verified, and is included the following steps:
The model trained using verify data set pair is verified, and is tested its lines and is simplified ability.At random from verification Some original images are taken out in data set, after step 2) processing, are input to the trained network model of step 5), are allowed the net Network model goes to simplify the sketch inputted, is compared with corresponding label data by the result of output, to judge the training The sketch of good network model simplifies ability.
The above embodiment is a preferred embodiment of the present invention, but embodiments of the present invention are not by above-described embodiment Limitation, it is other it is any without departing from the spirit and principles of the present invention made by changes, modifications, substitutions, combinations, simplifications, Equivalent substitute mode is should be, is included within the scope of the present invention.

Claims (9)

1. the sketch of the depth convolutional neural networks based on perception loss simplifies method, which is characterized in that include the following steps:
1) data acquisition;
Using rough draft figure and corresponding label data, training depth convolutional neural networks need training data, the label data It is exactly the correspondence line manuscript base manually drawn on the basis of rough draft figure, is then divided into training dataset and validation data set;
2) data processing;
The rough draft figure of image data set and corresponding line manuscript base are converted by pretreatment needed for trained depth convolutional neural networks to The format wanted;
3) model construction;
According to training objective and the input/output format of model, constructs one and be suitable for the depth convolution that sketch simplifies problem Neural network;
4) loss function is defined;
According to training objective and the framework of model, required loss function is defined;
5) model training;
The parameter of each layer network is initialized, continuous iteration inputs training sample, the loss of network is calculated according to loss function It is worth, then calculates the gradient of each network layer parameter by backpropagation, by stochastic gradient descent method to the parameter of each layer network It is updated;
6) model is verified;
The model trained using verify data set pair is verified, its Generalization Capability is tested.
2. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 2), data processing is specially:
2.1) image is changed into gray-scale map;
2.2) horizontal, vertical, diagonal turnover is carried out to the image that data are concentrated, corresponding line manuscript base is also same overturning behaviour Make;
2.3) picture after overturning is given to add tone reversal, noise and shade;
2.4) by treated, picture is cut to be sized, and corresponding line manuscript base is also cropped to correspondingly sized;
2.5) image concentrated to data is transformed into from [0,255] in the range of [- 1,1].
3. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 2.4), treated, and picture is cut to 384 × 384 pixel sizes.
4. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 3), model construction is specially:
3.1) sketch simplified network model is constructed;
The sketch simplified network model is responsible for the sketch of input to be simplified to line original text, and input is the grass of an arbitrary size Figure, output are the line manuscript bases after a same size size simplifies;Sketch simplified network model is by two concatenated encoder moulds Type and decoder model are constituted;The encoder model is for extracting high-rise semantic information and guarantor from the sketch of input It is stored in the coding of a low-dimensional;The decoder model is for after being disassembled in the coding of low-dimensional and restoring and be simplified Result figure;
3.2) construction feature extracts network model;
The feature extraction network model is used to extract high-rise semantic information from the image of input and in this high-rise language It is compared under adopted information, the input of feature extraction network is result figure and correspondence after the simplification that step 3.1) is handled Artificial marking line manuscript base need input picture being extended to 3 × 384 × 384, one before inputting pre-training VGG16 networks Under a high-layer semantic information, result figure and the corresponding artificial respective characteristic pattern of marking line manuscript base is taken to be compared.
5. the sketch of the depth convolutional neural networks according to claim 4 based on perception loss simplifies method, feature exists In in the step 3.1), the encoder model includes multiple cascade down-sampling layers, and the down-sampling layer is by concatenated volume Lamination, batch regularization layer and nonlinear activation layer composition, have been gradually reduced the ruler of image while increasing receptive field It is very little, make Web-based Self-regulated Learning down-sampling using the convolutional layer with step-length, batch regularization layer is by normalizing the same batch The mean value and standard deviation of input sample play the role of stablizing and acceleration model training, the addition of nonlinear activation layer prevent Model degradation is simple linear model, improves the descriptive power of model;
The decoder model includes multiple cascade up-sampling layers, the up-sampling layer by concatenated warp lamination, batch just Then change layer and nonlinear activation layer composition, plays the role of encoding enlarged drawing size, the deconvolution with step-length from low-dimensional Layer makes Web-based Self-regulated Learning up-sample, same in the effect of batch regularization layer and nonlinear activation layer and encoder, while I Also introduce residual block to accelerate training speed.
6. the sketch of the depth convolutional neural networks according to claim 4 based on perception loss simplifies method, feature exists In, in the step 3.2),
The feature extraction network model include multiple cascade down-sampling layers, the down-sampling layer by concatenated convolutional layer, with And nonlinear activation function layer, pond layer composition, wherein convolutional layer step-length is 1, and convolution kernel size is 3 × 3, is extracted corresponding Characteristic pattern, it is simple linear model that the addition of nonlinear activation layer, which prevents model degradation, improves the descriptive power of model, Chi Hua The effect of layer is to reduce the size of characteristic pattern, to increase the receptive field of convolution kernel.
7. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 4), defining loss function is specially:
The computational methods of the function are by the result figure after sketch simplifies network reduction and its corresponding prior people Input feature vector extraction network takes result figure and line manuscript base to the line manuscript base that work has marked under a high-rise semantic information together Corresponding characteristic pattern is compared, define loss function make result figure characteristic pattern and line manuscript base characteristic pattern as much as possible It is close;Therefore, perception loss can be defined as by simplifying the loss function of sketch, and formula is as follows:
Wherein, LpecTo perceive loss function,And IyIt indicates to simplify result figure line manuscript base corresponding with its that network generates respectively, Wi,j, Hi,jAnd Ci,jThe width of characteristic pattern, height, number of channels, φ are indicated respectivelyi,j(I) indicate that characteristic pattern function, i and j indicate Feature graph laplacian.
8. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 5), the specific method of model training is:
5.1) each layer parameter of initialization model
The initialization of each layer parameter is using commonly used method in depth convolutional neural networks, to the volume of feature extraction network Lamination parameter is using the convolutional layer parameter value in the good VGG16 network models of ImageNet pre-training as initial value, sketch letter Change network in convolutional layer, use mean value be 0, standard deviation use The Gaussian Profile being calculated carries out initial Change, wherein scale is result figure with respect to the multiple that artwork is amplified, faninIt is the picture number of input;
5.2) training network model
Stochastic inputs pass through the original image of step 2) processing, and the sketch by step 3.1) simplifies network and generates the knot after simplifying Fruit is schemed, and the result figure Jing Guo step 3.1) and corresponding artificial marking line manuscript base, the feature extraction network by step 3.2) are inputted Corresponding characteristic pattern is obtained, corresponding perception penalty values is calculated and by step 4.1), then leads to this perception penalty values The gradient for each layer parameter that backpropagation can obtain in step 3.1) sketch simplified network model is crossed, then by under stochastic gradient The gradient that drop algorithm makes optimizes each layer parameter, you can realizes the training of a wheel network model;
5.3) step 5.2) is repeated until network reaches the set goal about the ability that sketch simplifies.
9. the sketch of the depth convolutional neural networks according to claim 1 based on perception loss simplifies method, feature exists In in the step 6), the specific method of model verification is:
It is concentrated at random from verify data and takes out some original images, after step 2) processing, it is trained to be input to step 5) Network model allows the network model to go to simplify the sketch inputted, by the result of output and corresponding artificial marking line manuscript base into Row compares, to judge that the sketch of the trained network model simplifies ability.
CN201810259452.XA 2018-03-27 2018-03-27 Sketch simplification method of deep convolutional neural network based on perception loss Active CN108510456B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810259452.XA CN108510456B (en) 2018-03-27 2018-03-27 Sketch simplification method of deep convolutional neural network based on perception loss

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810259452.XA CN108510456B (en) 2018-03-27 2018-03-27 Sketch simplification method of deep convolutional neural network based on perception loss

Publications (2)

Publication Number Publication Date
CN108510456A true CN108510456A (en) 2018-09-07
CN108510456B CN108510456B (en) 2021-12-21

Family

ID=63378698

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810259452.XA Active CN108510456B (en) 2018-03-27 2018-03-27 Sketch simplification method of deep convolutional neural network based on perception loss

Country Status (1)

Country Link
CN (1) CN108510456B (en)

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109394205A (en) * 2018-09-30 2019-03-01 安徽心之声医疗科技有限公司 A kind of more illness analysis methods of electrocardiosignal based on deep neural network
CN109727253A (en) * 2018-11-14 2019-05-07 西安大数据与人工智能研究院 Divide the aided detection method of Lung neoplasm automatically based on depth convolutional neural networks
CN109784490A (en) * 2019-02-02 2019-05-21 北京地平线机器人技术研发有限公司 Training method, device and the electronic equipment of neural network
CN109801346A (en) * 2018-12-20 2019-05-24 武汉西山艺创文化有限公司 A kind of original painting neural network based auxiliary painting methods and device
CN109886986A (en) * 2019-01-23 2019-06-14 北京航空航天大学 A kind of skin lens image dividing method based on multiple-limb convolutional neural networks
CN110298899A (en) * 2019-06-10 2019-10-01 天津大学 One kind being based on the matched image texture synthetic method of convolutional neural networks characteristic pattern
CN110866469A (en) * 2019-10-30 2020-03-06 腾讯科技(深圳)有限公司 Human face facial features recognition method, device, equipment and medium
CN110930472A (en) * 2019-11-14 2020-03-27 三星电子(中国)研发中心 Picture generation method and device
CN111080605A (en) * 2019-12-12 2020-04-28 哈尔滨市科佳通用机电股份有限公司 Method for identifying railway wagon manual brake shaft chain falling fault image
CN111222519A (en) * 2020-01-16 2020-06-02 西北大学 Construction method, method and device of hierarchical colored drawing manuscript line extraction model
CN111311707A (en) * 2020-03-05 2020-06-19 云知声智能科技股份有限公司 Drawing method and device
WO2020177701A1 (en) * 2019-03-07 2020-09-10 腾讯科技(深圳)有限公司 Image processing method and apparatus, and computer device and storage medium
CN112101449A (en) * 2020-09-11 2020-12-18 南京邮电大学 Method for inverting radar data by satellite cloud picture based on semantic loss
CN112418197A (en) * 2021-01-22 2021-02-26 北京世纪好未来教育科技有限公司 Simplified image acquisition model training method, image text recognition method and related device
CN112800998A (en) * 2021-02-05 2021-05-14 南京邮电大学 Multi-mode emotion recognition method and system integrating attention mechanism and DMCCA
CN112837396A (en) * 2021-01-29 2021-05-25 深圳市天耀创想网络科技有限公司 Line draft generation method and device based on machine learning
CN112889084A (en) * 2018-11-08 2021-06-01 Oppo广东移动通信有限公司 Method, system and computer readable medium for improving color quality of image
CN113393546A (en) * 2021-05-17 2021-09-14 杭州电子科技大学 Fashion clothing image generation method based on clothing category and texture pattern control
CN114299184A (en) * 2021-12-30 2022-04-08 青海师范大学 Hidden building colored drawing line manuscript graph coloring method and device based on semantic matching
CN117649365A (en) * 2023-11-16 2024-03-05 西南交通大学 Paper book graph digital restoration method based on convolutional neural network and diffusion model

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103902988A (en) * 2014-04-21 2014-07-02 梁爽 Method for rough draft shape matching based on Modular product graph and maximum clique
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
CN107103590A (en) * 2017-03-22 2017-08-29 华南理工大学 A kind of image for resisting generation network based on depth convolution reflects minimizing technology
CN107220277A (en) * 2017-04-14 2017-09-29 西北大学 Image retrieval algorithm based on cartographical sketching

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103902988A (en) * 2014-04-21 2014-07-02 梁爽 Method for rough draft shape matching based on Modular product graph and maximum clique
CN106126581A (en) * 2016-06-20 2016-11-16 复旦大学 Cartographical sketching image search method based on degree of depth study
CN107103590A (en) * 2017-03-22 2017-08-29 华南理工大学 A kind of image for resisting generation network based on depth convolution reflects minimizing technology
CN107220277A (en) * 2017-04-14 2017-09-29 西北大学 Image retrieval algorithm based on cartographical sketching

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
EDGAR SIMO-SERRA.ET AL: "Mastering Sketching: Adversarial Augmentation for Structured Prediction", 《ARXIV》 *
JUSTIN JOHNSON.ET AL: "Perceptual Losses for Real-Time Style Transfer and Super-Resolution", 《ARXIV》 *

Cited By (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109394205B (en) * 2018-09-30 2022-06-17 安徽心之声医疗科技有限公司 Electrocardiosignal analysis method based on deep neural network
CN109394205A (en) * 2018-09-30 2019-03-01 安徽心之声医疗科技有限公司 A kind of more illness analysis methods of electrocardiosignal based on deep neural network
US11972543B2 (en) 2018-11-08 2024-04-30 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Method and terminal for improving color quality of images
CN112889084A (en) * 2018-11-08 2021-06-01 Oppo广东移动通信有限公司 Method, system and computer readable medium for improving color quality of image
CN109727253A (en) * 2018-11-14 2019-05-07 西安大数据与人工智能研究院 Divide the aided detection method of Lung neoplasm automatically based on depth convolutional neural networks
CN109801346A (en) * 2018-12-20 2019-05-24 武汉西山艺创文化有限公司 A kind of original painting neural network based auxiliary painting methods and device
CN109801346B (en) * 2018-12-20 2023-06-30 武汉西山艺创文化有限公司 Original painting auxiliary coloring method and device based on neural network
CN109886986A (en) * 2019-01-23 2019-06-14 北京航空航天大学 A kind of skin lens image dividing method based on multiple-limb convolutional neural networks
US11645537B2 (en) 2019-02-02 2023-05-09 Beijing Horizon Robotics Technology Research And Development Co., Ltd. Neural network training method, neural network training apparatus and electronic device
CN109784490B (en) * 2019-02-02 2020-07-03 北京地平线机器人技术研发有限公司 Neural network training method and device and electronic equipment
CN109784490A (en) * 2019-02-02 2019-05-21 北京地平线机器人技术研发有限公司 Training method, device and the electronic equipment of neural network
US11900567B2 (en) 2019-03-07 2024-02-13 Tencent Technology (Shenzhen) Company Limited Image processing method and apparatus, computer device, and storage medium
KR102509817B1 (en) 2019-03-07 2023-03-14 텐센트 테크놀로지(센젠) 컴퍼니 리미티드 Image processing method and apparatus, computer device, and storage medium
WO2020177701A1 (en) * 2019-03-07 2020-09-10 腾讯科技(深圳)有限公司 Image processing method and apparatus, and computer device and storage medium
KR20210107084A (en) * 2019-03-07 2021-08-31 텐센트 테크놀로지(센젠) 컴퍼니 리미티드 Image processing method and apparatus, computer device, and storage medium
CN110298899B (en) * 2019-06-10 2023-04-07 天津大学 Image texture synthesis method based on convolutional neural network feature map matching
CN110298899A (en) * 2019-06-10 2019-10-01 天津大学 One kind being based on the matched image texture synthetic method of convolutional neural networks characteristic pattern
CN110866469A (en) * 2019-10-30 2020-03-06 腾讯科技(深圳)有限公司 Human face facial features recognition method, device, equipment and medium
CN110930472A (en) * 2019-11-14 2020-03-27 三星电子(中国)研发中心 Picture generation method and device
CN111080605A (en) * 2019-12-12 2020-04-28 哈尔滨市科佳通用机电股份有限公司 Method for identifying railway wagon manual brake shaft chain falling fault image
CN111222519A (en) * 2020-01-16 2020-06-02 西北大学 Construction method, method and device of hierarchical colored drawing manuscript line extraction model
CN111222519B (en) * 2020-01-16 2023-03-24 西北大学 Construction method, method and device of hierarchical colored drawing manuscript line extraction model
CN111311707A (en) * 2020-03-05 2020-06-19 云知声智能科技股份有限公司 Drawing method and device
CN111311707B (en) * 2020-03-05 2023-05-05 云知声智能科技股份有限公司 Painting method and device
CN112101449A (en) * 2020-09-11 2020-12-18 南京邮电大学 Method for inverting radar data by satellite cloud picture based on semantic loss
CN112101449B (en) * 2020-09-11 2022-07-15 南京邮电大学 Method for inverting radar data by satellite cloud picture based on semantic loss
CN112418197A (en) * 2021-01-22 2021-02-26 北京世纪好未来教育科技有限公司 Simplified image acquisition model training method, image text recognition method and related device
CN112837396A (en) * 2021-01-29 2021-05-25 深圳市天耀创想网络科技有限公司 Line draft generation method and device based on machine learning
CN112837396B (en) * 2021-01-29 2024-05-07 深圳市天耀创想网络科技有限公司 Line manuscript generation method and device based on machine learning
CN112800998B (en) * 2021-02-05 2022-07-29 南京邮电大学 Multi-mode emotion recognition method and system integrating attention mechanism and DMCCA
CN112800998A (en) * 2021-02-05 2021-05-14 南京邮电大学 Multi-mode emotion recognition method and system integrating attention mechanism and DMCCA
CN113393546A (en) * 2021-05-17 2021-09-14 杭州电子科技大学 Fashion clothing image generation method based on clothing category and texture pattern control
CN113393546B (en) * 2021-05-17 2024-02-02 杭州电子科技大学 Fashion clothing image generation method based on clothing type and texture pattern control
CN114299184B (en) * 2021-12-30 2022-09-06 青海师范大学 Hidden building colored drawing line manuscript painting method and device based on semantic matching
CN114299184A (en) * 2021-12-30 2022-04-08 青海师范大学 Hidden building colored drawing line manuscript graph coloring method and device based on semantic matching
CN117649365A (en) * 2023-11-16 2024-03-05 西南交通大学 Paper book graph digital restoration method based on convolutional neural network and diffusion model

Also Published As

Publication number Publication date
CN108510456B (en) 2021-12-21

Similar Documents

Publication Publication Date Title
CN108510456A (en) The sketch of depth convolutional neural networks based on perception loss simplifies method
CN109191382B (en) Image processing method, device, electronic equipment and computer readable storage medium
CN105894045B (en) A kind of model recognizing method of the depth network model based on spatial pyramid pond
CN109064396A (en) A kind of single image super resolution ratio reconstruction method based on depth ingredient learning network
CN105426919B (en) The image classification method of non-supervisory feature learning is instructed based on conspicuousness
CN106845529A (en) Image feature recognition methods based on many visual field convolutional neural networks
US9449253B2 (en) Learning painting styles for painterly rendering
CN108304826A (en) Facial expression recognizing method based on convolutional neural networks
CN107464210A (en) A kind of image Style Transfer method based on production confrontation network
CN107644426A (en) Image, semantic dividing method based on pyramid pond encoding and decoding structure
CN108009629A (en) A kind of station symbol dividing method based on full convolution station symbol segmentation network
CN111242841A (en) Image background style migration method based on semantic segmentation and deep learning
CN108898138A (en) Scene text recognition methods based on deep learning
CN110321967A (en) Image classification innovatory algorithm based on convolutional neural networks
CN107679572A (en) A kind of image discriminating method, storage device and mobile terminal
CN110992374B (en) Hair refinement segmentation method and system based on deep learning
CN108537120A (en) A kind of face identification method and system based on deep learning
CN109598676A (en) A kind of single image super-resolution method based on Hadamard transform
CN106022355A (en) 3DCNN (three-dimensional convolutional neural network)-based high-spectral image space spectrum combined classification method
CN114066871B (en) Method for training new coronal pneumonia focus area segmentation model
CN107967497A (en) Manuscripted Characters Identification Method based on convolutional neural networks and extreme learning machine
CN110175248A (en) A kind of Research on face image retrieval and device encoded based on deep learning and Hash
CN109447897B (en) Real scene image synthesis method and system
CN107679539A (en) A kind of single convolutional neural networks local message wild based on local sensing and global information integration method
CN107958219A (en) Image scene classification method based on multi-model and Analysis On Multi-scale Features

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant