WO2023015610A1 - Artificial intelligence-based method and system for authenticating ancient and modern artwork - Google Patents

Artificial intelligence-based method and system for authenticating ancient and modern artwork Download PDF

Info

Publication number
WO2023015610A1
WO2023015610A1 PCT/CN2021/114254 CN2021114254W WO2023015610A1 WO 2023015610 A1 WO2023015610 A1 WO 2023015610A1 CN 2021114254 W CN2021114254 W CN 2021114254W WO 2023015610 A1 WO2023015610 A1 WO 2023015610A1
Authority
WO
WIPO (PCT)
Prior art keywords
distribution
image information
sample
samples
classification
Prior art date
Application number
PCT/CN2021/114254
Other languages
French (fr)
Chinese (zh)
Inventor
李應樵
马志雄
Original Assignee
万维数码智能有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 万维数码智能有限公司 filed Critical 万维数码智能有限公司
Publication of WO2023015610A1 publication Critical patent/WO2023015610A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/764Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks

Definitions

  • the invention belongs to the field of identification of ancient and modern works of art, in particular to a method and system for identifying ancient and modern works of art by using artificial intelligence.
  • CN107341461A discloses a method and system for identifying the authenticity of artworks with intelligent identification and analysis technology, classifying artworks into dead artworks and living artworks; establishing a database for artworks, storing all information of artworks, and analyzing the artist's
  • the picture system performs intelligent analysis and storage; enters the information of artworks, and the system performs self-learning evolution, and performs target source matching for artworks that need to be identified; if it exists, it compares the authenticity; if it does not exist Then compare the style of the artwork deduced by self-learning, and get the final result of authenticity identification; this solves the current situation that the art identification field cannot be systematized and standardized; through intelligent identification and analysis technology, and based on A self-learning evolution system developed on the basis of the database established by the artwork image information.
  • CN109191145A discloses a method for establishing a database for judging the age of artworks and a method for judging the age of artworks.
  • the method for establishing a database for judging the age of artworks includes the following steps: (1) selecting the same age, same type, At least two artwork specimens of the same style; (2) extracting the total field of view image; (3) establishing a database; a method for identifying the age of artworks which includes the following steps: I. using the database established above; II. extracting the art to be determined III. Analyze the extracted image on the artwork to be determined and the image saved in the database using image recognition technology; IV. Judgment.
  • CN111339974A discloses a method for identifying modern ceramics and ancient ceramics, by constructing positive samples corresponding to ancient ceramics and negative samples corresponding to antique porcelain, converting RGB images to HSV color space to obtain HSV images, and obtaining feature descriptors of HSV images , input the feature descriptor into the support vector machine for training, obtain the training parameters of the support vector machine, input the RGB image into the deep convolutional neural network architecture for training to obtain the network parameters of the convolutional neural network, according to the training parameters of the support vector machine and
  • the network parameters of the convolutional neural network determine the deep learning model, input the grayscale image of the positive sample and the grayscale image of the negative sample into the deep learning model for training, obtain the identification model, obtain the image of the porcelain to be identified, and convert the image to be identified
  • the picture is input into the identification model, and according to the output result of the identification model, it is determined whether the porcelain to be identified is modern ceramics or ancient ceramics, so as to improve the efficiency of corresponding ceramic identification.
  • the object of the present invention is to provide a method and system for appraising ancient and modern artworks by using artificial intelligence.
  • One aspect of the present invention provides a method for identifying ancient and modern works of art, including: inputting authentic image information; and inputting image information of artworks to be identified; combining the image information of artworks to be identified with the Authentic image information detects the samples in the distribution and the samples out of the distribution through the detector; classifies the samples in the distribution; classifies the samples in the distribution after classification and class image information similar to the artwork to be identified Granular classification: output classified samples in the distribution or samples after fine-grained classification, and obtain the confidence degree of the image information of the artwork to be identified compared with the authentic image information as the identification conclusion.
  • the step of detecting the in-distribution samples and out-of-distribution samples through the detector to detect the image information of the artwork to be identified and the image information of the authentic works further includes: using a pre-trained model (The maximum normalized index probability output by the pre-trained model) is used for statistical analysis; the distribution of the normalized index probability of the OOD sample and the ID sample is found statistically; the distribution gap between the two is increased; an appropriate threshold is selected to judge a sample Whether it is an out-of-distribution sample or a sample in-distribution.
  • a pre-trained model The maximum normalized index probability output by the pre-trained model
  • the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through a detector further includes: using a model to learn a pair The uncertainty attribute of the input sample; to judge the test data, if the model input is a sample in the distribution, the uncertainty is low; on the contrary, if the model input is an out-of-distribution sample, the uncertainty is high.
  • the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through a detector further includes: using variational automatic coding Variational Autoencoder (Variational Autoencoder) reconstruction error (reconstruction error) or other measurement methods to determine whether a sample belongs to the distribution or out-of-distribution samples; the latent space of the encoder can learn the obviousness of the data in the distribution feature (silent vector), but not for out-of-distribution samples, so out-of-distribution samples will produce higher reconstruction errors.
  • Variational Autoencoder Variational Autoencoder
  • reconstruction error reconstruction error
  • the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through the detector further includes: using a classifier to extract Classify the features of the distribution to determine whether it is an out-of-distribution sample; some modify the network structure to an n+1 classifier, n is the number of categories of the original classification task, and the n+1th class is an out-of-distribution class; some directly take Extract features for classification without modifying the structure of the network.
  • the step of fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified further includes: finding the image to be tested The feature area of the data; the feature area is input into the convolutional neural network; a part of the information of the feature area of the convolutional neural network enters the fully connected layer and the normalized exponential logistic regression layer for classification; through the volume Another part of the information of the feature region of the product neural network passes through the attention suggestion sub-network (APN) to obtain the candidate region; repeat the above-mentioned classification steps and APN steps, so that the feature region selected by the APN is the most discriminative region; introduce A loss function to obtain higher accuracy in identifying the image information of the class.
  • APN attention suggestion sub-network
  • the step of fine-grained classifying the classified samples in the distribution and the image-like information similar to the artwork to be identified further includes:
  • the local images (312) and (313) selected in the information (311) are input into two convolutional neural networks (314, A) and (315, B); the output of the convolutional neural network streams (A) and (B) is in Each location of the image is multiplied (318) using the outer product and combined to obtain a bilinear vector (316); the prediction is obtained through the classification layer (317).
  • classification layer (317) is a logistic regression or support vector machine classifier.
  • the step of fine-grained classifying the classified samples in the distribution and the image-like information similar to the artwork to be identified further includes: The information generates multiple candidate boxes on different scale feature maps, and the coordinates of each candidate box correspond to the pre-designed anchors; the "information content" of each candidate area is scored, and the area with a large amount of information has a high score;
  • the above feature map is followed by the feature extraction step, the fully connected layer (FC) and the normalized index step; the probability that the input area belongs to the target label is judged; the unnormalized probability extracted from each local area and the whole map is merged together Generates a long vector outputting the unnormalized probabilities for the 200 classes.
  • the present invention also provides a system for identifying ancient and modern works of art, including: an input module for inputting authentic image information; and inputting image information for artworks to be identified; The information and the authentic image information are detected by the detector to detect the samples in the distribution and the samples out of the distribution; the sample classification module classifies the samples in the distribution; the fine-grained classification module combines the classified samples in the distribution with the simulated Fine-grained classification of similar image information of the identified works of art; the output module outputs the samples in the distribution after classification or the samples after fine-grained classification, and obtains the image information of the artwork to be identified and the authentic The confidence level compared with the image information is used as the identification conclusion.
  • Fig. 1 is the flow chart of the ancient and modern works of art appraisal method of the present invention.
  • Figure 2(a)-(d) is a flow chart of the steps of detecting samples in distribution (in distribution) and out of distribution (out of distribution (OOD)) in the steps of the ancient and modern art identification method of the present invention.
  • Figure 2(a) is a flowchart of a normalized index based embodiment.
  • Fig. 2(b) is a flowchart of an embodiment of uncertainty.
  • Figure 2(c) is a flowchart of an embodiment of a probabilistic generative model.
  • Figure 2(d) is a flowchart of an embodiment of a classification model.
  • Fig. 3 (a) is a flow chart of the implementation of the attention convolutional neural network in the fine-grained classification step in the steps of the ancient and modern art identification method of the present invention.
  • Fig. 3(b) is a schematic diagram of the framework of a recurrent attention convolutional neural network (“RA-CNN”) for an implementation of the fine-grained classification step in the identification method of the present invention.
  • RA-CNN recurrent attention convolutional neural network
  • Fig. 3(c) is a schematic diagram of the bilinear vector network structure of another embodiment of the fine-grained classification step in the identification method of the present invention.
  • Fig. 3(d) is a flow chart of the implementation mode of bilinear vector network in the step of fine-grained classification in the steps of the ancient and modern art identification method of the present invention.
  • Fig. 3(e) is a flow chart of an embodiment in which the fine-grained classification step adopts the navigation-teaching-examination network (NTS-Net) classification in the steps of the ancient and modern artwork appraisal method of the present invention.
  • NTS-Net navigation-teaching-examination network
  • Fig. 4 is a structural diagram of the ancient and modern art identification system of the present invention.
  • Fig. 5 is a computer product diagram of the portable or fixed storage unit of the ancient and modern art identification system of the present invention.
  • Fig. 6(1) is an example of authentic image information involved in an embodiment of the ancient and modern artwork identification method of the present invention.
  • Fig. 6(2) is an example of authentic image information used to train the model involved in an embodiment of the ancient and modern artwork identification method of the present invention.
  • Fig. 6(3) is an example of the image information of the artwork to be authenticated involved in one embodiment of the ancient and modern artwork authentication method of the present invention.
  • Fig. 6 (4) is an example of the classification involved in an implementation of the ancient and modern artwork identification method of the present invention.
  • Fig. 1 is the flow chart of the ancient and modern works of art appraisal method of the present invention.
  • step 101 the image information and authentic image information of the artwork to be identified are input;
  • step 102 the image information of the artwork to be identified and the authentic image information are detected by a detector in distribution ) samples and distribution (out of distribution, OOD) samples;
  • step 103 the samples in the distribution are classified;
  • step 104 the samples in the distribution after classification are similar to the artwork to be identified Fine-grained classifier is performed on the class image information; in step 105, the classified samples in the distribution or samples after fine-grained classification are output to obtain the identification conclusion.
  • Figure 2(a)-(d) is a flow chart of the steps of detecting samples in distribution (in distribution) and out of distribution (out of distribution (OOD)) in the steps of the ancient and modern art identification method of the present invention.
  • Fig. 2 (a) is the flow chart of the embodiment based on normalized index
  • Fig. 2 (b) is the flow chart of the embodiment of uncertainty
  • Fig. 2 (c) is the flow chart of the embodiment of probability generation model
  • Figure 2(d) is a flowchart of an embodiment of a classification model.
  • the image-like data for model training and testing are independent and identically distributed (IID, Independent Identical Distribution) samples.
  • ID samples the data obtained after the model is deployed and launched is often not fully controlled, that is to say, the data received by the model may be out-of-distribution (OOD) samples, also known as outlier samples (outlier, abnormal).
  • OOD out-of-distribution
  • the depth model will consider an out-of-distribution (OOD) sample as a certain class in the distribution (ID) sample, and give a high degree of confidence.
  • the confidence degree described here is a normalized value of 0-1. Find out-of-distribution samples, but their settings may be different. For example, out-of-distribution detection (OOD detection) is modified on the model task, which requires not only to be able to effectively detect out-of-distribution (OOD) samples, but also to ensure that the performance of the model is not affected.
  • the detection distribution and out-of-distribution steps for the image-like data of ancient and modern artworks can be based on normalization index (Softmax-based), uncertainty (Uncertainty), probability generation model (Generative model), classification model (Classifier) method to detect in-distribution and out-of-distribution sample methods.
  • Softmax-based normalization index
  • Uncertainty uncertainty
  • Geneative model probability generation model
  • Classifier classification model
  • step 201 statistical analysis is performed using the maximum normalized index probability output by the pre-trained model (pre-trained model), and in step 202, statistically found OOD samples and For the distribution of the normalized index probability of the ID sample, in step 203, the distribution gap between the two is increased, and in step 204, an appropriate threshold is selected to determine whether a sample belongs to an out-of-distribution sample or an in-distribution sample.
  • This type of method is simple and effective, without modifying the structure of the classification model, and without training an out-of-distribution sample classifier.
  • the model is used to learn an uncertainty attribute for the input samples.
  • the test data is judged. If the model input is a sample in the distribution, the uncertainty is low; on the contrary, if the model input is an out-of-distribution sample, the uncertainty is high.
  • Such methods need to modify the network structure of the model to learn the uncertainty property.
  • step 221 use the reconstruction error (reconstruction error) of the variational autoencoder (Variational Autoencoder) or other measurement methods to judge whether a sample belongs to the sample in the distribution or out of the distribution;
  • the hidden space (latent space) of the encoder can learn the obvious features (silent vector) of the data in the distribution, but not for the out-of-distribution samples, so the out-of-distribution samples will generate higher reconstruction errors .
  • This method only focuses on out-of-distribution detection performance, and does not focus on the original task of the data in the distribution.
  • a classifier is used to classify the extracted features to determine whether it is an out-of-distribution sample; in step 232, some modify the network structure to be an n+1 classifier, n is the number of categories of the original classification task, and the n+1th category is an out-of-distribution category; in step 233, some features are directly extracted for classification without modifying the structure of the network.
  • Fig. 3 (a) is a flow chart of the implementation of the attention convolutional neural network in the fine-grained classification step in the steps of the ancient and modern art identification method of the present invention.
  • step 321 the characteristic area of the image data to be tested is searched, and in step 322, the characteristic area is input into the convolutional neural network; in step 323, a part of the information of the characteristic area through the convolutional neural network is entered Fully connected layer and normalized exponential logistic regression layer are classified; in step 324, another part of the information of the feature region through the convolutional neural network is passed through the attention suggestion subnetwork (APN), to obtain the candidate region; in step 325, repeating the step 323 and the step 324, so that the feature region selected by the APN is the most discriminative region; in step 326, introducing a loss function to obtain a higher accuracy of identifying the type of image information.
  • APN attention suggestion subnetwork
  • the "fine-grained" classification step is under the ordinary classification. For a more fine-grained division, it is necessary to explicitly find the most “discriminative" features in the picture. For ancient and modern works of art, it is necessary to find the characteristics of details, such as the degree of upturning of petals, the nuances of patterns, etc.
  • Fig. 3(b) is a schematic diagram of the framework of a recurrent attention convolutional neural network ("RA-CNN") for an implementation of the fine-grained classification step in the identification method of the present invention.
  • RA-CNN recurrent attention convolutional neural network
  • the symbol It means to cut a part of the characteristic area of the identified image-like information and enlarge it.
  • Each row 301, 302, 303 represents a common CNN network respectively.
  • the input ranges from coarse full-scale images to finer region attention (from top to bottom).
  • the picture (a 1 ) in the first row 301 is the roughest, and the picture (a 3 ) in the third row is finer.
  • a 1 After the image information a 1 enters b 1 (several convolutional layers), it is divided into two paths, all the way to c 1 and connected to fully connected layers (fully connected layers, FC) and softmax logistic regression layer for simple classification, and the other path enters d 1 is the attention proposal sub-network ("Attention Proposal Network", APN), get a candidate area.
  • APN Attention Proposal Network
  • the feature area is continuously enlarged and refined after two APNs.
  • a loss function (Ranking loss) is introduced: that is, the forced area a 1 , a 2 , a 3
  • the classification confidence level (confidence score) is getting higher and higher (that is, the corresponding P t probability of the last column of the picture is getting higher and higher), which means that the accuracy of identifying image information is getting higher and higher.
  • the network continuously refines the discriminative attention region.
  • Fig. 3(c) is a schematic diagram of the bilinear vector network structure of another embodiment of the fine-grained classification step in the identification method of the present invention.
  • Partial images 312 and 313 selected from the identified class image information 311 are input into two convolutional neural networks 314(A) and 315(B).
  • the outputs of the convolutional neural network streams A and B are multiplied 318 by outer product at each position of the image and combined to obtain a bilinear vector 316 , which is then passed through a classification layer 317 to obtain a prediction result.
  • f A and f B represent feature extraction functions, that is, convolutional network A and convolutional network B in Figure 3(c)
  • P is a pooling function (Pooling function)
  • C is a classification function.
  • the feature extraction function f( ) (i.e., the convolutional neural network stream CNN stream) consists of convolutional layers, pooling layers, and activation functions. This part of the network structure can be regarded as a function map:
  • the output of the pooling function P is an M ⁇ N matrix.
  • the feature matrix Stretched into a list of MN-sized feature vectors.
  • a classification function is used to classify the extracted features, and the classification layer 317 is implemented using a logistic regression or a support vector machine (support vector machine, SVM) classifier.
  • the CNN network can achieve high-level semantic feature acquisition for fine-grained images, and filter irrelevant background information in the image by iteratively training the convolution parameters in the network model.
  • the convolutional neural network flow A and the convolutional neural network flow B play complementary roles in the image recognition task, that is, the network A can locate the object in the image, and the network B can complete the positioning of the network A to Feature extraction of the object position.
  • the two networks can cooperate to complete the class detection and target feature removal process of the input fine-grained image, and better complete the fine-grained image recognition task.
  • Fig. 3(d) is a flow chart of the implementation mode of bilinear vector network in the step of fine-grained classification in the steps of the ancient and modern art identification method of the present invention.
  • the partial images 312 and 313 selected in the identified class image information 311 are input into two convolutional neural networks 314 (A) and 315 (B); in step 332, the convolutional neural network streams A and B
  • the output is multiplied 318 using an outer product at each location in the image and combined to obtain a bilinear vector 316 , and at step 333 the prediction is obtained through the classification layer 317 .
  • Fig. 3(e) is a flow chart of an embodiment in which the fine-grained classification step adopts the navigation-teaching-examination network (NTS-Net) classification in the steps of the ancient and modern artwork appraisal method of the present invention.
  • step 341 generate a plurality of candidate boxes on different scale feature maps (Feature maps) with the identified class image information, and the coordinates of each candidate box correspond to pre-designed anchors (Anchors); in step 342, give The "information content" of each candidate area is scored, and the area with a large amount of information has a high score; in step 343, a feature extraction step (Feature Extractor), a fully connected layer (FC) and a normalization index ( softmax) step; in step 344, determine the probability that the input region belongs to the target label (target label); in step 345, merge (concat) together the unnormalized probability (logits) extracted from each local area and the whole picture Generates a long vector outputting unnormalized probabilities (logits) corresponding to 200 categories.
  • the fine-grained classification step in the identification method of the present invention can also be adopted, and the navigation-teaching-examination network (NTS-Net) classification method of dividing the network subject into three components of navigation (Navigator), teaching (Teacher), and examination (Scrutinizer),
  • NTS-Net navigation-teaching-examination network
  • multiple candidate boxes are generated on feature maps of different scales, and the coordinates of each candidate box correspond to the pre-designed anchors (Anchors).
  • the Navigator scores the "information content" of each candidate area, and the area with a large amount of information has a higher score.
  • the teaching step is the commonly used feature extraction step (Feature Extractor), fully connected layer (FC) and normalized index (softmax) step, to judge the probability that the input area belongs to the target label (target label);
  • the review step is a fully connected layer,
  • the input is to combine (concat) the unnormalized probability (logits) extracted from each local area and the whole image together to generate a long vector, and output the unnormalized probability (logits) corresponding to 200 categories.
  • the specific steps of using this NTS method are: 1) The original image of size (448, 448, 3) enters the network, and after entering the Resnet-50 to extract features, it becomes a (14, 14, 2048) feature map, a A 2048-dimensional feature vector after the global pooling layer and a 200-dimensional unnormalized probability after the global pooling layer and the fully connected layer. 2)
  • the preset network (RPN) for generating candidate regions generates correspondences according to different sizes (Size) and aspect ratios on the three scales of (14, 14) (7, 7) (4, 4) There are 1614 anchors in total.
  • NMS Non-Maximum Suppression
  • Fig. 4 is a structural diagram of the ancient and modern art identification system of the present invention.
  • the server 401 of the ancient and modern artwork appraisal system includes a processor 410, where the processor can be a general-purpose or special-purpose chip (ASIC/eASIC) or FPGA or NPU, etc., and a computer program product in the form of a memory 420 or a computer-programmable Read media.
  • Memory 420 may be electronic memory such as flash memory, EEPROM (Electrically Erasable Programmable Read Only Memory), EPROM, hard disk, or ROM.
  • the memory 420 has a storage space 430 for program codes for performing any method steps in the methods described above.
  • the storage space 430 for program codes may include respective program codes 431 for respectively implementing various steps in the above methods.
  • These program codes can be read or written into the processor 410 .
  • These computer program products comprise program code carriers such as hard disks, compact disks (CDs), memory cards or floppy disks.
  • Such a computer program product is typically a portable or fixed storage unit as described with reference to FIG. 5 .
  • Fig. 5 is a computer product diagram of the portable or fixed storage unit of the ancient and modern art identification system of the present invention.
  • the storage unit may have storage segments, storage spaces, etc. arranged similarly to the memory 420 in the server of FIG. 4 .
  • the program code can eg be compressed in a suitable form.
  • the storage unit includes computer readable code 431', i.e. code readable by, for example, a processor such as 410, which code, when executed by the server, causes the server to perform the various steps in the methods described above. These codes, when executed by the server, cause the server to perform the steps of the methods described above.
  • Fig. 6(1) is an example of authentic image information involved in an embodiment of the ancient and modern artwork identification method of the present invention.
  • Figure 6(2) is an example of authentic image information used to train the model.
  • Figure 6(3) is an example of the image information of the artwork to be identified.
  • Figure 6(4) is an example of classification.
  • Fig. 6(1) shows an example of a certain image information of an authentic artwork. Taking the multiple image information obtained from 360 degrees of the authentic artwork given in Fig. 6 (2) as a standard, Fig.
  • references herein to "one embodiment,” “an embodiment,” or “one or more embodiments” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Additionally, please note that examples of the word “in one embodiment” herein do not necessarily all refer to the same embodiment.

Abstract

A method and system for authenticating ancient and modern artwork that uses artificial intelligence. The method comprises: inputting image information of a genuine work of art and image information of artwork to be authenticated (101); detecting, by means of a detector, in-distribution samples and out-of-distribution samples from the image information of the artwork to be authenticated and the image information of the genuine work of art (102); classifying the in-distribution samples (103); performing fine-grained classification on the classified in-distribution samples and class image information that is similar to the artwork to be authenticated (104); and outputting the classified in-distribution samples or the fine-grained classified samples (105, 105'), and obtaining the confidence of the image information of the artwork to be authenticated compared with the image information of the genuine work of art as the authentication conclusion. Thus, the accuracy of authentication is improved, and the amount of computation during model training is reduced.

Description

基于人工智能的古代及近现代艺术品鉴定方法和系统Artificial intelligence-based identification method and system for ancient and modern works of art 技术领域technical field
本发明属于古代及近现代艺术品鉴定领域,特别涉及一种运用人工智能对古代及近现代艺术品进行鉴定的方法和系统。The invention belongs to the field of identification of ancient and modern works of art, in particular to a method and system for identifying ancient and modern works of art by using artificial intelligence.
背景技术Background technique
随着古代及近现代艺术品交易的发展,对于被交易艺术品的真伪判别一直是该领域探讨的核心问题。为了保证鉴定结论的准确性,越来越多的人工智能技术研究成果被应用到该领域。With the development of ancient and modern art trade, the authenticity of the traded art has always been the core issue in this field. In order to ensure the accuracy of identification conclusions, more and more artificial intelligence technology research results have been applied to this field.
CN107341461A公开智能识别与分析技术鉴别艺术品真伪的方法及其系统,将艺术品进行分类,分为过世艺术品和在世艺术品;艺术品进行数据库建立,存储艺术品全部信息,并对艺术家的图片系统进行智能分析与存储;录入艺术品的信息、系统进行自学习演进,对需要鉴别的艺术品;针对需要鉴别的艺术品进行目标源匹配,如果存在,则进行真伪对比;如果不存在则根据自学习来推演出来的艺术品应该的风格进行对比,得出真伪鉴别的最终结果;本解决了目前艺术品鉴别领域无法系统化,标准化的现状;通过智能识别与分析技术,以及基于艺术品图像信息建立起来的数据库为基础发展起来的自学习演进系统。CN107341461A discloses a method and system for identifying the authenticity of artworks with intelligent identification and analysis technology, classifying artworks into dead artworks and living artworks; establishing a database for artworks, storing all information of artworks, and analyzing the artist's The picture system performs intelligent analysis and storage; enters the information of artworks, and the system performs self-learning evolution, and performs target source matching for artworks that need to be identified; if it exists, it compares the authenticity; if it does not exist Then compare the style of the artwork deduced by self-learning, and get the final result of authenticity identification; this solves the current situation that the art identification field cannot be systematized and standardized; through intelligent identification and analysis technology, and based on A self-learning evolution system developed on the basis of the database established by the artwork image information.
CN109191145A公开了一种用于对艺术品进行年代判别的数据库建立方法及艺术品年代判别方法,用于对艺术品进行年代判别的数据库建立方法它包括如下步骤:(1)选取同年代、同类型、同款式的至少两个艺术品标本;(2)提取总视场图像;(3)建立数据库;艺术品年代判别方法它包括如下步骤:Ⅰ.使用上述所建立的数据库;Ⅱ.提取待确定艺术品上的图像;Ⅲ.将提取的待确定艺术品上的图像与数据库中保存的图像使用图像识别技术进行分析;Ⅳ.判断。CN109191145A discloses a method for establishing a database for judging the age of artworks and a method for judging the age of artworks. The method for establishing a database for judging the age of artworks includes the following steps: (1) selecting the same age, same type, At least two artwork specimens of the same style; (2) extracting the total field of view image; (3) establishing a database; a method for identifying the age of artworks which includes the following steps: Ⅰ. using the database established above; Ⅱ. extracting the art to be determined Ⅲ. Analyze the extracted image on the artwork to be determined and the image saved in the database using image recognition technology; Ⅳ. Judgment.
CN111339974A公开了一种现代陶瓷与古陶瓷的鉴别方法,通过构建古陶瓷对应的正样本和仿古瓷对应的负样本,将RGB图像转换至HSV颜色空间,得到HSV图像,获取HSV图像的特征描述子,将特征描述子输入 支持向量机中进行训练,获得支持向量机的训练参数,将RGB图像输入深度卷积神经网络架构进行训练得到卷积神经网络的网络参数,根据支持向量机的训练参数和卷积神经网络的网络参数确定深度学习模型,将正样本的灰度图和负样本的灰度图分别输入深度学习模型进行训练,得到鉴别模型,获取待鉴别瓷器的待鉴别图片,将待鉴别图片输入所述鉴别模型,依据鉴别模型的输出结果确定待鉴别瓷器为现代陶瓷或者古陶瓷,以提高相应陶瓷鉴别的效率。CN111339974A discloses a method for identifying modern ceramics and ancient ceramics, by constructing positive samples corresponding to ancient ceramics and negative samples corresponding to antique porcelain, converting RGB images to HSV color space to obtain HSV images, and obtaining feature descriptors of HSV images , input the feature descriptor into the support vector machine for training, obtain the training parameters of the support vector machine, input the RGB image into the deep convolutional neural network architecture for training to obtain the network parameters of the convolutional neural network, according to the training parameters of the support vector machine and The network parameters of the convolutional neural network determine the deep learning model, input the grayscale image of the positive sample and the grayscale image of the negative sample into the deep learning model for training, obtain the identification model, obtain the image of the porcelain to be identified, and convert the image to be identified The picture is input into the identification model, and according to the output result of the identification model, it is determined whether the porcelain to be identified is modern ceramics or ancient ceramics, so as to improve the efficiency of corresponding ceramic identification.
可以看出,鉴定古近现代艺术品的技术手段经历了从数据库到模型训练的发展。然而为了提高鉴定的准确性,并降低鉴定方法的复杂度,需要更简单更精确的模型训练方法。It can be seen that the technical means of identifying ancient and modern artworks has experienced the development from database to model training. However, in order to improve the accuracy of identification and reduce the complexity of identification methods, simpler and more accurate model training methods are needed.
发明内容Contents of the invention
本发明的目的在于提供一种运用人工智能对古代及近现代艺术品进行鉴定的方法和系统。The object of the present invention is to provide a method and system for appraising ancient and modern artworks by using artificial intelligence.
本发明的一个方面提供一种古代及近现代艺术品鉴定方法,包括:输入真迹图像信息;和输入拟被鉴定的艺术品的图像信息;将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本;将所述分布中样本进行分类;将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类;输出分类后的分布中样本或细粒度分类后的样本,获得所述拟被鉴定的艺术品的图像信息与所述真迹图像信息相比的置信度作为鉴定结论。One aspect of the present invention provides a method for identifying ancient and modern works of art, including: inputting authentic image information; and inputting image information of artworks to be identified; combining the image information of artworks to be identified with the Authentic image information detects the samples in the distribution and the samples out of the distribution through the detector; classifies the samples in the distribution; classifies the samples in the distribution after classification and class image information similar to the artwork to be identified Granular classification: output classified samples in the distribution or samples after fine-grained classification, and obtain the confidence degree of the image information of the artwork to be identified compared with the authentic image information as the identification conclusion.
本发明的另一个方面的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:利用预训练模型(pre-trained model)输出的最大归一化指数概率进行统计分析;统计发现OOD样本和ID样本归一化指数概率的分布情况;将二者的分布差距加大;选取合适的阈值来判断一个样本属于分布外样本还是分布中样本。Another aspect of the identification method of the present invention, wherein the step of detecting the in-distribution samples and out-of-distribution samples through the detector to detect the image information of the artwork to be identified and the image information of the authentic works further includes: using a pre-trained model ( The maximum normalized index probability output by the pre-trained model) is used for statistical analysis; the distribution of the normalized index probability of the OOD sample and the ID sample is found statistically; the distribution gap between the two is increased; an appropriate threshold is selected to judge a sample Whether it is an out-of-distribution sample or a sample in-distribution.
本发明的再一个方面的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:利用模型学习一个对输入样本的不确定性属性;判断测试数据,如果模型输入为分布中样本,则不确定性低,相反,如果模型输入为分布外样本,则不确定性高。In another aspect of the identification method of the present invention, the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through a detector further includes: using a model to learn a pair The uncertainty attribute of the input sample; to judge the test data, if the model input is a sample in the distribution, the uncertainty is low; on the contrary, if the model input is an out-of-distribution sample, the uncertainty is high.
本发明的再一个方面的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步 骤还包括:利用变分自动编码器(Variational Autoencoder)的重构误差(reconstruction error)或者其他度量方式来判断一个样本是否属于分布中或分布外样本;所述编码器的隐含空间(latent space)能够学习出分布中数据的明显特征(silent vector),而对于分布外样本则不行,因此分布外样本会产生较高的重构误差。Another aspect of the identification method of the present invention, wherein the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through a detector further includes: using variational automatic coding Variational Autoencoder (Variational Autoencoder) reconstruction error (reconstruction error) or other measurement methods to determine whether a sample belongs to the distribution or out-of-distribution samples; the latent space of the encoder can learn the obviousness of the data in the distribution feature (silent vector), but not for out-of-distribution samples, so out-of-distribution samples will produce higher reconstruction errors.
本发明的再一个方面的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:使用分类器对提取的特征进行分类来判断是否为分布外样本;有的修改网络结构为一个n+1类分类器,n为原本分类任务的类别数,第n+1类则是分布外类;有的直接取提取特征来进行分类,不需要修改网络的结构。According to another aspect of the identification method of the present invention, the step of detecting the samples in the distribution and the samples out of the distribution by using the image information of the artwork to be identified and the image information of the authentic works through the detector further includes: using a classifier to extract Classify the features of the distribution to determine whether it is an out-of-distribution sample; some modify the network structure to an n+1 classifier, n is the number of categories of the original classification task, and the n+1th class is an out-of-distribution class; some directly take Extract features for classification without modifying the structure of the network.
本发明的再一个方面的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:寻找拟被测类图像数据的特征区域;将所述特征区域输入卷积神经网络;经过所述卷积神经网络的所述特征区域的一部分信息进入全连接层和归一化指数逻辑回归层进行分类;经过所述卷积神经网络的所述特征区域的另一部分信息经过注意力建议子网络(APN),得到候选区域;重复上述分类步骤和APN步骤,使得通过APN选取的特征区域为最具有判别性的区域;引入损失函数,获得更高地识别所述类图像信息的准确性。In another aspect of the identification method of the present invention, the step of fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified further includes: finding the image to be tested The feature area of the data; the feature area is input into the convolutional neural network; a part of the information of the feature area of the convolutional neural network enters the fully connected layer and the normalized exponential logistic regression layer for classification; through the volume Another part of the information of the feature region of the product neural network passes through the attention suggestion sub-network (APN) to obtain the candidate region; repeat the above-mentioned classification steps and APN steps, so that the feature region selected by the APN is the most discriminative region; introduce A loss function to obtain higher accuracy in identifying the image information of the class.
本发明的再一个方面的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:将被鉴定的类图像信息(311)中选取的局部图像(312)和(313)输入两个卷积神经网络(314,A)和(315,B);卷积神经网络流(A)和(B)的输出在图像的每个位置使用外积相乘(318)并合并以获得双线性向量(316);通过分类层(317)获得预测结果。In another aspect of the identification method of the present invention, the step of fine-grained classifying the classified samples in the distribution and the image-like information similar to the artwork to be identified further includes: The local images (312) and (313) selected in the information (311) are input into two convolutional neural networks (314, A) and (315, B); the output of the convolutional neural network streams (A) and (B) is in Each location of the image is multiplied (318) using the outer product and combined to obtain a bilinear vector (316); the prediction is obtained through the classification layer (317).
本发明的再一个方面的鉴定方法,其中所述分类层(317)为逻辑回归或者支持向量机分类器。The identification method according to another aspect of the present invention, wherein the classification layer (317) is a logistic regression or support vector machine classifier.
本发明的再一个方面的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:将被鉴定的类图像信息在不同尺度特征图上生成多个候选框,每个候选框的坐标与预先设计好的锚相对应;给每个候选区域的“信息量”打分,信息量大的区域分数高;对所述特征图依次实施特征提取步骤,全连接层(FC)以及归一化指数步骤;判断输入区域属于目标标签的概率;将各个局部区域和全图提取出来的未归一化的概率合并到一起生成一个长向量,输出对应200个类别的未归一化的概率。In another aspect of the identification method of the present invention, the step of fine-grained classifying the classified samples in the distribution and the image-like information similar to the artwork to be identified further includes: The information generates multiple candidate boxes on different scale feature maps, and the coordinates of each candidate box correspond to the pre-designed anchors; the "information content" of each candidate area is scored, and the area with a large amount of information has a high score; The above feature map is followed by the feature extraction step, the fully connected layer (FC) and the normalized index step; the probability that the input area belongs to the target label is judged; the unnormalized probability extracted from each local area and the whole map is merged together Generates a long vector outputting the unnormalized probabilities for the 200 classes.
本发明还提供一种古代及近现代艺术品鉴定系统,包括:输入模块,输入真迹图像信息;和输入拟被鉴定的艺术品的图像信息;检测模块,将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本;样本分类模块,将所述分布中样本进行分类;细粒度分类模块,将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类;输出模块,输出分类后的分布中样本或细粒度分类后的样本,获得所述拟被鉴定的艺术品的图像信息与所述真迹图像信息相比的置信度作为鉴定结论。The present invention also provides a system for identifying ancient and modern works of art, including: an input module for inputting authentic image information; and inputting image information for artworks to be identified; The information and the authentic image information are detected by the detector to detect the samples in the distribution and the samples out of the distribution; the sample classification module classifies the samples in the distribution; the fine-grained classification module combines the classified samples in the distribution with the simulated Fine-grained classification of similar image information of the identified works of art; the output module outputs the samples in the distribution after classification or the samples after fine-grained classification, and obtains the image information of the artwork to be identified and the authentic The confidence level compared with the image information is used as the identification conclusion.
采用上述的运用人工智能对古代及近现代艺术品进行鉴定的方法和系统,因为采取了对数据进行多层分类的方法,鉴定的精确度得到了提高,并且降低了模型训练过程中的运算量。Using the above-mentioned method and system for using artificial intelligence to identify ancient and modern artworks, because of the multi-layer classification method for data, the accuracy of identification has been improved, and the amount of computation in the model training process has been reduced. .
附图说明Description of drawings
为了更清楚地说明本发明实施例中的技术方案,下面将对实施例中所需要使用的附图作简单地介绍。显而易见地,下面描述中的附图仅仅是本发明的一些实例,对于本领域普通技术人员来讲,在不付出创新性劳动的前提下,还可以根据这些附图获得其他的附图。In order to illustrate the technical solutions in the embodiments of the present invention more clearly, the following will briefly introduce the drawings that are used in the embodiments. Apparently, the drawings in the following description are only some examples of the present invention, and those skilled in the art can also obtain other drawings according to these drawings without any innovative work.
图1为本发明的古代及近现代艺术品鉴定方法的流程图。Fig. 1 is the flow chart of the ancient and modern works of art appraisal method of the present invention.
图2(a)-(d)为本发明的古代及近现代艺术品鉴定方法步骤中检测分布中(in distribution)和分布外(out of distribution,OOD)样本步骤的流程图。Figure 2(a)-(d) is a flow chart of the steps of detecting samples in distribution (in distribution) and out of distribution (out of distribution (OOD)) in the steps of the ancient and modern art identification method of the present invention.
图2(a)为基于归一化指数的实施方式的流程图。Figure 2(a) is a flowchart of a normalized index based embodiment.
图2(b)为不确定性的实施方式的流程图。Fig. 2(b) is a flowchart of an embodiment of uncertainty.
图2(c)为概率生成模型的实施方式的流程图。Figure 2(c) is a flowchart of an embodiment of a probabilistic generative model.
图2(d)为分类模型的实施方式的流程图。Figure 2(d) is a flowchart of an embodiment of a classification model.
图3(a)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用注意力卷积神经网络实施方式的流程图。Fig. 3 (a) is a flow chart of the implementation of the attention convolutional neural network in the fine-grained classification step in the steps of the ancient and modern art identification method of the present invention.
图3(b)为本发明的鉴定方法中细粒度分类步骤一个实施方式循环注意力卷积神经网络(“RA-CNN”)的框架示意图。Fig. 3(b) is a schematic diagram of the framework of a recurrent attention convolutional neural network ("RA-CNN") for an implementation of the fine-grained classification step in the identification method of the present invention.
图3(c)为本发明的鉴定方法中细粒度分类步骤另一个实施方式的双线性向量网络结构示意图。Fig. 3(c) is a schematic diagram of the bilinear vector network structure of another embodiment of the fine-grained classification step in the identification method of the present invention.
图3(d)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用双线性向量网络实施方式的流程图。Fig. 3(d) is a flow chart of the implementation mode of bilinear vector network in the step of fine-grained classification in the steps of the ancient and modern art identification method of the present invention.
图3(e)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用导航-教学-审查网络(NTS-Net)分类的实施方式的流程图。Fig. 3(e) is a flow chart of an embodiment in which the fine-grained classification step adopts the navigation-teaching-examination network (NTS-Net) classification in the steps of the ancient and modern artwork appraisal method of the present invention.
图4为本发明的古代及近现代艺术品鉴定系统的结构图。Fig. 4 is a structural diagram of the ancient and modern art identification system of the present invention.
图5为本发明的古代及近现代艺术品鉴定系统的便携式或者固定存储单元的计算机产品图。Fig. 5 is a computer product diagram of the portable or fixed storage unit of the ancient and modern art identification system of the present invention.
图6(1)为本发明的古代及近现代艺术品鉴定方法的一个实施方式中涉及的真迹图像信息示例。Fig. 6(1) is an example of authentic image information involved in an embodiment of the ancient and modern artwork identification method of the present invention.
图6(2)为本发明的古代及近现代艺术品鉴定方法的一个实施方式中涉及的用来训练模型的真迹图像信息示例。Fig. 6(2) is an example of authentic image information used to train the model involved in an embodiment of the ancient and modern artwork identification method of the present invention.
图6(3)为本发明的古代及近现代艺术品鉴定方法的一个实施方式中涉及的拟被鉴定的艺术品的图像信息示例。Fig. 6(3) is an example of the image information of the artwork to be authenticated involved in one embodiment of the ancient and modern artwork authentication method of the present invention.
图6(4)为本发明的古代及近现代艺术品鉴定方法的一个实施方式中涉及的分类示例。Fig. 6 (4) is an example of the classification involved in an implementation of the ancient and modern artwork identification method of the present invention.
具体实施方式Detailed ways
现结合相应的附图,对本发明的具体实施例进行描述。然而,本发明可以以多种不同的形式实施,而不应被解释为局限于此处展示的实施例。提供这些实施例只是为了本发明可以详尽和全面,从而可以将本发明的范围完全地描述给本领域的技术人员。附图中说明的实施例的详细描述中使用的措辞不应对本发明造成限制。Specific embodiments of the present invention will now be described in conjunction with the corresponding drawings. However, this invention may be embodied in many different forms and should not be construed as limited to the embodiments set forth herein. These embodiments are provided only so that the present invention will be thorough and complete so that those skilled in the art can fully describe the scope of the present invention. Wording used in the detailed description of the embodiments illustrated in the drawings should not limit the invention.
图1为本发明的古代及近现代艺术品鉴定方法的流程图。在步骤101,输入拟被鉴定的艺术品的图像信息和真迹图像信息;在步骤102,将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中(in distribution)样本和分布外(out of distribution,OOD)样本;在步骤103,将所述分布中样本进行分类;在步骤104,将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类(Fine-grained classifier);在步骤105,输出分类后的分布中样本或细粒度分类后的样本,获得鉴定结论。Fig. 1 is the flow chart of the ancient and modern works of art appraisal method of the present invention. In step 101, the image information and authentic image information of the artwork to be identified are input; in step 102, the image information of the artwork to be identified and the authentic image information are detected by a detector in distribution ) samples and distribution (out of distribution, OOD) samples; in step 103, the samples in the distribution are classified; in step 104, the samples in the distribution after classification are similar to the artwork to be identified Fine-grained classifier is performed on the class image information; in step 105, the classified samples in the distribution or samples after fine-grained classification are output to obtain the identification conclusion.
图2(a)-(d)为本发明的古代及近现代艺术品鉴定方法步骤中检测分布中(in distribution)和分布外(out of distribution,OOD)样本步骤的流程图。其中图2(a)为基于归一化指数的实施方式的流程图;图2(b)为不确定性的实施方式的流程图;图2(c)为概率生成模型的实施方式的流程图;图2(d)为分类模型的实施方式的流程图。Figure 2(a)-(d) is a flow chart of the steps of detecting samples in distribution (in distribution) and out of distribution (out of distribution (OOD)) in the steps of the ancient and modern art identification method of the present invention. Wherein Fig. 2 (a) is the flow chart of the embodiment based on normalized index; Fig. 2 (b) is the flow chart of the embodiment of uncertainty; Fig. 2 (c) is the flow chart of the embodiment of probability generation model ; Figure 2(d) is a flowchart of an embodiment of a classification model.
在模型训练和测试的类图像数据是独立同分布的情况下(IID,Independent Identical Distribution),将训练和测试的类图像数据称为分布中(In Distribution(ID))样本。除了ID样本外,在实际应用当中,模型部署上线后得到的数据往往不能被完全控制的,也就是说模型接收的数据有可能是分布外(OOD)样本,也被称为异常样本(outlier,abnormal)。深度模型会对一个分布外(OOD)样本认为是分布中(ID)样本中的某一个类,并给出高的置信度,这里所述的置信度为0-1的归一化数值,找出分布外样本,但它们的设置可能不同。比如分布外检测(OOD detection)是在模型任务上修改,要求不仅能够有效地检测出分布外(OOD)样本,而且要保证模型的性能不受影响。When the image-like data for model training and testing are independent and identically distributed (IID, Independent Identical Distribution), the image-like data for training and testing are called In Distribution (ID) samples. In addition to ID samples, in practical applications, the data obtained after the model is deployed and launched is often not fully controlled, that is to say, the data received by the model may be out-of-distribution (OOD) samples, also known as outlier samples (outlier, abnormal). The depth model will consider an out-of-distribution (OOD) sample as a certain class in the distribution (ID) sample, and give a high degree of confidence. The confidence degree described here is a normalized value of 0-1. Find out-of-distribution samples, but their settings may be different. For example, out-of-distribution detection (OOD detection) is modified on the model task, which requires not only to be able to effectively detect out-of-distribution (OOD) samples, but also to ensure that the performance of the model is not affected.
本发明针对古代及现代艺术品的类图像数据的检测分布中和分布外步骤可采用基于归一化指数(Softmax-based)、不确定性(Uncertainty)、概率生成模型(Generative model)、分类模型(Classifier)方法来检测分布中和分布外样本方法。In the present invention, the detection distribution and out-of-distribution steps for the image-like data of ancient and modern artworks can be based on normalization index (Softmax-based), uncertainty (Uncertainty), probability generation model (Generative model), classification model (Classifier) method to detect in-distribution and out-of-distribution sample methods.
其中,在利用基于归一化指数方法的实施方式中,在步骤201,利用预训练模型(pre-trained model)输出的最大归一化指数概率进行统计分析,在步骤202,统计发现OOD样本和ID样本归一化指数概率的分布情况,在步骤203,将二者的分布差距加大,在步骤204,选取合适的阈值来判断一个样本属于分布外样本还是分布中样本。这类方法简单且有效,不用修改分类模型的结构,也不需要训练一个分布外样本分类器。Wherein, in the embodiment based on the normalized index method, in step 201, statistical analysis is performed using the maximum normalized index probability output by the pre-trained model (pre-trained model), and in step 202, statistically found OOD samples and For the distribution of the normalized index probability of the ID sample, in step 203, the distribution gap between the two is increased, and in step 204, an appropriate threshold is selected to determine whether a sample belongs to an out-of-distribution sample or an in-distribution sample. This type of method is simple and effective, without modifying the structure of the classification model, and without training an out-of-distribution sample classifier.
在利用不确定性方法的实施方式中,由于模型的概率输出并不能直接表示模型的置信度(confidence)。在步骤211,利用模型学习一个对输入样本的不确定性属性。在步骤212,判断测试数据,如果模型输入为分布中样本,则不确定性低,相反,如果模型输入为分布外样本,则不确定性高。这类方法需要修改模型的网络结构来学习不确定性属性。In the embodiment using the uncertainty method, since the probability output of the model cannot directly represent the confidence of the model. In step 211, the model is used to learn an uncertainty attribute for the input samples. In step 212, the test data is judged. If the model input is a sample in the distribution, the uncertainty is low; on the contrary, if the model input is an out-of-distribution sample, the uncertainty is high. Such methods need to modify the network structure of the model to learn the uncertainty property.
在利用概率生成模型方法的实施方式中,在步骤221,利用变分自动编码器(Variational Autoencoder)的重构误差(reconstruction error)或者其他度量方式来判断一个样本是否属于分布中或分布外样本;在步骤222, 所述编码器的隐含空间(latent space)能够学习出分布中数据的明显特征(silent vector),而对于分布外样本则不行,因此分布外样本会产生较高的重构误差。这种方法只关注分布外检测性能,不关注分布中数据本来的任务。In the embodiment of using the probability generation model method, in step 221, use the reconstruction error (reconstruction error) of the variational autoencoder (Variational Autoencoder) or other measurement methods to judge whether a sample belongs to the sample in the distribution or out of the distribution; In step 222, the hidden space (latent space) of the encoder can learn the obvious features (silent vector) of the data in the distribution, but not for the out-of-distribution samples, so the out-of-distribution samples will generate higher reconstruction errors . This method only focuses on out-of-distribution detection performance, and does not focus on the original task of the data in the distribution.
在利用分类模型方法的实施方式中,在步骤231,使用分类器对提取的特征进行分类来判断是否为分布外样本;在步骤232,有的修改网络结构为一个n+1类分类器,n为原本分类任务的类别数,第n+1类则是分布外类;在步骤233,有的直接取提取特征来进行分类,不需要修改网络的结构。In the implementation of the classification model method, in step 231, a classifier is used to classify the extracted features to determine whether it is an out-of-distribution sample; in step 232, some modify the network structure to be an n+1 classifier, n is the number of categories of the original classification task, and the n+1th category is an out-of-distribution category; in step 233, some features are directly extracted for classification without modifying the structure of the network.
图3(a)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用注意力卷积神经网络实施方式的流程图。在步骤321,寻找拟被测类图像数据的特征区域,在步骤322,将所述特征区域输入卷积神经网络;在步骤323,经过所述卷积神经网络的所述特征区域的一部分信息进入全连接层和归一化指数逻辑回归层进行分类;在步骤324,经过所述卷积神经网络的所述特征区域的另一部分信息经过注意力建议子网络(APN),得到候选区域;在步骤325,重复所述步骤323和所述步骤324,使得通过APN选取的特征区域为最具有判别性的区域;在步骤326,引入损失函数,获得更高地识别所述类图像信息的准确性。Fig. 3 (a) is a flow chart of the implementation of the attention convolutional neural network in the fine-grained classification step in the steps of the ancient and modern art identification method of the present invention. In step 321, the characteristic area of the image data to be tested is searched, and in step 322, the characteristic area is input into the convolutional neural network; in step 323, a part of the information of the characteristic area through the convolutional neural network is entered Fully connected layer and normalized exponential logistic regression layer are classified; in step 324, another part of the information of the feature region through the convolutional neural network is passed through the attention suggestion subnetwork (APN), to obtain the candidate region; in step 325, repeating the step 323 and the step 324, so that the feature region selected by the APN is the most discriminative region; in step 326, introducing a loss function to obtain a higher accuracy of identifying the type of image information.
“细粒度”分类步骤是在普通的分类之下,更加精细的划分,就需要显式地找到图片中最具有“判别性(discriminative)”的特征。对古代及近现代艺术品来说,需要找到细节上的特征,例如花瓣的上翘程度,花纹的细微差别等。The "fine-grained" classification step is under the ordinary classification. For a more fine-grained division, it is necessary to explicitly find the most "discriminative" features in the picture. For ancient and modern works of art, it is necessary to find the characteristics of details, such as the degree of upturning of petals, the nuances of patterns, etc.
图3(b)为本发明的鉴定方法中细粒度分类步骤一个实施方式循环注意力卷积神经网络(“RA-CNN”)的框架示意图。其中符号
Figure PCTCN2021114254-appb-000001
代表在被鉴定的类图像信息的特征区域上剪切一部分并将其放大。每一行301,302,303分别代表一个普通的CNN网络。如图3(b)所示,输入从粗糙的全尺寸图像到更精细的区域注意力(从上到下)。其中第一行301的图片(a 1)为最粗糙的,第三行的图片(a 3)更精细。图像信息a 1进入b 1(几个卷积层)之后分成两路,一路走到c 1接进入全连接层(fully connected layers,FC)和softmax逻辑回归层进行简单的分类,另外一路进入d 1即注意力建议子网络(“Attention Proposal Network”,APN),得到一个候选区域。在原图上利用d 1提出的候选区域,在原图上剪切(crop)出一个更有判别性的小区域,插值之后得到a 2,同样的道理得到a 3
Fig. 3(b) is a schematic diagram of the framework of a recurrent attention convolutional neural network ("RA-CNN") for an implementation of the fine-grained classification step in the identification method of the present invention. where the symbol
Figure PCTCN2021114254-appb-000001
It means to cut a part of the characteristic area of the identified image-like information and enlarge it. Each row 301, 302, 303 represents a common CNN network respectively. As shown in Figure 3(b), the input ranges from coarse full-scale images to finer region attention (from top to bottom). The picture (a 1 ) in the first row 301 is the roughest, and the picture (a 3 ) in the third row is finer. After the image information a 1 enters b 1 (several convolutional layers), it is divided into two paths, all the way to c 1 and connected to fully connected layers (fully connected layers, FC) and softmax logistic regression layer for simple classification, and the other path enters d 1 is the attention proposal sub-network ("Attention Proposal Network", APN), get a candidate area. Using the candidate area proposed by d 1 on the original image, a more discriminative small area is cropped on the original image, and a 2 is obtained after interpolation, and a 3 is obtained in the same way.
特征区域经过两次APN之后不断放大和细化,为了使得APN选取的特征区域是图像中最具有判别性的区域,引入损失函数(Ranking loss):即强迫区域a 1、a 2、a 3的分类信心程度(confidence score)越来越高(即图片最后一列的对应P t概率越来越大),也就是说识别类图像信息的准确性越来越高。这样一来,结合普通的分类损失,使网络不断细化判别注意力区域(discriminative attention region)。 The feature area is continuously enlarged and refined after two APNs. In order to make the feature area selected by APN the most discriminative area in the image, a loss function (Ranking loss) is introduced: that is, the forced area a 1 , a 2 , a 3 The classification confidence level (confidence score) is getting higher and higher (that is, the corresponding P t probability of the last column of the picture is getting higher and higher), which means that the accuracy of identifying image information is getting higher and higher. In this way, combined with ordinary classification loss, the network continuously refines the discriminative attention region.
图3(c)为本发明的鉴定方法中细粒度分类步骤另一个实施方式的双线性向量网络结构示意图。Fig. 3(c) is a schematic diagram of the bilinear vector network structure of another embodiment of the fine-grained classification step in the identification method of the present invention.
将被鉴定的类图像信息311中选取的局部图像312和313输入两个卷积神经网络314(A)和315(B)。卷积神经网络流A和B的输出在图像的每个位置使用外积相乘318并合并以获得双线性向量316,再通过分类层317获得预测结果。双线性模型M由一个四元组组成:M=(f A,f B;P;C)。其中,f A,f B代表特征提取函数,即图3(c)中的卷积网络A和卷积网络B,P是一个池化函数(Pooling function),C则是分类函数。 Partial images 312 and 313 selected from the identified class image information 311 are input into two convolutional neural networks 314(A) and 315(B). The outputs of the convolutional neural network streams A and B are multiplied 318 by outer product at each position of the image and combined to obtain a bilinear vector 316 , which is then passed through a classification layer 317 to obtain a prediction result. The bilinear model M consists of a quadruple: M = (f A , f B ; P; C). Among them, f A and f B represent feature extraction functions, that is, convolutional network A and convolutional network B in Figure 3(c), P is a pooling function (Pooling function), and C is a classification function.
特征提取函数f(·)(即卷积神经网络流CNN stream)由卷积层,池化层和激活函数组成。这一部分网络结构可以看作函数映射:The feature extraction function f( ) (i.e., the convolutional neural network stream CNN stream) consists of convolutional layers, pooling layers, and activation functions. This part of the network structure can be regarded as a function map:
f:L×I→RK×D      (1)f: L×I→RK×D (1)
将输入的鉴定的类图像信息与位置区域映射为一个维的特征,其中K为卷积网络输出特征图的通道数,D为每个通道中的二维特征图展开成的一维特征向量的大小。而两个特征提取函数输出的卷积特征,通过双线性操作进行汇聚,得到双线性特征316:bilinear(l;T;f A,f B)=f A(L;T) Tf B(L;T) T。而池化函数P的作用则是将所有位置的双线性特征汇聚成一个特征。所采用的池化函数是将所有位置的双线性特征累加起来,得到图像的全局特征表示Φ'(I): Map the input identified class image information and location area into a one-dimensional feature, where K is the number of channels of the convolutional network output feature map, and D is the number of one-dimensional feature vectors expanded from the two-dimensional feature map in each channel. size. The convolution features output by the two feature extraction functions are converged through bilinear operations to obtain bilinear features 316: bilinear(l; T; f A , f B ) = f A (L; T) T f B (L;T) T . The function of the pooling function P is to aggregate the bilinear features of all positions into one feature. The pooling function adopted is to accumulate the bilinear features of all positions to obtain the global feature representation of the image Φ'(I):
Φ`(I)=Σ l∈Lbilinear(l;T,f A;f B)=Σ l∈Lf A(l;I) Tf B(l;I)   (2) Φ`(I)=Σ l∈L bilinear(l; T, f A ; f B )=Σ l∈L f A (l; I) T f B (l; I) (2)
在两个特征函数f A,f B提取的特征维度分别是K×M与K×N的情况下,池化函数P的输出为M×N的矩阵,在对其进行分类之前先把特征矩阵拉伸成一列MN大小的特征向量。最后,利用分类函数对提取的特征进行分类,所述分类层317采用逻辑回归或者支持向量机(support vector machine,SVM)分类器实现。 In the case that the feature dimensions extracted by the two feature functions f A and f B are K×M and K×N respectively, the output of the pooling function P is an M×N matrix. Before classifying it, the feature matrix Stretched into a list of MN-sized feature vectors. Finally, a classification function is used to classify the extracted features, and the classification layer 317 is implemented using a logistic regression or a support vector machine (support vector machine, SVM) classifier.
CNN网络能实现对细粒度图像进行高层语义特征获取,通过迭代训练网络模型中的卷积参数,过滤图像中不相关的背景信息。另一方面,卷 积神经网络流A和卷积神经网络流B在图像识别任务中扮演着互补的角色,即网络A能够对图像中的物体进行定位,网络B则是完成对网络A定位到的物体位置进行特征提取。通过这种方式,两个网络能够配合完成对输入细粒度图像的类检测和目标特征去的过程,较好地完成细粒度图像识别任务。The CNN network can achieve high-level semantic feature acquisition for fine-grained images, and filter irrelevant background information in the image by iteratively training the convolution parameters in the network model. On the other hand, the convolutional neural network flow A and the convolutional neural network flow B play complementary roles in the image recognition task, that is, the network A can locate the object in the image, and the network B can complete the positioning of the network A to Feature extraction of the object position. In this way, the two networks can cooperate to complete the class detection and target feature removal process of the input fine-grained image, and better complete the fine-grained image recognition task.
图3(d)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用双线性向量网络实施方式的流程图。在步骤331,将被鉴定的类图像信息311中选取的局部图像312和313输入两个卷积神经网络314(A)和315(B);在步骤332,卷积神经网络流A和B的输出在图像的每个位置使用外积相乘318并合并以获得双线性向量316,在步骤333,通过分类层317获得预测结果。Fig. 3(d) is a flow chart of the implementation mode of bilinear vector network in the step of fine-grained classification in the steps of the ancient and modern art identification method of the present invention. In step 331, the partial images 312 and 313 selected in the identified class image information 311 are input into two convolutional neural networks 314 (A) and 315 (B); in step 332, the convolutional neural network streams A and B The output is multiplied 318 using an outer product at each location in the image and combined to obtain a bilinear vector 316 , and at step 333 the prediction is obtained through the classification layer 317 .
图3(e)为本发明的古代及近现代艺术品鉴定方法步骤中细粒度分类步骤采用导航-教学-审查网络(NTS-Net)分类的实施方式的流程图。在步骤341,将被鉴定的类图像信息在不同尺度特征图(Feature maps)上生成多个候选框,每个候选框的坐标与预先设计好的锚(Anchors)相对应;在步骤342,给每个候选区域的“信息量”打分,信息量大的区域分数高;在步骤343,对所述特征图依次实施特征提取步骤(Feature Extractor),全连接层(FC)以及归一化指数(softmax)步骤;在步骤344,判断输入区域属于目标标签(target label)的概率;在步骤345,将各个局部区域和全图提取出来的未归一化的概率(logits)合并(concat)到一起生成一个长向量,输出对应200个类别的未归一化的概率(logits)。Fig. 3(e) is a flow chart of an embodiment in which the fine-grained classification step adopts the navigation-teaching-examination network (NTS-Net) classification in the steps of the ancient and modern artwork appraisal method of the present invention. In step 341, generate a plurality of candidate boxes on different scale feature maps (Feature maps) with the identified class image information, and the coordinates of each candidate box correspond to pre-designed anchors (Anchors); in step 342, give The "information content" of each candidate area is scored, and the area with a large amount of information has a high score; in step 343, a feature extraction step (Feature Extractor), a fully connected layer (FC) and a normalization index ( softmax) step; in step 344, determine the probability that the input region belongs to the target label (target label); in step 345, merge (concat) together the unnormalized probability (logits) extracted from each local area and the whole picture Generates a long vector outputting unnormalized probabilities (logits) corresponding to 200 categories.
本发明的鉴定方法中细粒度分类步骤还可采用,将网络主体分成导航(Navigator),教学(Teacher),审查(Scrutinizer)三个组件的导航-教学-审查网络(NTS-Net)分类方法,其中导航步骤,在不同尺度特征图(Feature maps)上生成多个候选框,每个候选框的坐标与预先设计好的锚(Anchors)相对应。导航(Navigator)给每一个候选区域的“信息量”打分,信息量大的区域分数高。教学步骤是常用的特征提取步骤(Feature Extractor),全连接层(FC)以及归一化指数(softmax)步骤,判断输入区域属于目标标签(target label)的概率;审查步骤就是一个全连接层,输入是把各个局部区域和全图提取出来的未归一化的概率(logits)合并(concat)到一起生成一个长向量,输出对应200个类别的未归一化的概率(logits)。The fine-grained classification step in the identification method of the present invention can also be adopted, and the navigation-teaching-examination network (NTS-Net) classification method of dividing the network subject into three components of navigation (Navigator), teaching (Teacher), and examination (Scrutinizer), In the navigation step, multiple candidate boxes are generated on feature maps of different scales, and the coordinates of each candidate box correspond to the pre-designed anchors (Anchors). The Navigator scores the "information content" of each candidate area, and the area with a large amount of information has a higher score. The teaching step is the commonly used feature extraction step (Feature Extractor), fully connected layer (FC) and normalized index (softmax) step, to judge the probability that the input area belongs to the target label (target label); the review step is a fully connected layer, The input is to combine (concat) the unnormalized probability (logits) extracted from each local area and the whole image together to generate a long vector, and output the unnormalized probability (logits) corresponding to 200 categories.
采用这种NTS方法的具体步骤为:1)尺寸(448,448,3)的原图进入网络,进过Resnet-50提取特征以后,变成一个(14,14,2048)的特征图,一个经过全局池化层之后2048维的特征向量和一个经过全局池 化层以及全连接层之后的200维的未归一化的概率。2)预设的生成候选区域的网络(RPN)在(14,14)(7,7)(4,4)这三种尺度上根据不同的尺寸(Size),纵横比(aspect ration)生成对应的锚(Anchors)一共1614个。3)用步骤1中的特征图,到导航中打分,用非极大值抑制(Non-Maximum Suppression,NMS)根据打分结果只保留N个信息量最多的局部候选框。4)把那N个局部区域双线性插值到(224,224),输入教学(Teacher)网络,得到这些局部区域的特征向量和未归一化的概率(logits)。5)把步骤1和4中的全图特征向量feature vector和局部特征向量合并(concat)在一起,之后接FC层,得到联合分类logits用于最终决策。The specific steps of using this NTS method are: 1) The original image of size (448, 448, 3) enters the network, and after entering the Resnet-50 to extract features, it becomes a (14, 14, 2048) feature map, a A 2048-dimensional feature vector after the global pooling layer and a 200-dimensional unnormalized probability after the global pooling layer and the fully connected layer. 2) The preset network (RPN) for generating candidate regions generates correspondences according to different sizes (Size) and aspect ratios on the three scales of (14, 14) (7, 7) (4, 4) There are 1614 anchors in total. 3) Use the feature map in step 1 to score in the navigation, and use Non-Maximum Suppression (NMS) to retain only N local candidate boxes with the most information according to the scoring results. 4) Bilinearly interpolate the N local regions to (224, 224), input them into the teacher network, and obtain the feature vectors and unnormalized probabilities (logits) of these local regions. 5) Merge (concat) the full-image feature vector feature vector and local feature vector in steps 1 and 4, and then connect to the FC layer to obtain the joint classification logits for final decision-making.
图4为本发明的古代及近现代艺术品鉴定系统的结构图。例如古代及近现代艺术品鉴定系统的服务器401。该古代及近现代艺术品鉴定系统的服务器包括处理器410,此处的处理器可以为通用或专用芯片(ASIC/eASIC)或FPGA或NPU等,和以存储器420形式的计算机程序产品或者计算机可读介质。存储器420可以是诸如闪存、EEPROM(电可擦除可编程只读存储器)、EPROM、硬盘或者ROM之类的电子存储器。存储器420具有用于执行上述方法中的任何方法步骤的程序代码的存储空间430。例如,用于程序代码的存储空间430可以包括分别用于实现上面的方法中的各种步骤的各个程序代码431。这些程序代码可以被读出或者写入到所述处理器410中。这些计算机程序产品包括诸如硬盘,紧致盘(CD)、存储卡或者软盘之类的程序代码载体。这样的计算机程序产品通常为如参考图5所述的便携式或者固定存储单元。图5为本发明的古代及近现代艺术品鉴定系统的便携式或者固定存储单元的计算机产品图。该存储单元可以具有与图4的服务器中的存储器420类似布置的存储段、存储空间等。程序代码可以例如以适当形式进行压缩。通常,存储单元包括计算机可读代码431’,即可以由例如诸如410之类的处理器读取的代码,这些代码当由服务器运行时,导致该服务器执行上面所描述的方法中的各个步骤。这些代码当由服务器运行时,导致该服务器执行上面所描述的方法中的各个步骤。Fig. 4 is a structural diagram of the ancient and modern art identification system of the present invention. For example, the server 401 of the ancient and modern artwork appraisal system. The server of this ancient and modern art identification system includes a processor 410, where the processor can be a general-purpose or special-purpose chip (ASIC/eASIC) or FPGA or NPU, etc., and a computer program product in the form of a memory 420 or a computer-programmable Read media. Memory 420 may be electronic memory such as flash memory, EEPROM (Electrically Erasable Programmable Read Only Memory), EPROM, hard disk, or ROM. The memory 420 has a storage space 430 for program codes for performing any method steps in the methods described above. For example, the storage space 430 for program codes may include respective program codes 431 for respectively implementing various steps in the above methods. These program codes can be read or written into the processor 410 . These computer program products comprise program code carriers such as hard disks, compact disks (CDs), memory cards or floppy disks. Such a computer program product is typically a portable or fixed storage unit as described with reference to FIG. 5 . Fig. 5 is a computer product diagram of the portable or fixed storage unit of the ancient and modern art identification system of the present invention. The storage unit may have storage segments, storage spaces, etc. arranged similarly to the memory 420 in the server of FIG. 4 . The program code can eg be compressed in a suitable form. Typically, the storage unit includes computer readable code 431', i.e. code readable by, for example, a processor such as 410, which code, when executed by the server, causes the server to perform the various steps in the methods described above. These codes, when executed by the server, cause the server to perform the steps of the methods described above.
图6(1)为本发明的古代及近现代艺术品鉴定方法的一个实施方式中涉及的真迹图像信息示例。图6(2)为用来训练模型的真迹图像信息示例。图6(3)为拟被鉴定的艺术品的图像信息示例。图6(4)为分类示例。其中,图6(1)展示了一个艺术品真迹的某一图像信息示例,以图6(2)中给出的艺术品真迹从360度获得的多张图像信息为标准,将图6(3)中的拟被鉴定的艺术品图像信息,即包含不同特征区域的类图像信息,进行艺术品模型识别,获得0至1的置信度,评估待识别艺术品与真 迹之间的相似度,越接近1,越相似,当结果为1时,二者一致。图6(4)给出了基于不同拟被鉴定的艺术品图像信息,得出的不同置信度数值的分类示例,图6(4)给出示例均为接近二者一致的数值1。Fig. 6(1) is an example of authentic image information involved in an embodiment of the ancient and modern artwork identification method of the present invention. Figure 6(2) is an example of authentic image information used to train the model. Figure 6(3) is an example of the image information of the artwork to be identified. Figure 6(4) is an example of classification. Among them, Fig. 6(1) shows an example of a certain image information of an authentic artwork. Taking the multiple image information obtained from 360 degrees of the authentic artwork given in Fig. 6 (2) as a standard, Fig. 6 (3 ) in the artwork image information to be identified, that is, image-like information containing different characteristic regions, to identify the artwork model, obtain a confidence level of 0 to 1, and evaluate the similarity between the artwork to be identified and the authentic work, the more The closer to 1, the more similar, when the result is 1, the two are consistent. Figure 6(4) shows the classification examples of different confidence values based on different artwork image information to be identified.
本文中所称的“一个实施例”、“实施例”或者“一个或者多个实施例”意味着,结合实施例描述的特定特征、结构或者特性包括在本发明的至少一个实施例中。此外,请注意,这里“在一个实施例中”的词语例子不一定全指同一个实施例。Reference herein to "one embodiment," "an embodiment," or "one or more embodiments" means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Additionally, please note that examples of the word "in one embodiment" herein do not necessarily all refer to the same embodiment.
以上所述仅用于说明本发明的技术方案,任何本领域普通技术人员均可在不违背本发明的精神及范畴下,对上述实施例进行修饰与改变。因此,本发明的权利保护范围应视权利要求范围为准。本发明已结合例子在上面进行了阐述。然而,在本发明公开范围以内的上述实施例以外的其它实施例也同样可行。本发明的不同的特点和步骤可以以不同于所描述的其它方法进行组合。本发明的范围仅受限于所附的权利要求书。更一般地,本领域普通技术人员可以轻易地理解此处描述的所有的参数,尺寸,材料和配置是为示范目的而实际的参数,尺寸,材料和/或配置将取决于特定应用或本发明教导所用于的应用。The above description is only used to illustrate the technical solutions of the present invention, and any person skilled in the art can modify and change the above embodiments without departing from the spirit and scope of the present invention. Therefore, the protection scope of the present invention should be determined by the scope of the claims. The present invention has been described above with reference to examples. However, other embodiments than those described above are equally possible within the disclosed scope of the present invention. The different features and steps of the invention may be combined in other ways than described. The scope of the present invention is limited only by the appended claims. More generally, one of ordinary skill in the art can readily understand that all parameters, dimensions, materials and configurations described herein are for exemplary purposes and actual parameters, dimensions, materials and/or configurations will depend on the particular application or invention Teach the application for which it is used.

Claims (18)

  1. 一种古代及近现代艺术品鉴定方法,包括:A method for identifying ancient and modern works of art, comprising:
    输入真迹图像信息;和输入拟被鉴定的艺术品的图像信息;Input the image information of the authentic work; and input the image information of the artwork to be identified;
    将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本;Detecting the in-distribution sample and the out-of-distribution sample through the detector to detect the image information of the artwork to be identified and the image information of the authentic work;
    将所述分布中样本进行分类;Classify the samples in the distribution;
    将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类;Carrying out fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified;
    输出分类后的分布中样本或细粒度分类后的样本,获得所述拟被鉴定的艺术品的图像信息与所述真迹图像信息相比的置信度作为鉴定结论。Output the samples in the classified distribution or the samples after fine-grained classification, and obtain the confidence degree of the image information of the artwork to be identified compared with the authentic image information as the identification conclusion.
  2. 如权利要求1所述的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:The identification method according to claim 1, wherein the step of detecting the in-distribution sample and the out-of-distribution sample through the detector to detect the image information of the artwork to be identified and the authentic image information further comprises:
    利用预训练模型(pre-trained model)输出的最大归一化指数概率进行统计分析;Statistical analysis using the maximum normalized exponential probability output by the pre-trained model;
    统计发现OOD样本和ID样本归一化指数概率的分布情况;Statistically discover the distribution of normalized index probability of OOD samples and ID samples;
    将二者的分布差距加大;Increase the distribution gap between the two;
    选取合适的阈值来判断一个样本属于分布外样本还是分布中样本。Select an appropriate threshold to judge whether a sample belongs to an out-of-distribution sample or an in-distribution sample.
  3. 如权利要求1所述的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:The identification method according to claim 1, wherein the step of detecting the in-distribution sample and the out-of-distribution sample through the detector to detect the image information of the artwork to be identified and the authentic image information further comprises:
    利用模型学习一个对输入样本的不确定性属性;Use the model to learn an uncertainty attribute for the input sample;
    判断测试数据,如果模型输入为分布中样本,则不确定性低,相反,如果模型输入为分布外样本,则不确定性高。Judging the test data, if the model input is a sample in the distribution, the uncertainty is low. On the contrary, if the model input is an out-of-distribution sample, the uncertainty is high.
  4. 如权利要求1所述的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:The identification method according to claim 1, wherein the step of detecting the in-distribution sample and the out-of-distribution sample through the detector to detect the image information of the artwork to be identified and the authentic image information further comprises:
    利用变分自动编码器(Variational Autoencoder)的重构误差(reconstruction error)或者其他度量方式来判断一个样本是否属于分布中或分布外样本;Use the reconstruction error of the Variational Autoencoder (Variational Autoencoder) or other metrics to determine whether a sample belongs to a sample in the distribution or out of the distribution;
    所述编码器的隐含空间(latent space)能够学习出分布中数据的明显特征(silent vector),而对于分布外样本则不行,因此分布外样本会产生较高的重构误差。The latent space (latent space) of the encoder can learn the obvious features (silent vector) of the data in the distribution, but not for the out-of-distribution samples, so the out-of-distribution samples will generate higher reconstruction errors.
  5. 如权利要求1所述的鉴定方法,其中将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本的步骤还包括:The identification method according to claim 1, wherein the step of detecting the in-distribution sample and the out-of-distribution sample through the detector to detect the image information of the artwork to be identified and the authentic image information further comprises:
    使用分类器对提取的特征进行分类来判断是否为分布外样本;Use a classifier to classify the extracted features to determine whether it is an out-of-distribution sample;
    有的修改网络结构为一个n+1类分类器,n为原本分类任务的类别数,第n+1类则是分布外类;Some modify the network structure into an n+1 classifier, n is the number of categories of the original classification task, and the n+1th class is an out-of-distribution class;
    有的直接取提取特征来进行分类,不需要修改网络的结构。Some directly take the extracted features for classification without modifying the structure of the network.
  6. 如权利要求1-5任意一个所述的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:The identification method according to any one of claims 1-5, wherein the step of fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified further includes:
    寻找拟被测类图像数据的特征区域;Find the characteristic area of the image data to be tested;
    将所述特征区域输入卷积神经网络;inputting the feature region into a convolutional neural network;
    经过所述卷积神经网络的所述特征区域的一部分信息进入全连接层和归一化指数逻辑回归层进行分类;Part of the information of the feature region through the convolutional neural network enters the fully connected layer and the normalized exponential logistic regression layer for classification;
    经过所述卷积神经网络的所述特征区域的另一部分信息经过注意力建议子网络(APN),得到候选区域;Another part of the information of the feature region through the convolutional neural network passes through the attention suggestion sub-network (APN) to obtain a candidate region;
    重复上述分类步骤和APN步骤,使得通过APN选取的特征区域为最具有判别性的区域;Repeat the above classification steps and APN steps, so that the feature region selected by APN is the most discriminative region;
    引入损失函数,获得更高地识别所述类图像信息的准确性。A loss function is introduced to obtain higher accuracy in identifying the image information.
  7. 如权利要求1-5任意一个所述的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:The identification method according to any one of claims 1-5, wherein the step of fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified further includes:
    将被鉴定的类图像信息(311)中选取的局部图像(312)和(313)输入两个卷积神经网络(314,A)和(315,B);Input two convolutional neural networks (314, A) and (315, B) into partial images (312) and (313) selected in the identified class image information (311);
    卷积神经网络流(A)和(B)的输出在图像的每个位置使用外积相乘(318)并合并以获得双线性向量(316);The outputs of the convolutional neural network streams (A) and (B) are multiplied (318) using an outer product at each location in the image and combined to obtain a bilinear vector (316);
    通过分类层(317)获得预测结果。Predictions are obtained through the classification layer (317).
  8. 如权利要求7所述的鉴定方法,其中所述分类层(317)为逻辑回归或者支持向量机分类器。The authentication method according to claim 7, wherein the classification layer (317) is a logistic regression or a support vector machine classifier.
  9. 如权利要求1-5任意一个所述的鉴定方法,其中将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类的步骤,还包括:The identification method according to any one of claims 1-5, wherein the step of fine-grained classification of the classified samples in the distribution and image-like information similar to the artwork to be identified further includes:
    将被鉴定的类图像信息在不同尺度特征图上生成多个候选框,每个候选框的坐标与预先设计好的锚相对应;Generate multiple candidate boxes on feature maps of different scales from the identified class image information, and the coordinates of each candidate box correspond to the pre-designed anchors;
    给每个候选区域的“信息量”打分,信息量大的区域分数高;Score the "information content" of each candidate area, and the area with a large amount of information has a high score;
    对所述特征图依次实施特征提取步骤,全连接层(FC)以及归一化指数步骤;Implement feature extraction step, fully connected layer (FC) and normalized index step to described feature map successively;
    判断输入区域属于目标标签的概率;Determine the probability that the input area belongs to the target label;
    将各个局部区域和全图提取出来的未归一化的概率合并到一起生成一个长向量,输出对应200个类别的未归一化的概率。The unnormalized probabilities extracted from each local area and the whole image are combined to generate a long vector, and the unnormalized probabilities corresponding to 200 categories are output.
  10. 一种古代及近现代艺术品鉴定系统,包括:An ancient and modern art identification system, including:
    输入模块,输入真迹图像信息;和输入拟被鉴定的艺术品的图像信息;Input module, input authentic image information; and input the image information of the artwork to be identified;
    检测模块,将所述拟被鉴定的艺术品图像信息和所述真迹图像信息通过检测器检测出分布中样本和分布外样本;The detection module detects the samples in the distribution and the samples out of the distribution by using the detector to detect the image information of the artwork to be identified and the image information of the authentic work;
    样本分类模块,将所述分布中样本进行分类;A sample classification module, classifying the samples in the distribution;
    细粒度分类模块,将分类后的所述分布中样本和与拟被鉴定的艺术品相近似的类图像信息进行细粒度分类;The fine-grained classification module performs fine-grained classification on the classified samples in the distribution and similar image information similar to the artwork to be identified;
    输出模块,输出分类后的分布中样本或细粒度分类后的样本,获得所述拟被鉴定的艺术品的图像信息与所述真迹图像信息相比的置信度作为鉴定结论。The output module outputs the classified samples in the distribution or the fine-grained classified samples, and obtains the confidence degree of the image information of the artwork to be identified compared with the authentic image information as the identification conclusion.
  11. 如权利要求10所述的鉴定系统,其中所述检测模块还包括:The identification system according to claim 10, wherein said detection module further comprises:
    分析模块,利用预训练模型(pre-trained model)输出的最大归一化指数概率进行统计分析;The analysis module utilizes the maximum normalized index probability output by the pre-trained model (pre-trained model) to carry out statistical analysis;
    统计发现OOD样本和ID样本归一化指数概率的分布情况;Statistically discover the distribution of normalized index probability of OOD samples and ID samples;
    将二者的分布差距加大;Increase the distribution gap between the two;
    选择模块,选取合适的阈值来判断一个样本属于分布外样本还是分布中样本。Select the module and select an appropriate threshold to judge whether a sample belongs to an out-of-distribution sample or an in-distribution sample.
  12. 如权利要求10所述的鉴定系统,其中所述检测模块还包括:The identification system according to claim 10, wherein said detection module further comprises:
    学习模块,利用模型学习一个对输入样本的不确定性属性;The learning module uses the model to learn an uncertainty attribute of the input sample;
    判断模块,判断测试数据,如果模型输入为分布中样本,则不确定性低,相反,如果模型输入为分布外样本,则不确定性高。The judgment module judges the test data. If the model input is a sample in the distribution, the uncertainty is low. On the contrary, if the model input is an out-of-distribution sample, the uncertainty is high.
  13. 如权利要求10所述的鉴定系统,其中所述检测模块还包括:The identification system according to claim 10, wherein said detection module further comprises:
    判断模块,利用变分自动编码器(Variational Autoencoder)的重构误差(reconstruction error)或者其他度量方式来判断一个样本是否属于分布中或分布外样本;The judging module uses the reconstruction error (reconstruction error) of the variational autoencoder (Variational Autoencoder) or other measurement methods to judge whether a sample belongs to the sample in the distribution or out of the distribution;
    所述编码器的隐含空间(latent space)能够学习出分布中数据的明显特征(silent vector),而对于分布外样本则不行,因此分布外样本会产生较高的重构误差。The latent space of the encoder can learn the obvious features (silent vector) of the data in the distribution, but not for the out-of-distribution samples, so the out-of-distribution samples will generate higher reconstruction errors.
  14. 如权利要求10所述的鉴定系统,其中所述检测模块还包括:The identification system according to claim 10, wherein said detection module further comprises:
    使用分类器对提取的特征进行分类来判断是否为分布外样本;Use a classifier to classify the extracted features to determine whether it is an out-of-distribution sample;
    有的修改网络结构为一个n+1类分类器,n为原本分类任务的类别数,第n+1类则是分布外类;Some modify the network structure into an n+1 classifier, n is the number of categories of the original classification task, and the n+1th class is an out-of-distribution class;
    有的直接取提取特征来进行分类,不需要修改网络的结构。Some directly take the extracted features for classification without modifying the structure of the network.
  15. 如权利要求10-14任意一个所述的鉴定系统,其中细粒度分类模块还包括:The identification system according to any one of claims 10-14, wherein the fine-grained classification module further includes:
    特征找寻模块,寻找拟被测类图像数据的特征区域;The feature finding module is used to find the feature area of the image data to be tested;
    特征训练模块,将所述特征区域输入卷积神经网络;A feature training module, which inputs the feature region into a convolutional neural network;
    部分信息分类模块,经过所述卷积神经网络的所述特征区域的一部分信息进入全连接层和归一化指数逻辑回归层进行分类;Part of the information classification module, through a part of the information of the feature area of the convolutional neural network, enters the fully connected layer and the normalized exponential logistic regression layer for classification;
    候选区域获得模块,经过所述卷积神经网络的所述特征区域的另一部分信息经过注意力建议子网络(APN),得到候选区域;Candidate region obtaining module, another part of information of the feature region through the convolutional neural network passes through the attention suggestion sub-network (APN), to obtain the candidate region;
    重复上述分类步骤和APN步骤,使得通过APN选取的特征区域为最具有判别性的区域;Repeat the above classification steps and APN steps, so that the feature region selected by APN is the most discriminative region;
    识别模块,引入损失函数,获得更高地识别所述类图像信息的准确性。The identification module introduces a loss function to obtain higher accuracy in identifying the type of image information.
  16. 如权利要求10-14任意一个所述的鉴定系统,其中细粒度分类模块还包括:The identification system according to any one of claims 10-14, wherein the fine-grained classification module further includes:
    局部图像训练模块,将被鉴定的类图像信息(311)中选取的局部图像(312)和(313)输入两个卷积神经网络(314,A)和(315,B);Partial image training module, input two convolutional neural networks (314, A) and (315, B) into the partial images (312) and (313) selected in the class image information (311) to be identified;
    卷积神经网络流(A)和(B)的输出在图像的每个位置使用外积相乘(318)并合并以获得双线性向量(316);The outputs of the convolutional neural network streams (A) and (B) are multiplied (318) using an outer product at each location in the image and combined to obtain a bilinear vector (316);
    预测模块,通过分类层(317)获得预测结果。The prediction module obtains the prediction result through the classification layer (317).
  17. 如权利要求16所述的鉴定系统,其中所述分类层(317)为逻辑回归或者支持向量机分类器。The authentication system according to claim 16, wherein said classification layer (317) is a logistic regression or support vector machine classifier.
  18. 如权利要求10-14任意一个所述的鉴定系统,其中细粒度分类模块还包括:The identification system according to any one of claims 10-14, wherein the fine-grained classification module further includes:
    候选框锚定模块,将被鉴定的类图像信息在不同尺度特征图上生成多个候选框,每个候选框的坐标与预先设计好的锚相对应;Candidate box anchoring module generates multiple candidate boxes on feature maps of different scales from the identified class image information, and the coordinates of each candidate box correspond to the pre-designed anchors;
    打分模块,给每个候选区域的“信息量”打分,信息量大的区域分数高;The scoring module scores the "information volume" of each candidate area, and the area with a large amount of information has a high score;
    特征图处理模块,对所述特征图依次实施特征提取,全连接层(FC)以及归一化指数;The feature map processing module implements feature extraction, fully connected layer (FC) and normalized index to the feature map in sequence;
    概率判断模块,判断输入区域属于目标标签的概率;The probability judgment module judges the probability that the input area belongs to the target label;
    概率合并输出模块,将各个局部区域和全图提取出来的未归一化的概率合并到一起生成一个长向量,输出对应200个类别的未归一化的概率。The probability merge output module combines the unnormalized probabilities extracted from each local area and the whole image to generate a long vector, and outputs the unnormalized probabilities corresponding to 200 categories.
PCT/CN2021/114254 2021-08-10 2021-08-24 Artificial intelligence-based method and system for authenticating ancient and modern artwork WO2023015610A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110916187.XA CN115705688A (en) 2021-08-10 2021-08-10 Ancient and modern artwork identification method and system based on artificial intelligence
CN202110916187.X 2021-08-10

Publications (1)

Publication Number Publication Date
WO2023015610A1 true WO2023015610A1 (en) 2023-02-16

Family

ID=85179636

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/114254 WO2023015610A1 (en) 2021-08-10 2021-08-24 Artificial intelligence-based method and system for authenticating ancient and modern artwork

Country Status (2)

Country Link
CN (1) CN115705688A (en)
WO (1) WO2023015610A1 (en)

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160140438A1 (en) * 2014-11-13 2016-05-19 Nec Laboratories America, Inc. Hyper-class Augmented and Regularized Deep Learning for Fine-grained Image Classification
CN106446874A (en) * 2016-10-28 2017-02-22 王友炎 Authentic artwork identification instrument and identification method
CN109657527A (en) * 2017-10-12 2019-04-19 上海友福文化艺术有限公司 A kind of paintings style of writing identification systems and method
CN109670365A (en) * 2017-10-12 2019-04-23 上海友福文化艺术有限公司 A kind of calligraphy identification systems and method
CN110232445A (en) * 2019-06-18 2019-09-13 清华大学深圳研究生院 A kind of historical relic authenticity identification method of knowledge based distillation
CN111539469A (en) * 2020-04-20 2020-08-14 东南大学 Weak supervision fine-grained image identification method based on vision self-attention mechanism
CN111898577A (en) * 2020-08-10 2020-11-06 腾讯科技(深圳)有限公司 Image detection method, device, equipment and computer readable storage medium

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160140438A1 (en) * 2014-11-13 2016-05-19 Nec Laboratories America, Inc. Hyper-class Augmented and Regularized Deep Learning for Fine-grained Image Classification
CN106446874A (en) * 2016-10-28 2017-02-22 王友炎 Authentic artwork identification instrument and identification method
CN109657527A (en) * 2017-10-12 2019-04-19 上海友福文化艺术有限公司 A kind of paintings style of writing identification systems and method
CN109670365A (en) * 2017-10-12 2019-04-23 上海友福文化艺术有限公司 A kind of calligraphy identification systems and method
CN110232445A (en) * 2019-06-18 2019-09-13 清华大学深圳研究生院 A kind of historical relic authenticity identification method of knowledge based distillation
CN111539469A (en) * 2020-04-20 2020-08-14 东南大学 Weak supervision fine-grained image identification method based on vision self-attention mechanism
CN111898577A (en) * 2020-08-10 2020-11-06 腾讯科技(深圳)有限公司 Image detection method, device, equipment and computer readable storage medium

Also Published As

Publication number Publication date
CN115705688A (en) 2023-02-17

Similar Documents

Publication Publication Date Title
US11868394B2 (en) Analyzing content of digital images
Wu et al. Self-critical reasoning for robust visual question answering
Zhou et al. Interpretable basis decomposition for visual explanation
Baró et al. Traffic sign recognition using evolutionary adaboost detection and forest-ECOC classification
WO2019015246A1 (en) Image feature acquisition
CN109919252B (en) Method for generating classifier by using few labeled images
CN108492298B (en) Multispectral image change detection method based on generation countermeasure network
CN108334805B (en) Method and device for detecting document reading sequence
CN112633382A (en) Mutual-neighbor-based few-sample image classification method and system
CN111324765A (en) Fine-grained sketch image retrieval method based on depth cascade cross-modal correlation
CN112990282B (en) Classification method and device for fine-granularity small sample images
Valavanis et al. Ipl at clef 2016 medical task.
CN111325237A (en) Image identification method based on attention interaction mechanism
Soumya et al. Emotion recognition from partially occluded facial images using prototypical networks
Das et al. Determining attention mechanism for visual sentiment analysis of an image using svm classifier in deep learning based architecture
CN113792686A (en) Vehicle weight identification method based on cross-sensor invariance of visual representation
Al-Qudah et al. Synthetic blood smears generation using locality sensitive hashing and deep neural networks
CN112597997A (en) Region-of-interest determining method, image content identifying method and device
CN116935411A (en) Radical-level ancient character recognition method based on character decomposition and reconstruction
WO2023015610A1 (en) Artificial intelligence-based method and system for authenticating ancient and modern artwork
Nithya et al. A review on automatic image captioning techniques
US20170293863A1 (en) Data analysis system, and control method, program, and recording medium therefor
Pryor et al. Deepfake Detection Analyzing Hybrid Dataset Utilizing CNN and SVM
CN113724261A (en) Fast image composition method based on convolutional neural network
Bhoomeshwar et al. Random Forest Classifier For Classifying Birds Species using Scikitlearn

Legal Events

Date Code Title Description
NENP Non-entry into the national phase

Ref country code: DE