WO2023072094A1 - Visualization and quantitative analysis method and system for expression capability of layer feature in neural network - Google Patents

Visualization and quantitative analysis method and system for expression capability of layer feature in neural network Download PDF

Info

Publication number
WO2023072094A1
WO2023072094A1 PCT/CN2022/127435 CN2022127435W WO2023072094A1 WO 2023072094 A1 WO2023072094 A1 WO 2023072094A1 CN 2022127435 W CN2022127435 W CN 2022127435W WO 2023072094 A1 WO2023072094 A1 WO 2023072094A1
Authority
WO
WIPO (PCT)
Prior art keywords
sample
low
level features
region
neural network
Prior art date
Application number
PCT/CN2022/127435
Other languages
French (fr)
Chinese (zh)
Inventor
张拳石
Original Assignee
上海交通大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 上海交通大学 filed Critical 上海交通大学
Publication of WO2023072094A1 publication Critical patent/WO2023072094A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24133Distances to prototypes
    • G06F18/24137Distances to cluster centroïds
    • G06F18/2414Smoothing the distance, e.g. radial basis function networks [RBFN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • This application relates to the technical field of machine learning, and in particular to a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks.
  • the visualization method is the most widely used method in the field of artificial intelligence interpretability, but it is impossible to use the existing visualization method to quantitatively analyze the expressive ability of the layer features in the neural network.
  • the purpose of the present invention is to provide a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks, which can automatically visualize and quantitatively analyze the expressive ability of middle-level features of neural networks under unsupervised conditions.
  • This application discloses a method for visualizing and quantitatively analyzing the expressive ability of the middle layer of the neural network, including the following steps
  • the model includes a middle-level expression, including: a neural network, a hierarchical graph model;
  • the dimensionality reduction of the sample-level features is performed to obtain the visualization results of the sample-level features in the low-dimensional space;
  • the dimensionality reduction of the regional-level features is performed to obtain the regional-level features Visualization results in low-dimensional space;
  • the quantity and quality of knowledge points in the features are quantitatively analyzed.
  • said step (1) further includes the following steps:
  • the neural network is a classification neural network.
  • said step (2) further includes the following sub-steps:
  • said step (3) further includes the following sub-steps:
  • the radial distribution radial distribution
  • the probability density function of g in low-dimensional space can be calculated by the following formula.
  • y ⁇ 1,2,...,C ⁇ represents different categories in the classification task;
  • ⁇ y represents the prior probability of the yth class;
  • ⁇ y represents the mean direction of the yth class;
  • ⁇ ( ⁇ ) is a monotonically increasing function.
  • y) indicates the prior probability of l g on category y
  • ⁇ y , ⁇ (l g )) indicates that the average direction is ⁇ y
  • the concentration parameter is ⁇ ( l g ) vMF distribution (von Mises-Fisher distribution).
  • x) between the low-dimensional representation g and the yth category can be calculated by the following formula.
  • the "closeness between the sample and each category" is calculated as the output probability of the sample, that is, the closeness P(y
  • the optimization of the projection matrix M further includes the calculation results of the closeness Q M (y
  • the similarity between samples is split into regions corresponding to low-dimensional
  • the weighted product of the similarity between representations and based on the vMF distribution, the similarity between the corresponding low-dimensional representations of each region is quantified.
  • ⁇ ( ⁇ ) is a monotonically increasing function.
  • cos( ⁇ , ⁇ ) represents the cosine similarity between two vectors
  • ⁇ p is a non-negative constant
  • aligning the low-dimensional representation of region-level features with the low-dimensional representation of sample-level features is equivalent to optimizing the following loss function.
  • MI( ⁇ ; ⁇ ) represents mutual information
  • g represents the low-dimensional representation of the sample-level features of sample x
  • h (r) represents the low-dimensional representation of the r-th region-level features of sample x
  • w ( r) indicates the importance of the r-th region-level feature of sample x for classification.
  • the optimization of the projection matrix ⁇ includes the following steps: respectively calculating the above two loss functions and Then calculate the total loss function Optimizing ⁇ based on the total loss function such that the total loss function minimize.
  • is a positive constant, and in a preferred example, the value range of ⁇ is 0.01 to 100. In one embodiment, the value of ⁇ is taken as 0.1.
  • said step (4) further includes the following steps:
  • a knowledge point is defined as a set of region-level features such that the following formula is greater than a certain threshold.
  • h (r) )> ⁇ , that is, the set ⁇ h (r) : max c p(y c
  • is a positive constant, and in a preferred example, the value range of ⁇ is 0.3-0.8. In one embodiment, the value of ⁇ is 0.4.
  • reliable knowledge points are the set of knowledge points that further satisfy the following formula.
  • h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x
  • the ratio of reliable knowledge points to total knowledge points measures the quality of knowledge points, which can be calculated by the following formula.
  • the second aspect of the present invention provides a visualization and quantitative analysis system for the expressive ability of the middle layer of the neural network, which is characterized in that it includes:
  • the input module is configured as a pre-trained classification neural network and contains input samples of all possible categories;
  • a feature extraction module configured to extract sample-level features and region-level features of the input samples
  • a visualization module configured to, based on the extracted sample-level features and region-level features, reduce its dimension to obtain a low-dimensional representation, and visualize the low-dimensional representation in the low-dimensional space;
  • the quantitative analysis module is configured to quantitatively analyze the quantity and quality of knowledge points in the feature based on the visualization result of the feature at the region level.
  • Fig. 1 is a schematic flow chart of a method for visualization and quantitative analysis of layer feature expression capabilities in a neural network according to a first embodiment of the present invention
  • Fig. 2 is a schematic diagram of visualization of sample-level features in low-dimensional space obtained according to the present invention
  • Fig. 3 is a schematic diagram of visualization of region-level features obtained according to the present invention in low-dimensional space at different training stages of the neural network;
  • Fig. 4 is a schematic diagram of visualization of region-level features obtained according to the present invention in low-dimensional space at different stages of forward propagation of the neural network;
  • Fig. 5 is according to the quantification to knowledge point quantity and quality in the present invention, obtains the knowledge point and reliable knowledge point in the different inter-layer features of neural network, with the change graph of different training stages of neural network;
  • Fig. 6 is the quantification of knowledge points according to the present invention, and the visualized diagram of knowledge points in different inter-layer features of the neural network obtained;
  • Fig. 7 is a schematic diagram of the system structure of the visualization and quantitative analysis of the feature expression ability of the middle layer of the neural network according to the second embodiment of the present invention.
  • the present inventor has first developed a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks.
  • the visual interpretation of the neural network is closely linked with the quantitative analysis of the feature expression ability of the middle layer of the neural network; through visualization, the emergence process of the feature expression ability of the middle layer of the neural network can be clearly displayed over time and space; this kind of The method can quantitatively analyze the quantity and quality of the knowledge points in the middle layer of the neural network, and then analyze the reliability of the model to be explained; based on this method and system, it can be used for many existing deep learning algorithms, such as adversarial attacks and knowledge distillation. An interpretive framework from a new perspective.
  • the present invention includes the following steps:
  • a neural network to be analyzed which is a pre-trained deep neural network on a certain data set
  • This method can quantitatively analyze the quantity and quality of knowledge points in the middle layer of the neural network, and then analyze the reliability of the model to be explained;
  • the first embodiment of the present invention designs a kind of visualization and quantitative analysis method to the feature expression ability of neural network middle layer, and its flow process is as shown in Figure 1, and this method comprises the following steps:
  • step 101 based on a certain data set, train a neural network as the neural network to be analyzed.
  • the neural network is a classification neural network.
  • step 102 this step can be further divided into following two sub-steps:
  • step 103 this step can be further divided into following two sub-steps:
  • the radial distribution radial distribution
  • the probability density function of g in low-dimensional space can be calculated by the following formula.
  • y ⁇ 1,2,...,C ⁇ represents different categories in the classification task;
  • ⁇ y represents the prior probability of the yth class;
  • ⁇ y represents the mean direction of the yth class;
  • ⁇ ( ⁇ ) is a monotonically increasing function.
  • y) indicates the prior probability of l g on category y
  • ⁇ y , ⁇ (l g )) indicates that the average direction is ⁇ y
  • the concentration parameter is ⁇ ( l g ) vMF distribution (von Mises-Fisher distribution).
  • x) between the low-dimensional representation g and the yth category can be calculated by the following formula.
  • the "closeness between the sample and each category" is calculated as the output probability of the sample, that is, the closeness P(y
  • the optimization of the projection matrix M further includes the calculation results of the closeness Q M (y
  • a VGG-16 network pre-trained on the Tiny ImageNet image classification data set is given, wherein only the steel arch is used in the neural network during training
  • the ten categories of bridge, school bus, sports car, tabby cat, desk, golden retriever, tailed frog, iPod, lifeboat, orange are extracted from the second-to-last fully connected layer of the VGG-16 network as sample-level features, using The aforementioned method performs dimensionality reduction to a low-dimensional representation scatter diagram obtained in a three-dimensional space.
  • different colors represent different categories shown in the legend, and the arrows corresponding to the colors of each category indicate the average direction of each category.
  • the similarity between samples is split into regions corresponding to low-dimensional
  • the weighted product of the similarity between representations and based on the vMF distribution, the similarity between the corresponding low-dimensional representations of each region is quantified.
  • ⁇ ( ⁇ ) is a monotonically increasing function.
  • cos( ⁇ , ⁇ ) represents the cosine similarity between two vectors
  • ⁇ p is a non-negative constant
  • aligning the low-dimensional representation of region-level features with the low-dimensional representation of sample-level features is equivalent to optimizing the following loss function.
  • MI( ⁇ ; ⁇ ) represents mutual information
  • g represents the low-dimensional representation of the sample-level features of sample x
  • h (r) represents the low-dimensional representation of the r-th region-level features of sample x
  • w ( r) indicates the importance of the r-th region-level feature of sample x for classification.
  • the optimization of the projection matrix ⁇ includes the following steps: respectively calculating the above two loss functions and Then calculate the total loss function Optimizing ⁇ based on the total loss function such that the total loss function minimize.
  • is a positive constant, and in a preferred example, the value range of ⁇ is 0.01 to 100. In one embodiment, the value of ⁇ is taken as 0.1.
  • the output features of the conv_53 layer are extracted as region-level features, and the dimensionality reduction to three dimensions is performed using the aforementioned method to obtain a low-dimensional representation. , the distribution in three-dimensional space.
  • the scatter points of different colors represent the low-dimensional representations of the region-level features corresponding to different types of samples
  • the ellipsoids in the figure represent the approximate distribution of the low-dimensional representations of the region-level features corresponding to different types of samples.
  • Figure 4 shows the distribution of low-dimensional representations of regional-level features corresponding to different samples as the number of layers of forward propagation increases, where the vertical upward arrow represents the correct average direction of the sample, and the scatter points of different colors represent different The distribution of low-dimensional representations of region-level features corresponding to layers in three-dimensional space.
  • step 104 this step can be further divided into following sub-steps:
  • a knowledge point is defined as a set of region-level features such that the following formula is greater than a certain threshold.
  • h (r) )> ⁇ , that is, the set ⁇ h (r) : max c p(y c
  • is a positive constant, and in a preferred example, the value range of ⁇ is 0.3-0.8. In one embodiment, the value of ⁇ is 0.4.
  • reliable knowledge points are the set of knowledge points that further satisfy the following formula.
  • h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x
  • the ratio of reliable knowledge points to total knowledge points measures the quality of knowledge points, which can be calculated by the following formula.
  • Figure 5 shows the change curve of the total amount of knowledge points and the number of reliable knowledge points at different layers of the neural network with the number of neural network training iterations.
  • the second embodiment of the present invention designs a kind of visualization and quantitative analysis system to the feature expression ability of neural network, its structure is shown in Figure 7, and this system comprises:
  • the input module is configured as a pre-trained classification neural network and contains input samples of all possible categories;
  • a visualization module configured to, based on the extracted sample-level features and region-level features, reduce its dimension to obtain a low-dimensional representation, and visualize the low-dimensional representation in the low-dimensional space;
  • the quantitative analysis module is configured to quantitatively analyze the quantity and quality of knowledge points in the feature based on the visualization result of the feature at the region level.
  • each module shown in the implementation of the above-mentioned visualization and quantitative analysis system for the middle-level feature expression ability of the neural network can refer to the aforementioned description of the middle-level feature expression ability of the neural network. Visualization and related descriptions of quantitative analysis methods can be understood.
  • the functions of the modules shown in the implementation of the above-mentioned visualization and quantitative analysis system for the middle-level feature expression ability of the neural network can be realized by a program (executable instruction) running on the processor, or can be realized by a specific logic circuit. accomplish.
  • the above-mentioned visualization and quantitative analysis system for the expression ability of neural network middle layer is implemented in the form of software function modules and sold or used as an independent product, it can also be stored in a computer-readable storage medium.
  • the computer software products are stored in a storage medium and include several instructions for Make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the method in each embodiment of the present invention.
  • the aforementioned storage medium includes: various media that can store program codes such as U disk, mobile hard disk, read-only memory (ROM, Read Only Memory), magnetic disk or optical disk.
  • embodiments of the invention are not limited to any specific combination of hardware and software.
  • the embodiments of the present invention also provide a computer-readable storage medium, in which computer-executable instructions are stored, and when the computer-executable instructions are executed by a processor, various method embodiments of the present invention are implemented.
  • Computer-readable storage media includes both volatile and non-permanent, removable and non-removable media by any method or technology for storage of information.
  • Information may be computer readable instructions, data structures, modules of a program, or other data.
  • Examples of storage media for computers include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Flash memory or other memory technology, Compact Disc Read-Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridge, tape magnetic disk storage or other magnetic storage device or any other non-transmission medium that can be used to store information that can be accessed by a computing device.
  • computer-readable storage media does not include transitory computer-readable media, such as modulated data signals and carrier waves.
  • the embodiment of the present invention also provides a visualization and quantitative analysis system for the expression ability of the layer features in the neural network, which includes a memory for storing computer-executable instructions, and a processor; the processor is used to execute the memory in the memory
  • the processor can be a central processing unit (Central Processing Unit, referred to as "CPU"), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, referred to as "DSP"), application specific integrated circuits (Application Specific Integrated Circuit, referred to as "ASIC”) and so on.
  • CPU Central Processing Unit
  • DSP Digital Signal Processor
  • ASIC Application Specific Integrated Circuit
  • the aforementioned memory may be a read-only memory ("ROM” for short), a random access memory (random access memory, "RAM” for short), a flash memory (Flash), a hard disk or a solid-state hard disk, and the like.
  • ROM read-only memory
  • RAM random access memory
  • flash flash memory
  • hard disk or a solid-state hard disk and the like.
  • an action is performed according to a certain element, it means that the action is performed based on at least the element, which includes two situations: performing the action only based on the element, and performing the action based on the element and Other elements perform the behavior.
  • Expressions such as multiple, multiple, and multiple include 2, 2 times, 2 types, and 2 or more, 2 or more times, or 2 or more types.

Abstract

The present application relates to the technical field of machine learning. Disclosed are a visualization and quantitative analysis method and system for an expression capability of a layer feature in a neural network. The automatic visualization and quantitative analysis of an expression capability of a layer feature in a neural network can be realized under an unsupervised condition. The method comprises: providing a neural network to be analyzed, wherein the neural network is a deep neural network, which is pre-trained on a certain data set; providing a group of input samples, inputting the samples into the neural network, and extracting sample-wise features and region-wise features corresponding to the samples; respectively performing dimension reduction on the sample-wise features and the region-wise features, so as to obtain visualization results in a low-dimensional space; and quantitatively analyzing the number and quality of knowledge points in the features on the basis of the visualization result of the region-wise features.

Description

一种对神经网络中层特征表达能力的可视化及定量分析方法和系统A method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features in neural networks 技术领域technical field
本申请涉及机器学习技术领域,特别涉及一种对神经网络中层特征表达能力的可视化及定量分析方法和系统。This application relates to the technical field of machine learning, and in particular to a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks.
背景技术Background technique
目前,深度神经网络已经在各个领域展现出强大性能,但神经网络的黑盒本质让人们很难理解其内部的行为。在现有的技术中,可视化方法是人工智能可解释性领域最为广泛应用的方法,但无法利用现有的可视化方法去定量分析神经网络中层特征的表达能力。At present, deep neural networks have demonstrated powerful performance in various fields, but the black-box nature of neural networks makes it difficult for people to understand their internal behavior. Among the existing technologies, the visualization method is the most widely used method in the field of artificial intelligence interpretability, but it is impossible to use the existing visualization method to quantitatively analyze the expressive ability of the layer features in the neural network.
因此,将神经网络的可视化解释与中层特征表达能力的定量分析相结合,是人工智能可解释性领域亟待解决的问题。Therefore, combining the visual interpretation of neural networks with the quantitative analysis of the expressive power of mid-level features is an urgent problem to be solved in the field of artificial intelligence interpretability.
发明内容Contents of the invention
本发明的目的在于提供一种对神经网络中层特征表达能力的可视化及定量分析方法和系统,可实现在无监督的条件下,自动可视化以及定量分析神经网络中层特征的表达能力。The purpose of the present invention is to provide a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks, which can automatically visualize and quantitatively analyze the expressive ability of middle-level features of neural networks under unsupervised conditions.
本申请公开了一种对神经网络中层特征表达能力的可视化及定量分析方法,包括以下步骤This application discloses a method for visualizing and quantitatively analyzing the expressive ability of the middle layer of the neural network, including the following steps
(1)选取特征解释对象:(1) Select the feature interpretation object:
选取待分析的模型,其中,所述模型包括有中层表达,包括:神经网络,层次化图模型;Select the model to be analyzed, wherein the model includes a middle-level expression, including: a neural network, a hierarchical graph model;
(2)提取神经网络特征:(2) Extract neural network features:
提供一组输入样本,将这些样本输入上述神经网络,提取这些样本的特征,其中,所述特征包括:样本级别特征(sample-wise feature)和区域级 别特征(regional feature);Provide a set of input samples, input these samples into the above-mentioned neural network, and extract the features of these samples, wherein the features include: sample-wise feature and regional feature;
(3)特征降维,得到可视化结果:(3) Feature dimensionality reduction to obtain visualization results:
首先对样本级别特征进行降维,得到样本级别特征在低维空间中的可视化结果;其次,基于样本级别特征的低维表征和区域级别特征,将区域级别特征进行降维,以得到区域级别特征在低维空间中的可视化结果;First, the dimensionality reduction of the sample-level features is performed to obtain the visualization results of the sample-level features in the low-dimensional space; secondly, based on the low-dimensional representation of the sample-level features and the regional-level features, the dimensionality reduction of the regional-level features is performed to obtain the regional-level features Visualization results in low-dimensional space;
(4)根据可视化结果对特征进行定量分析:(4) Quantitative analysis of features based on visualization results:
基于所述可视化结果,定量分析特征中知识点(knowledge point)的数量与质量。Based on the visualization results, the quantity and quality of knowledge points in the features are quantitatively analyzed.
在一个优选例中,所述步骤(1)进一步包括以下步骤:In a preferred example, said step (1) further includes the following steps:
基于某一数据集,训练一神经网络,作为待分析的神经网络。可选地,该神经网络为一个分类神经网络。Based on a certain data set, train a neural network as the neural network to be analyzed. Optionally, the neural network is a classification neural network.
在一个优选例中,所述步骤(2)进一步包括以下子步骤:In a preferred example, said step (2) further includes the following sub-steps:
(a)对样本级别特征的提取:将给定的一组样本输入待分析的神经网络,对于每个样本,都提取该神经网络某一中间层的输出特征,从而得到每个输入样本对应的样本级别特征,即可得到该组输入样本对应的样本级别特征。(a) Extraction of sample-level features: Input a given set of samples into the neural network to be analyzed, and for each sample, extract the output features of an intermediate layer of the neural network, so as to obtain the corresponding Sample-level features, the sample-level features corresponding to the set of input samples can be obtained.
(b)对区域级别特征的提取:将给定的一组输入样本输入待分析的神经网络,对于每个样本,都提取该神经网络某一卷积层的输出特征,从而得到每个输入样本对应的特征图(feature map),其中,特征图的每一个位置对应的高维向量即为该样本在这一区域的区域级别特征。当这一特征图的高与宽分别为H和W,且共有K个通道时,那么,这一特征图包含HW个区域级别特征,其中,每个区域级别特征为一个K维向量。(b) Extraction of region-level features: Input a given set of input samples into the neural network to be analyzed, and for each sample, extract the output features of a certain convolutional layer of the neural network to obtain each input sample The corresponding feature map (feature map), where the high-dimensional vector corresponding to each position of the feature map is the region-level feature of the sample in this region. When the height and width of this feature map are H and W respectively, and there are K channels in total, then this feature map contains HW region-level features, where each region-level feature is a K-dimensional vector.
在一个优选例中,所述步骤(3)进一步包括以下子步骤:In a preferred example, said step (3) further includes the following sub-steps:
(a)对样本级别特征进行降维,并在低维空间中进行可视化;(a) Dimensionality reduction of sample-level features and visualization in low-dimensional space;
(b)对区域级别特征进行降维,并在低维空间中进行可视化。(b) Dimensionality reduction is performed on region-level features and visualized in a low-dimensional space.
上述子步骤(a)中,对于每个样本x对应的样本级别特征
Figure PCTCN2022127435-appb-000001
通过一个投影矩阵
Figure PCTCN2022127435-appb-000002
将其映射到一个低维空间中,得到样本级别特征的低维表征
Figure PCTCN2022127435-appb-000003
并且,优化M使得该低维表征应当满足,低维表征g和各个类别的接近程度应当与该样本x和各个类别的接近程度尽可能地保持一致。
In the above sub-step (a), for each sample x corresponding to the sample level features
Figure PCTCN2022127435-appb-000001
through a projection matrix
Figure PCTCN2022127435-appb-000002
Map it into a low-dimensional space to obtain a low-dimensional representation of sample-level features
Figure PCTCN2022127435-appb-000003
And, optimize M so that the low-dimensional representation should satisfy that the closeness of the low-dimensional representation g to each category should be consistent with the closeness of the sample x to each category as much as possible.
可选地,在“低维表征和各个类别的接近程度”的计算中,先利用径向分布(radial distribution)建模样本级别特征的低维表征g在低维空间中的分布,再基于此计算低维表征与各个类别的接近程度。Optionally, in the calculation of the "closeness between low-dimensional representation and each category", first use the radial distribution (radial distribution) to model the distribution of the low-dimensional representation g of the sample-level feature in the low-dimensional space, and then based on this Computes how close a low-dimensional representation is to each class.
基于径向分布,g在低维空间的概率密度函数可由如下公式计算得到。Based on the radial distribution, the probability density function of g in low-dimensional space can be calculated by the following formula.
Figure PCTCN2022127435-appb-000004
Figure PCTCN2022127435-appb-000004
其中,y∈{1,2,…,C}表示分类任务中的不同类别;π y表示第y类的先验概率;l g=‖g‖表示g的L2范数,称为g的强度(strength);o g=g/l g表示g的方向(orientation);μ y表示第y类的平均方向(mean direction);κ(·)是一个单调递增的函数。p(l g|y)表示在类别y上l g的先验概率,p vMF(o gy,κ(l g))表示平均方向为μ y,聚集参数(concentration parameter)为κ(l g)的vMF分布(von Mises-Fisher distribution)。 Among them, y∈{1,2,…,C} represents different categories in the classification task; π y represents the prior probability of the yth class; l g =‖g‖ represents the L2 norm of g, which is called the strength of g (strength); o g = g/l g represents the orientation of g; μ y represents the mean direction of the yth class; κ(·) is a monotonically increasing function. p(l g |y) indicates the prior probability of l g on category y, p vMF (o gy ,κ(l g )) indicates that the average direction is μ y , and the concentration parameter is κ( l g ) vMF distribution (von Mises-Fisher distribution).
可选地,κ(·)可以是任意一个单调递增的函数,更佳地,κ(·)函数可以通过如下方法生成:给定一个非负的常数κ m,以及维数d,从平均方向为
Figure PCTCN2022127435-appb-000005
聚集参数为κ=κ m的vMF分布中采样得到N个样本
Figure PCTCN2022127435-appb-000006
将这些样本都缩放到长度为l,而不改变方向,此处的l是一个任意非负数,将缩放得到的样本记作
Figure PCTCN2022127435-appb-000007
从标准正态分布中采样得到N个高斯噪声样本
Figure PCTCN2022127435-appb-000008
将前述缩放得到的样本与高斯噪声样本对应相加,得到
Figure PCTCN2022127435-appb-000009
将κ(l)定义为
Figure PCTCN2022127435-appb-000010
其中
Figure PCTCN2022127435-appb-000011
N的范围为1000-100000,在一个实施例中,N取为10000。
Optionally, κ(·) can be any monotonically increasing function, more preferably, the κ(·) function can be generated by the following method: Given a non-negative constant κ m and dimension d, from the average direction for
Figure PCTCN2022127435-appb-000005
N samples are obtained by sampling from the vMF distribution with aggregation parameter κ=κ m
Figure PCTCN2022127435-appb-000006
Scale these samples to a length of l without changing the direction, where l is an arbitrary non-negative number, and record the scaled samples as
Figure PCTCN2022127435-appb-000007
Sample N samples of Gaussian noise from a standard normal distribution
Figure PCTCN2022127435-appb-000008
Add the sample obtained by the aforementioned scaling to the Gaussian noise sample correspondingly, and obtain
Figure PCTCN2022127435-appb-000009
Define κ(l) as
Figure PCTCN2022127435-appb-000010
in
Figure PCTCN2022127435-appb-000011
The range of N is 1000-100000, and in one embodiment, N is taken as 10000.
基于上述径向分布,并假设l g的先验概率与类别y是独立的,那么,低维表征g和第y类的接近程度Q M(y|x)可以由如下公式计算得到。 Based on the radial distribution above, and assuming that the prior probability of l g is independent of category y, then the closeness Q M (y|x) between the low-dimensional representation g and the yth category can be calculated by the following formula.
Figure PCTCN2022127435-appb-000012
Figure PCTCN2022127435-appb-000012
可选地,在“样本和各个类别的接近程度”计算为样本的输出概率,即样本x和第y类的接近程度P(y|x)为神经网络网络输出中对应第y类的输出概率值。Optionally, the "closeness between the sample and each category" is calculated as the output probability of the sample, that is, the closeness P(y|x) between the sample x and the yth category is the output probability corresponding to the yth category in the neural network output value.
可选地,对于投影矩阵M的优化,进一步包括,低维表征g和各个类别的接近程度Q M(y|x),以及样本x和各个类别的接近程度P(y|x)的计算结果,优化投影矩阵M,使得P(y|x)与Q M(y|x)间的KL散度(Kullback–Leibler divergence)最小化。 Optionally, the optimization of the projection matrix M further includes the calculation results of the closeness Q M (y|x) between the low-dimensional representation g and each category, and the closeness P(y|x) between the sample x and each category , optimize the projection matrix M so that the KL divergence (Kullback–Leibler divergence) between P(y|x) and Q M (y|x) is minimized.
Figure PCTCN2022127435-appb-000013
Figure PCTCN2022127435-appb-000013
可选地,在得到样本级别特征的低维表征的过程中,交替优化投影矩阵M与径向分布中的参数{π,μ}={π yy} y∈Y。其中,优化投影矩阵M时,固定径向分布中的参数{π,μ},并且更新M,使得KL[P(Y|X)‖Q M(Y|X)]的值最小化;优化径向分布中的参数{π,μ}时,固定投影矩阵M,并且更新{π,μ},使得似然∏ gp(g)=∏ gy′π y′·p vMF(o gy′,κ(l g))的值最大。 Optionally, in the process of obtaining the low-dimensional representation of sample-level features, alternately optimize the projection matrix M and the parameters {π,μ}={π yy } y∈Y in the radial distribution. Among them, when optimizing the projection matrix M, the parameters {π, μ} in the radial distribution are fixed, and M is updated to minimize the value of KL[P(Y|X)‖Q M (Y|X)]; the optimized path When the parameters in the distribution are {π, μ}, the projection matrix M is fixed, and {π, μ} is updated, so that the likelihood ∏ g p(g)=∏ gy′ π y′ ·p vMF (o g | μ y′ ,κ(l g )) has the largest value.
上述子步骤(b)中,对于每个样本x的HW个区域级别特征
Figure PCTCN2022127435-appb-000014
Figure PCTCN2022127435-appb-000015
通过一个投影矩阵
Figure PCTCN2022127435-appb-000016
将它们映射到一个低维空间中,得到HW个区域级别特征的低维表征
Figure PCTCN2022127435-appb-000017
Figure PCTCN2022127435-appb-000018
并且,优化Λ使得该低维表征应当满足,基于低维表征h={h (1),h (2),…,h (HW)}所推断出的样本间相似度与基于网络输出所推断出的样本间相似度尽可能保持一致,进一步,区域级别特征的低维表征需要与样本级别特征的低维特征表征对齐。
In the above sub-step (b), for each sample x's HW region-level features
Figure PCTCN2022127435-appb-000014
Figure PCTCN2022127435-appb-000015
through a projection matrix
Figure PCTCN2022127435-appb-000016
Map them into a low-dimensional space to obtain a low-dimensional representation of HW region-level features
Figure PCTCN2022127435-appb-000017
Figure PCTCN2022127435-appb-000018
And, optimize Λ so that the low-dimensional representation should satisfy that the similarity between samples inferred based on the low-dimensional representation h={h (1) ,h (2) ,…,h (HW) } is the same as that inferred based on the network output The similarity between samples obtained should be as consistent as possible. Furthermore, the low-dimensional representation of region-level features needs to be aligned with the low-dimensional feature representation of sample-level features.
可选地,在“基于低维表征所推断出的样本间相似度”的计算中,基于词袋模型(the bag-of-words model),将样本间相似度拆分为各个区域对应低维表征间相似度的加权乘积;并基于vMF分布,量化各个区域对应低维表征间的相似度。Optionally, in the calculation of "similarity between samples inferred based on low-dimensional representation", based on the bag-of-words model (the bag-of-words model), the similarity between samples is split into regions corresponding to low-dimensional The weighted product of the similarity between representations; and based on the vMF distribution, the similarity between the corresponding low-dimensional representations of each region is quantified.
可选地,在以上描述中,设x 1和x 2为任意两个样本,所对应区域级别特征的低维表征分别为
Figure PCTCN2022127435-appb-000019
Figure PCTCN2022127435-appb-000020
基于词袋模型,将x 1和x 2之间的相似度Q Λ(x 2|x 1)拆分为各个区域对应低维表征间相似度的加权乘积,如下所示。
Optionally, in the above description, let x1 and x2 be any two samples, and the low-dimensional representations of the corresponding region-level features are respectively
Figure PCTCN2022127435-appb-000019
and
Figure PCTCN2022127435-appb-000020
Based on the bag-of-words model, the similarity Q Λ (x 2 |x 1 ) between x 1 and x 2 is split into the weighted product of the similarity between the corresponding low-dimensional representations of each region, as shown below.
Figure PCTCN2022127435-appb-000021
Figure PCTCN2022127435-appb-000021
其中,
Figure PCTCN2022127435-appb-000022
表示样本x 2中第r个区域特征对于分类的重要程度,在一个优选例中,
Figure PCTCN2022127435-appb-000023
是一个非负数。并且,
Figure PCTCN2022127435-appb-000024
进一步量化为如下形式。
in,
Figure PCTCN2022127435-appb-000022
Indicates the importance of the rth regional feature in the sample x 2 for classification. In a preferred example,
Figure PCTCN2022127435-appb-000023
is a non-negative number. and,
Figure PCTCN2022127435-appb-000024
It is further quantified into the following form.
Figure PCTCN2022127435-appb-000025
Figure PCTCN2022127435-appb-000025
其中,
Figure PCTCN2022127435-appb-000026
Figure PCTCN2022127435-appb-000027
在平均方向为
Figure PCTCN2022127435-appb-000028
聚集参数为
Figure PCTCN2022127435-appb-000029
的vMF分布中的概率密度;同权利要求6中所述,κ(·)是一个单调递增的函数。
in,
Figure PCTCN2022127435-appb-000026
for
Figure PCTCN2022127435-appb-000027
in the mean direction of
Figure PCTCN2022127435-appb-000028
The aggregation parameters are
Figure PCTCN2022127435-appb-000029
The probability density in the distribution of vMF; as described in claim 6, κ(·) is a monotonically increasing function.
可选地,在“基于网络输出所推断出的样本间相似度”的计算中,设x 1和x 2为任意两个样本,
Figure PCTCN2022127435-appb-000030
分别为这两个样本对应的网络输出概率,基于网络输出所推断出的样本间相似度P(x 2|x 1)可进一步计算为如下形式。
Optionally, in the calculation of "similarity between samples inferred based on network output", let x1 and x2 be any two samples,
Figure PCTCN2022127435-appb-000030
are the network output probabilities corresponding to the two samples, and the similarity P(x 2 |x 1 ) between samples inferred based on the network output can be further calculated as the following form.
Figure PCTCN2022127435-appb-000031
Figure PCTCN2022127435-appb-000031
其中,
Figure PCTCN2022127435-appb-000032
cos(·,·)表示两个向量之间的余弦相似度,κ p为一个非负常数。
in,
Figure PCTCN2022127435-appb-000032
cos(·,·) represents the cosine similarity between two vectors, and κ p is a non-negative constant.
可选地,使得“基于低维表征所推断出的样本间相似度与基于网络输出 所推断出的样本间相似度尽可能保持一致”,等价于最小化如下损失函数。Optionally, making "the similarity between samples inferred based on low-dimensional representations and the similarity between samples inferred based on network output as consistent as possible" is equivalent to minimizing the following loss function.
Figure PCTCN2022127435-appb-000033
Figure PCTCN2022127435-appb-000033
其中,P(x 2|x 1)和Q Λ(x 2|x 1)的计算如权利要求16和权利要求17分别所示。 Wherein, the calculations of P(x 2 |x 1 ) and Q Λ (x 2 |x 1 ) are as shown in claim 16 and claim 17 respectively.
可选地,将区域级别特征的低维表征与样本级别特征的低维表征对齐,等价于优化如下损失函数。Optionally, aligning the low-dimensional representation of region-level features with the low-dimensional representation of sample-level features is equivalent to optimizing the following loss function.
Figure PCTCN2022127435-appb-000034
Figure PCTCN2022127435-appb-000034
其中,MI(·;·)表示互信息(mutual information);g表示样本x的样本级别特征的低维表征;h (r)表示样本x的第r个区域级别特征的低维表征,w (r)表示样本x的第r个区域级别特征对于分类的重要程度。 Among them, MI(·;·) represents mutual information; g represents the low-dimensional representation of the sample-level features of sample x; h (r) represents the low-dimensional representation of the r-th region-level features of sample x, w ( r) indicates the importance of the r-th region-level feature of sample x for classification.
可选地,对投影矩阵Λ的优化包括以下步骤:分别计算上述两个损失函数
Figure PCTCN2022127435-appb-000035
Figure PCTCN2022127435-appb-000036
进而计算总损失函数
Figure PCTCN2022127435-appb-000037
基于总损失函数优化Λ,使得总损失函数
Figure PCTCN2022127435-appb-000038
最小化。其中,α是一个正的常数,在一个优选例中,α的取值范围是0.01至100。在一个实施例中,α的值取为0.1。
Optionally, the optimization of the projection matrix Λ includes the following steps: respectively calculating the above two loss functions
Figure PCTCN2022127435-appb-000035
and
Figure PCTCN2022127435-appb-000036
Then calculate the total loss function
Figure PCTCN2022127435-appb-000037
Optimizing Λ based on the total loss function such that the total loss function
Figure PCTCN2022127435-appb-000038
minimize. Wherein, α is a positive constant, and in a preferred example, the value range of α is 0.01 to 100. In one embodiment, the value of α is taken as 0.1.
在一个优选例中,所述步骤(4)进一步包括以下步骤:In a preferred example, said step (4) further includes the following steps:
(a)量化区域级别特征中的知识点;(a) Quantify knowledge points in region-level features;
(b)进一步量化可靠知识点,以及可靠知识点的比例。(b) Further quantify reliable knowledge points and the proportion of reliable knowledge points.
可选地,知识点定义为使得如下式子大于某个阈值的区域级别特征的集合。Optionally, a knowledge point is defined as a set of region-level features such that the following formula is greater than a certain threshold.
Figure PCTCN2022127435-appb-000039
Figure PCTCN2022127435-appb-000039
其中,h (r)为某一样本x对应的第r个区域级别特征的低维表征。所以,知识点表示使得max c p(y=c|h (r))>τ的区域级别特征,即集合{h (r):max c p(y=c|h (r))>τ}中所包含的区域级别特征,其中,τ是一个正的常数,在一个优选例中,τ的取值范围是0.3-0.8。在一个实施例中,τ的取值为0.4。 Among them, h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x. Therefore, knowledge points represent region-level features such that max c p(y=c|h (r) )>τ, that is, the set {h (r) : max c p(y=c|h (r) )>τ} The region-level features included in , where τ is a positive constant, and in a preferred example, the value range of τ is 0.3-0.8. In one embodiment, the value of τ is 0.4.
可选地,进一步可量化知识点中的可靠知识点与不可靠知识点,从而量化可靠知识点占总知识点的比例。其中,可靠知识点为进一步满足如下式子的知识点的集合。Optionally, it is further possible to quantify reliable knowledge points and unreliable knowledge points in the knowledge points, so as to quantify the proportion of reliable knowledge points to the total knowledge points. Among them, reliable knowledge points are the set of knowledge points that further satisfy the following formula.
Figure PCTCN2022127435-appb-000040
Figure PCTCN2022127435-appb-000040
其中,h (r)为某一样本x对应的第r个区域级别特征的低维表征,c truth表示该样本的真实类别标签。即,可靠知识点为集合{h (r):c truth=argmax cp(y=c|h (r))}中所包含的区域级别特征。 Among them, h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x, and c truth represents the true category label of the sample. That is, reliable knowledge points are region-level features included in the set {h (r) :c truth =argmax c p(y=c|h (r) )}.
进一步,可靠知识点占总知识点的比例衡量了知识点的质量,可以由如下式子计算得到。Further, the ratio of reliable knowledge points to total knowledge points measures the quality of knowledge points, which can be calculated by the following formula.
Figure PCTCN2022127435-appb-000041
Figure PCTCN2022127435-appb-000041
本发明的第二方面,提供了一种对神经网络中层特征表达能力的可视化及定量分析系统,其特征在于,包括:The second aspect of the present invention provides a visualization and quantitative analysis system for the expressive ability of the middle layer of the neural network, which is characterized in that it includes:
(1)输入模块,被配置为一个预训练好的分类神经网络和包含所有可能类别的输入样本;(1) The input module is configured as a pre-trained classification neural network and contains input samples of all possible categories;
(2)特征提取模块,被配置为提取所述输入样本的样本级别特征和区域级别特征;(2) a feature extraction module configured to extract sample-level features and region-level features of the input samples;
(3)可视化模块,被配置为,基于提取得到的样本级别特征和区域级别特征,将其降维得到低维表征,并且在低维空间中将低维表征进行可视化;(3) a visualization module configured to, based on the extracted sample-level features and region-level features, reduce its dimension to obtain a low-dimensional representation, and visualize the low-dimensional representation in the low-dimensional space;
(3)定量分析模块,被配置为,基于区域级别特征的可视化结果,定量分析特征中知识点的数量与质量。(3) The quantitative analysis module is configured to quantitatively analyze the quantity and quality of knowledge points in the feature based on the visualization result of the feature at the region level.
应理解,在本发明范围内中,本发明的上述各技术特征和在下文(如实施例)中具体描述的各技术特征之间都可以互相组合,从而构成新的或优选的技术方案。限于篇幅,在此不再一一累述。It should be understood that within the scope of the present invention, the above-mentioned technical features of the present invention and the technical features specifically described in the following (such as embodiments) can be combined with each other to form new or preferred technical solutions. Due to space limitations, we will not repeat them here.
通过阅读参照以下附图对非限制性实施例所作的详细描述,本发明的其他特征、目的和优点将会变得更加明显。Other features, objects and advantages of the present invention will become more apparent by reading the detailed description of non-limiting embodiments with reference to the following drawings.
附图说明Description of drawings
图1是根据本发明第一实施方式的对神经网络中层特征表达能力的可视化及定量分析方法流程示意图;Fig. 1 is a schematic flow chart of a method for visualization and quantitative analysis of layer feature expression capabilities in a neural network according to a first embodiment of the present invention;
图2是根据本发明得到的样本级别特征在低维空间中的可视化示意图;Fig. 2 is a schematic diagram of visualization of sample-level features in low-dimensional space obtained according to the present invention;
图3是在神经网络不同训练阶段,根据本发明得到的区域级别特征在低维空间中的可视化示意图;Fig. 3 is a schematic diagram of visualization of region-level features obtained according to the present invention in low-dimensional space at different training stages of the neural network;
图4是在神经网络不同前向传播的阶段,根据本发明得到的区域级别特征在低维空间中的可视化示意图;Fig. 4 is a schematic diagram of visualization of region-level features obtained according to the present invention in low-dimensional space at different stages of forward propagation of the neural network;
图5是根据本发明中对知识点数量与质量的量化,得到神经网络不同层间特征中知识点与可靠知识点,随神经网络不同训练阶段的变化曲线图;Fig. 5 is according to the quantification to knowledge point quantity and quality in the present invention, obtains the knowledge point and reliable knowledge point in the different inter-layer features of neural network, with the change graph of different training stages of neural network;
图6是根据本发明对知识点的量化,得到的神经网络不同层间特征中知识点的可视化示意图;Fig. 6 is the quantification of knowledge points according to the present invention, and the visualized diagram of knowledge points in different inter-layer features of the neural network obtained;
图7是根据本发明第二实施方式的对神经网络中层特征表达能力的可视化及定量分析的系统结构示意图。Fig. 7 is a schematic diagram of the system structure of the visualization and quantitative analysis of the feature expression ability of the middle layer of the neural network according to the second embodiment of the present invention.
具体实施方式Detailed ways
本发明人经过细致深入的研究,首次开发了一种对神经网络中层特征表达能力的可视化及定量分析方法和系统。通过这种方法与系统,将神经网络的可视化解释与神经网络中层特征表达能力的量化分析紧密地联系在一起;通过可视化,能够清晰地展示神经网络中层特征表达能力随时空的涌现过程;这种方法能够定量地分析神经网络中层知识点的数量与质量,进而分析待解释模型的可靠程度;基于这种方法和系统,可以为现有的诸多深度学习算法,例如对抗攻击、知识蒸馏,提供一个全新角度的解释框架。After careful and in-depth research, the present inventor has first developed a method and system for visualizing and quantitatively analyzing the expressive ability of middle-level features of neural networks. Through this method and system, the visual interpretation of the neural network is closely linked with the quantitative analysis of the feature expression ability of the middle layer of the neural network; through visualization, the emergence process of the feature expression ability of the middle layer of the neural network can be clearly displayed over time and space; this kind of The method can quantitatively analyze the quantity and quality of the knowledge points in the middle layer of the neural network, and then analyze the reliability of the model to be explained; based on this method and system, it can be used for many existing deep learning algorithms, such as adversarial attacks and knowledge distillation. An interpretive framework from a new perspective.
基于这种方法和系统,可以为现有的诸多深度学习算法,例如对抗攻击、知识蒸馏,提供一个全新角度的解释框架。Based on this method and system, it can provide a new interpretation framework for many existing deep learning algorithms, such as adversarial attacks and knowledge distillation.
通用方法general method
典型地,本发明包括以下步骤:Typically, the present invention includes the following steps:
(1)提供一待分析的神经网络,该神经网络为在某一数据集上预训练好的深度神经网络;(1) Provide a neural network to be analyzed, which is a pre-trained deep neural network on a certain data set;
(2)提供一组输入样本,将这些样本输入上述神经网络,提取这些样本对应的样本级别特征(sample-wise feature),以及区域级别特征(regional feature);(2) Provide a set of input samples, input these samples into the above neural network, and extract the sample-wise features and regional features corresponding to these samples;
(3)分别对样本级别特征与区域级别特征进行降维,以得到在低维空间中的可视化结果;(3) Dimensionality reduction is performed on sample-level features and region-level features to obtain visualization results in low-dimensional space;
(4)基于区域级别特征的可视化结果,定量分析特征中知识点(knowledge point)的数量与质量。(4) Quantitatively analyze the quantity and quality of knowledge points in features based on the visualization results of region-level features.
本发明的主要优点在于:The main advantages of the present invention are:
(1)一种对神经网络中层特征表达能力的可视化及定量分析方法和系统;(1) A method and system for visualization and quantitative analysis of the feature expression ability of the middle layer of the neural network;
(2)通过这种方法与系统,将神经网络的可视化解释与神经网络中层特征表达能力的量化分析紧密地联系在一起;(2) Through this method and system, the visual interpretation of the neural network is closely linked with the quantitative analysis of the expressive ability of the layer features in the neural network;
(3)通过可视化,能够清晰地展示神经网络中层特征表达能力随时空的涌现过程;(3) Through visualization, it is possible to clearly show the emergence process of the feature expression ability of the middle layer of the neural network over time and space;
(4)这种方法能够定量地分析神经网络中层知识点的数量与质量,进而分析待解释模型的可靠程度;(4) This method can quantitatively analyze the quantity and quality of knowledge points in the middle layer of the neural network, and then analyze the reliability of the model to be explained;
(5)基于这种方法和系统,可以为现有的诸多深度学习算法,例如对抗攻击、知识蒸馏,提供一个全新角度的解释框架。(5) Based on this method and system, it can provide a new interpretation framework for many existing deep learning algorithms, such as adversarial attacks and knowledge distillation.
实施例Example
为使本申请的目的、技术方案和优点更加清楚,下面将结合附图对本申请的实施方式作进一步地详细描述。In order to make the purpose, technical solution and advantages of the present application clearer, the implementation manner of the present application will be further described in detail below in conjunction with the accompanying drawings.
本发明的第一实施方式设计一种对神经网络中层特征表达能力的可视化及定量分析方法,其流程如图1所示,该方法包括以下步骤:The first embodiment of the present invention designs a kind of visualization and quantitative analysis method to the feature expression ability of neural network middle layer, and its flow process is as shown in Figure 1, and this method comprises the following steps:
在步骤101中:基于某一数据集,训练一神经网络,作为待分析的神经网络。可选地,该神经网络为一个分类神经网络。In step 101: based on a certain data set, train a neural network as the neural network to be analyzed. Optionally, the neural network is a classification neural network.
之后,进入步骤102,该步骤可以进一步分为如下两个子步骤:Afterwards, enter step 102, this step can be further divided into following two sub-steps:
(a)对样本级别特征的提取:将给定的一组样本输入待分析的神经网络,对于每个样本,都提取该神经网络某一中间层的输出特征,从而得到每个输入样本对应的样本级别特征,即可得到该组输入样本对应的样本级别特征。(a) Extraction of sample-level features: Input a given set of samples into the neural network to be analyzed, and for each sample, extract the output features of an intermediate layer of the neural network, so as to obtain the corresponding Sample-level features, the sample-level features corresponding to the set of input samples can be obtained.
(b)对区域级别特征的提取:将给定的一组输入样本输入待分析的神 经网络,对于每个样本,都提取该神经网络某一卷积层的输出特征,从而得到每个输入样本对应的特征图(feature map),其中,特征图的每一个位置对应的高维向量即为该样本在这一区域的区域级别特征。当这一特征图的高与宽分别为H和W,且共有K个通道时,那么,这一特征图包含HW个区域级别特征,其中,每个区域级别特征为一个K维向量。(b) Extraction of region-level features: Input a given set of input samples into the neural network to be analyzed, and for each sample, extract the output features of a certain convolutional layer of the neural network to obtain each input sample The corresponding feature map (feature map), where the high-dimensional vector corresponding to each position of the feature map is the region-level feature of the sample in this region. When the height and width of this feature map are H and W respectively, and there are K channels in total, then this feature map contains HW region-level features, where each region-level feature is a K-dimensional vector.
之后,进入步骤103,该步骤可以进一步分为如下两个子步骤:Afterwards, enter step 103, this step can be further divided into following two sub-steps:
(a)对样本级别特征进行降维,并在低维空间中进行可视化;(a) Dimensionality reduction of sample-level features and visualization in low-dimensional space;
(b)对区域级别特征进行降维,并在低维空间中进行可视化。(b) Dimensionality reduction is performed on region-level features and visualized in a low-dimensional space.
上述子步骤(a)中,对于每个样本x对应的样本级别特征
Figure PCTCN2022127435-appb-000042
通过一个投影矩阵
Figure PCTCN2022127435-appb-000043
将其映射到一个低维空间中,得到样本级别特征的低维表征
Figure PCTCN2022127435-appb-000044
并且,优化M使得该低维表征应当满足,低维表征g和各个类别的接近程度应当与该样本x和各个类别的接近程度尽可能地保持一致。
In the above sub-step (a), for each sample x corresponding to the sample level features
Figure PCTCN2022127435-appb-000042
through a projection matrix
Figure PCTCN2022127435-appb-000043
Map it into a low-dimensional space to obtain a low-dimensional representation of sample-level features
Figure PCTCN2022127435-appb-000044
And, optimize M so that the low-dimensional representation should satisfy that the closeness of the low-dimensional representation g to each category should be consistent with the closeness of the sample x to each category as much as possible.
可选地,在“低维表征和各个类别的接近程度”的计算中,先利用径向分布(radial distribution)建模样本级别特征的低维表征g在低维空间中的分布,再基于此计算低维表征与各个类别的接近程度。Optionally, in the calculation of the "closeness between low-dimensional representation and each category", first use the radial distribution (radial distribution) to model the distribution of the low-dimensional representation g of the sample-level feature in the low-dimensional space, and then based on this Computes how close a low-dimensional representation is to each class.
基于径向分布,g在低维空间的概率密度函数可由如下公式计算得到。Based on the radial distribution, the probability density function of g in low-dimensional space can be calculated by the following formula.
Figure PCTCN2022127435-appb-000045
Figure PCTCN2022127435-appb-000045
其中,y∈{1,2,…,C}表示分类任务中的不同类别;π y表示第y类的先验概率;l g=‖g‖表示g的L2范数,称为g的强度(strength);o g=g/l g表示g的方向(orientation);μ y表示第y类的平均方向(mean direction);κ(·)是一个单调递增的函数。p(l g|y)表示在类别y上l g的先验概率,p vMF(o gy,κ(l g))表示平均方向为μ y,聚集参数(concentration parameter) 为κ(l g)的vMF分布(von Mises-Fisher distribution)。 Among them, y∈{1,2,…,C} represents different categories in the classification task; π y represents the prior probability of the yth class; l g =‖g‖ represents the L2 norm of g, which is called the strength of g (strength); o g = g/l g represents the orientation of g; μ y represents the mean direction of the yth class; κ(·) is a monotonically increasing function. p(l g |y) indicates the prior probability of l g on category y, p vMF (o gy ,κ(l g )) indicates that the average direction is μ y , and the concentration parameter is κ( l g ) vMF distribution (von Mises-Fisher distribution).
可选地,κ(·)可以是任意一个单调递增的函数,更佳地,κ(·)函数可以通过如下方法生成:给定一个非负的常数κ m,以及维数d,从平均方向为
Figure PCTCN2022127435-appb-000046
聚集参数为κ=κ m的vMF分布中采样得到N个样本
Figure PCTCN2022127435-appb-000047
将这些样本都缩放到长度为l,而不改变方向,此处的l是一个任意非负数,将缩放得到的样本记作
Figure PCTCN2022127435-appb-000048
从标准正态分布中采样得到N个高斯噪声样本
Figure PCTCN2022127435-appb-000049
将前述缩放得到的样本与高斯噪声样本对应相加,得到
Figure PCTCN2022127435-appb-000050
将κ(l)定义为
Figure PCTCN2022127435-appb-000051
其中
Figure PCTCN2022127435-appb-000052
N的范围为1000-100000,在一个实施例中,N取为10000。
Optionally, κ(·) can be any monotonically increasing function, more preferably, the κ(·) function can be generated by the following method: Given a non-negative constant κ m and dimension d, from the average direction for
Figure PCTCN2022127435-appb-000046
N samples are obtained by sampling from the vMF distribution with aggregation parameter κ=κ m
Figure PCTCN2022127435-appb-000047
Scale these samples to a length of l without changing the direction, where l is an arbitrary non-negative number, and record the scaled samples as
Figure PCTCN2022127435-appb-000048
Sample N samples of Gaussian noise from a standard normal distribution
Figure PCTCN2022127435-appb-000049
Add the sample obtained by the aforementioned scaling to the Gaussian noise sample correspondingly, and obtain
Figure PCTCN2022127435-appb-000050
Define κ(l) as
Figure PCTCN2022127435-appb-000051
in
Figure PCTCN2022127435-appb-000052
The range of N is 1000-100000, and in one embodiment, N is taken as 10000.
基于上述径向分布,并假设l g的先验概率与类别y是独立的,那么,低维表征g和第y类的接近程度Q M(y|x)可以由如下公式计算得到。 Based on the radial distribution above, and assuming that the prior probability of l g is independent of category y, then the closeness Q M (y|x) between the low-dimensional representation g and the yth category can be calculated by the following formula.
Figure PCTCN2022127435-appb-000053
Figure PCTCN2022127435-appb-000053
可选地,在“样本和各个类别的接近程度”计算为样本的输出概率,即样本x和第y类的接近程度P(y|x)为神经网络网络输出中对应第y类的输出概率值。Optionally, the "closeness between the sample and each category" is calculated as the output probability of the sample, that is, the closeness P(y|x) between the sample x and the yth category is the output probability corresponding to the yth category in the neural network output value.
可选地,对于投影矩阵M的优化,进一步包括,低维表征g和各个类别的接近程度Q M(y|x),以及样本x和各个类别的接近程度P(y|x)的计算结果,优化投影矩阵M,使得P(y|x)与Q M(y|x)间的KL散度(Kullback–Leibler divergence)最小化。 Optionally, the optimization of the projection matrix M further includes the calculation results of the closeness Q M (y|x) between the low-dimensional representation g and each category, and the closeness P(y|x) between the sample x and each category , optimize the projection matrix M so that the KL divergence (Kullback–Leibler divergence) between P(y|x) and Q M (y|x) is minimized.
Figure PCTCN2022127435-appb-000054
Figure PCTCN2022127435-appb-000054
可选地,在得到样本级别特征的低维表征的过程中,交替优化投影矩阵M与径向分布中的参数{π,μ}={π yy} y∈Y。其中,优化投影矩阵M时,固定 径向分布中的参数{π,μ},并且更新M,使得KL[P(Y|X)‖Q M(Y|X)]的值最小化;优化径向分布中的参数{π,μ}时,固定投影矩阵M,并且更新{π,μ},使得似然∏ gp(g)=∏ gy′π y′·p vMF(o gy′,κ(l g))的值最大。 Optionally, in the process of obtaining the low-dimensional representation of sample-level features, alternately optimize the projection matrix M and the parameters {π,μ}={π yy } y∈Y in the radial distribution. Among them, when optimizing the projection matrix M, the parameters {π, μ} in the radial distribution are fixed, and M is updated to minimize the value of KL[P(Y|X)‖Q M (Y|X)]; the optimized path When the parameters in the distribution are {π, μ}, the projection matrix M is fixed, and {π, μ} is updated, so that the likelihood ∏ g p(g)=∏ gy′ π y′ ·p vMF (o g | μ y′ ,κ(l g )) has the largest value.
如图2所示,在本发明的一个实施例中,给定一个在Tiny ImageNet图像分类数据集上预训练好的VGG-16网络,其中,在训练时神经网络时只采用了其中的steel arch bridge,school bus,sports car,tabby cat,desk,golden retriever,tailed frog,iPod,lifeboat,orange这十个类别,抽取VGG-16网络倒数第二个全连接层后的特征作为样本级别特征,用前述方法进行降维至三维空间中得到的低维表征散点图。图中,不同色彩表示图例中所示的不同类别,各个类别色彩对应的箭头表示各个类别的平均方向。As shown in Figure 2, in one embodiment of the present invention, a VGG-16 network pre-trained on the Tiny ImageNet image classification data set is given, wherein only the steel arch is used in the neural network during training The ten categories of bridge, school bus, sports car, tabby cat, desk, golden retriever, tailed frog, iPod, lifeboat, orange are extracted from the second-to-last fully connected layer of the VGG-16 network as sample-level features, using The aforementioned method performs dimensionality reduction to a low-dimensional representation scatter diagram obtained in a three-dimensional space. In the figure, different colors represent different categories shown in the legend, and the arrows corresponding to the colors of each category indicate the average direction of each category.
上述子步骤(b)中,对于每个样本x的HW个区域级别特征
Figure PCTCN2022127435-appb-000055
Figure PCTCN2022127435-appb-000056
通过一个投影矩阵
Figure PCTCN2022127435-appb-000057
将它们映射到一个低维空间中,得到HW个区域级别特征的低维表征
Figure PCTCN2022127435-appb-000058
Figure PCTCN2022127435-appb-000059
并且,优化Λ使得该低维表征应当满足,基于低维表征h={h (1),h (2),…,h (HW)}所推断出的样本间相似度与基于网络输出所推断出的样本间相似度尽可能保持一致,进一步,区域级别特征的低维表征需要与样本级别特征的低维特征表征对齐。
In the above sub-step (b), for each sample x's HW region-level features
Figure PCTCN2022127435-appb-000055
Figure PCTCN2022127435-appb-000056
through a projection matrix
Figure PCTCN2022127435-appb-000057
Map them into a low-dimensional space to obtain a low-dimensional representation of HW region-level features
Figure PCTCN2022127435-appb-000058
Figure PCTCN2022127435-appb-000059
And, optimize Λ so that the low-dimensional representation should satisfy that the similarity between samples inferred based on the low-dimensional representation h={h (1) ,h (2) ,…,h (HW) } is the same as that inferred based on the network output The similarity between samples obtained should be as consistent as possible. Furthermore, the low-dimensional representation of region-level features needs to be aligned with the low-dimensional feature representation of sample-level features.
可选地,在“基于低维表征所推断出的样本间相似度”的计算中,基于词袋模型(the bag-of-words model),将样本间相似度拆分为各个区域对应低维表征间相似度的加权乘积;并基于vMF分布,量化各个区域对应低维表征间的相似度。Optionally, in the calculation of "similarity between samples inferred based on low-dimensional representation", based on the bag-of-words model (the bag-of-words model), the similarity between samples is split into regions corresponding to low-dimensional The weighted product of the similarity between representations; and based on the vMF distribution, the similarity between the corresponding low-dimensional representations of each region is quantified.
可选地,在以上描述中,设x 1和x 2为任意两个样本,所对应区域级别特征的低维表征分别为
Figure PCTCN2022127435-appb-000060
Figure PCTCN2022127435-appb-000061
基于词袋模型,将x 1和x 2之间的相似度Q Λ(x 2|x 1)拆分为各个区域对应低维表征间相似度的加权乘积,如下所示。
Optionally, in the above description, let x1 and x2 be any two samples, and the low-dimensional representations of the corresponding region-level features are respectively
Figure PCTCN2022127435-appb-000060
and
Figure PCTCN2022127435-appb-000061
Based on the bag-of-words model, the similarity Q Λ (x 2 |x 1 ) between x 1 and x 2 is split into the weighted product of the similarity between the corresponding low-dimensional representations of each region, as shown below.
Figure PCTCN2022127435-appb-000062
Figure PCTCN2022127435-appb-000062
其中,
Figure PCTCN2022127435-appb-000063
表示样本x 2中第r个区域特征对于分类的重要程度,在一个优选例中,
Figure PCTCN2022127435-appb-000064
是一个非负数。并且,
Figure PCTCN2022127435-appb-000065
进一步量化为如下形式。
in,
Figure PCTCN2022127435-appb-000063
Indicates the importance of the rth regional feature in the sample x 2 for classification. In a preferred example,
Figure PCTCN2022127435-appb-000064
is a non-negative number. and,
Figure PCTCN2022127435-appb-000065
It is further quantified into the following form.
Figure PCTCN2022127435-appb-000066
Figure PCTCN2022127435-appb-000066
其中,
Figure PCTCN2022127435-appb-000067
Figure PCTCN2022127435-appb-000068
在平均方向为
Figure PCTCN2022127435-appb-000069
聚集参数为
Figure PCTCN2022127435-appb-000070
的vMF分布中的概率密度;同权利要求6中所述,κ(·)是一个单调递增的函数。
in,
Figure PCTCN2022127435-appb-000067
for
Figure PCTCN2022127435-appb-000068
in the mean direction of
Figure PCTCN2022127435-appb-000069
The aggregation parameters are
Figure PCTCN2022127435-appb-000070
The probability density in the distribution of vMF; as described in claim 6, κ(·) is a monotonically increasing function.
可选地,在“基于网络输出所推断出的样本间相似度”的计算中,设x 1和x 2为任意两个样本,
Figure PCTCN2022127435-appb-000071
分别为这两个样本对应的网络输出概率,基于网络输出所推断出的样本间相似度P(x 2|x 1)可进一步计算为如下形式。
Optionally, in the calculation of "similarity between samples inferred based on network output", let x1 and x2 be any two samples,
Figure PCTCN2022127435-appb-000071
are the network output probabilities corresponding to the two samples, and the similarity P(x 2 |x 1 ) between samples inferred based on the network output can be further calculated as the following form.
Figure PCTCN2022127435-appb-000072
Figure PCTCN2022127435-appb-000072
其中,
Figure PCTCN2022127435-appb-000073
cos(·,·)表示两个向量之间的余弦相似度,κ p为一个非负常数。
in,
Figure PCTCN2022127435-appb-000073
cos(·,·) represents the cosine similarity between two vectors, and κ p is a non-negative constant.
可选地,使得“基于低维表征所推断出的样本间相似度与基于网络输出所推断出的样本间相似度尽可能保持一致”,等价于最小化如下损失函数。Optionally, making "the similarity between samples inferred based on the low-dimensional representation and the similarity between samples inferred based on the network output as consistent as possible" is equivalent to minimizing the following loss function.
Figure PCTCN2022127435-appb-000074
Figure PCTCN2022127435-appb-000074
其中,P(x 2|x 1)和Q Λ(x 2|x 1)的计算如权利要求16和权利要求17分别所示。 Wherein, the calculations of P(x 2 |x 1 ) and Q Λ (x 2 |x 1 ) are as shown in claim 16 and claim 17 respectively.
可选地,将区域级别特征的低维表征与样本级别特征的低维表征对齐,等价于优化如下损失函数。Optionally, aligning the low-dimensional representation of region-level features with the low-dimensional representation of sample-level features is equivalent to optimizing the following loss function.
Figure PCTCN2022127435-appb-000075
Figure PCTCN2022127435-appb-000075
Figure PCTCN2022127435-appb-000076
Figure PCTCN2022127435-appb-000076
其中,MI(·;·)表示互信息(mutual information);g表示样本x的样本级别特征的低维表征;h (r)表示样本x的第r个区域级别特征的低维表征,w (r)表示样本x的第r个区域级别特征对于分类的重要程度。 Among them, MI(·;·) represents mutual information; g represents the low-dimensional representation of the sample-level features of sample x; h (r) represents the low-dimensional representation of the r-th region-level features of sample x, w ( r) indicates the importance of the r-th region-level feature of sample x for classification.
可选地,对投影矩阵Λ的优化包括以下步骤:分别计算上述两个损失函数
Figure PCTCN2022127435-appb-000077
Figure PCTCN2022127435-appb-000078
进而计算总损失函数
Figure PCTCN2022127435-appb-000079
基于总损失函数优化Λ,使得总损失函数
Figure PCTCN2022127435-appb-000080
最小化。其中,α是一个正的常数,在一个优选例中,α的取值范围是0.01至100。在一个实施例中,α的值取为0.1。
Optionally, the optimization of the projection matrix Λ includes the following steps: respectively calculating the above two loss functions
Figure PCTCN2022127435-appb-000077
and
Figure PCTCN2022127435-appb-000078
Then calculate the total loss function
Figure PCTCN2022127435-appb-000079
Optimizing Λ based on the total loss function such that the total loss function
Figure PCTCN2022127435-appb-000080
minimize. Wherein, α is a positive constant, and in a preferred example, the value range of α is 0.01 to 100. In one embodiment, the value of α is taken as 0.1.
如图3所示,给定一个与前述相同的在Tiny ImageNet数据集上预训练的VGG-16网络,抽取conv_53层的输出特征作为区域级别特征,用前述方法进行降维到三维得到低维表征,在三维空间中的分布。其中,和前述相同,不同色彩的散点表示不同类别的样本对应的区域级别特征的低维表征,图中的椭球表示不同类别样本所对应区域级别特征的低维表征的大致分布。As shown in Figure 3, given the same VGG-16 network pre-trained on the Tiny ImageNet dataset as mentioned above, the output features of the conv_53 layer are extracted as region-level features, and the dimensionality reduction to three dimensions is performed using the aforementioned method to obtain a low-dimensional representation. , the distribution in three-dimensional space. Among them, the same as above, the scatter points of different colors represent the low-dimensional representations of the region-level features corresponding to different types of samples, and the ellipsoids in the figure represent the approximate distribution of the low-dimensional representations of the region-level features corresponding to different types of samples.
如图4所示,给定一个与前述相同的在Tiny ImageNet数据集上预训练的VGG-16网络,抽取conv_12,conv_22,conv_33,conv_43,conv_53层的输出特征作为区域级别特征,分别运用前述方法进行降维到三维空间得到低维表征。图4显示了不同样本对应的区域级别特征的低维表征随前向传播的层数增加的分布变化图,其中,竖直向上的箭头代表该样本的正确平均方向,不同色彩的散点表示不同层所对应区域级别特征的低维表征在三维空间中的分布。As shown in Figure 4, given a VGG-16 network pre-trained on the Tiny ImageNet dataset as mentioned above, the output features of conv_12, conv_22, conv_33, conv_43, conv_53 layers are extracted as region-level features, and the aforementioned methods are used respectively Perform dimensionality reduction to three-dimensional space to obtain low-dimensional representation. Figure 4 shows the distribution of low-dimensional representations of regional-level features corresponding to different samples as the number of layers of forward propagation increases, where the vertical upward arrow represents the correct average direction of the sample, and the scatter points of different colors represent different The distribution of low-dimensional representations of region-level features corresponding to layers in three-dimensional space.
之后,进入步骤104,该步骤可以进一步分为如下子步骤:Afterwards, enter step 104, this step can be further divided into following sub-steps:
(a)量化区域级别特征中的知识点;(a) Quantify knowledge points in region-level features;
(b)进一步量化可靠知识点,以及可靠知识点的比例。(b) Further quantify reliable knowledge points and the proportion of reliable knowledge points.
可选地,知识点定义为使得如下式子大于某个阈值的区域级别特征的集 合。Optionally, a knowledge point is defined as a set of region-level features such that the following formula is greater than a certain threshold.
Figure PCTCN2022127435-appb-000081
Figure PCTCN2022127435-appb-000081
其中,h (r)为某一样本x对应的第r个区域级别特征的低维表征。所以,知识点表示使得max c p(y=c|h (r))>τ的区域级别特征,即集合{h (r):max c p(y=c|h (r))>τ}中所包含的区域级别特征,其中,τ是一个正的常数,在一个优选例中,τ的取值范围是0.3-0.8。在一个实施例中,τ的取值为0.4。 Among them, h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x. Therefore, knowledge points represent region-level features such that max c p(y=c|h (r) )>τ, that is, the set {h (r) : max c p(y=c|h (r) )>τ} The region-level features included in , where τ is a positive constant, and in a preferred example, the value range of τ is 0.3-0.8. In one embodiment, the value of τ is 0.4.
可选地,进一步可量化知识点中的可靠知识点与不可靠知识点,从而量化可靠知识点占总知识点的比例。其中,可靠知识点为进一步满足如下式子的知识点的集合。Optionally, it is further possible to quantify reliable knowledge points and unreliable knowledge points in the knowledge points, so as to quantify the proportion of reliable knowledge points to the total knowledge points. Among them, reliable knowledge points are the set of knowledge points that further satisfy the following formula.
Figure PCTCN2022127435-appb-000082
Figure PCTCN2022127435-appb-000082
其中,h (r)为某一样本x对应的第r个区域级别特征的低维表征,c truth表示该样本的真实类别标签。即,可靠知识点为集合{h (r):c truth=argmax cp(y=c|h (r))}中所包含的区域级别特征。 Among them, h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x, and c truth represents the true category label of the sample. That is, reliable knowledge points are region-level features included in the set {h (r) :c truth =argmax c p(y=c|h (r) )}.
进一步,可靠知识点占总知识点的比例衡量了知识点的质量,可以由如下式子计算得到。Further, the ratio of reliable knowledge points to total knowledge points measures the quality of knowledge points, which can be calculated by the following formula.
Figure PCTCN2022127435-appb-000083
Figure PCTCN2022127435-appb-000083
如图5所示,给定一个与前述相同的在Tiny ImageNet数据集上预训练的VGG-16网络,用前述方法分别计算conv_33,conv_43,conv_53层对应区域级别特征中所有知识点的个数,以及可靠知识点的个数。图5显示了神经网络不同层的知识点总量与可靠知识点数量随神经网络训练迭代次数的变化曲线。As shown in Figure 5, given a VGG-16 network pre-trained on the Tiny ImageNet data set as mentioned above, the number of all knowledge points in the corresponding region-level features of the conv_33, conv_43, and conv_53 layers are calculated using the aforementioned method, And the number of reliable knowledge points. Figure 5 shows the change curve of the total amount of knowledge points and the number of reliable knowledge points at different layers of the neural network with the number of neural network training iterations.
如图6所示,给定一个与前述相同的在Tiny ImageNet数据集上预训练的VGG-16网络,用前述方法分别得到conv_33,conv_43,conv_53层对应区 域级别特征中所有知识点。图6显示了不同层所对应的知识点在图中的区域,如图中高亮部分所示。As shown in Figure 6, given the same VGG-16 network pre-trained on the Tiny ImageNet dataset as mentioned above, all knowledge points in the corresponding region-level features of the conv_33, conv_43, and conv_53 layers are obtained using the aforementioned method. Figure 6 shows the areas of knowledge points corresponding to different layers in the figure, as shown in the highlighted part of the figure.
本发明的第二实施方式设计一种对神经网络中层特征表达能力的可视化及定量分析系统,其结构如图7所示,该系统包括:The second embodiment of the present invention designs a kind of visualization and quantitative analysis system to the feature expression ability of neural network, its structure is shown in Figure 7, and this system comprises:
(1)输入模块,被配置为一个预训练好的分类神经网络和包含所有可能类别的输入样本;(1) The input module is configured as a pre-trained classification neural network and contains input samples of all possible categories;
(2)可视化模块,被配置为,基于提取得到的样本级别特征和区域级别特征,将其降维得到低维表征,并且在低维空间中将低维表征进行可视化;(2) a visualization module configured to, based on the extracted sample-level features and region-level features, reduce its dimension to obtain a low-dimensional representation, and visualize the low-dimensional representation in the low-dimensional space;
(3)定量分析模块,被配置为,基于区域级别特征的可视化结果,定量分析特征中知识点的数量与质量。(3) The quantitative analysis module is configured to quantitatively analyze the quantity and quality of knowledge points in the feature based on the visualization result of the feature at the region level.
需要说明的是,本领域技术人员应当理解,上述对神经网络中层特征表达能力的可视化及定量分析系统的实施方式中所示的各模块的实现功能,可参照前述对神经网络中层特征表达能力的可视化及定量分析方法的相关描述而理解。上述对神经网络中层特征表达能力的可视化及定量分析系统的实施方式中所示的各模块的功能可通过运行于处理器上的程序(可执行指令)而实现,也可通过具体的逻辑电路而实现。本发明实施例上述对神经网络中层特征表达能力的可视化及定量分析系统如果以软件功能模块的形式实现并作为独立的产品销售或使用时,也可以存储在一个计算机可读取存储介质中。基于这样的理解,本发明实施例的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机、服务器、或者网络设备等)执行本发明各个实施例该方法的全部或部分。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read Only Memory)、磁碟或者光盘等各种可以存储程序代码的介质。这样,本发明实施例不限制 于任何特定的硬件和软件结合。It should be noted that those skilled in the art should understand that the functions of each module shown in the implementation of the above-mentioned visualization and quantitative analysis system for the middle-level feature expression ability of the neural network can refer to the aforementioned description of the middle-level feature expression ability of the neural network. Visualization and related descriptions of quantitative analysis methods can be understood. The functions of the modules shown in the implementation of the above-mentioned visualization and quantitative analysis system for the middle-level feature expression ability of the neural network can be realized by a program (executable instruction) running on the processor, or can be realized by a specific logic circuit. accomplish. In the embodiment of the present invention, if the above-mentioned visualization and quantitative analysis system for the expression ability of neural network middle layer is implemented in the form of software function modules and sold or used as an independent product, it can also be stored in a computer-readable storage medium. Based on this understanding, the technical solutions of the embodiments of the present invention can be embodied in the form of software products in essence or the part that contributes to the prior art. The computer software products are stored in a storage medium and include several instructions for Make a computer device (which may be a personal computer, a server, or a network device, etc.) execute all or part of the method in each embodiment of the present invention. The aforementioned storage medium includes: various media that can store program codes such as U disk, mobile hard disk, read-only memory (ROM, Read Only Memory), magnetic disk or optical disk. Thus, embodiments of the invention are not limited to any specific combination of hardware and software.
相应地,本发明实施方式还提供一种计算机可读存储介质,其中存储有计算机可执行指令,该计算机可执行指令被处理器执行时实现本发明的各方法实施方式。计算机可读存储介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括但不限于,相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。按照本文中的界定,计算机可读存储介质不包括暂存电脑可读媒体(transitory media),如调制的数据信号和载波。Correspondingly, the embodiments of the present invention also provide a computer-readable storage medium, in which computer-executable instructions are stored, and when the computer-executable instructions are executed by a processor, various method embodiments of the present invention are implemented. Computer-readable storage media includes both volatile and non-permanent, removable and non-removable media by any method or technology for storage of information. Information may be computer readable instructions, data structures, modules of a program, or other data. Examples of storage media for computers include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read only memory (ROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Flash memory or other memory technology, Compact Disc Read-Only Memory (CD-ROM), Digital Versatile Disc (DVD) or other optical storage, Magnetic tape cartridge, tape magnetic disk storage or other magnetic storage device or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, computer-readable storage media does not include transitory computer-readable media, such as modulated data signals and carrier waves.
此外,本发明实施方式还提供一种对神经网络中层特征表达能力的可视化及定量分析系统,其中包括用于存储计算机可执行指令的存储器,以及,处理器;该处理器用于在执行该存储器中的计算机可执行指令时实现上述各方法实施方式中的步骤。其中,该处理器可以是中央处理单元(Central Processing Unit,简称“CPU”),还可以是其他通用处理器、数字信号处理器(Digital Signal Processor,简称“DSP”)、专用集成电路(Application Specific Integrated Circuit,简称“ASIC”)等。前述的存储器可以是只读存储器(read-only memory,简称“ROM”)、随机存取存储器(random access memory,简称“RAM”)、快闪存储器(Flash)、硬盘或者固态硬盘等。本发明各实施方式所公开的方法的步骤可以直接体现为硬件处理器执行完成,或者用处理器中的硬件及软件模块组合执行完成。In addition, the embodiment of the present invention also provides a visualization and quantitative analysis system for the expression ability of the layer features in the neural network, which includes a memory for storing computer-executable instructions, and a processor; the processor is used to execute the memory in the memory The steps in the above-mentioned method implementations are implemented when computer-executable instructions are used. Among them, the processor can be a central processing unit (Central Processing Unit, referred to as "CPU"), and can also be other general-purpose processors, digital signal processors (Digital Signal Processor, referred to as "DSP"), application specific integrated circuits (Application Specific Integrated Circuit, referred to as "ASIC") and so on. The aforementioned memory may be a read-only memory ("ROM" for short), a random access memory (random access memory, "RAM" for short), a flash memory (Flash), a hard disk or a solid-state hard disk, and the like. The steps of the methods disclosed in the various embodiments of the present invention can be directly implemented by a hardware processor, or implemented by a combination of hardware and software modules in the processor.
需要说明的是,在本专利的发明文件中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定 要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个”限定的要素,并不排除在包括该要素的过程、方法、物品或者设备中还存在另外的相同要素。本专利的发明文件中,如果提到根据某要素执行某行为,则是指至少根据该要素执行该行为的意思,其中包括了两种情况:仅根据该要素执行该行为、和根据该要素和其它要素执行该行为。多个、多次、多种等表达包括2个、2次、2种以及2个以上、2次以上、2种以上。It should be noted that in the invention documents of this patent, relative terms such as first and second are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply these No such actual relationship or order exists between entities or operations. Furthermore, the term "comprises", "comprises" or any other variation thereof is intended to cover a non-exclusive inclusion such that a process, method, article, or apparatus comprising a set of elements includes not only those elements, but also includes elements not expressly listed. other elements of or also include elements inherent in such a process, method, article, or device. Without more limitations, an element defined by the statement "comprising a" does not exclude the presence of additional same elements in the process, method, article or device comprising the element. In the invention documents of this patent, if it is mentioned that an action is performed according to a certain element, it means that the action is performed based on at least the element, which includes two situations: performing the action only based on the element, and performing the action based on the element and Other elements perform the behavior. Expressions such as multiple, multiple, and multiple include 2, 2 times, 2 types, and 2 or more, 2 or more times, or 2 or more types.
在本发明提及的所有文献都被认为是整体性地包括在本发明的公开内容中,以便在必要时可以作为修改的依据。此外应理解,以上该仅为本说明书的较佳实施例而已,并非用于限定本说明书的保护范围。凡在本说明书一个或多个实施例的精神和原则之内,所作的任何修改、等同替换、改进等,均应包含在本说明书一个或多个实施例的保护范围之内。All documents mentioned in the present invention are considered to be included in the disclosure content of the present invention in their entirety so as to serve as a basis for amendments when necessary. In addition, it should be understood that the above descriptions are only preferred embodiments of the present specification, and are not intended to limit the protection scope of the present specification. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of one or more embodiments of this specification shall be included in the protection scope of one or more embodiments of this specification.

Claims (10)

  1. 一种对神经网络中层特征表达能力的可视化及定量分析方法,其特征在于,包括以下步骤A method for visualizing and quantitatively analyzing neural network middle layer feature expression capabilities, characterized in that it includes the following steps
    (1)选取特征解释对象:(1) Select the feature interpretation object:
    选取待分析的模型,其中,所述模型包括有中层表达,包括:神经网络,层次化图模型;Select the model to be analyzed, wherein the model includes a middle-level expression, including: a neural network, a hierarchical graph model;
    (2)提取神经网络特征:(2) Extract neural network features:
    提供一组输入样本,将这些样本输入上述神经网络,提取这些样本的特征,其中,所述特征包括:样本级别特征和区域级别特征;Provide a set of input samples, input these samples into the above neural network, and extract the features of these samples, wherein the features include: sample-level features and region-level features;
    (3)特征降维,得到可视化结果:(3) Feature dimensionality reduction to obtain visualization results:
    首先对样本级别特征进行降维,得到样本级别特征在低维空间中的可视化结果;其次,基于样本级别特征的低维表征和区域级别特征,将区域级别特征进行降维,以得到区域级别特征在低维空间中的可视化结果;First, the dimensionality reduction of the sample-level features is performed to obtain the visualization results of the sample-level features in the low-dimensional space; secondly, based on the low-dimensional representation of the sample-level features and the regional-level features, the dimensionality reduction of the regional-level features is performed to obtain the regional-level features Visualization results in low-dimensional space;
    (4)根据可视化结果对特征进行定量分析:(4) Quantitative analysis of features based on visualization results:
    基于所述可视化结果,定量分析特征中知识点的数量与质量。Based on the visualization results, the quantity and quality of knowledge points in the features are quantitatively analyzed.
  2. 根据权利要求1所述的方法,其特征在于,所述步骤(2)中对样本级别特征的提取,进一步包括以下步骤:The method according to claim 1, wherein the extraction of sample level features in the step (2) further comprises the following steps:
    将给定的一组样本输入待分析的神经网络,对于每个样本,都提取该神经网络某一中间层的输出特征,从而得到每个输入样本对应的样本级别特征,即可得到该组输入样本对应的样本级别特征。Input a given set of samples into the neural network to be analyzed, and for each sample, extract the output features of an intermediate layer of the neural network to obtain the sample-level features corresponding to each input sample, and then the set of input The sample-level features that the sample corresponds to.
  3. 根据权利要求1中所述的方法,其特征在于,所述步骤(2)中对区域级别特征的提取,进一步包括以下步骤:According to the method described in claim 1, it is characterized in that, in described step (2), to the extraction of regional level feature, further comprise the following steps:
    将给定的一组输入样本输入待分析的神经网络,对于每个样本,都提取该神经网络某一卷积层的输出特征,从而得到每个输入样本对应的特征图,其中,特征图的每一个位置对应的高维向量即为该样本在这一区域的区域级 别特征;当这一特征图的高与宽分别为H和W,且共有K个通道时,那么,这一特征图包含HW个区域级别特征,其中,每个区域级别特征为一个K维向量。Input a given set of input samples into the neural network to be analyzed, and for each sample, extract the output features of a certain convolutional layer of the neural network, so as to obtain the feature map corresponding to each input sample, where the feature map The high-dimensional vector corresponding to each position is the region-level feature of the sample in this region; when the height and width of this feature map are H and W respectively, and there are K channels in total, then this feature map contains HW region-level features, wherein each region-level feature is a K-dimensional vector.
  4. 根据权利要求1中所述的方法,其特征在于,所述步骤(3)中对样本级别特征的降维过程,包括以下步骤:According to the method described in claim 1, it is characterized in that, in the described step (3), the dimensionality reduction process to sample level feature comprises the following steps:
    对于每个样本x对应的样本级别特征
    Figure PCTCN2022127435-appb-100001
    通过一个投影矩阵
    Figure PCTCN2022127435-appb-100002
    将其映射到一个低维空间中,得到样本级别特征的低维表征
    Figure PCTCN2022127435-appb-100003
    并且,优化M使得该低维表征应当满足,低维表征g和各个类。
    For each sample x corresponding to the sample-level features
    Figure PCTCN2022127435-appb-100001
    through a projection matrix
    Figure PCTCN2022127435-appb-100002
    Map it into a low-dimensional space to obtain a low-dimensional representation of sample-level features
    Figure PCTCN2022127435-appb-100003
    And, optimize M such that the low-dimensional representation should satisfy the low-dimensional representation g and each class.
  5. 根据权利要求4中所述的方法,其特征在于,所述低维表征和各个类别的接近程度的计算,包括:According to the method described in claim 4, it is characterized in that, the calculation of the degree of proximity between the low-dimensional representation and each category includes:
    (a)利用径向分布建模样本级别特征的低维表征g在低维空间中的分布;(a) Use radial distribution to model the distribution of low-dimensional representation g of sample-level features in low-dimensional space;
    (b)计算低维表征与各个类别的接近程度。(b) Calculate how close the low-dimensional representation is to each category.
  6. 根据权利要求5所述的方法,其特征在于,所述步骤(a)包括:The method according to claim 5, wherein said step (a) comprises:
    基于径向分布,g在低维空间的概率密度函数可写为如下形式:Based on radial distribution, the probability density function of g in low-dimensional space can be written as follows:
    Figure PCTCN2022127435-appb-100004
    Figure PCTCN2022127435-appb-100004
    其中,y∈{1,2,…,C}表示分类任务中的不同类别;π y表示第y类的先验概率;l g=‖g‖表示g的L2范数,称为g的强度;o g=g/l g表示g的方向(orientation);μ y表示第y类的平均方向;κ(·)是一个单调递增的函数,p(l g|y)表示在类别y上l g的先验概率,p vMF(o gy,κ(l g))表示平均方向为μ y,聚集参数为κ(l g)的vMF分布。 Among them, y∈{1,2,…,C} represents different categories in the classification task; π y represents the prior probability of the yth class; l g =‖g‖ represents the L2 norm of g, which is called the strength of g ; o g = g/l g represents the orientation of g; μ y represents the average direction of the yth class; κ( ) is a monotonically increasing function, and p(l g |y) represents the l The prior probability of g , p vMF (o gy ,κ(l g )) represents the vMF distribution with mean direction μ y and aggregation parameter κ(l g ).
  7. 根据权利要求5中所述的方法,其特征在于,所述步骤(b)包括:The method according to claim 5, wherein said step (b) comprises:
    基于上述径向分布,并假设l g的先验概率与类别y是独立的,那么,低维表征g和第y类的接近程度Q M(y|x)表示为如下形式: Based on the above radial distribution, and assuming that the prior probability of l g is independent of the category y, then the closeness Q M (y|x) between the low-dimensional representation g and the yth category is expressed as follows:
    Figure PCTCN2022127435-appb-100005
    Figure PCTCN2022127435-appb-100005
  8. 根据权利要求1中所述的方法,其特征在于,所述步骤(3)中对区域级别特征的降维过程,进一步包括以下步骤:According to the method described in claim 1, it is characterized in that, in the described step (3), the dimensionality reduction process to region-level features further comprises the following steps:
    对于每个样本x的HW个区域级别特征
    Figure PCTCN2022127435-appb-100006
    通过一个投影矩阵
    Figure PCTCN2022127435-appb-100007
    将它们映射到一个低维空间中,得到HW个区域级别特征的低维表征
    Figure PCTCN2022127435-appb-100008
    并且,优化Λ使得该低维表征应当满足,基于低维表征h={h (1),h (2),…,h (HW)}所推断出的样本间相似度与基于网络输出所推断出的样本间相似度尽可能保持一致,进一步,区域级别特征的低维表征需要与样本级别特征的低维特征表征对齐。
    For each sample x HW region-level features
    Figure PCTCN2022127435-appb-100006
    through a projection matrix
    Figure PCTCN2022127435-appb-100007
    Map them into a low-dimensional space to obtain a low-dimensional representation of HW region-level features
    Figure PCTCN2022127435-appb-100008
    And, optimize Λ so that the low-dimensional representation should satisfy that the similarity between samples inferred based on the low-dimensional representation h={h (1) ,h (2) ,…,h (HW) } is the same as that inferred based on the network output The similarity between samples obtained should be as consistent as possible. Furthermore, the low-dimensional representation of region-level features needs to be aligned with the low-dimensional feature representation of sample-level features.
  9. 如权利要求1中所述的方法,其特征在于,所述步骤(4)中的所述知识点为使得如下式子大于某个阈值的区域级别特征的集合:The method as claimed in claim 1, characterized in that, the knowledge point in the step (4) is a collection of region-level features that make the following formula greater than a certain threshold:
    Figure PCTCN2022127435-appb-100009
    Figure PCTCN2022127435-appb-100009
    其中,h (r)为某一样本x对应的第r个区域级别特征的低维表征;所以,知识点表示使得max c p(y=c|h (r))>τ的区域级别特征,即集合{h (r):max c p(y=c|h (r))>τ}中所包含的区域级别特征,其中,τ是一个正的常数,在一个优选例中,τ的取值范围是0.3-0.8。 Among them, h (r) is the low-dimensional representation of the r-th region-level feature corresponding to a certain sample x; therefore, the knowledge point represents the region-level feature that makes max c p(y=c|h (r) )>τ, That is, the region-level features contained in the set {h (r) :max c p(y=c|h (r) )>τ}, where τ is a positive constant, and in a preferred example, the value of τ is The value range is 0.3-0.8.
  10. 一种对神经网络中层特征表达能力的可视化及定量分析系统,其特征在于,所述系统包括如下模块:A visualization and quantitative analysis system for neural network middle layer feature expression ability, characterized in that the system includes the following modules:
    (1)输入模块,被配置为一个预训练好的分类神经网络和包含所有可能类别的输入样本;(1) The input module is configured as a pre-trained classification neural network and contains input samples of all possible categories;
    (2)特征提取模块,被配置为提取所述输入样本的样本级别特征和区域级别特征;(2) a feature extraction module configured to extract sample-level features and region-level features of the input samples;
    (3)可视化模块,被配置为,基于提取得到的样本级别特征和区域级别特征,将其降维得到低维表征,并且在低维空间中将低维表征进行可视化;(3) a visualization module configured to, based on the extracted sample-level features and region-level features, reduce its dimension to obtain a low-dimensional representation, and visualize the low-dimensional representation in the low-dimensional space;
    (4)定量分析模块,被配置为,基于区域级别特征的可视化结果,定量分析特征中知识点的数量与质量。(4) The quantitative analysis module is configured to quantitatively analyze the quantity and quality of knowledge points in the feature based on the visualization result of the feature at the region level.
PCT/CN2022/127435 2021-10-25 2022-10-25 Visualization and quantitative analysis method and system for expression capability of layer feature in neural network WO2023072094A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202111240906.7 2021-10-25
CN202111240906.7A CN113963185A (en) 2021-10-25 2021-10-25 Visualization and quantitative analysis method and system for layer feature expression capability in neural network

Publications (1)

Publication Number Publication Date
WO2023072094A1 true WO2023072094A1 (en) 2023-05-04

Family

ID=79466728

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/127435 WO2023072094A1 (en) 2021-10-25 2022-10-25 Visualization and quantitative analysis method and system for expression capability of layer feature in neural network

Country Status (2)

Country Link
CN (1) CN113963185A (en)
WO (1) WO2023072094A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113963185A (en) * 2021-10-25 2022-01-21 上海交通大学 Visualization and quantitative analysis method and system for layer feature expression capability in neural network

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109460812A (en) * 2017-09-06 2019-03-12 富士通株式会社 Average information analytical equipment, the optimization device, feature visualization device of neural network
CN110781933A (en) * 2019-10-14 2020-02-11 杭州电子科技大学 Visual analysis method for understanding graph convolution neural network
WO2020104252A1 (en) * 2018-11-19 2020-05-28 Siemens Aktiengesellschaft Verification of classification decisions in convolutional neural networks
CN111695590A (en) * 2020-04-24 2020-09-22 浙江大学 Deep neural network feature visualization method for constraint optimization class activation mapping
WO2020244941A1 (en) * 2019-06-06 2020-12-10 Koninklijke Philips N.V. Deep neural network visualisation
CN112884021A (en) * 2021-01-29 2021-06-01 之江实验室 Visual analysis system oriented to deep neural network interpretability
CN113963185A (en) * 2021-10-25 2022-01-21 上海交通大学 Visualization and quantitative analysis method and system for layer feature expression capability in neural network

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109460812A (en) * 2017-09-06 2019-03-12 富士通株式会社 Average information analytical equipment, the optimization device, feature visualization device of neural network
WO2020104252A1 (en) * 2018-11-19 2020-05-28 Siemens Aktiengesellschaft Verification of classification decisions in convolutional neural networks
WO2020244941A1 (en) * 2019-06-06 2020-12-10 Koninklijke Philips N.V. Deep neural network visualisation
CN110781933A (en) * 2019-10-14 2020-02-11 杭州电子科技大学 Visual analysis method for understanding graph convolution neural network
CN111695590A (en) * 2020-04-24 2020-09-22 浙江大学 Deep neural network feature visualization method for constraint optimization class activation mapping
CN112884021A (en) * 2021-01-29 2021-06-01 之江实验室 Visual analysis system oriented to deep neural network interpretability
CN113963185A (en) * 2021-10-25 2022-01-21 上海交通大学 Visualization and quantitative analysis method and system for layer feature expression capability in neural network

Also Published As

Publication number Publication date
CN113963185A (en) 2022-01-21

Similar Documents

Publication Publication Date Title
TWI766855B (en) A character recognition method and device
Baştanlar et al. Introduction to machine learning
RU2656708C1 (en) Method for separating texts and illustrations in images of documents using a descriptor of document spectrum and two-level clustering
WO2017143914A1 (en) Method for training model using training data, and training system
US8458194B1 (en) System and method for content-based document organization and filing
CN114830133A (en) Supervised contrast learning with multiple positive examples
US11671241B2 (en) Privacy-preserving fast approximate K-means clustering with hamming vectors
US11514369B2 (en) Systems and methods for machine learning model interpretation
Anwar et al. Measurement of data complexity for classification problems with unbalanced data
US11914963B2 (en) Systems and methods for determining and using semantic relatedness to classify segments of text
CN111178533B (en) Method and device for realizing automatic semi-supervised machine learning
WO2023072094A1 (en) Visualization and quantitative analysis method and system for expression capability of layer feature in neural network
CN110909540B (en) Method and device for identifying new words of short message spam and electronic equipment
Yakovlev et al. Abstraction-based outlier detection for image data
US11048711B1 (en) System and method for automated classification of structured property description extracted from data source using numeric representation and keyword search
US20220058431A1 (en) Semantic input sampling for explanation (sise) of convolutional neural networks
McLachlan Discriminant analysis
CN112200666A (en) Feature vector processing method and related device
Kazemi et al. FEM-DBSCAN: AN efficient density-based clustering approach
Channoufi et al. Spatially constrained mixture model with feature selection for image and video segmentation
KR102472447B1 (en) A system and method for automatically blocking specific content in complex documents using machine learning
US20170293863A1 (en) Data analysis system, and control method, program, and recording medium therefor
US20210406693A1 (en) Data sample analysis in a dataset for a machine learning model
JP5946949B1 (en) DATA ANALYSIS SYSTEM, ITS CONTROL METHOD, PROGRAM, AND RECORDING MEDIUM
Congleton et al. Tracing Political Positioning of Dutch Newspapers

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22885971

Country of ref document: EP

Kind code of ref document: A1