CN118690824A - Performance prediction model training method and device, performance prediction method and device - Google Patents
Performance prediction model training method and device, performance prediction method and device Download PDFInfo
- Publication number
- CN118690824A CN118690824A CN202411186366.2A CN202411186366A CN118690824A CN 118690824 A CN118690824 A CN 118690824A CN 202411186366 A CN202411186366 A CN 202411186366A CN 118690824 A CN118690824 A CN 118690824A
- Authority
- CN
- China
- Prior art keywords
- target
- mask
- matrix
- task
- neural network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/0895—Weakly supervised learning, e.g. semi-supervised or self-supervised learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
- G06N3/0455—Auto-encoder networks; Encoder-decoder networks
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
技术领域Technical Field
本发明涉及数据处理领域,更具体地涉及一种性能预测模型的训练方法及装置、性能预测方法及装置。The present invention relates to the field of data processing, and more specifically to a performance prediction model training method and device, and a performance prediction method and device.
背景技术Background Art
神经网络结构设计方法缺少完备且有效的理论指导,针对不同的任务,利用神经网络结构搜索技术对给定任务和数据集自动搜索高性能神经网络结构。The neural network structure design method lacks complete and effective theoretical guidance. For different tasks, the neural network structure search technology is used to automatically search for high-performance neural network structures for given tasks and data sets.
在传统神经网络结构搜索方法中,通常需要将神经网络结构在给定任务上进行实际训练或在不同硬件设备上进行实际部署,以获取不同候选网络的精度或时延等性能数据,并基于此对大量候选网络结构进行比较。在这过程中,执行一次搜索算法通常需要消耗上千个GPU天数才可以得到性能超越人工设计的神经网络结构。In traditional neural network structure search methods, it is usually necessary to actually train the neural network structure on a given task or actually deploy it on different hardware devices to obtain performance data such as accuracy or latency of different candidate networks, and compare a large number of candidate network structures based on this. In this process, executing a search algorithm usually takes thousands of GPU days to obtain a neural network structure with performance that exceeds that of manually designed ones.
相关技术为改善传统神经网络结构搜索方法,基于性能预测模型来减少搜索过程中的评估耗时,从而加速整个搜索进程。然而,性能预测模型采用传统的有监督学习方式进行训练,训练数据又需要通过采样获得,导致预测器的训练成本较高,并且在标签数据有限的情况下无法获得较高的预测准确度。The related technology aims to improve the traditional neural network structure search method and reduce the evaluation time in the search process based on the performance prediction model, thereby accelerating the entire search process. However, the performance prediction model is trained using the traditional supervised learning method, and the training data needs to be obtained through sampling, resulting in a high training cost for the predictor, and it is impossible to obtain a high prediction accuracy when the label data is limited.
发明内容Summary of the invention
鉴于上述问题,本发明提供了一种性能预测模型的训练方法及装置、性能预测方法及装置。In view of the above problems, the present invention provides a performance prediction model training method and device, and a performance prediction method and device.
根据本发明的第一个方面,提供了一种性能预测模型的训练方法,包括:利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵,掩码节点特征矩阵和掩码标识矩阵的矩阵维度相同;基于可见节点特征矩阵,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征;利用目标计算单元对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,利用目标计算单元对性能预测模型执行训练任务,得到训练后的性能预测模型。According to a first aspect of the present invention, a method for training a performance prediction model is provided, comprising: using a target computing unit to perform a mask processing task on unlabeled neural network structure data to obtain a visible node feature matrix, a mask node feature matrix and a mask identification matrix, wherein the mask node feature matrix and the mask identification matrix have the same matrix dimension; based on the visible node feature matrix, using the target computing unit to perform a prediction task on the mask identification matrix to obtain predicted latent space features of the mask identification matrix; using the target computing unit to perform an encoding task on the mask node feature matrix to obtain latent space features of the mask node feature matrix; using the target computing unit to perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, using the target computing unit to perform a training task on the performance prediction model to obtain a trained performance prediction model.
根据本发明的实施例,基于可见节点特征矩阵,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征,包括:利用目标计算单元对可见节点特征矩阵执行编码处理任务,得到可见节点特征矩阵的隐空间特征;基于可见节点特征矩阵的隐空间特征,利用目标计算单元对掩码标识矩阵执行预测处理任务,得到掩码标识矩阵的预测隐空间特征。According to an embodiment of the present invention, based on the visible node feature matrix, a target computing unit is used to perform a prediction task on a mask identification matrix to obtain predicted latent space features of the mask identification matrix, including: using the target computing unit to perform an encoding processing task on the visible node feature matrix to obtain latent space features of the visible node feature matrix; based on the latent space features of the visible node feature matrix, using the target computing unit to perform a prediction processing task on the mask identification matrix to obtain predicted latent space features of the mask identification matrix.
根据本发明的实施例,利用目标计算单元对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征,包括:利用目标计算单元确定编码参数;利用目标计算单元对编码参数执行指数移动平均处理任务,得到目标编码参数;基于目标编码参数,利用目标计算单元对掩码节点特征矩阵执行编码处理任务,得到掩码节点特征矩阵的隐空间特征。According to an embodiment of the present invention, a target computing unit is used to perform an encoding task on a mask node feature matrix to obtain latent space features of the mask node feature matrix, including: determining encoding parameters using the target computing unit; performing an exponential moving average processing task on the encoding parameters using the target computing unit to obtain target encoding parameters; based on the target encoding parameters, performing an encoding processing task on the mask node feature matrix using the target computing unit to obtain latent space features of the mask node feature matrix.
根据本发明的实施例,利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值,包括:利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行缩放余弦误差计算任务,得到第一初始损失值;利用目标计算单元确定掩码标识矩阵的预测隐空间特征对应的操作索引,操作索引表征与掩码标识矩阵对应的结构组成;基于操作索引,利用目标计算单元对掩码标识矩阵的预测隐空间特征执行构建任务,得到中间神经网络结构数据;利用目标计算单元对中间神经网络结构数据和无标签神经网络结构数据执行交叉熵计算任务,得到第二初始损失值;基于第一预设权重 ,利用目标计算单元对第一初始损失值和第二初始损失值执行求和任务,得到第一目标损失值。According to an embodiment of the present invention, a target computing unit is used to perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value, including: using the target computing unit to perform a scaled cosine error calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first initial loss value; using the target computing unit to determine the operation index corresponding to the predicted latent space features of the mask identification matrix, the operation index characterizing the structural composition corresponding to the mask identification matrix; based on the operation index, using the target computing unit to perform a construction task on the predicted latent space features of the mask identification matrix to obtain intermediate neural network structure data; using the target computing unit to perform a cross entropy calculation task on the intermediate neural network structure data and the unlabeled neural network structure data to obtain a second initial loss value; based on the first preset weight , using the target computing unit to perform a summation task on the first initial loss value and the second initial loss value to obtain a first target loss value.
根据本发明的实施例,上述方法还包括:利用目标计算单元对有标签神经网络结构数据执行编码任务,确定第二目标损失值。According to an embodiment of the present invention, the above method also includes: using the target computing unit to perform an encoding task on the labeled neural network structure data to determine a second target loss value.
根据本发明的实施例,对利用目标计算单元对有标签神经网络结构数据执行编码处理任务,确定第二目标损失值,包括:利用目标计算单元对有标签神经网络结构数据执行编码任务,得到有标签神经网络结构数据的隐空间特征;利用目标计算单元对有标签神经网络结构数据的隐空间特征执行性能预测任务,确定预测性能值;利用目标计算单元确定与有标签神经网络结构数据对应的标签性能值;利用目标计算单元对预测性能值和标签性能值执行均方差计算任务,得到第三初始损失值;利用目标计算单元对预测性能值和标签性能值执行排序误差计算任务,得到第四初始损失值;基于第二预设权重,利用目标计算单元对第三初始损失值和第四初始损失值执行求和任务,得到第二目标损失值。According to an embodiment of the present invention, a target computing unit is used to perform an encoding processing task on labeled neural network structure data to determine a second target loss value, including: performing an encoding task on the labeled neural network structure data using the target computing unit to obtain latent space features of the labeled neural network structure data; performing a performance prediction task on the latent space features of the labeled neural network structure data using the target computing unit to determine a predicted performance value; determining a label performance value corresponding to the labeled neural network structure data using the target computing unit; performing a mean square error calculation task on the predicted performance value and the label performance value using the target computing unit to obtain a third initial loss value; performing a sorting error calculation task on the predicted performance value and the label performance value using the target computing unit to obtain a fourth initial loss value; and based on a second preset weight, performing a summation task on the third initial loss value and the fourth initial loss value using the target computing unit to obtain a second target loss value.
根据本发明的实施例,基于第一目标损失值,利用目标计算单元对神经网络预测模型执行训练任务,得到训练后的神经网络预测模型,包括:基于第一目标损失值,利用目标计算单元对性能预测模型执行训练任务,得到中间性能预测模型;基于第二目标损失值,利用目标计算单元对中间性能预测模型执行训练任务,得到训练后的性能预测模型。According to an embodiment of the present invention, based on the first target loss value, a target computing unit is used to perform a training task on a neural network prediction model to obtain a trained neural network prediction model, including: based on the first target loss value, a target computing unit is used to perform a training task on a performance prediction model to obtain an intermediate performance prediction model; based on the second target loss value, a target computing unit is used to perform a training task on the intermediate performance prediction model to obtain a trained performance prediction model.
本发明的第二方面提供了一种性能预测方法,包括:利用目标预测单元获取待预测神经网络结构数据;利用目标预测单元,基于上述性能预测模型对待预测神经网络结构数据执行预测处理任务,得到与待预测神经网络结构数据对应的目标性能。The second aspect of the present invention provides a performance prediction method, including: using a target prediction unit to obtain the neural network structure data to be predicted; using the target prediction unit to perform a prediction processing task on the neural network structure data to be predicted based on the above-mentioned performance prediction model, and obtaining the target performance corresponding to the neural network structure data to be predicted.
本发明的第三方面提供了一种性能预测模型的训练装置,包括:目标计算单元,配置为:对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵,掩码节点特征矩阵和掩码标识矩阵的矩阵维度相同;基于可见节点特征矩阵,对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征;对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,对性能预测模型执行训练任务,得到训练后的性能预测模型。The third aspect of the present invention provides a training device for a performance prediction model, comprising: a target calculation unit, configured to: perform a mask processing task on unlabeled neural network structure data to obtain a visible node feature matrix, a mask node feature matrix and a mask identification matrix, wherein the mask node feature matrix and the mask identification matrix have the same matrix dimension; based on the visible node feature matrix, perform a prediction task on the mask identification matrix to obtain predicted latent space features of the mask identification matrix; perform an encoding task on the mask node feature matrix to obtain latent space features of the mask node feature matrix; perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, perform a training task on the performance prediction model to obtain a trained performance prediction model.
本发明的第四方面提供了一种目标预测单元,配置为:获取待预测神经网络结构数据;基于上述性能预测模型对待预测神经网络结构数据执行预测处理任务,得到与待预测神经网络结构数据对应的目标性能。The fourth aspect of the present invention provides a target prediction unit configured to: obtain the neural network structure data to be predicted; perform a prediction processing task on the neural network structure data to be predicted based on the above-mentioned performance prediction model to obtain the target performance corresponding to the neural network structure data to be predicted.
本发明的第五方面提供了一种电子设备,包括:一个或多个处理器;存储器,用于存储一个或多个计算机程序,其中,上述一个或多个处理器执行上述一个或多个计算机程序以实现上述方法的步骤。A fifth aspect of the present invention provides an electronic device, comprising: one or more processors; a memory for storing one or more computer programs, wherein the one or more processors execute the one or more computer programs to implement the steps of the above method.
本发明的第六方面还提供了一种计算机可读存储介质,其上存储有计算机程序或指令,上述计算机程序或指令被处理器执行时实现上述方法的步骤。The sixth aspect of the present invention further provides a computer-readable storage medium on which a computer program or instruction is stored, and the steps of the above method are implemented when the above computer program or instruction is executed by a processor.
本发明的第七方面还提供了一种计算机程序产品,包括计算机程序或指令,上述计算机程序或指令被处理器执行时实现上述方法的步骤。The seventh aspect of the present invention also provides a computer program product, including a computer program or instructions, which implement the steps of the above method when executed by a processor.
根据本发明的实施例,利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,对掩码处理得到的掩码标识矩阵执行预测任务,得到述掩码标识矩阵的预测隐空间特征;对掩码处理得到的掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,对性能预测模型执行训练任务,得到训练后的性能预测模型。由于对大量易获得的无标签神经网络结构数据进行掩码处理,并对掩码处理后的数据进行预测和编码等操作,得到预测隐空间特征和所述掩码节点特征矩阵的隐空间特征,以确定第一目标损失值的技术手段,避免了标签数据有限的情况下无法获得较高的预测准确度的问题,实现了在训练过程能够更好的挖掘神经网络结构数据的上下文信息,以减少有标签神经网络结构数据的训练,减小性能预测模型训练成本的技术效果。According to an embodiment of the present invention, a target computing unit is used to perform a mask processing task on unlabeled neural network structure data, and a prediction task is performed on the mask identification matrix obtained by the mask processing to obtain the predicted latent space features of the mask identification matrix; an encoding task is performed on the mask node feature matrix obtained by the mask processing to obtain the latent space features of the mask node feature matrix; a loss calculation task is performed on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, a training task is performed on the performance prediction model to obtain a trained performance prediction model. Since a large amount of easily available unlabeled neural network structure data is masked, and the masked data is predicted and encoded, the predicted latent space features and the latent space features of the mask node feature matrix are obtained to determine the technical means of the first target loss value, which avoids the problem of not being able to obtain a high prediction accuracy when the label data is limited, and realizes the technical effect of being able to better mine the context information of the neural network structure data during the training process to reduce the training of labeled neural network structure data and reduce the training cost of the performance prediction model.
附图说明BRIEF DESCRIPTION OF THE DRAWINGS
通过以下参照附图对本发明实施例的描述,本发明的上述内容以及其他目的、特征和优点将更为清楚,在附图中:The above contents and other objects, features and advantages of the present invention will become more apparent through the following description of the embodiments of the present invention with reference to the accompanying drawings, in which:
图1示出了根据本发明实施例的性能预测模型的训练方法、装置、设备、介质和程序产品的应用场景图;FIG1 shows an application scenario diagram of a training method, apparatus, device, medium, and program product for a performance prediction model according to an embodiment of the present invention;
图2示出了根据本发明实施例的性能预测模型的训练方法的流程图;FIG2 shows a flow chart of a method for training a performance prediction model according to an embodiment of the present invention;
图3示出了根据本发明实施例的性能预测模型的训练方法的示意图;FIG3 is a schematic diagram showing a method for training a performance prediction model according to an embodiment of the present invention;
图4示出了根据本发明实施例的性能预测模型的示意图;FIG4 shows a schematic diagram of a performance prediction model according to an embodiment of the present invention;
图5示出了根据本发明实施例的自注意力层的示意图;FIG5 shows a schematic diagram of a self-attention layer according to an embodiment of the present invention;
图6示出了根据本发明实施例的性能预测方法的流程图;FIG6 shows a flow chart of a performance prediction method according to an embodiment of the present invention;
图7示出了根据本发明实施例的性能预测模型的训练装置的结构框图;FIG7 shows a structural block diagram of a training device for a performance prediction model according to an embodiment of the present invention;
图8示出了根据本发明实施例的性能预测装置的结构框图;FIG8 shows a structural block diagram of a performance prediction device according to an embodiment of the present invention;
图9示出了根据本发明实施例的适于实现性能预测模型的训练方法的电子设备的方框图。FIG9 shows a block diagram of an electronic device suitable for implementing a training method for a performance prediction model according to an embodiment of the present invention.
具体实施方式DETAILED DESCRIPTION
以下,将参照附图来描述本发明的实施例。但是应该理解,这些描述只是示例性的,而并非要限制本发明的范围。在下面的详细描述中,为便于解释,阐述了许多具体的细节以提供对本发明实施例的全面理解。然而,明显地,一个或多个实施例在没有这些具体细节的情况下也可以被实施。此外,在以下说明中,省略了对公知结构和技术的描述,以避免不必要地混淆本发明的概念。Below, embodiments of the present invention will be described with reference to the accompanying drawings. However, it should be understood that these descriptions are exemplary only and are not intended to limit the scope of the present invention. In the following detailed description, for ease of explanation, many specific details are set forth to provide a comprehensive understanding of embodiments of the present invention. However, it is apparent that one or more embodiments may also be implemented without these specific details. In addition, in the following description, descriptions of known structures and technologies are omitted to avoid unnecessary confusion of concepts of the present invention.
在此使用的术语仅仅是为了描述具体实施例,而并非意在限制本发明。在此使用的术语“包括”、“包含”等表明了特征、步骤、操作和/或部件的存在,但是并不排除存在或添加一个或多个其他特征、步骤、操作或部件。The terms used herein are only for describing specific embodiments and are not intended to limit the present invention. The terms "comprise", "include", etc. used herein indicate the existence of features, steps, operations and/or components, but do not exclude the existence or addition of one or more other features, steps, operations or components.
在此使用的所有术语(包括技术和科学术语)具有本领域技术人员通常所理解的含义,除非另外定义。应注意,这里使用的术语应解释为具有与本说明书的上下文相一致的含义,而不应以理想化或过于刻板的方式来解释。All terms (including technical and scientific terms) used herein have the meanings commonly understood by those skilled in the art unless otherwise defined. It should be noted that the terms used herein should be interpreted as having a meaning consistent with the context of this specification and should not be interpreted in an idealized or overly rigid manner.
在使用类似于“A、B和C等中至少一个”这样的表述的情况下,一般来说应该按照本领域技术人员通常理解该表述的含义来予以解释(例如,“具有A、B和C中至少一个的系统”应包括但不限于单独具有A、单独具有B、单独具有C、具有A和B、具有A和C、具有B和C、和/或具有A、B、C的系统等)。When using expressions such as "at least one of A, B, and C, etc.", they should generally be interpreted according to the meaning of the expression commonly understood by those skilled in the art (for example, "a system having at least one of A, B, and C" should include but is not limited to a system having A alone, B alone, C alone, A and B, A and C, B and C, and/or A, B, C, etc.).
本发明的实施例提供了一种性能预测模型的训练方法,利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵,掩码节点特征矩阵和掩码标识矩阵的矩阵维度相同;基于可见节点特征矩阵,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征;利用目标计算单元对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,利用目标计算单元对性能预测模型执行训练任务,得到训练后的性能预测模型。An embodiment of the present invention provides a training method for a performance prediction model, which uses a target computing unit to perform a mask processing task on unlabeled neural network structure data to obtain a visible node feature matrix, a mask node feature matrix and a mask identification matrix, wherein the matrix dimensions of the mask node feature matrix and the mask identification matrix are the same; based on the visible node feature matrix, the target computing unit is used to perform a prediction task on the mask identification matrix to obtain predicted latent space features of the mask identification matrix; the target computing unit is used to perform an encoding task on the mask node feature matrix to obtain latent space features of the mask node feature matrix; the target computing unit is used to perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, the target computing unit is used to perform a training task on the performance prediction model to obtain a trained performance prediction model.
图1示出了根据本发明实施例的性能预测模型的训练方法、装置、设备、介质和程序产品的应用场景图。FIG1 shows an application scenario diagram of a training method, apparatus, device, medium, and program product for a performance prediction model according to an embodiment of the present invention.
如图1所示,根据该实施例的应用场景100可以包括第一终端设备101、第二终端设备102、第三终端设备103、网络104和服务器105。网络104用以在第一终端设备101、第二终端设备102、第三终端设备103和服务器105之间提供通信链路的介质。网络104可以包括各种连接类型,例如有线、无线通信链路或者光纤电缆等等。As shown in Fig. 1, the application scenario 100 according to this embodiment may include a first terminal device 101, a second terminal device 102, a third terminal device 103, a network 104, and a server 105. The network 104 is used to provide a medium for a communication link between the first terminal device 101, the second terminal device 102, the third terminal device 103, and the server 105. The network 104 may include various connection types, such as wired, wireless communication links, or optical fiber cables, etc.
用户可以使用第一终端设备101、第二终端设备102、第三终端设备103通过网络104与服务器105交互,以接收或发送消息等。第一终端设备101、第二终端设备102、第三终端设备103上可以安装有各种通讯客户端应用,例如购物类应用、网页浏览器应用、搜索类应用、即时通信工具、邮箱客户端、社交平台软件等(仅为示例)。The user can use the first terminal device 101, the second terminal device 102, and the third terminal device 103 to interact with the server 105 through the network 104 to receive or send messages, etc. Various communication client applications can be installed on the first terminal device 101, the second terminal device 102, and the third terminal device 103, such as shopping applications, web browser applications, search applications, instant messaging tools, email clients, social platform software, etc. (only for example).
第一终端设备101、第二终端设备102、第三终端设备103可以是具有显示屏并且支持网页浏览的各种电子设备,包括但不限于智能手机、平板电脑、膝上型便携计算机和台式计算机等等。The first terminal device 101, the second terminal device 102, and the third terminal device 103 may be various electronic devices having display screens and supporting web browsing, including but not limited to smart phones, tablet computers, laptop computers, desktop computers, and the like.
服务器105可以是提供各种服务的服务器,例如对用户利用第一终端设备101、第二终端设备102、第三终端设备103所浏览的网站提供支持的后台管理服务器(仅为示例)。后台管理服务器可以对接收到的用户请求等数据进行分析等处理,并将处理结果(例如根据用户请求获取或生成的网页、信息、或数据等)反馈给终端设备。The server 105 may be a server that provides various services, such as a background management server (only as an example) that provides support for websites browsed by users using the first terminal device 101, the second terminal device 102, and the third terminal device 103. The background management server may analyze and process the received data such as user requests, and feed back the processing results (such as web pages, information, or data obtained or generated according to user requests) to the terminal device.
需要说明的是,本发明实施例所提供的性能预测模型的训练方法一般可以由服务器105执行。相应地,本发明实施例所提供的性能预测模型的训练装置一般可以设置于服务器105中。本发明实施例所提供的性能预测模型的训练方法也可以由不同于服务器105且能够与第一终端设备101、第二终端设备102、第三终端设备103和/或服务器105通信的服务器或服务器集群执行。相应地,本发明实施例所提供的性能预测模型的训练装置也可以设置于不同于服务器105且能够与第一终端设备101、第二终端设备102、第三终端设备103和/或服务器105通信的服务器或服务器集群中。It should be noted that the training method of the performance prediction model provided in the embodiment of the present invention can generally be executed by the server 105. Accordingly, the training device of the performance prediction model provided in the embodiment of the present invention can generally be set in the server 105. The training method of the performance prediction model provided in the embodiment of the present invention can also be executed by a server or server cluster that is different from the server 105 and can communicate with the first terminal device 101, the second terminal device 102, the third terminal device 103 and/or the server 105. Accordingly, the training device of the performance prediction model provided in the embodiment of the present invention can also be set in a server or server cluster that is different from the server 105 and can communicate with the first terminal device 101, the second terminal device 102, the third terminal device 103 and/or the server 105.
应该理解,图1中的终端设备、网络和服务器的数目仅仅是示意性的。根据实现需要,可以具有任意数目的终端设备、网络和服务器。It should be understood that the number of terminal devices, networks and servers in Figure 1 is only illustrative. Any number of terminal devices, networks and servers may be provided according to implementation requirements.
图2示出了根据本发明实施例的性能预测模型的训练方法的流程图。FIG. 2 shows a flow chart of a method for training a performance prediction model according to an embodiment of the present invention.
如图2所示,该实施例的性能预测模型的训练方法包括操作S210~操作S250。As shown in FIG. 2 , the training method of the performance prediction model of this embodiment includes operations S210 to S250 .
在操作S210,利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵,掩码节点特征矩阵和掩码标识矩阵的矩阵维度相同。In operation S210, a mask processing task is performed on the unlabeled neural network structure data using a target computing unit to obtain a visible node feature matrix, a mask node feature matrix, and a mask identification matrix, wherein the mask node feature matrix and the mask identification matrix have the same matrix dimensions.
根据本发明的实施例,目标计算单元可以包括中央处理单元、神经网络处理器、分布式计算单元等至少一种。According to an embodiment of the present invention, the target computing unit may include at least one of a central processing unit, a neural network processor, a distributed computing unit, and the like.
根据本发明的实施例,无标签神经网络结构数据可以是神经网络层类型、权重、偏置等参数中的至少一种。According to an embodiment of the present invention, the unlabeled neural network structure data may be at least one of parameters such as neural network layer type, weight, bias, etc.
根据本发明的实施例,利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,按照一定的掩码率通过完全随机的方式选取部分无标签神经网络结构数据的节点作为可见节点特征矩阵,其余为掩码节点特征矩阵。According to an embodiment of the present invention, a target computing unit is used to perform a mask processing task on unlabeled neural network structure data, and some nodes of the unlabeled neural network structure data are selected in a completely random manner according to a certain mask rate as visible node feature matrices, and the rest are mask node feature matrices.
根据本发明的实施例,确定掩码标识,每个掩码标识与每个掩码节点一一对应,得到的掩码标识矩阵与掩码节点特征矩阵的矩阵维度相同,其中,掩码标识为一个可学习的随机向量。According to an embodiment of the present invention, a mask identifier is determined, each mask identifier corresponds to each mask node one by one, and the obtained mask identifier matrix has the same matrix dimension as the mask node feature matrix, wherein the mask identifier is a learnable random vector.
根据本发明的实施例,上述无标签神经网络结构数据的节点可以是该无标签神经网络结构数据的输入参数、权重或函数等至少一种。According to an embodiment of the present invention, the node of the above-mentioned unlabeled neural network structure data can be at least one of the input parameters, weights or functions of the unlabeled neural network structure data.
根据本发明的实施例,在执行掩码处理任务时,只对无标签神经网络结构数据的节点进行掩码,而不改变无标签神经网络结构数据的结构信息。According to an embodiment of the present invention, when performing a mask processing task, only the nodes of the unlabeled neural network structure data are masked, and the structural information of the unlabeled neural network structure data is not changed.
在操作S220,基于可见节点特征矩阵,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征。In operation S220, based on the visible node feature matrix, a prediction task is performed on the mask identification matrix using a target computing unit to obtain predicted latent space features of the mask identification matrix.
根据本发明的实施例,预测任务可以是利用目标计算单元调用性能预测模型的交叉注意力层实现的。According to an embodiment of the present invention, the prediction task can be implemented by using the target computing unit to call the cross-attention layer of the performance prediction model.
根据本发明的实施例,利用可见节点特征矩阵的特征信息,对掩码标识矩阵进行预测,以得到掩码标识矩阵的预测隐空间特征。According to an embodiment of the present invention, the mask identification matrix is predicted using feature information of the visible node feature matrix to obtain predicted latent space features of the mask identification matrix.
在操作S230,利用目标计算单元对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征。In operation S230, the target computing unit is used to perform an encoding task on the mask node feature matrix to obtain latent space features of the mask node feature matrix.
根据本发明的实施例,编码任务可以是利用目标计算单元调用性能预测模型的自注意力层实现的。According to an embodiment of the present invention, the encoding task can be implemented by using the target computing unit to call the self-attention layer of the performance prediction model.
根据本发明的实施例,对掩码节点特征矩阵执行编码任务,根据掩码节点特征矩阵中元素之间的相似关系提取掩码节点特征矩阵的隐空间特征。According to an embodiment of the present invention, an encoding task is performed on a mask node feature matrix, and latent space features of the mask node feature matrix are extracted based on similarity relationships between elements in the mask node feature matrix.
在操作S240,利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值。In operation S240, a target calculation unit is used to perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value.
根据本发明的实施例,通过根据掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征,确定第一目标损失值,使得性能预测模型在训练过程中实现各个网络层的功能解耦,增强性能预测模型特征提取和学习能力,为下游任务迁移打好基础。According to an embodiment of the present invention, by determining the first target loss value based on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix, the performance prediction model can achieve functional decoupling of each network layer during the training process, enhance the feature extraction and learning capabilities of the performance prediction model, and lay a good foundation for downstream task migration.
在操作S250,基于第一目标损失值,利用目标计算单元对性能预测模型执行训练任务,得到训练后的性能预测模型。In operation S250, based on the first target loss value, a training task is performed on the performance prediction model using a target computing unit to obtain a trained performance prediction model.
根据本发明的实施例,基于第一目标损失值,对性能预测模型的相关参数进行调整,以得到训练后的性能预测模型。According to an embodiment of the present invention, relevant parameters of the performance prediction model are adjusted based on the first target loss value to obtain a trained performance prediction model.
根据本发明的实施例,训练后的性能预测模型能够根据神经网络结构数据自身规律对神经网络结构数据进行有效编码,因此仅需使用少量有标签神经网络结构数据,通过有监督训练方式对性能预测模型进行微调,即可使性能预测模型适应多种下游性能预测任务,从而降低性能预测模型训练过程中对有标签数据的依赖。According to an embodiment of the present invention, the trained performance prediction model can effectively encode the neural network structure data according to the rules of the neural network structure data itself. Therefore, only a small amount of labeled neural network structure data is needed to fine-tune the performance prediction model through supervised training. This can make the performance prediction model adapt to a variety of downstream performance prediction tasks, thereby reducing the dependence on labeled data during the performance prediction model training process.
根据本发明的实施例,利用目标计算单元对无标签神经网络结构数据执行掩码处理任务,对掩码处理得到的掩码标识矩阵执行预测任务,得到述掩码标识矩阵的预测隐空间特征;对掩码处理得到的掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,对性能预测模型执行训练任务,得到训练后的性能预测模型。由于对大量易获得的无标签神经网络结构数据进行掩码处理,并对掩码处理后的数据进行预测和编码等操作,得到预测隐空间特征和所述掩码节点特征矩阵的隐空间特征,以确定第一目标损失值的技术手段,避免了标签数据有限的情况下无法获得较高的预测准确度的问题,实现了在训练过程能够更好的挖掘神经网络结构数据的上下文信息,以减少有标签神经网络结构数据的训练,减小性能预测模型训练成本的技术效果。According to an embodiment of the present invention, a target computing unit is used to perform a mask processing task on unlabeled neural network structure data, and a prediction task is performed on the mask identification matrix obtained by the mask processing to obtain the predicted latent space features of the mask identification matrix; an encoding task is performed on the mask node feature matrix obtained by the mask processing to obtain the latent space features of the mask node feature matrix; a loss calculation task is performed on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, a training task is performed on the performance prediction model to obtain a trained performance prediction model. Since a large amount of easily available unlabeled neural network structure data is masked, and the masked data is predicted and encoded, the predicted latent space features and the latent space features of the mask node feature matrix are obtained to determine the technical means of the first target loss value, which avoids the problem of not being able to obtain a high prediction accuracy when the label data is limited, and realizes the technical effect of being able to better mine the context information of the neural network structure data in the training process to reduce the training of the labeled neural network structure data and reduce the training cost of the performance prediction model.
根据本发明的实施例,基于可见节点特征矩阵,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征,包括:利用目标计算单元对可见节点特征矩阵执行编码任务,得到可见节点特征矩阵的隐空间特征;基于可见节点特征矩阵的隐空间特征,利用目标计算单元对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征。According to an embodiment of the present invention, based on the visible node feature matrix, a target computing unit is used to perform a prediction task on a mask identification matrix to obtain predicted latent space features of the mask identification matrix, including: using the target computing unit to perform an encoding task on the visible node feature matrix to obtain latent space features of the visible node feature matrix; based on the latent space features of the visible node feature matrix, using the target computing unit to perform a prediction task on the mask identification matrix to obtain predicted latent space features of the mask identification matrix.
根据本发明的实施例,利用目标计算单元调用性能预测模型的自注意力层对可见节点特征矩阵进行编码,对可见节点特征矩阵中的每个矩阵元素之间的相似关系进行特征提取,以获得可见节点特征矩阵的隐空间特征。According to an embodiment of the present invention, the target computing unit is used to call the self-attention layer of the performance prediction model to encode the visible node feature matrix, and feature extraction is performed on the similarity relationship between each matrix element in the visible node feature matrix to obtain the latent space features of the visible node feature matrix.
根据本发明的实施例,利用目标计算单元调用性能预测模型的交叉注意力层,对可见节点特征矩阵的隐空间特征和掩码标识矩阵进行交叉注意力计算,其中,在交叉注意力计算过程中,交叉注意力层由键矩阵K、值矩阵V和查询矩阵Q构成,查询矩阵Q由掩码标识矩阵线性变换获得,对可见节点和掩码节点的特征学习过程解耦,以得到掩码标识矩阵的预测隐空间特征。According to an embodiment of the present invention, the cross-attention layer of the performance prediction model is called by the target computing unit to perform cross-attention calculations on the latent space features of the visible node feature matrix and the mask identification matrix. In the cross-attention calculation process, the cross-attention layer is composed of a key matrix K, a value matrix V and a query matrix Q. The query matrix Q is obtained by linear transformation of the mask identification matrix. The feature learning processes of the visible nodes and the mask nodes are decoupled to obtain the predicted latent space features of the mask identification matrix.
根据本发明的实施例,上述解耦过程中可以对掩码标识进行删除处理,使得到的掩码标识矩阵的预测隐空间特征可以作为掩码节点特征矩阵的预测隐空间特征。According to an embodiment of the present invention, the mask identifiers may be deleted during the above decoupling process, so that the predicted latent space features of the obtained mask identifier matrix can be used as the predicted latent space features of the mask node feature matrix.
根据本发明的实施例,利用目标计算单元对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征,包括:利用目标计算单元确定编码参数;利用目标计算单元对编码参数执行指数移动平均处理任务,得到目标编码参数;基于目标编码参数,利用目标计算单元对掩码节点特征矩阵执行编码处理任务,得到掩码节点特征矩阵的隐空间特征。According to an embodiment of the present invention, a target computing unit is used to perform an encoding task on a mask node feature matrix to obtain latent space features of the mask node feature matrix, including: determining encoding parameters using the target computing unit; performing an exponential moving average processing task on the encoding parameters using the target computing unit to obtain target encoding parameters; based on the target encoding parameters, performing an encoding processing task on the mask node feature matrix using the target computing unit to obtain latent space features of the mask node feature matrix.
根据本发明的实施例,利用目标计算单元对掩码节点特征矩阵执行的编码任务,与上述对可见节点特征矩阵执行的编码任务的操作流程相同,以使得消除了编码任务差异,对实现高效的任务迁移具有积极的影响。According to an embodiment of the present invention, the encoding task performed on the mask node feature matrix using the target computing unit has the same operation flow as the encoding task performed on the visible node feature matrix mentioned above, so that the difference in encoding tasks is eliminated, which has a positive impact on achieving efficient task migration.
根据本发明的实施例,编码参数用于避免所有节点都收敛到相同的特征,导致性能预测模型出现模型坍塌的问题,以维持性能预测模型的稳定。According to an embodiment of the present invention, the encoding parameters are used to prevent all nodes from converging to the same features, which would cause the performance prediction model to collapse, so as to maintain the stability of the performance prediction model.
根据本发明的实施例,编码参数不基于梯度的反向传播更新,而是通过对编码参数计算指数移动平均(Exponential Moving Average,EMA)得到目标编码参数,具体可以如公式(1)所示。According to an embodiment of the present invention, the coding parameters are not updated based on the back propagation of the gradient, but the target coding parameters are obtained by calculating the exponential moving average (EMA) of the coding parameters, which can be specifically shown in formula (1).
(1); (1);
其中,为目标编码参数,为编码参数,为超参数,的值可以根据实际情况进行设定,优选为0.99。in, is the target encoding parameter, is the encoding parameter, is a hyperparameter, The value can be set according to actual conditions, and is preferably 0.99.
根据本发明的实施例,基于目标编码参数,利用目标计算单元调用性能预测模型的自注意力层对掩码节点特征矩阵进行编码,得到掩码节点特征矩阵的隐空间特征。According to an embodiment of the present invention, based on the target encoding parameters, the target computing unit is used to call the self-attention layer of the performance prediction model to encode the mask node feature matrix to obtain the latent space features of the mask node feature matrix.
根据本发明的实施例,利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值,包括:利用目标计算单元对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行缩放余弦误差计算任务,得到第一初始损失值;利用目标计算单元确定掩码标识矩阵的预测隐空间特征对应的操作索引,操作索引表征与掩码标识矩阵对应的结构组成;基于操作索引,利用目标计算单元对掩码标识矩阵的预测隐空间特征执行构建任务,得到中间神经网络结构数据;利用目标计算单元对中间神经网络结构数据和无标签神经网络结构数据执行交叉熵计算任务,得到第二初始损失值;基于第一预设权重,利用目标计算单元对第一初始损失值和第二初始损失值执行求和任务,得到第一目标损失值。According to an embodiment of the present invention, a target computing unit is used to perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value, including: using the target computing unit to perform a scaled cosine error calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first initial loss value; using the target computing unit to determine the operation index corresponding to the predicted latent space features of the mask identification matrix, the operation index characterizing the structural composition corresponding to the mask identification matrix; based on the operation index, using the target computing unit to perform a construction task on the predicted latent space features of the mask identification matrix to obtain intermediate neural network structure data; using the target computing unit to perform a cross entropy calculation task on the intermediate neural network structure data and the unlabeled neural network structure data to obtain a second initial loss value; based on the first preset weight, using the target computing unit to perform a summation task on the first initial loss value and the second initial loss value to obtain a first target loss value.
根据本发明的实施例,对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行缩放余弦误差计算,得到第一初始损失值,具体如公式(2)所示。According to an embodiment of the present invention, a scaled cosine error calculation is performed on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first initial loss value, as specifically shown in formula (2).
(2); (2);
其中,为第一初始损失值,为掩码节点特征矩阵,为第i个掩码标识矩阵的预测隐空间特征,为第i个掩码节点特征矩阵的隐空间特征,表示向量内积运算,为可调节缩放参数。in, is the first initial loss value, is the mask node feature matrix, is the predicted latent space feature of the i-th mask identification matrix, is the latent space feature of the feature matrix of the i-th mask node, represents the vector inner product operation, is an adjustable scaling parameter.
根据本发明的实施例,掩码标识矩阵的预测隐空间特征对应的操作索引可以包括赋值索引、切片索引或步长索引等多种索引方式中的至少一种。According to an embodiment of the present invention, the operation index corresponding to the predicted latent space feature of the mask identification matrix may include at least one of multiple indexing methods such as an assignment index, a slice index or a step index.
根据本发明的实施例,以该操作索引作为重建目标,对掩码标识矩阵的预测隐空间特征进行重建,得到中间神经网络结构数据。According to an embodiment of the present invention, the predicted latent space features of the mask identification matrix are reconstructed using the operation index as a reconstruction target to obtain intermediate neural network structure data.
根据本发明的实施例,将中间神经网络结构数据作为掩码节点所表示操作的分类问题,计算中间神经网络结构数据和无标签神经网络结构数据之间的交叉熵损失,得到第二初始损失值,具体如公式(3)所示。According to an embodiment of the present invention, the intermediate neural network structure data is used as the classification problem of the operation represented by the mask node, and the cross entropy loss between the intermediate neural network structure data and the unlabeled neural network structure data is calculated to obtain a second initial loss value, as specifically shown in formula (3).
(3); (3);
其中,为第二初始损失值,为掩码节点特征矩阵,为无标签神经网络结构数据中可操作节点的数量,c为操作索引,的取值为0或1,当第i个掩码节点的实际操作索引为c时取值为1,否则取值为0,为中间神经网络结构数据的第i个掩码节点的操作索引为c的预测概率。in, is the second initial loss value, is the mask node feature matrix, is the number of operable nodes in the unlabeled neural network structure data, c is the operation index, The value of is 0 or 1. When the actual operation index of the i-th mask node is c, the value is 1, otherwise the value is 0. is the predicted probability of the operation index c of the i-th mask node of the intermediate neural network structure data.
根据本发明的实施例,第一预设权重可以是相应损失占比的权重。According to an embodiment of the present invention, the first preset weight may be a weight of a corresponding loss ratio.
根据本发明的实施例,基于第一预设权重,对第一初始损失值和第二初始损失值执行求和任务,得到第一目标损失值,具体如公式(4)所示。According to an embodiment of the present invention, based on the first preset weight, a summing task is performed on the first initial loss value and the second initial loss value to obtain a first target loss value, as specifically shown in formula (4).
(4); (4);
其中,为第一目标损失值,和为第一预设权重,和的值可以根据实际需要进行选择,优选为1。in, is the first target loss value, and is the first preset weight, and The value of can be selected according to actual needs, and is preferably 1.
根据本发明的实施例,通过确定第一目标损失值,以实现训练过程中性能预测模型不同模块的功能解耦,增强性能预测模型的特征提取和学习能力,为下游任务迁移打好基础。According to an embodiment of the present invention, by determining the first target loss value, the functional decoupling of different modules of the performance prediction model during the training process is achieved, the feature extraction and learning capabilities of the performance prediction model are enhanced, and a good foundation is laid for downstream task migration.
根据本发明的实施例,上述方法还包括:利用目标计算单元对有标签神经网络结构数据执行编码任务,确定第二目标损失值。According to an embodiment of the present invention, the above method also includes: using the target computing unit to perform an encoding task on the labeled neural network structure data to determine a second target loss value.
根据本发明的实施例,在利用第一目标损失值对性能预测模型进行训练后,对少量有标签神经网络结构数据进行编码,以得到第二目标损失值,基于第二目标损失值对性能预测模型进行进一步训练。According to an embodiment of the present invention, after the performance prediction model is trained using the first target loss value, a small amount of labeled neural network structure data is encoded to obtain a second target loss value, and the performance prediction model is further trained based on the second target loss value.
根据本发明的实施例,对利用目标计算单元对有标签神经网络结构数据执行编码处理任务,确定第二目标损失值,包括:利用目标计算单元对有标签神经网络结构数据执行编码任务,得到有标签神经网络结构数据的隐空间特征;利用目标计算单元对有标签神经网络结构数据的隐空间特征执行性能预测任务,确定预测性能值;利用目标计算单元确定与有标签神经网络结构数据对应的标签性能值;利用目标计算单元对预测性能值和标签性能值执行均方差计算任务,得到第三初始损失值;利用目标计算单元对预测性能值和标签性能值执行排序误差计算任务,得到第四初始损失值;基于第二预设权重,利用目标计算单元对第三初始损失值和第四初始损失值执行求和任务,得到第二目标损失值。According to an embodiment of the present invention, a target computing unit is used to perform an encoding processing task on labeled neural network structure data to determine a second target loss value, including: performing an encoding task on the labeled neural network structure data using the target computing unit to obtain latent space features of the labeled neural network structure data; performing a performance prediction task on the latent space features of the labeled neural network structure data using the target computing unit to determine a predicted performance value; determining a label performance value corresponding to the labeled neural network structure data using the target computing unit; performing a mean square error calculation task on the predicted performance value and the label performance value using the target computing unit to obtain a third initial loss value; performing a sorting error calculation task on the predicted performance value and the label performance value using the target computing unit to obtain a fourth initial loss value; and based on a second preset weight, performing a summation task on the third initial loss value and the fourth initial loss value using the target computing unit to obtain a second target loss value.
根据本发明的实施例,利用目标计算单元调用性能预测模型的自注意力层对有标签神经网络结构数据进行编码,对有标签神经网络结构数据中的每个矩阵元素之间的相似关系进行特征提取,以获得有标签神经网络结构数据的隐空间特征。According to an embodiment of the present invention, the target computing unit is used to call the self-attention layer of the performance prediction model to encode the labeled neural network structure data, and feature extraction is performed on the similarity relationship between each matrix element in the labeled neural network structure data to obtain the latent space features of the labeled neural network structure data.
根据本发明的实施例,利用目标计算单元调用性能预测模型的全连接网络对有标签神经网络结构数据的隐空间特征进行预测,以建立有标签神经网络结构数据的隐空间特征中的每个节点特征与神经网络性能之间的映射关系,得到预测性能值。According to an embodiment of the present invention, a fully connected network of a performance prediction model is called by a target computing unit to predict latent space features of labeled neural network structure data, so as to establish a mapping relationship between each node feature in the latent space features of the labeled neural network structure data and the neural network performance, and obtain a predicted performance value.
根据本发明的实施例,有标签神经网络结构数据对应的标签性能值可以是该有标签神经网络结构数据实际的性能需求。According to an embodiment of the present invention, the label performance value corresponding to the labeled neural network structure data may be the actual performance requirement of the labeled neural network structure data.
根据本发明的实施例,计算预测性能值和标签性能值之间的均方误差,得到第三初始损失值,具体如公式(5)所示。According to an embodiment of the present invention, the mean square error between the predicted performance value and the label performance value is calculated to obtain a third initial loss value, as specifically shown in formula (5).
(5); (5);
其中,为第三初始损失值,为预测性能值,为标签性能值。in, is the third initial loss value, To predict the performance value, is the label performance value.
根据本发明的实施例,对神经网络结构数据的性能需求进行预测本质上可以看作一个回归任务,对性能预测模型输入神经网络结构数据并预测得到神经网络结构数据的性能数值,因此,训练性能预测模型最直接的方式即采用预测性能值和标签性能值的均方误差作为损失值。According to an embodiment of the present invention, predicting the performance requirements of neural network structure data can essentially be regarded as a regression task. The neural network structure data is input into the performance prediction model and the performance value of the neural network structure data is predicted. Therefore, the most direct way to train the performance prediction model is to use the mean square error of the predicted performance value and the label performance value as the loss value.
根据本发明的实施例,计算预测性能值和标签性能值之间的排序误差,得到第四初始损失值,具体如公式(6)所示。According to an embodiment of the present invention, the ranking error between the predicted performance value and the label performance value is calculated to obtain a fourth initial loss value, as specifically shown in formula (6).
(6); (6);
其中,为第四初始损失值,m为超参数。in, is the fourth initial loss value, and m is a hyperparameter.
根据本发明的实施例,在神经网络结构搜索任务中,为确定哪种神经网络结构数据具有更好的性能数值。在衡量用于神经网络结构搜索任务中性能预测模型的准确程度时,不同神经网络结构数据之间预测性能的相对排名准确性,往往比各自绝对性能数值的预测准确性更为重要,而传统的均方误差损失无法感知排序相关的预测误差,则进一步确定预测性能值和标签性能值之间的排序误差。According to an embodiment of the present invention, in a neural network structure search task, in order to determine which neural network structure data has a better performance value. When measuring the accuracy of the performance prediction model used in the neural network structure search task, the relative ranking accuracy of the prediction performance between different neural network structure data is often more important than the prediction accuracy of their respective absolute performance values, and the traditional mean square error loss cannot perceive the ranking-related prediction error, so the ranking error between the prediction performance value and the label performance value is further determined.
根据本发明的实施例,第二预设权重可以是不同损失函数之间重要性的权重参数。According to an embodiment of the present invention, the second preset weight may be a weight parameter of importance between different loss functions.
根据本发明的实施例,基于第二预设权重,对第三初始损失值和第四初始损失值执行求和任务,得到第二目标损失值,具体如公式(7)所示。According to an embodiment of the present invention, based on the second preset weight, a summing task is performed on the third initial loss value and the fourth initial loss value to obtain a second target loss value, as specifically shown in formula (7).
(7); (7);
其中,为第二目标损失值,和为第二预设权重,和的值可以根据实际需要进行选择,优选为1。in, is the second target loss value, and is the second preset weight, and The value of can be selected according to actual needs, and is preferably 1.
根据本发明的实施例,第二预设权重结合了预测性能值和标签性能值之间的绝对误差和相对误差,可以帮助性能预测模型在具有一定的预测精度的同时提高预测结果的排序准确性,从而提高性能预测模型的准确性。According to an embodiment of the present invention, the second preset weight combines the absolute error and relative error between the predicted performance value and the label performance value, which can help the performance prediction model improve the ranking accuracy of the prediction results while maintaining a certain prediction accuracy, thereby improving the accuracy of the performance prediction model.
根据本发明的实施例,基于第一目标损失值,利用目标计算单元对神经网络预测模型执行训练任务,得到训练后的神经网络预测模型,包括:基于第一目标损失值,利用目标计算单元对性能预测模型执行训练任务,得到中间性能预测模型;基于第二目标损失值,利用目标计算单元对中间性能预测模型执行训练任务,得到训练后的性能预测模型。According to an embodiment of the present invention, based on the first target loss value, a target computing unit is used to perform a training task on a neural network prediction model to obtain a trained neural network prediction model, including: based on the first target loss value, a target computing unit is used to perform a training task on a performance prediction model to obtain an intermediate performance prediction model; based on the second target loss value, a target computing unit is used to perform a training task on the intermediate performance prediction model to obtain a trained performance prediction model.
图3示出了根据本发明实施例的性能预测模型的训练方法的示意图。FIG3 shows a schematic diagram of a method for training a performance prediction model according to an embodiment of the present invention.
如图3所示,在操作S310,对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵。在操作S320,对可见节点特征矩阵进行编码,对可见节点特征矩阵中的每个矩阵元素之间的相似关系进行特征提取,以获得可见节点特征矩阵的隐空间特征。在操作S330,基于可见节点特征矩阵的隐空间特征,对掩码标识矩阵进行预测,得到掩码标识矩阵的预测隐空间特征。在操作S340,对掩码节点特征矩阵进行编码,得到掩码节点特征矩阵的隐空间特征。在操作S350,基于掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征,确定第一目标损失值。在操作S360,对有标签神经网络结构数据执行编码处理任务,以得到有标签神经网络结构数据对应的预测性能值。在操作S370,根据预测性能值与标签性能值,确定第二目标损失值。在操作S380,基于第一目标损失值和第二目标损失值,对性能预测模型进行训练,得到训练后的性能预测模型。As shown in FIG3 , in operation S310, a mask processing task is performed on the unlabeled neural network structure data to obtain a visible node feature matrix, a mask node feature matrix, and a mask identification matrix. In operation S320, the visible node feature matrix is encoded, and the similarity relationship between each matrix element in the visible node feature matrix is feature extracted to obtain the latent space features of the visible node feature matrix. In operation S330, based on the latent space features of the visible node feature matrix, the mask identification matrix is predicted to obtain the predicted latent space features of the mask identification matrix. In operation S340, the mask node feature matrix is encoded to obtain the latent space features of the mask node feature matrix. In operation S350, a first target loss value is determined based on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix. In operation S360, an encoding processing task is performed on the labeled neural network structure data to obtain a predicted performance value corresponding to the labeled neural network structure data. In operation S370, a second target loss value is determined based on the predicted performance value and the label performance value. In operation S380, the performance prediction model is trained based on the first target loss value and the second target loss value to obtain a trained performance prediction model.
根据本发明的实施例,由于基于第一目标损失值对性能预测模型进行训练,使得在有监督学习时,仅需使用少量有标签神经网络结构数据确定第二目标损失值,并基于第二目标损失值对性能预测模型进行二次训练。与直接使用有标签神经网络结构数据训练得到的性能预测模型相比,加入基于无标签神经网络结构数据的训练,使得性能预测模型具有更高的准确性和较强的泛化性,即使加入基于无标签神经网络结构数据的训练会引入额外的时间和计算资源开销,但对于相同的搜索空间而言,只需执行一次基于第一目标损失值的训练过程,随后通过多次基于第二目标损失值的训练过程即可获得适用于不同性能指标预测任务的性能预测模型。According to an embodiment of the present invention, since the performance prediction model is trained based on the first target loss value, during supervised learning, only a small amount of labeled neural network structure data is needed to determine the second target loss value, and the performance prediction model is trained a second time based on the second target loss value. Compared with the performance prediction model obtained by directly training with labeled neural network structure data, adding training based on unlabeled neural network structure data makes the performance prediction model have higher accuracy and stronger generalization. Even if adding training based on unlabeled neural network structure data will introduce additional time and computing resource overhead, for the same search space, only one training process based on the first target loss value needs to be performed, and then a performance prediction model suitable for different performance indicator prediction tasks can be obtained through multiple training processes based on the second target loss value.
图4示出了根据本发明实施例的性能预测模型的示意图;图5示出了根据本发明实施例的自注意力层的示意图。FIG. 4 shows a schematic diagram of a performance prediction model according to an embodiment of the present invention; FIG. 5 shows a schematic diagram of a self-attention layer according to an embodiment of the present invention.
根据本发明的实施例,如图4所示,性能预测模型包括编码器410、解码器413、隐空间特征预测器411、目标编码器412和回归模块414。其中,编码器410、解码器413和目标编码器412构造相同。According to an embodiment of the present invention, as shown in Fig. 4, the performance prediction model includes an encoder 410, a decoder 413, a latent space feature predictor 411, a target encoder 412 and a regression module 414. The encoder 410, the decoder 413 and the target encoder 412 have the same structure.
根据本发明的实施例,编码器410、解码器413和目标编码器412均包括自注意力层、前馈神经网络和归一化层,该编码器410和目标编码器412均用于对输入的神经网络结构数据或隐空间特征进行编码;该解码器413用于对掩码节点执行构建任务。According to an embodiment of the present invention, the encoder 410, the decoder 413 and the target encoder 412 all include a self-attention layer, a feedforward neural network and a normalization layer. The encoder 410 and the target encoder 412 are both used to encode the input neural network structure data or latent space features; the decoder 413 is used to perform construction tasks on mask nodes.
自注意力层,用于根据输入数据中每个元素之间的相似关系提取特征信息。The self-attention layer is used to extract feature information based on the similarity relationship between each element in the input data.
其中,自注意力层的计算方式如图5所示,对输入的键矩阵K、值矩阵V、查询矩阵Q和神经网络结构数据进行掩码、归一化操作和与权重矩阵相乘等操作。自注意力层对经典自注意力计算过程中的掩码机制进行改进,基于输入神经网络结构数据的邻接矩阵和节点间最短路径矩阵构造掩码矩阵,从而增强模型对于输入数据的拓扑结构信息的感知能力。The calculation method of the self-attention layer is shown in Figure 5, which performs masking, normalization, and multiplication with the weight matrix on the input key matrix K, value matrix V, query matrix Q, and neural network structure data. The self-attention layer improves the masking mechanism in the classic self-attention calculation process, and constructs a masking matrix based on the adjacency matrix of the input neural network structure data and the shortest path matrix between nodes, thereby enhancing the model's perception of the topological structure information of the input data.
前馈神经网络,用于对自注意力层的输出进行非线性激活,该前馈神经网络包括两个全连接层和一次ReLU激活函数。A feedforward neural network is used to perform nonlinear activation on the output of the self-attention layer. The feedforward neural network includes two fully connected layers and a ReLU activation function.
归一化层,用于将上述每层网络的输入转换为近似独立同分布的数据,以提高性能预测模型训练时的收敛速度。The normalization layer is used to convert the input of each layer of the above network into approximately independent and identically distributed data to improve the convergence speed during performance prediction model training.
根据本发明的实施例,隐空间特征预测器411用于对输入的神经网络结构数据进行预测,以得到隐空间特征,该隐空间特征预测器411包括交叉注意力层、前馈神经网络和归一化层。According to an embodiment of the present invention, the latent space feature predictor 411 is used to predict the input neural network structure data to obtain the latent space features, and the latent space feature predictor 411 includes a cross attention layer, a feedforward neural network and a normalization layer.
交叉注意力层,用于根据可见节点的特征信息预测掩码节点的特征信息。The cross-attention layer is used to predict the feature information of the masked nodes based on the feature information of the visible nodes.
前馈神经网络,用于对交叉注意力层的输出进行非线性激活,该前馈神经网络包括两个全连接层和一次ReLU激活函数。A feedforward neural network is used to perform nonlinear activation on the output of the cross-attention layer. The feedforward neural network includes two fully connected layers and a ReLU activation function.
归一化层,用于将上述每层网络的输入转换为近似独立同分布的数据,以提高性能预测模型训练时的收敛速度。The normalization layer is used to convert the input of each layer of the above network into approximately independent and identically distributed data to improve the convergence speed during performance prediction model training.
根据本发明的实施例,回归模块414用于对输入的神经网络结构数据进行性能预测,以得到预测性能值。该回归模块414包括全连接网络。According to an embodiment of the present invention, the regression module 414 is used to perform performance prediction on the input neural network structure data to obtain a predicted performance value. The regression module 414 includes a fully connected network.
全连接网络,用于建立隐空间特征同神经网络性能之间的映射关系。A fully connected network is used to establish a mapping relationship between latent space features and neural network performance.
根据本发明的实施例,神经网络结构数据401中的无标签神经网络结构数据对应的可见节点特征矩阵输入至编码器410,输出可见节点特征矩阵的隐空间特征403;将可见节点特征矩阵的隐空间特征403和掩码标识矩阵404输入至隐空间特征预测器411,输出掩码标识矩阵的预测隐空间特征405;神经网络结构数据401中的掩码节点特征矩阵输入至目标编码器412,输出掩码节点特征矩阵的隐空间特征402,基于掩码标识矩阵的预测隐空间特征405和掩码节点特征矩阵的隐空间特征402,确定第一初始损失值415。掩码标识矩阵的预测隐空间特征405输入至解码器413,输出中间神经网络结构数据406,基于中间神经网络结构数据406和神经网络结构数据401中的无标签神经网络结构数据,确定第二初始损失值416,基于第一初始损失值415和第二初始损失值416,可以确定第一目标损失值。神经网络结构数据401中的有标签神经网络结构数据输入至编码器410,输出有标签神经网络结构数据的隐空间特征407;将有标签神经网络结构数据的隐空间特征407输入至回归模块414,输出预测性能值408,基于预测性能值408和标签性能值409,可以确定第二目标损失值。According to an embodiment of the present invention, the visible node feature matrix corresponding to the unlabeled neural network structure data in the neural network structure data 401 is input to the encoder 410, and the latent space features 403 of the visible node feature matrix are output; the latent space features 403 of the visible node feature matrix and the mask identification matrix 404 are input to the latent space feature predictor 411, and the predicted latent space features 405 of the mask identification matrix are output; the mask node feature matrix in the neural network structure data 401 is input to the target encoder 412, and the latent space features 402 of the mask node feature matrix are output, and the first initial loss value 415 is determined based on the predicted latent space features 405 of the mask identification matrix and the latent space features 402 of the mask node feature matrix. The predicted latent space features 405 of the mask identification matrix are input to the decoder 413, and the intermediate neural network structure data 406 is output. Based on the intermediate neural network structure data 406 and the unlabeled neural network structure data in the neural network structure data 401, the second initial loss value 416 is determined, and the first target loss value can be determined based on the first initial loss value 415 and the second initial loss value 416. The labeled neural network structure data in the neural network structure data 401 is input to the encoder 410, which outputs the latent space features 407 of the labeled neural network structure data; the latent space features 407 of the labeled neural network structure data are input to the regression module 414, which outputs the predicted performance value 408. Based on the predicted performance value 408 and the labeled performance value 409, the second target loss value can be determined.
图6示出了根据本发明实施例的性能预测方法的流程图。FIG. 6 shows a flow chart of a performance prediction method according to an embodiment of the present invention.
本发明的第二方面提供了一种性能预测方法,该性能预测方法包括:A second aspect of the present invention provides a performance prediction method, the performance prediction method comprising:
在操作S610,利用目标预测单元获取待预测神经网络结构数据。In operation S610, the target prediction unit is used to obtain the neural network structure data to be predicted.
在操作S620,利用目标预测单元,基于上述性能预测模型对待预测神经网络结构数据执行预测处理任务,得到与待预测神经网络结构数据对应的目标性能。In operation S620, a target prediction unit is used to perform a prediction processing task on the neural network structure data to be predicted based on the performance prediction model to obtain a target performance corresponding to the neural network structure data to be predicted.
根据本发明的实施例,目标预测单元可以包括中央处理单元、神经网络处理器、分布式计算单元等至少一种。According to an embodiment of the present invention, the target prediction unit may include at least one of a central processing unit, a neural network processor, a distributed computing unit, and the like.
根据本发明的实施例,基于上述性能预测模型对待预测神经网络结构数据执行预测处理任务,得到与待预测神经网络结构数据对应的目标性能,使得神经网络结构搜索中,能够减小系统负担的同时,提高搜索效率。According to an embodiment of the present invention, a prediction processing task is performed on the neural network structure data to be predicted based on the above-mentioned performance prediction model to obtain the target performance corresponding to the neural network structure data to be predicted, so that in the neural network structure search, the system burden can be reduced while improving the search efficiency.
根据本发明的实施例,对本发明性能预测方法进行测试,选择NAS-Bench-101和NAS-Bench-201这两个神经网络结构搜索数据集进行测试实验。在基于第一目标损失值的训练阶段,编码模块和目标编码器均由4层Transformer网络层构成,隐空间预测模块和解码模块分别包括使用了2层Transformer网络,其中自注意力层和交叉注意力层中的注意力头数均设置为8。除前馈神经网络层中的隐藏层特征维度大小为512外,其余各层中隐层特征维度大小均设置为128。训练过程中,针对所有数据集的训练批大小均为2048,在NAS-Bench-101和NAS-Bench-201数据集上预训练各需200轮。在预训练中,使用NAS-Bench-101和NAS-Bench-201搜索空间中全部神经网络结构数据作为训练数据。实验中均选择最大学习率为0.001的AdamW优化器,并采用带有预热的余弦衰减策略对学习率进行调整。默认掩码率r为40%,超参数τ设置为0.99,损失值中第一预设权重和均设置为1以实现不同优化目标之间的权衡。另外,为了防止过拟合,对于所有数据集,Dropout和权重衰减分别设置为0.1和1e^(-3)。在NAS-Bench-101数据集上的预训练时间平均为9个小时,在NAS-Bench-201数据集上的预训练时间平均为0.5个小时。According to an embodiment of the present invention, the performance prediction method of the present invention is tested, and two neural network structure search data sets, NAS-Bench-101 and NAS-Bench-201, are selected for testing experiments. In the training phase based on the first target loss value, the encoding module and the target encoder are both composed of 4 layers of Transformer network layers, and the latent space prediction module and the decoding module respectively include 2 layers of Transformer networks, in which the number of attention heads in the self-attention layer and the cross-attention layer are set to 8. Except for the hidden layer feature dimension size of 512 in the feedforward neural network layer, the hidden layer feature dimension size in the remaining layers is set to 128. During the training process, the training batch size for all data sets is 2048, and 200 rounds are required for pre-training on the NAS-Bench-101 and NAS-Bench-201 data sets. In pre-training, all neural network structure data in the search space of NAS-Bench-101 and NAS-Bench-201 are used as training data. In the experiment, the AdamW optimizer with a maximum learning rate of 0.001 is selected, and the learning rate is adjusted using a cosine decay strategy with preheating. The default mask rate r is 40%, the hyperparameter τ is set to 0.99, and the loss value The first preset weight and Both are set to 1 to achieve a trade-off between different optimization objectives. In addition, to prevent overfitting, Dropout and weight decay are set to 0.1 and 1e^(-3) for all datasets, respectively. The average pre-training time on the NAS-Bench-101 dataset is 9 hours, and the average pre-training time on the NAS-Bench-201 dataset is 0.5 hours.
在基于第二目标损失值的训练阶段,回归模块由隐层维度为128的2层全连接网络构成。针对所有数据集的最大训练轮数均为300轮,训练批大小根据不同数据集中训练样本的数量进行具体调整,其余参数设置与基于第一目标损失值的训练过程中保持一致。在基于第二目标损失值的训练过程中,编码器和回归模块中的所有模型参数均会随着训练过程进行更新。由于微调过程中仅需少量有标签神经网络结构数据,因此,对于每个数据集来说,训练耗时只有数分钟。During the training phase based on the second objective loss value, the regression module consists of a 2-layer fully connected network with a hidden dimension of 128. The maximum number of training rounds for all datasets is 300 rounds. The training batch size is adjusted according to the number of training samples in different datasets. The rest of the parameter settings are consistent with the training process based on the first objective loss value. During the training process based on the second objective loss value, all model parameters in the encoder and regression modules are updated as the training process progresses. Since only a small amount of labeled neural network structure data is required during fine-tuning, training takes only a few minutes for each dataset.
针对性能预测测试实验,选择肯德尔等级相关系数KTau作为性能预测模型的评估指标。KTau可以有效衡量两个有序变量之间的单调关系强弱,该指标取值范围为[-1,1],越接近1表明两组变量的排序一致性越强。因此,在实验过程中,将大量网络结构的预测性能和实际性能视作两组变量并计算二者之间的KTau数值,即可评估不同预测器的预测效果。表1给出了在NAS-Bench-101数据集上的本发明性能预测模型的性能预测表现和现有预测模型的比较结果。For the performance prediction test experiment, the Kendall rank correlation coefficient KTau was selected as the evaluation indicator of the performance prediction model. KTau can effectively measure the strength of the monotonic relationship between two ordered variables. The value range of this indicator is [-1,1]. The closer it is to 1, the stronger the ranking consistency of the two groups of variables. Therefore, in the experimental process, the predicted performance and actual performance of a large number of network structures are regarded as two groups of variables and the KTau value between the two is calculated to evaluate the prediction effect of different predictors. Table 1 shows the performance prediction performance of the performance prediction model of the present invention on the NAS-Bench-101 data set and the comparison results of the existing prediction model.
表2给出了在NAS-Bench-201数据集上的本发明性能预测模型的性能预测表现和现有预测模型的比较结果。Table 2 shows the performance prediction performance of the performance prediction model of the present invention and the comparison results of the existing prediction models on the NAS-Bench-201 dataset.
由表1可以看出,相较于其他性能预测模型,本发明性能预测模型在所有情况下都获得了较高的性能表现KTau结果,体现出了本发明性能预测模型对于神经网络性能预测任务的显著优势,而这种优势在训练数据较少时体现更为明显。在NAS-Bench-101上,本发明性能预测模型相较于同类型的先进对比模型取得了最高9.5%的KTau相对提升。As can be seen from Table 1, compared with other performance prediction models, the performance prediction model of the present invention has obtained higher performance KTau results in all cases, reflecting the significant advantages of the performance prediction model of the present invention for neural network performance prediction tasks, and this advantage is more obvious when there is less training data. On NAS-Bench-101, the performance prediction model of the present invention has achieved a relative improvement of up to 9.5% in KTau compared with the same type of advanced comparison models.
表2的实验结果与表1较为一致,本发明性能预测模型在多种性能预测任务中又进一步拉开了与其他对比模型之间的差距,获得了平均4.1%的KTau提升。以上实验充分表明本发明性能预测模型利用大量原始神经网络结构数据中蕴含的信息学习更好的表征,从而使得性能预测模型在实际训练数据有限的情况下依然具有良好的预测准确性和较强的泛化能力。另外,通过对NAS-Bench-201中网络结构在3种不同分类数据集上的任务精度预测实验,可以看出该性能预测模型对于多种精度预测任务都能够获得良好的预测效果,表明其具有较强的泛化能力。The experimental results in Table 2 are relatively consistent with those in Table 1. The performance prediction model of the present invention has further widened the gap with other comparison models in a variety of performance prediction tasks, and achieved an average KTau improvement of 4.1%. The above experiments fully demonstrate that the performance prediction model of the present invention uses the information contained in a large amount of original neural network structure data to learn better representations, so that the performance prediction model still has good prediction accuracy and strong generalization ability when the actual training data is limited. In addition, through the task accuracy prediction experiment of the network structure in NAS-Bench-201 on three different classification data sets, it can be seen that the performance prediction model can obtain good prediction results for a variety of accuracy prediction tasks, indicating that it has strong generalization ability.
基于上述性能预测模型的训练方法,本发明还提供了一种性能预测模型的训练装置。以下将结合图7对该装置进行详细描述。Based on the above performance prediction model training method, the present invention also provides a performance prediction model training device, which will be described in detail below in conjunction with FIG.
图7示意性示出了根据本发明实施例的性能预测模型的训练装置的结构框图。FIG. 7 schematically shows a structural block diagram of a training device for a performance prediction model according to an embodiment of the present invention.
如图7所示,该实施例的性能预测模型的训练装置700包括目标计算单元710,配置为:对无标签神经网络结构数据执行掩码处理任务,得到可见节点特征矩阵、掩码节点特征矩阵和掩码标识矩阵,掩码节点特征矩阵和掩码标识矩阵的矩阵维度相同;基于可见节点特征矩阵,对掩码标识矩阵执行预测任务,得到掩码标识矩阵的预测隐空间特征;对掩码节点特征矩阵执行编码任务,得到掩码节点特征矩阵的隐空间特征;对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行损失计算任务,得到第一目标损失值;基于第一目标损失值,对性能预测模型执行训练任务,得到训练后的性能预测模型。As shown in Figure 7, the training device 700 of the performance prediction model of this embodiment includes a target calculation unit 710, which is configured to: perform a mask processing task on the unlabeled neural network structure data to obtain a visible node feature matrix, a mask node feature matrix and a mask identification matrix, and the matrix dimensions of the mask node feature matrix and the mask identification matrix are the same; based on the visible node feature matrix, perform a prediction task on the mask identification matrix to obtain the predicted latent space features of the mask identification matrix; perform an encoding task on the mask node feature matrix to obtain the latent space features of the mask node feature matrix; perform a loss calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first target loss value; based on the first target loss value, perform a training task on the performance prediction model to obtain a trained performance prediction model.
根据本发明的实施例,目标计算单元710还被配置为对可见节点特征矩阵执行编码处理任务,得到可见节点特征矩阵的隐空间特征;基于可见节点特征矩阵的隐空间特征,对掩码标识矩阵执行预测处理任务,得到掩码标识矩阵的预测隐空间特征。According to an embodiment of the present invention, the target computing unit 710 is also configured to perform encoding processing tasks on the visible node feature matrix to obtain latent space features of the visible node feature matrix; based on the latent space features of the visible node feature matrix, perform prediction processing tasks on the mask identification matrix to obtain predicted latent space features of the mask identification matrix.
根据本发明的实施例,目标计算单元710还被配置为确定编码参数;对编码参数执行指数移动平均处理任务,得到目标编码参数;基于目标编码参数,对掩码节点特征矩阵执行编码处理任务,得到掩码节点特征矩阵的隐空间特征。According to an embodiment of the present invention, the target calculation unit 710 is also configured to determine the coding parameters; perform an exponential moving average processing task on the coding parameters to obtain target coding parameters; based on the target coding parameters, perform a coding processing task on the mask node feature matrix to obtain latent space features of the mask node feature matrix.
根据本发明的实施例,目标计算单元710还被配置为对掩码标识矩阵的预测隐空间特征和掩码节点特征矩阵的隐空间特征执行缩放余弦误差计算任务,得到第一初始损失值;确定掩码标识矩阵的预测隐空间特征对应的操作索引,操作索引表征与掩码标识矩阵对应的结构组成;基于操作索引,对掩码标识矩阵的预测隐空间特征执行构建任务,得到中间神经网络结构数据;对中间神经网络结构数据和无标签神经网络结构数据执行交叉熵计算任务,得到第二初始损失值;基于第一预设权重,对第一初始损失值和第二初始损失值执行求和任务,得到第一目标损失值。According to an embodiment of the present invention, the target calculation unit 710 is also configured to perform a scaled cosine error calculation task on the predicted latent space features of the mask identification matrix and the latent space features of the mask node feature matrix to obtain a first initial loss value; determine the operation index corresponding to the predicted latent space features of the mask identification matrix, the operation index characterizing the structural composition corresponding to the mask identification matrix; based on the operation index, perform a construction task on the predicted latent space features of the mask identification matrix to obtain intermediate neural network structure data; perform a cross entropy calculation task on the intermediate neural network structure data and the unlabeled neural network structure data to obtain a second initial loss value; based on the first preset weight, perform a summation task on the first initial loss value and the second initial loss value to obtain a first target loss value.
根据本发明的实施例,目标计算单元710还被配置为对有标签神经网络结构数据执行编码任务,确定第二目标损失值。According to an embodiment of the present invention, the target calculation unit 710 is also configured to perform an encoding task on the labeled neural network structure data to determine a second target loss value.
根据本发明的实施例,目标计算单元710还被配置为对有标签神经网络结构数据执行编码任务,得到有标签神经网络结构数据的隐空间特征;对有标签神经网络结构数据的隐空间特征执行性能预测任务,确定预测性能值;确定与有标签神经网络结构数据对应的标签性能值;对预测性能值和标签性能值执行均方差计算任务,得到第三初始损失值;对预测性能值和标签性能值执行排序误差计算任务,得到第四初始损失值;基于第二预设权重,对第三初始损失值和第四初始损失值执行求和任务,得到第二目标损失值。According to an embodiment of the present invention, the target calculation unit 710 is also configured to perform an encoding task on the labeled neural network structure data to obtain latent space features of the labeled neural network structure data; perform a performance prediction task on the latent space features of the labeled neural network structure data to determine a predicted performance value; determine a label performance value corresponding to the labeled neural network structure data; perform a mean square error calculation task on the predicted performance value and the label performance value to obtain a third initial loss value; perform a sorting error calculation task on the predicted performance value and the label performance value to obtain a fourth initial loss value; based on a second preset weight, perform a summation task on the third initial loss value and the fourth initial loss value to obtain a second target loss value.
根据本发明的实施例,目标计算单元710还被配置为基于第一目标损失值,对性能预测模型执行训练任务,得到中间性能预测模型;基于第二目标损失值,对中间性能预测模型执行训练任务,得到训练后的性能预测模型。According to an embodiment of the present invention, the target calculation unit 710 is also configured to perform a training task on the performance prediction model based on the first target loss value to obtain an intermediate performance prediction model; and to perform a training task on the intermediate performance prediction model based on the second target loss value to obtain a trained performance prediction model.
基于上述性能预测方法,本发明还提供了一种性能预测装置。以下将结合图8对该装置进行详细描述。Based on the above performance prediction method, the present invention further provides a performance prediction device, which will be described in detail below in conjunction with FIG.
图8示意性示出了根据本发明实施例的性能预测装置的结构框图。FIG8 schematically shows a structural block diagram of a performance prediction device according to an embodiment of the present invention.
如图8所示,该实施例的性能预测模型的训练装置800包括目标预测单元810,配置为:获取待预测神经网络结构数据;基于上述性能预测模型对待预测神经网络结构数据执行预测处理任务,得到与待预测神经网络结构数据对应的目标性能。As shown in FIG8 , the training device 800 of the performance prediction model of this embodiment includes a target prediction unit 810, which is configured to: obtain the neural network structure data to be predicted; perform a prediction processing task on the neural network structure data to be predicted based on the above-mentioned performance prediction model to obtain the target performance corresponding to the neural network structure data to be predicted.
图9示意性示出了根据本发明实施例的适于实现性能预测模型的训练方法的电子设备的方框图。FIG9 schematically shows a block diagram of an electronic device suitable for implementing a training method for a performance prediction model according to an embodiment of the present invention.
如图9所示,根据本发明实施例的电子设备900包括处理器901,其可以根据存储在只读存储器(ROM)902中的程序或者从存储部分908加载到随机访问存储器(RAM)903中的程序而执行各种适当的动作和处理。处理器901例如可以包括通用微处理器(例如CPU)、指令集处理器和/或相关芯片组和/或专用微处理器(例如,专用集成电路(ASIC))等等。处理器901还可以包括用于缓存用途的板载存储器。处理器901可以包括用于执行根据本发明实施例的方法流程的不同动作的单一处理单元或者是多个处理单元。As shown in Figure 9, the electronic device 900 according to an embodiment of the present invention includes a processor 901, which can perform various appropriate actions and processes according to the program stored in the read-only memory (ROM) 902 or the program loaded from the storage part 908 to the random access memory (RAM) 903. The processor 901 may include, for example, a general-purpose microprocessor (such as a CPU), an instruction set processor and/or a related chipset and/or a special-purpose microprocessor (for example, an application-specific integrated circuit (ASIC)), etc. The processor 901 may also include an onboard memory for caching purposes. The processor 901 may include a single processing unit or multiple processing units for performing different actions of the method flow according to an embodiment of the present invention.
在RAM 903中,存储有电子设备900操作所需的各种程序和数据。处理器 901、ROM902以及RAM 903通过总线904彼此相连。处理器901通过执行ROM 902和/或RAM 903中的程序来执行根据本发明实施例的方法流程的各种操作。需要注意,程序也可以存储在除ROM902和RAM 903以外的一个或多个存储器中。处理器901也可以通过执行存储在一个或多个存储器中的程序来执行根据本发明实施例的方法流程的各种操作。In RAM 903, various programs and data required for the operation of electronic device 900 are stored. Processor 901, ROM 902 and RAM 903 are connected to each other through bus 904. Processor 901 performs various operations of the method flow according to the embodiment of the present invention by executing the program in ROM 902 and/or RAM 903. It should be noted that the program can also be stored in one or more memories other than ROM 902 and RAM 903. Processor 901 can also perform various operations of the method flow according to the embodiment of the present invention by executing the program stored in one or more memories.
根据本发明的实施例,电子设备900还可以包括输入/输出(I/O)接口905,输入/输出(I/O)接口905也连接至总线904。电子设备900还可以包括连接至I/O接口905的以下部件中的一项或多项:包括键盘、鼠标等的输入部分906;包括诸如阴极射线管(CRT)、液晶显示器(LCD)等以及扬声器等的输出部分907;包括硬盘等的存储部分908;以及包括诸如LAN卡、调制解调器等的网络接口卡的通信部分909。通信部分909经由诸如因特网的网络执行通信处理。驱动器910也根据需要连接至I/O接口905。可拆卸介质911,诸如磁盘、光盘、磁光盘、半导体存储器等等,根据需要安装在驱动器910上,以便于从其上读出的计算机程序根据需要被安装入存储部分908。According to an embodiment of the present invention, the electronic device 900 may further include an input/output (I/O) interface 905, which is also connected to the bus 904. The electronic device 900 may further include one or more of the following components connected to the I/O interface 905: an input portion 906 including a keyboard, a mouse, etc.; an output portion 907 including a cathode ray tube (CRT), a liquid crystal display (LCD), etc., and a speaker, etc.; a storage portion 908 including a hard disk, etc.; and a communication portion 909 including a network interface card such as a LAN card, a modem, etc. The communication portion 909 performs communication processing via a network such as the Internet. A drive 910 is also connected to the I/O interface 905 as needed. A removable medium 911, such as a magnetic disk, an optical disk, a magneto-optical disk, a semiconductor memory, etc., is installed on the drive 910 as needed, so that a computer program read therefrom is installed into the storage portion 908 as needed.
本发明还提供了一种计算机可读存储介质,该计算机可读存储介质可以是上述实施例中描述的设备/装置/系统中所包含的;也可以是单独存在,而未装配入该设备/装置/系统中。上述计算机可读存储介质承载有一个或者多个程序,当上述一个或者多个程序被执行时,实现根据本发明实施例的方法。The present invention also provides a computer-readable storage medium, which may be included in the device/apparatus/system described in the above embodiment; or may exist independently without being assembled into the device/apparatus/system. The above computer-readable storage medium carries one or more programs, and when the above one or more programs are executed, the method according to the embodiment of the present invention is implemented.
根据本发明的实施例,计算机可读存储介质可以是非易失性的计算机可读存储介质,例如可以包括但不限于:便携式计算机磁盘、硬盘、随机访问存储器(RAM)、只读存储器(ROM)、可擦式可编程只读存储器(EPROM或闪存)、便携式紧凑磁盘只读存储器(CD-ROM)、光存储器件、磁存储器件、或者上述的任意合适的组合。在本发明中,计算机可读存储介质可以是任何包含或存储程序的有形介质,该程序可以被指令执行系统、装置或者器件使用或者与其结合使用。例如,根据本发明的实施例,计算机可读存储介质可以包括上文描述的ROM 902和/或RAM 903和/或ROM 902和RAM 903以外的一个或多个存储器。According to an embodiment of the present invention, the computer-readable storage medium may be a non-volatile computer-readable storage medium, for example, it may include but is not limited to: a portable computer disk, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or flash memory), a portable compact disk read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination thereof. In the present invention, the computer-readable storage medium may be any tangible medium containing or storing a program, which may be used by or in combination with an instruction execution system, an apparatus or a device. For example, according to an embodiment of the present invention, the computer-readable storage medium may include the ROM 902 and/or RAM 903 described above and/or one or more memories other than ROM 902 and RAM 903.
本发明的实施例还包括一种计算机程序产品,其包括计算机程序,该计算机程序包含用于执行流程图所示的方法的程序代码。当计算机程序产品在计算机系统中运行时,该程序代码用于使计算机系统实现本发明实施例所提供的上述方法。The embodiment of the present invention also includes a computer program product, which includes a computer program, and the computer program contains program code for executing the method shown in the flowchart. When the computer program product is run in a computer system, the program code is used to enable the computer system to implement the above method provided by the embodiment of the present invention.
在该计算机程序被处理器901执行时执行本发明实施例的系统/装置中限定的上述功能。根据本发明的实施例,上文描述的系统、装置、模块、单元等可以通过计算机程序模块来实现。The computer program executes the above functions defined in the system/device of the embodiment of the present invention when it is executed by the processor 901. According to the embodiment of the present invention, the system, device, module, unit, etc. described above can be implemented by a computer program module.
在一种实施例中,该计算机程序可以依托于光存储器件、磁存储器件等有形存储介质。在另一种实施例中,该计算机程序也可以在网络介质上以信号的形式进行传输、分发,并通过通信部分909被下载和安装,和/或从可拆卸介质911被安装。该计算机程序包含的程序代码可以用任何适当的网络介质传输,包括但不限于:无线、有线等等,或者上述的任意合适的组合。In one embodiment, the computer program may rely on tangible storage media such as optical storage devices, magnetic storage devices, etc. In another embodiment, the computer program may also be transmitted and distributed in the form of signals on a network medium, and downloaded and installed through the communication part 909, and/or installed from a removable medium 911. The program code contained in the computer program may be transmitted using any appropriate network medium, including but not limited to: wireless, wired, etc., or any suitable combination of the above.
在这样的实施例中,该计算机程序可以通过通信部分909从网络上被下载和安装,和/或从可拆卸介质911被安装。在该计算机程序被处理器901执行时,执行本发明实施例的系统中限定的上述功能。根据本发明的实施例,上文描述的系统、设备、装置、模块、单元等可以通过计算机程序模块来实现。In such an embodiment, the computer program can be downloaded and installed from the network through the communication part 909, and/or installed from the removable medium 911. When the computer program is executed by the processor 901, the above functions defined in the system of the embodiment of the present invention are performed. According to the embodiment of the present invention, the system, device, means, module, unit, etc. described above can be implemented by a computer program module.
根据本发明的实施例,可以以一种或多种程序设计语言的任意组合来编写用于执行本发明实施例提供的计算机程序的程序代码,具体地,可以利用高级过程和/或面向对象的编程语言、和/或汇编/机器语言来实施这些计算程序。程序设计语言包括但不限于诸如Java,C++,python,“C”语言或类似的程序设计语言。程序代码可以完全地在用户计算设备上执行、部分地在用户设备上执行、部分在远程计算设备上执行、或者完全在远程计算设备或服务器上执行。在涉及远程计算设备的情形中,远程计算设备可以通过任意种类的网络,包括局域网(LAN)或广域网(WAN),连接到用户计算设备,或者,可以连接到外部计算设备(例如利用因特网服务提供商来通过因特网连接)。According to an embodiment of the present invention, the program code for executing the computer program provided by the embodiment of the present invention can be written in any combination of one or more programming languages. Specifically, these computing programs can be implemented using high-level process and/or object-oriented programming languages, and/or assembly/machine languages. Programming languages include, but are not limited to, such as Java, C++, python, "C" language or similar programming languages. The program code can be executed entirely on the user computing device, partially on the user device, partially on the remote computing device, or entirely on the remote computing device or server. In the case of a remote computing device, the remote computing device can be connected to the user computing device through any type of network, including a local area network (LAN) or a wide area network (WAN), or can be connected to an external computing device (e.g., using an Internet service provider to connect through the Internet).
附图中的流程图和框图,图示了按照本发明各种实施例的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段、或代码的一部分,上述模块、程序段、或代码的一部分包含一个或多个用于实现规定的逻辑功能的可执行指令。也应当注意,在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个接连地表示的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图或流程图中的每个方框、以及框图或流程图中的方框的组合,可以用执行规定的功能或操作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。The flow chart and block diagram in the accompanying drawings illustrate the possible architecture, function and operation of the system, method and computer program product according to various embodiments of the present invention. In this regard, each box in the flow chart or block diagram can represent a module, a program segment, or a part of a code, and the above-mentioned module, program segment, or a part of a code contains one or more executable instructions for realizing the specified logical function. It should also be noted that in some alternative implementations, the functions marked in the box can also occur in a different order from the order marked in the accompanying drawings. For example, two boxes represented in succession can actually be executed substantially in parallel, and they can sometimes be executed in the opposite order, depending on the functions involved. It should also be noted that each box in the block diagram or flow chart, and the combination of the boxes in the block diagram or flow chart can be implemented with a dedicated hardware-based system that performs a specified function or operation, or can be implemented with a combination of dedicated hardware and computer instructions.
本领域技术人员可以理解,本发明的各个实施例和/或权利要求中记载的特征可以进行多种组合或/或结合,即使这样的组合或结合没有明确记载于本发明中。特别地,在不脱离本发明精神和教导的情况下,本发明的各个实施例和/或权利要求中记载的特征可以进行多种组合和/或结合。所有这些组合和/或结合均落入本发明的范围。It will be appreciated by those skilled in the art that the features described in the various embodiments and/or claims of the present invention may be combined and/or combined in various ways, even if such combinations and/or combinations are not explicitly described in the present invention. In particular, the features described in the various embodiments and/or claims of the present invention may be combined and/or combined in various ways without departing from the spirit and teachings of the present invention. All of these combinations and/or combinations fall within the scope of the present invention.
以上对本发明的实施例进行了描述。但是,这些实施例仅仅是为了说明的目的,而并非为了限制本发明的范围。尽管在以上分别描述了各实施例,但是这并不意味着各个实施例中的措施不能有利地结合使用。本发明的范围由所附权利要求及其等同物限定。不脱离本发明的范围,本领域技术人员可以做出多种替代和修改,这些替代和修改都应落在本发明的范围之内。The embodiments of the present invention are described above. However, these embodiments are only for the purpose of illustration, and are not intended to limit the scope of the present invention. Although the embodiments are described above, this does not mean that the measures in the various embodiments cannot be used in combination. The scope of the present invention is defined by the appended claims and their equivalents. Without departing from the scope of the present invention, those skilled in the art may make various substitutions and modifications, which should all fall within the scope of the present invention.
Claims (10)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202411186366.2A CN118690824B (en) | 2024-08-28 | 2024-08-28 | Training method and device of performance prediction model, and performance prediction method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202411186366.2A CN118690824B (en) | 2024-08-28 | 2024-08-28 | Training method and device of performance prediction model, and performance prediction method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN118690824A true CN118690824A (en) | 2024-09-24 |
CN118690824B CN118690824B (en) | 2024-12-31 |
Family
ID=92768795
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202411186366.2A Active CN118690824B (en) | 2024-08-28 | 2024-08-28 | Training method and device of performance prediction model, and performance prediction method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN118690824B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN119090880A (en) * | 2024-11-06 | 2024-12-06 | 东北大学 | Industrial defect image detection method based on dynamic regularized entropy differentiable architecture search |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114021836A (en) * | 2021-11-16 | 2022-02-08 | 电子科技大学 | Multivariable reservoir water inflow prediction system, training method and application based on fusion of different angles |
US20220134435A1 (en) * | 2020-10-30 | 2022-05-05 | Ut Battelle, Llc | Systems and methods for powder bed additive manufacturing anomaly detection |
-
2024
- 2024-08-28 CN CN202411186366.2A patent/CN118690824B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220134435A1 (en) * | 2020-10-30 | 2022-05-05 | Ut Battelle, Llc | Systems and methods for powder bed additive manufacturing anomaly detection |
CN114021836A (en) * | 2021-11-16 | 2022-02-08 | 电子科技大学 | Multivariable reservoir water inflow prediction system, training method and application based on fusion of different angles |
Non-Patent Citations (1)
Title |
---|
QIAOCHU LIANG 等: "NeuralMAE: Data-Efficient Neural Architecture Predictor with Masked Autoencoder", 《PATTERN RECOGNITION AND COMPUTER VISION》, 29 December 2023 (2023-12-29), pages 142 - 154, XP047679062, DOI: 10.1007/978-981-99-8543-2_12 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN119090880A (en) * | 2024-11-06 | 2024-12-06 | 东北大学 | Industrial defect image detection method based on dynamic regularized entropy differentiable architecture search |
Also Published As
Publication number | Publication date |
---|---|
CN118690824B (en) | 2024-12-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10649794B2 (en) | Aggregate features for machine learning | |
US9934260B2 (en) | Streamlined analytic model training and scoring system | |
WO2022007438A1 (en) | Emotional voice data conversion method, apparatus, computer device, and storage medium | |
US20200356729A1 (en) | Generation of text from structured data | |
CN114298417A (en) | Anti-fraud risk assessment method, anti-fraud risk training method, anti-fraud risk assessment device, anti-fraud risk training device and readable storage medium | |
CN110390408B (en) | Transaction object prediction method and device | |
CN112270547A (en) | Financial risk assessment method and device based on feature construction and electronic equipment | |
WO2022152161A1 (en) | Training and prediction of hybrid graph neural network model | |
CN117312325B (en) | Quantitative index construction method, device and equipment based on knowledge distillation | |
US12229815B2 (en) | Method, device, and computer program product for user behavior prediction | |
CN114547267A (en) | Intelligent question-answering model generation method and device, computing equipment and storage medium | |
CN118690824A (en) | Performance prediction model training method and device, performance prediction method and device | |
CN116127164B (en) | Codebook quantization model training method, search data quantization method and device thereof | |
CN113987154A (en) | Similar sentence generation model training method and related equipment based on UniLM and contrastive learning | |
CN113961720A (en) | Method for predicting entity relationship and method and device for training relationship prediction model | |
CN117670366A (en) | Risk prediction method, apparatus, device, medium, and program product | |
CN115098783A (en) | Inquiry for recommended methods, apparatus, electronic equipment, media and program products | |
CN114529309A (en) | Information auditing method and device, electronic equipment and computer readable medium | |
CN116155628B (en) | Network security detection method, training device, electronic equipment and medium | |
CN116681467A (en) | Sales predicting method based on improved WOA algorithm and related equipment thereof | |
CN116932697A (en) | Service data processing method based on rule engine optimization and related equipment | |
Zhang et al. | Small files storing and computing optimization in Hadoop parallel rendering | |
CN116228419A (en) | Information processing method, device, equipment and storage medium | |
CN115048561A (en) | Recommendation information determination method and device, electronic equipment and readable storage medium | |
CN113901278A (en) | Data search method and device based on global multi-detection and adaptive termination |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |