WO2023130656A1 - Method for generating heterougeneous multi-node interconnection topology, and storage medium - Google Patents

Method for generating heterougeneous multi-node interconnection topology, and storage medium Download PDF

Info

Publication number
WO2023130656A1
WO2023130656A1 PCT/CN2022/096236 CN2022096236W WO2023130656A1 WO 2023130656 A1 WO2023130656 A1 WO 2023130656A1 CN 2022096236 W CN2022096236 W CN 2022096236W WO 2023130656 A1 WO2023130656 A1 WO 2023130656A1
Authority
WO
WIPO (PCT)
Prior art keywords
heterogeneous multi
node
graph
interconnection topology
node interconnection
Prior art date
Application number
PCT/CN2022/096236
Other languages
French (fr)
Chinese (zh)
Inventor
杨宏斌
金良
胡克坤
赵雅倩
董刚
刘海威
蒋东东
晁银银
Original Assignee
苏州浪潮智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 苏州浪潮智能科技有限公司 filed Critical 苏州浪潮智能科技有限公司
Publication of WO2023130656A1 publication Critical patent/WO2023130656A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Definitions

  • the present application relates to the technical field of computer networks, in particular to a heterogeneous multi-node interconnection topology generation method and storage medium.
  • the computing power of a single computing node is different, as large as a server, including multiple CPUs and multiple GPU computing cards, as small as a single dedicated computing chip, including hundreds or thousands of PEs; the interface is different, such as the QPI interface of Intel CPU, Nvidia GPU Nvlink of FPGA, SRIO of FPGA; multi-node interconnection application scenarios are different, such as supercomputer, distributed computing, ultra-heterogeneous platform, network-on-chip, many-core CPU multi-core architecture and multi-PE interconnection of heterogeneous acceleration chip, among which supercomputer, Distributed computing and ultra-heterogeneous platforms use device nodes, such as CPUs, GPUs, FPGAs, and dedicated ICs; network-on-chips, many-core CPU multi-core architectures, and heterogeneous acceleration chips use core nodes, such as CPU cores, CUDA cores,
  • each computing node is also various, such as: shared bus, Crossbar switching matrix (as shown in Figure 1 and Figure 2), Ring (as shown in Figure 3 is the topology diagram of the bus and Ring), Star connection, Mesh and Torus (as shown in Figure 4 is a schematic diagram of a 2D Mesh and 2D Torus distributed switching matrix, and Figure 5 is a schematic diagram of a 2D Torus structure), etc.
  • shared bus As shown in Figure 1 and Figure 2 and Figure 2D
  • Ring is the topology diagram of the bus and Ring
  • Star connection a Mesh and Torus (as shown in Figure 4 is a schematic diagram of a 2D Mesh and 2D Torus distributed switching matrix
  • Figure 5 is a schematic diagram of a 2D Torus structure
  • the network connecting multiple independent computers is called Network
  • the interconnection network inside a chip or between multiple chips is called Fabric
  • NoC noC
  • QoS Quality of Service
  • the receiving and sending ports of each node Crossbar can increase the queue buffer to realize QoS priority control; realize more advanced flow control strategy, make full use of queues; use more advanced routing algorithm and congestion judgment algorithm to calculate which way to go to the target node more smoothly.
  • FIG. 7 shows a schematic diagram of the internal architecture of an image processing chip.
  • Local Router in the figure refers to a local router
  • Global Router refers to a global router
  • Hierarchical Star refers to a hierarchical star
  • RISC Reduced Instruction-Set Computing (reduced instruction Set computer)
  • IIE refers to Integrated Information Environment (integrated information environment)
  • ME refers to Motion Estimation (motion estimation)
  • ST refers to store (storage)
  • SM refers to Shared Memory (shared memory)
  • PMC Power Management Controller (power management controller)
  • VAE refers to Variational autoEncoder (variational automatic encoder)
  • GTMU refers to GSM Transmission Timing Management Unit for BBU (GSM main control transmission unit)
  • OGW refers to Originating Gateway device (originating gateway device )
  • FMP refers to Functional Multiprocessor Architecture (functional multiprocessor structure)
  • VPE refers to Vector Processing Element (vector processing unit)
  • LTMU refers to Local Task Management Units (local task management unit)
  • SPE refers to
  • the chip internally uses 6 Crossbar Switches to form a star network. At the same time, 4 13x13 Crossbars are connected in series to form a Ring.
  • the overall hybrid topology is adopted.
  • the inventor realized that due to the different computing power provided by different computing devices or computing cores, the interface types, numbers and bandwidths are different, and the interconnection topologies between computing nodes are various. High computing power, multiple interfaces, and multiple interconnect lines correspond to high power consumption and cost. If the computing nodes and the interconnection do not match, the computing nodes or interconnect lines will be idle.
  • the heterogeneous multi-node topology selects computing units and connection relationships according to computing tasks. Since the performance of each computing node is different, the cost is also different. If there are many interconnecting lines between two nodes, the communication bandwidth will be high, but the cost will also be high. Larger, it will also increase the probability of channel congestion.
  • the present application provides a method for generating a heterogeneous multi-node interconnection topology, including the following steps:
  • Input feature integration information into a preset generation network to generate a heterogeneous multi-node interconnection topology
  • the eigenvalues of the heterogeneous multi-node interconnection topology are obtained to ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements.
  • the generating network specifically includes: an upsampling layer, a convolutional layer, a fully connected layer, a batch normalization layer, a modified linear unit, and a Sigmoid function.
  • the low-dimensional vector representation includes a node embedding vector and a connection embedding vector; based on the following formula, the node embedding vector of the graph convolutional network model is obtained:
  • v i represents the node Represents the neighbor nodes of node v i
  • e ij represents the connection between nodes v i and v j
  • mean represents the average function
  • f c0 and f c1 represent two feed-forward networks of different sizes
  • w ij e represents the learnable 1x1 weight of the corresponding adjacent edge
  • concat represents the concatenation function, and creates a node vector based on node features
  • v i and v j both represent node.
  • the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges;
  • the first adjacency matrix is trained to obtain the optimal value A ⁇ of the first adjacency matrix.
  • the vertex features of the graph S'(V', E') are obtained based on the following formula:
  • f in is the vertex feature of graph S'(V',E')
  • k ij represents the geodesic distance between vertex j and vertex i of graph S'(V',E')
  • N*n The optimal value f j of the graph S'(V',E') obtained after the weight of any vertex in the vertices represents the vertex characteristics of the graph S(V,E).
  • the convolutional layer generates a global graph and an independent graph based on the upsampling result of the upsampling layer, and performs a convolution operation based on the global graph and the independent graph; the convolution operation specifically includes the following steps: initializing the independent graph For graph S(V,E), an independent graph is generated based on:
  • C k represents the independent graph
  • fin represents the vertex feature W ⁇ k of the graph S'(V',E') and are embedding functions ⁇ and Parameters
  • SoftMax is the normalization function
  • N represents the number of vertices in the graph S'(V',E'), ⁇ (v i ), Respectively represent two 1 ⁇ 1 convolutional layers with different initial values.
  • the eigenvalues of the heterogeneous multi-node interconnection topology are obtained based on the following formula:
  • B k represents the global graph
  • C k represents the independent graph
  • represents the parameter to adjust the weight of the independent graph
  • fin represents the vertex features of the graph S'(V', E')
  • K v represents the kernel size of the spatial dimension
  • W k A vector of weights representing the 1x1 convolution operation.
  • ensuring that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements specifically includes: combining the heterogeneous multi-node interconnection topology with the preset real heterogeneous multi-node interconnection topology based on the following formula Perform the cross entropy loss operation:
  • E represents the expected value of the distribution function
  • P data represents the distribution of actual topological samples
  • x is the real sample in P data
  • P z represents the distribution of input noise
  • D(x) represents the probability of judging the sample as correct
  • G(z ) represents the heterogeneous multi-node interconnection topology
  • z represents the input noise
  • P t represents the heterogeneous multi-node interconnection topology
  • P t ' represents the real heterogeneous multi-node interconnection topology
  • L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology
  • is the weight of the reconstruction item
  • L rec is the topology reconstruction loss
  • L cGAN is the cross-entropy loss
  • the performance parameters include noise signal, performance requirements, power consumption requirements and cost requirements.
  • ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements includes:
  • the cross-entropy loss results and topology reconstruction loss results of the heterogeneous multi-node interconnection topology and the preset real heterogeneous multi-node interconnection topology are obtained;
  • a non-volatile computer-readable storage medium storing computer-readable instructions, when the computer-readable instructions are executed by one or more processors, one or more processors are made to execute the method provided in any one of the above-mentioned embodiments.
  • the steps of the heterogeneous multi-node interconnection topology generation method are made to execute the method provided in any one of the above-mentioned embodiments.
  • Fig. 1 is a schematic diagram of Crossbar in the prior art
  • Fig. 2 is a schematic diagram of Crossbar cascading in the prior art
  • Fig. 3 is a topological structure diagram of a bus and a Ring in the prior art
  • Fig. 4 is the schematic diagram of 2D Mesh and 2D Torus distributed switch matrix in the prior art
  • Fig. 5 is the structural representation of 2D Torus in the prior art
  • Fig. 6 is a schematic diagram of the topological structure of banded rings and triple rings in the prior art
  • Fig. 7 is a schematic diagram of the internal architecture of an image processing dedicated chip in the prior art.
  • FIG. 8 is a flowchart of a method for generating a heterogeneous multi-node interconnection topology provided in one or more embodiments of the present application;
  • Fig. 9 is a schematic diagram of the application environment of the method provided in one or more embodiments of the present application.
  • FIG. 10 is a schematic diagram of the internal operation structure of the discriminant network provided in one or more embodiments of the present application.
  • FIG. 11 is a schematic diagram of the internal operation structure of the convolutional layer provided in one or more embodiments of the present application.
  • Fig. 12 is a schematic diagram of the internal operation structure of the generating network provided in one or more embodiments of the present application.
  • FIG. 8 is a flowchart of a method for generating a heterogeneous multi-node interconnection topology provided in Embodiment 1. Referring to FIG. 8 , FIG. 8 is a flowchart of a method for generating a heterogeneous multi-node interconnection topology provided in Embodiment 1. Referring to FIG. 8
  • the method provided in this embodiment is applied in the application environment shown in FIG. 9 .
  • the method of this embodiment includes the following steps:
  • Step S1 Based on the graph convolutional network model, feature extraction of node information and topological structure to obtain a low-dimensional vector representation of node information and topological structure.
  • the graph convolutional network model is based on node information library and topology library, feature node information and topology.
  • the node information library includes node models established based on characteristics such as node computing power, node core number, node interface number, and node interface broadband;
  • the topology model established by characteristics such as density and connection line length; performance parameters include performance, power consumption, cost, etc.
  • the node embedding vector of the graph convolutional network model is obtained based on the following formula:
  • v i represents the node Represents the neighbor nodes of node v i
  • e ij represents the connection between nodes v i and v j
  • mean represents the average function
  • Step S2 based on the node embedding vector, connection embedding vector and performance parameters input to the fully connected layer, the node embedding vector, connection embedding vector and performance parameters are fused in the fully connected layer to form feature integration information, and the feature integration information is input into the preset Generate networks to generate heterogeneous multi-node interconnection topologies.
  • the generation network includes, but is not limited to, an upsampling layer and a convolutional layer, as shown in FIG. 10 , which is a schematic diagram of the generation network of this application.
  • the spatial upsampling layer operates using an aggregation function defined by a graph A ⁇ that maps a graph S(V,E) with V vertices and E edges to a larger graph S'(V' , E'), by assigning different importances to new sets of vertices, the network can learn the optimal value of A ⁇ for good upsampling of the graph.
  • the upsampling process of the upsampling layer specifically includes: mapping the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V',E') Generate the first adjacency matrix and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the optimal value of the first adjacency matrix A ⁇ .
  • the graph S'(V', E') vertex features are obtained based on the following formula:
  • f in is the vertex feature of the graph S'(V', E')
  • kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E')
  • is the calculation N* Any of the n vertices
  • f j represents the vertex feature of the graph S(V, E).
  • the convolution layer performs convolution operations based on the global graph and the independent graph.
  • the operation process of the convolutional layer is shown in Figure 11. Specifically, the independent graph is initialized as graph S(V,E); the independent graph is generated based on the following formula:
  • C k represents the independent graph
  • fin represents the vertex features W ⁇ k of the graph S'(V', E') and are embedding functions ⁇ and Parameters
  • SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
  • C k represents the independent graph
  • fin represents the vertex features W ⁇ k of the graph S'(V', E') and are embedding functions ⁇ and Parameters
  • SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
  • B k represents the global graph
  • C k represents the independent graph
  • represents the parameter to adjust the weight of the independent graph
  • fin represents the vertex features of the graph S'(V', E')
  • K v represents the kernel size of the spatial dimension
  • W k A vector of weights representing the 1x1 convolution operation.
  • Step S3 Based on the pre-built discriminant network, ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements. Among them, there is a similarity between the operation structure of the discrimination network and the generation network, as shown in FIG. 12 , which is the internal structure of the discrimination network of the present application. Specifically, the discriminative network uses an aggregation matrix B ⁇ with trainable weights ⁇ that is different from the weights learned by the generator network, since the aggregation is mapped from a larger graph S'(V', E') to A smaller graph S 1 (V 1 , E 1 ), based on the following formula, obtains the vertex features of graph S 1 (V 1 , E 1 ):
  • f i is the vertex feature of graph S 1 (V 1 , E 1 );
  • f j ' is the vertex feature of graph S'(V', E');
  • k is the graph S'(V', E')
  • the geodesic distance B ⁇ between vertex j and vertex i is an aggregation matrix B ⁇ with trainable weights ⁇ .
  • ensuring that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements specifically includes: intersecting the heterogeneous multi-node interconnection topology with the preset real heterogeneous multi-node interconnection topology based on the following formula entropy loss operation;
  • E represents the expected value of the distribution function
  • P data represents the distribution of actual topological samples
  • x is the real sample in P data
  • P z represents the distribution of input noise
  • D(x) represents the probability of judging the sample as correct
  • G(z ) represents the heterogeneous multi-node interconnection topology
  • z represents the input noise
  • P t represents the heterogeneous multi-node interconnection topology
  • P t ' represents the real heterogeneous multi-node interconnection topology
  • L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology
  • is the weight of the reconstruction item
  • L rec is the topology reconstruction loss
  • L cGAN is the cross-entropy loss
  • the method of this embodiment includes the following steps: based on the graph convolutional network model, feature extraction of node information and topology to obtain low-dimensional vector representations of node information and topology; input performance parameters and low-dimensional vector representations to the fully connected layer , to generate feature integration information; input feature integration information into the preset generation network to generate heterogeneous multi-node interconnection topology; based on the pre-built discriminant network, ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements .
  • the generating network specifically includes: an upsampling layer and a convolutional layer; the convolutional layer generates a heterogeneous multi-node interconnection topology based on the upsampling processing result of the upsampling layer.
  • the low-dimensional vector representation includes a node embedding vector and a connection embedding vector; based on the following formula, the node embedding vector of the graph convolutional network model is obtained:
  • v i represents the node Represents the neighbor nodes of node v i
  • e ij represents the connection between nodes v i and v j
  • mean represents the average function
  • the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges; based on the graph S(V, E), sequentially Perform the following operations: map the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V', E' ), generate the first adjacency matrix, and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the optimal value A ⁇ of the first adjacency matrix.
  • the graph S'(V', E') vertex features are obtained based on the following formula:
  • f in is the vertex feature of the graph S'(V', E')
  • kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E')
  • is the calculation N* Any of the n vertices
  • f j represents the vertex feature of the graph S(V, E).
  • the convolutional layer generates a global graph and an independent graph based on the upsampling result of the upsampling layer, and performs a convolution operation based on the global graph and the independent graph;
  • the convolution operation specifically includes the following steps: initializing the independent graph as a graph S(V,E), which generates an independent graph based on:
  • C k represents the independent graph
  • fin represents the vertex features W ⁇ k of the graph S'(V', E') and are embedding functions ⁇ and Parameters
  • SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
  • N represents the number of vertices ⁇ (v i ) of the graph S'(V', E') Respectively represent two 1x1 convolutional layers with different initial values.
  • the eigenvalues of the heterogeneous multi-node interconnection topology are obtained based on the following formula:
  • B k represents the global graph
  • C k represents the independent graph
  • represents the parameter to adjust the weight of the independent graph
  • fin represents the vertex features of the graph S'(V', E')
  • K v represents the kernel size of the spatial dimension
  • W k A vector of weights representing the 1x1 convolution operation.
  • B k is a global map, which is unique to each layer.
  • Bk is an independent graph for learning a per-sample specific topology.
  • ⁇ and are two embedding functions, here a 1x1 convolutional layer.
  • Kv denotes the number of subgraphs, represents the residual operation
  • Represents the matrix multiplication operation is the gate that controls the importance weights of the two graphs. The importance of independent graphs in different layers is adjusted through a gating mechanism, using a different ⁇ value for each layer that is learned and updated through training.
  • ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements specifically includes:
  • heterogeneous multi-node interconnection topology and the preset real heterogeneous multi-node interconnection topology are calculated based on the following formula for cross-entropy loss;
  • E represents the expected value of the distribution function
  • P data represents the distribution of actual topological samples
  • x is the real sample in P data
  • P z represents the distribution of input noise
  • D(x) represents the probability of judging the sample as correct
  • G(z ) represents the heterogeneous multi-node interconnection topology
  • z represents the input noise
  • P t represents the heterogeneous multi-node interconnection topology
  • P t ' represents the real heterogeneous multi-node interconnection topology
  • L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology
  • is the weight of the reconstruction item
  • L rec is the topology reconstruction loss
  • L cGAN is the cross-entropy loss
  • the upsampling operation includes: the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges; Perform the following operations: map the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V', E' ), generate the first adjacency matrix, and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the best value A ⁇ of the first adjacency matrix; Get the best value A ⁇ , get the vertex features of the graph S'(V', E').
  • f in is the vertex feature of the graph S'(V', E')
  • kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E')
  • is the calculation N* The optimal value of the graph S'(V', E') obtained after the weight of any vertex in the n vertices
  • f j represents the vertex feature of the graph S(V, E).
  • the graph convolution operation includes the convolution layer generating a global graph and an independent graph based on the upsampling results of the upsampling layer, and performing convolution operations based on the global graph and the independent graph; the convolution operation specifically includes the following steps: initializing the independent graph as a graph S( V,E), generating an independent graph based on:
  • C k represents the independent graph
  • fin represents the vertex features W ⁇ k of the graph S'(V', E') and are embedding functions ⁇ and Parameters
  • SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
  • N represents the number of vertices ⁇ (v i ) of the graph S'(V', E'), Respectively represent two 1x1 convolutional layers with different initial values.
  • B k represents the global graph
  • C k represents the independent graph
  • represents the parameter to adjust the weight of the independent graph
  • fin represents the vertex features of the graph S'(V', E')
  • K v represents the kernel size of the spatial dimension
  • W k represents 1x1 A vector of weights for the convolution operation.
  • the performance parameters include noise signal, performance requirements, power consumption requirements and cost requirements.
  • This embodiment provides a non-volatile computer-readable storage medium storing computer-readable instructions.
  • the program When the program is executed by one or more processors, one or more processors execute the program provided in any of the above-mentioned embodiments.
  • the steps of the heterogeneous multi-node interconnect topology generation method are described in detail below.
  • the embodiments in the embodiments of the present application may be provided as methods, systems, or computer program products. Therefore, the embodiment of the present application may be in the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Moreover, the embodiments of the present application may take the form of computer program products implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes therein. .
  • Embodiments of the present application are described with reference to flowcharts and/or block diagrams of methods, devices (systems), and computer program products according to embodiments of the present application. It should be understood that each procedure and/or block in the flowchart and/or block diagram, and a combination of procedures and/or blocks in the flowchart and/or block diagram can be realized by computer program instructions. These computer program instructions may be provided to a general purpose computer, special purpose computer, embedded processor, or processor of other programmable data processing equipment to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing equipment produce a An apparatus for realizing the functions specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.
  • These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to operate in a specific manner, such that the instructions stored in the computer-readable memory produce an article of manufacture comprising instruction means, the instructions
  • the device realizes the function specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Design And Manufacture Of Integrated Circuits (AREA)

Abstract

The present application relates to a method for generating a heterogeneous multi-node interconnection topology, and a storage medium. The method comprises the following steps: on the basis of a graph convolutional network model, feature extraction node information, and a topological structure, acquiring node information, and a low-dimensional vector representation of the topological structure; inputting a performance parameter and the low-dimensional vector representation to a fully connected layer, so as to generate feature integration information; inputting the feature integration information into a preset generation network, so as to generate a heterogeneous multi-node interconnection topological structure; and acquiring a feature value of the heterogeneous multi-node interconnection topological structure, so as to ensure that the heterogeneous multi-node interconnection topological structure meets a preset accuracy requirement.

Description

一种异构多节点互联拓扑生成方法和存储介质A heterogeneous multi-node interconnection topology generation method and storage medium
相关申请的交叉引用Cross References to Related Applications
本申请要求于2022年01月10日提交中国专利局,申请号为202210024578.5,申请名称为“一种异构多节点互联拓扑生成方法和存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application submitted to the China Patent Office on January 10, 2022, with the application number 202210024578.5, and the application name is "a heterogeneous multi-node interconnection topology generation method and storage medium", the entire content of which is passed References are incorporated in this application.
技术领域technical field
本申请涉及计算机网络技术领域,尤其是指一种异构多节点互联拓扑生成方法和存储介质。The present application relates to the technical field of computer networks, in particular to a heterogeneous multi-node interconnection topology generation method and storage medium.
背景技术Background technique
在目前的异构计算领域中,存在多种类型的计算设备,如CPU、GPU、FPGA和专用IC等。单个计算节点的算力不同,大到服务器,包含多个CPU和多张GPU计算卡,小到单个专用计算芯片,包含几百上千个PE;接口不同,如Intel CPU的QPI接口,Nvidia GPU的Nvlink,FPGA的SRIO;多节点的互联应用场景不同,比如超级计算机、分布式计算、超异构平台、片上网络、众核CPU多核架构和异构加速芯片多PE互联等,其中超级计算机、分布式计算和超异构平台这些使用设备节点,如CPU、GPU、FPGA和专用IC等;片上网络、众核CPU多核架构和异构加速芯片使用核心节点,如CPU核心、CUDA核心和PE阵列等。In the current heterogeneous computing field, there are various types of computing devices, such as CPUs, GPUs, FPGAs, and dedicated ICs. The computing power of a single computing node is different, as large as a server, including multiple CPUs and multiple GPU computing cards, as small as a single dedicated computing chip, including hundreds or thousands of PEs; the interface is different, such as the QPI interface of Intel CPU, Nvidia GPU Nvlink of FPGA, SRIO of FPGA; multi-node interconnection application scenarios are different, such as supercomputer, distributed computing, ultra-heterogeneous platform, network-on-chip, many-core CPU multi-core architecture and multi-PE interconnection of heterogeneous acceleration chip, among which supercomputer, Distributed computing and ultra-heterogeneous platforms use device nodes, such as CPUs, GPUs, FPGAs, and dedicated ICs; network-on-chips, many-core CPU multi-core architectures, and heterogeneous acceleration chips use core nodes, such as CPU cores, CUDA cores, and PE arrays wait.
各计算节点的互联方式也多种多样,典型的如:共享总线、Crossbar交换矩阵(如图1和图2所示)、Ring(如图3所示即为总线和Ring的拓扑结构图)、星形连接、Mesh和Torus(如图4所示即为2D Mesh和2D Torus分布式交换矩阵示意图,如图5所示即为2D Torus的结构示意图)等。通常将连接多台独立计算机的网络称为Network,将芯片内部或者多个芯片之间的互联网络称为Fabric,将嵌入到单个芯片内部的用于连接片内多个不同模块的大规模互联网络称为NoC。不同于计算机网络,在NoC中可以实现更多特色技术,如QoS(Quality of Service),可以决定先转发哪个请求,后转发哪个,每个节点Crossbar的接收和发送端口可以增加队列缓冲,以实现QoS优先级控制;实现更先进的流量控制策略,更加充分地利用队列;利用更先进的路由算法和拥塞判断算法,算出从哪条路走到目标节点更加顺畅。常用的Fabric拓扑如HyperCube超立方也是Intel QPI所使用的拓扑,胖树(Fat Tree)拓扑也是天河II超级计算机连接大量计算机节点使用的拓扑,金字塔(Pyramid)拓扑,蝴蝶(Butterfly)拓扑,Intel在其12核心的Ivy Bridge CPU微架构中串接其12核心使用的带弦环和三重环(Triple Ring),带弦环(如图6所示即为带弦环和三重环拓扑结构示意图),带环立方体和ClosNetwork拓扑等。如图7所示的图像处理专用芯片内部架构示意图,图中的Local Router是指局部路由器,Global Router是指全局路由器,Hierarchical Star是指分级星形,RISC是指Reduced Instruction-Set Computing(精简指令集计算机),IIE是指Integrated Information Environment(综合信息环境),ME是指Motion Estimation(运动估测),ST是指store(存储),SM是指Shared Memory(共享内存),PMC是指Power Management Controller(电源管理控制器),VAE是指Variational autoEncoder(变分自动编码器),GTMU是指GSM Transmission Timing Management Unit for BBU(GSM主控传输单元),OGW是指Originating Gateway device(始发网关设备),FMP是指Functional  Multiprocessor Architecture(功能型多处理器结构),VPE是指Vector Processing Element(向量处理单元),LTMU是指Local Task Management Units(本地任务管理单元),SPE是指Scalar Processing Element(标量处理单元),FEC是指Forwarding Equivalence Class(转发等价类)。该芯片内部使用了6个Crossbar Switch形成了一个星形网络,同时又将4个13x13的Crossbar串接起来形成了一个Ring,总体采用混合拓扑方式。其中,发明人意识到,由于不同的计算设备或计算核心能提供的算力不同,其接口类型、数量和带宽不同,各计算节点之间的互联拓扑又多种多样。算力高、接口多、互联线多对应的功耗和成本高,如果计算节点和互联不匹配会造成计算节点空闲或互联线空闲,因此在部署一个计算任务或设计异构多节点硬件电路时,节点类型、数量如何选择,采用哪种拓扑,各节点如何互联,使其在满足计算性能的情况下尽量低功耗低成本,是一个优化问题。The interconnection methods of each computing node are also various, such as: shared bus, Crossbar switching matrix (as shown in Figure 1 and Figure 2), Ring (as shown in Figure 3 is the topology diagram of the bus and Ring), Star connection, Mesh and Torus (as shown in Figure 4 is a schematic diagram of a 2D Mesh and 2D Torus distributed switching matrix, and Figure 5 is a schematic diagram of a 2D Torus structure), etc. Usually, the network connecting multiple independent computers is called Network, the interconnection network inside a chip or between multiple chips is called Fabric, and the large-scale interconnection network embedded in a single chip to connect multiple different modules in a chip Called NoC. Different from computer networks, more characteristic technologies can be implemented in NoC, such as QoS (Quality of Service), which can decide which request to forward first and which one to forward later. The receiving and sending ports of each node Crossbar can increase the queue buffer to realize QoS priority control; realize more advanced flow control strategy, make full use of queues; use more advanced routing algorithm and congestion judgment algorithm to calculate which way to go to the target node more smoothly. Commonly used Fabric topologies such as HyperCube are also used by Intel QPI, Fat Tree topology is also used by Tianhe II supercomputer to connect a large number of computer nodes, Pyramid topology, Butterfly topology, Intel in Its 12-core Ivy Bridge CPU micro-architecture is connected in series with stringed rings and triple rings (Triple Ring) used by its 12 cores. Ringed Cube and ClosNetwork topology etc. Figure 7 shows a schematic diagram of the internal architecture of an image processing chip. Local Router in the figure refers to a local router, Global Router refers to a global router, Hierarchical Star refers to a hierarchical star, and RISC refers to Reduced Instruction-Set Computing (reduced instruction Set computer), IIE refers to Integrated Information Environment (integrated information environment), ME refers to Motion Estimation (motion estimation), ST refers to store (storage), SM refers to Shared Memory (shared memory), PMC refers to Power Management Controller (power management controller), VAE refers to Variational autoEncoder (variational automatic encoder), GTMU refers to GSM Transmission Timing Management Unit for BBU (GSM main control transmission unit), OGW refers to Originating Gateway device (originating gateway device ), FMP refers to Functional Multiprocessor Architecture (functional multiprocessor structure), VPE refers to Vector Processing Element (vector processing unit), LTMU refers to Local Task Management Units (local task management unit), SPE refers to Scalar Processing Element ( Scalar processing unit), FEC refers to Forwarding Equivalence Class (forwarding equivalence class). The chip internally uses 6 Crossbar Switches to form a star network. At the same time, 4 13x13 Crossbars are connected in series to form a Ring. The overall hybrid topology is adopted. Among them, the inventor realized that due to the different computing power provided by different computing devices or computing cores, the interface types, numbers and bandwidths are different, and the interconnection topologies between computing nodes are various. High computing power, multiple interfaces, and multiple interconnect lines correspond to high power consumption and cost. If the computing nodes and the interconnection do not match, the computing nodes or interconnect lines will be idle. Therefore, when deploying a computing task or designing a heterogeneous multi-node hardware circuit , how to choose the type and number of nodes, which topology to use, and how to interconnect each node to make it as low as possible in the case of satisfying the computing performance is an optimization problem.
也就是说,不同于芯片的布局布线任务其网表中确定的单元和连接关系,异构多节点拓扑是根据计算任务来选取计算单元和连接关系。由于每个计算节点的性能不同,成本也不同,两节点间互联线多,其通信带宽就高,但成本也高;相反,如果采用级联方式,其级数越多,对应的延迟可能越大,也会增大通道拥塞的概率。That is to say, different from the units and connection relationships determined in the netlist of chip placement and routing tasks, the heterogeneous multi-node topology selects computing units and connection relationships according to computing tasks. Since the performance of each computing node is different, the cost is also different. If there are many interconnecting lines between two nodes, the communication bandwidth will be high, but the cost will also be high. Larger, it will also increase the probability of channel congestion.
发明内容Contents of the invention
本申请提供了一种异构多节点互联拓扑生成方法,包括以下步骤:The present application provides a method for generating a heterogeneous multi-node interconnection topology, including the following steps:
基于图卷积网络模型,特征提取节点信息和拓扑结构,以获取节点信息和拓扑结构的低维向量表示;Based on the graph convolutional network model, feature extraction node information and topology to obtain low-dimensional vector representation of node information and topology;
输入性能参数和低维向量表示至全连接层,以生成特征整合信息;Input performance parameters and low-dimensional vector representations to fully connected layers to generate feature integration information;
将特征整合信息输入预设的生成网络,以生成异构多节点互联拓扑结构;及Input feature integration information into a preset generation network to generate a heterogeneous multi-node interconnection topology; and
获取异构多节点互联拓扑结构的特征值,确保异构多节点互联拓扑结构满足预设的准确性要求。The eigenvalues of the heterogeneous multi-node interconnection topology are obtained to ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements.
在本申请的一个实施例中,生成网络具体包括:上采样层、卷积层、全连接层、批归一化层、修正线性单元和S型函数。In one embodiment of the present application, the generating network specifically includes: an upsampling layer, a convolutional layer, a fully connected layer, a batch normalization layer, a modified linear unit, and a Sigmoid function.
在本申请的一个实施例中,低维向量表示包括节点嵌入向量和连接嵌入向量;基于下式,获取得到所述图卷积网络模型的节点嵌入向量:In one embodiment of the present application, the low-dimensional vector representation includes a node embedding vector and a connection embedding vector; based on the following formula, the node embedding vector of the graph convolutional network model is obtained:
Figure PCTCN2022096236-appb-000001
Figure PCTCN2022096236-appb-000001
其中,v i表示节点
Figure PCTCN2022096236-appb-000002
表示节点v i的邻居节点,e ij表示节点v i和v j的连接,mean表示平均值函数;基于下式,获取得到图卷积网络模型的连接嵌入向量:
Among them, v i represents the node
Figure PCTCN2022096236-appb-000002
Represents the neighbor nodes of node v i , e ij represents the connection between nodes v i and v j , mean represents the average function; based on the following formula, the connection embedding vector of the graph convolutional network model is obtained:
Figure PCTCN2022096236-appb-000003
Figure PCTCN2022096236-appb-000003
其中,f c0和f c1表示两个不同大小的前馈网络,w ij e表示对应邻边的可学习的1x1权重,concat表示拼接函数,基于节点特征创建节点向量,v i和v j均表示节点。 Among them, f c0 and f c1 represent two feed-forward networks of different sizes, w ij e represents the learnable 1x1 weight of the corresponding adjacent edge, concat represents the concatenation function, and creates a node vector based on node features, and v i and v j both represent node.
在本申请的一个实施例中,上采样层的上采样过程具体包括:假设特征整合信息为一个包含V个顶点和E个邻边的图S(V,E);In one embodiment of the present application, the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges;
基于图S(V,E)依次进行如下操作:将图S(V,E)映射到一个包含N*n个顶点和E*m个邻边的图 S'(V',E');Based on the graph S(V,E), the following operations are performed sequentially: map the graph S(V,E) to a graph S'(V',E') containing N*n vertices and E*m adjacent edges;
基于图S'(V',E'),生成第一邻接矩阵,并获得第一邻接矩阵的初始值;及Based on the graph S'(V', E'), generate a first adjacency matrix, and obtain an initial value of the first adjacency matrix; and
基于第一邻接矩阵的初始值,对第一邻接矩阵进行训练,以获得第一邻接矩阵的最佳值A ωBased on the initial value of the first adjacency matrix, the first adjacency matrix is trained to obtain the optimal value A ω of the first adjacency matrix.
在本申请的一个实施例中,基于下式获得图S'(V',E')的顶点特征:In one embodiment of the present application, the vertex features of the graph S'(V', E') are obtained based on the following formula:
Figure PCTCN2022096236-appb-000004
Figure PCTCN2022096236-appb-000004
其中,f in为图S'(V',E')的顶点特征,k ij表示图S'(V',E')的顶点j和顶点i之间的测地距离,为计算N*n个顶点中任一顶点的权重后得到的图S'(V',E')的最佳值f j表示图S(V,E)的顶点特征。 Among them, f in is the vertex feature of graph S'(V',E'), k ij represents the geodesic distance between vertex j and vertex i of graph S'(V',E'), for calculating N*n The optimal value f j of the graph S'(V',E') obtained after the weight of any vertex in the vertices represents the vertex characteristics of the graph S(V,E).
在本申请的一个实施例中,卷积层基于上采样层的上采样结果生成全局图和独立图,并基于全局图和独立图进行卷积运算;卷积运算具体包括以下步骤:初始化独立图为图S(V,E),基于下式生成独立图:In one embodiment of the present application, the convolutional layer generates a global graph and an independent graph based on the upsampling result of the upsampling layer, and performs a convolution operation based on the global graph and the independent graph; the convolution operation specifically includes the following steps: initializing the independent graph For graph S(V,E), an independent graph is generated based on:
Figure PCTCN2022096236-appb-000005
Figure PCTCN2022096236-appb-000005
其中,C k表示独立图,f in表示图S'(V',E')的顶点特征W θk
Figure PCTCN2022096236-appb-000006
分别为嵌入函数θ和
Figure PCTCN2022096236-appb-000007
的参数,SoftMax为归一化函数;其中,归一化函数为:
Among them, C k represents the independent graph, fin represents the vertex feature W θk of the graph S'(V',E') and
Figure PCTCN2022096236-appb-000006
are embedding functions θ and
Figure PCTCN2022096236-appb-000007
Parameters, SoftMax is the normalization function; where, the normalization function is:
Figure PCTCN2022096236-appb-000008
Figure PCTCN2022096236-appb-000008
其中,N表示图S'(V',E')的顶点数量,θ(v i)、
Figure PCTCN2022096236-appb-000009
分别表示两个初始值不等的1×1卷积层。
Among them, N represents the number of vertices in the graph S'(V',E'), θ(v i ),
Figure PCTCN2022096236-appb-000009
Respectively represent two 1×1 convolutional layers with different initial values.
在本申请的一个实施例中,基于下式获取异构多节点互联拓扑结构的特征值:In one embodiment of the present application, the eigenvalues of the heterogeneous multi-node interconnection topology are obtained based on the following formula:
Figure PCTCN2022096236-appb-000010
Figure PCTCN2022096236-appb-000010
其中,B k表示全局图,C k表示独立图α表示调整独立图权重的参数,f in表示图S'(V',E')的顶点特征,K v表示空间维的核大小,W k表示1x1卷积操作的权重向量。 Among them, B k represents the global graph, C k represents the independent graph α represents the parameter to adjust the weight of the independent graph, fin represents the vertex features of the graph S'(V', E'), K v represents the kernel size of the spatial dimension, W k A vector of weights representing the 1x1 convolution operation.
在本申请的一个实施例中,确保异构多节点互联拓扑结构满足预设的准确性要求具体包括:将异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构基于下式进行交叉熵损失运算:In an embodiment of the present application, ensuring that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements specifically includes: combining the heterogeneous multi-node interconnection topology with the preset real heterogeneous multi-node interconnection topology based on the following formula Perform the cross entropy loss operation:
Figure PCTCN2022096236-appb-000011
Figure PCTCN2022096236-appb-000011
其中,E表示分布函数的期望值,P data代表实际拓扑样本的分布,x是P data中的真实样本,P z表示输入噪音的分布,D(x)表示判别样本为正确的概率,G(z)表示异构多节点互联拓扑图;z表示输入噪音;基于下式,获取拓扑重建损失结果: Among them, E represents the expected value of the distribution function, P data represents the distribution of actual topological samples, x is the real sample in P data , P z represents the distribution of input noise, D(x) represents the probability of judging the sample as correct, G(z ) represents the heterogeneous multi-node interconnection topology; z represents the input noise; based on the following formula, the topology reconstruction loss result is obtained:
Figure PCTCN2022096236-appb-000012
Figure PCTCN2022096236-appb-000012
其中,P t表示异构多节点互联拓扑结构P t'表示真实异构多节点互联拓扑结构,L topo表示异构多节点互联拓扑结构与真实异构多节点互联拓扑结构对应节点的拓扑距离;根据拓扑重建损失与交叉熵损失运算结果,基于下式获取异构多节点互联拓扑结构的最终损失: Among them, P t represents the heterogeneous multi-node interconnection topology P t 'represents the real heterogeneous multi-node interconnection topology, L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology; According to the calculation results of topology reconstruction loss and cross-entropy loss, the final loss of the heterogeneous multi-node interconnection topology is obtained based on the following formula:
L=L cGAN+λL rec L=L cGAN +λL rec
其中,λ为重建项加权L rec为拓扑重建损失L cGAN为交叉熵损失;将异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若异构多节点互联拓扑结构的最终损失大于预设的异构多节点互联拓扑结构的损失,则重复执行生成异构多节点互联拓扑结构直至异构多节点互联拓扑结构的最终损失不大于预设的异构多节点互联拓扑结构的损失,确保异构多节点互联拓扑结构满足预设的准确性要求。 Among them, λ is the weight of the reconstruction item L rec is the topology reconstruction loss L cGAN is the cross-entropy loss; compare the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the heterogeneous The final loss of the multi-node interconnection topology is greater than the preset loss of the heterogeneous multi-node interconnection topology, then repeat the execution to generate the heterogeneous multi-node interconnection topology until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset heterogeneous The loss of the heterogeneous multi-node interconnection topology ensures that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements.
在本申请的一个实施例中,性能参数包括噪音信号、性能要求、功耗要求和成本要求。In one embodiment of the present application, the performance parameters include noise signal, performance requirements, power consumption requirements and cost requirements.
在本申请的一个实施例中,确保异构多节点互联拓扑结构满足预设的准确性要求,包括:In an embodiment of the present application, ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements includes:
基于异构多节点互联拓扑结构的特征值,获得异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构的交叉熵损失结果及拓扑重建损失结果;Based on the eigenvalues of the heterogeneous multi-node interconnection topology, the cross-entropy loss results and topology reconstruction loss results of the heterogeneous multi-node interconnection topology and the preset real heterogeneous multi-node interconnection topology are obtained;
基于交叉熵损失结果及拓扑重建损失结果,获得异构多节点互联拓扑结构的最终损失;及Based on the cross-entropy loss result and the topology reconstruction loss result, the final loss of the heterogeneous multi-node interconnection topology is obtained; and
将异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若异构多节点互联拓扑结构的最终损失大于预设的异构多节点互联拓扑结构的损失,则重复执行上采样层的上采样过程和卷积层的卷积运算直至异构多节点互联拓扑结构的最终损失不大于预设的异构多节点互联拓扑结构的损失。Comparing the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the final loss of the heterogeneous multi-node interconnection topology is greater than the preset heterogeneous multi-node interconnection topology loss, the upsampling process of the upsampling layer and the convolution operation of the convolutional layer are repeated until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset loss of the heterogeneous multi-node interconnection topology.
为实现上述目的,本申请还提出第二技术方案:In order to achieve the above object, the application also proposes a second technical solution:
一种存储有计算机可读指令的非易失性计算机可读存储介质,当计算机可读指令被一个或多个处理器执行时,使得一个或多个处理器执行上述任一实施例中提供的异构多节互联拓扑生成方法的步骤。A non-volatile computer-readable storage medium storing computer-readable instructions, when the computer-readable instructions are executed by one or more processors, one or more processors are made to execute the method provided in any one of the above-mentioned embodiments. The steps of the heterogeneous multi-node interconnection topology generation method.
本申请的一个或多个实施例的细节在下面的附图和描述中提出。本申请的其它特征和优点将从说明书、附图以及权利要求书变得明显。The details of one or more embodiments of the application are set forth in the accompanying drawings and the description below. Other features and advantages of the application will be apparent from the description, drawings, and claims.
附图说明Description of drawings
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其它的附图。In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings that need to be used in the description of the embodiments will be briefly introduced below. Obviously, the drawings in the following description are only some embodiments of the present application. For those skilled in the art, other drawings can also be obtained based on these drawings without creative effort.
图1是现有技术中Crossbar示意图;Fig. 1 is a schematic diagram of Crossbar in the prior art;
图2是现有技术中Crossbar级联示意图;Fig. 2 is a schematic diagram of Crossbar cascading in the prior art;
图3是现有技术中总线和Ring的拓扑结构图;Fig. 3 is a topological structure diagram of a bus and a Ring in the prior art;
图4是现有技术中2D Mesh和2D Torus分布式交换矩阵示意图;图5是现有技术中2D Torus的结构 示意图;Fig. 4 is the schematic diagram of 2D Mesh and 2D Torus distributed switch matrix in the prior art; Fig. 5 is the structural representation of 2D Torus in the prior art;
图6是现有技术中带弦环和三重环拓扑结构示意图;Fig. 6 is a schematic diagram of the topological structure of banded rings and triple rings in the prior art;
图7是现有技术中图像处理专用芯片内部架构示意图;Fig. 7 is a schematic diagram of the internal architecture of an image processing dedicated chip in the prior art;
图8是本申请一个或多个实施例中提供的异构多节互联拓扑生成方法的流程图;FIG. 8 is a flowchart of a method for generating a heterogeneous multi-node interconnection topology provided in one or more embodiments of the present application;
图9是本申请一个或多个实施例中提供的方法的应用环境示意图;Fig. 9 is a schematic diagram of the application environment of the method provided in one or more embodiments of the present application;
图10是的本申请一个或多个实施例中提供判别网络的内部运算结构示意图;FIG. 10 is a schematic diagram of the internal operation structure of the discriminant network provided in one or more embodiments of the present application;
图11是的本申请一个或多个实施例中提供卷积层的内部运算结构示意图;FIG. 11 is a schematic diagram of the internal operation structure of the convolutional layer provided in one or more embodiments of the present application;
图12是的本申请一个或多个实施例中提供生成网络的内部运算结构示意图。Fig. 12 is a schematic diagram of the internal operation structure of the generating network provided in one or more embodiments of the present application.
具体实施方式Detailed ways
为使本申请的目的、技术方案和优点更加清楚,下面将结合本申请的实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其它实施例,都属于本申请保护的范围。In order to make the purpose, technical solutions and advantages of the application clearer, the technical solutions in the embodiments of the application will be clearly and completely described below in conjunction with the accompanying drawings in the embodiments of the application. Obviously, the described embodiments are only It is a part of the embodiments of this application, not all of them. Based on the embodiments in this application, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the scope of protection of this application.
实施例一:Embodiment one:
参照图8所示,图8为实施例一提供给的异构多节互联拓扑生成方法流程图。Referring to FIG. 8 , FIG. 8 is a flowchart of a method for generating a heterogeneous multi-node interconnection topology provided in Embodiment 1. Referring to FIG.
本实施例提供的该方法应用于如图9所示的应用环境中。本实施例的方法,包括以下步骤:The method provided in this embodiment is applied in the application environment shown in FIG. 9 . The method of this embodiment includes the following steps:
步骤S1、基于图卷积网络模型,特征提取节点信息和拓扑结构,以获取节点信息和拓扑结构的低维向量表示。Step S1. Based on the graph convolutional network model, feature extraction of node information and topological structure to obtain a low-dimensional vector representation of node information and topological structure.
在其中一个实施方式中,图卷积网络模型基于节点信息库和拓扑结构库,特征节点信息和拓扑结构。其中,节点信息库包括分别基于节点算力、节点核心数量、节点接口数量和节点接口宽带等特征建立的节点模型;所述拓扑结构库包括基于连接方式、连接点数量、连接级数、连接线密度、连接线长度等特征建立的拓扑结构模型;性能参数包括性能、功耗、成本等。In one of the implementations, the graph convolutional network model is based on node information library and topology library, feature node information and topology. Among them, the node information library includes node models established based on characteristics such as node computing power, node core number, node interface number, and node interface broadband; The topology model established by characteristics such as density and connection line length; performance parameters include performance, power consumption, cost, etc.
在其中一个实施方式中,基于性能参数、节点信息和拓扑结构的低维向量表示,基于下式,获取得到图卷积网络模型的节点嵌入向量:In one of the implementations, based on the low-dimensional vector representation of performance parameters, node information and topology, the node embedding vector of the graph convolutional network model is obtained based on the following formula:
Figure PCTCN2022096236-appb-000013
Figure PCTCN2022096236-appb-000013
其中,v i表示节点
Figure PCTCN2022096236-appb-000014
表示节点v i的邻居节点,e ij表示节点v i和v j的连接,mean表示平均值函数;基于下式,获取得到图卷积网络模型的连接嵌入向量:
Among them, v i represents the node
Figure PCTCN2022096236-appb-000014
Represents the neighbor nodes of node v i , e ij represents the connection between nodes v i and v j , mean represents the average function; based on the following formula, the connection embedding vector of the graph convolutional network model is obtained:
Figure PCTCN2022096236-appb-000015
Figure PCTCN2022096236-appb-000015
其中,
Figure PCTCN2022096236-appb-000016
Figure PCTCN2022096236-appb-000017
表示两个不同大小的前馈网络,
Figure PCTCN2022096236-appb-000018
是对应于邻边的可学习的1x1权重,concat表示拼接函数,基于节点特征创建节点向量,v i和v j均表示节点。
in,
Figure PCTCN2022096236-appb-000016
and
Figure PCTCN2022096236-appb-000017
Denotes two feed-forward networks of different sizes,
Figure PCTCN2022096236-appb-000018
is the learnable 1x1 weight corresponding to the adjacent edge, concat represents the concatenation function, and creates a node vector based on node features, and v i and v j both represent nodes.
步骤S2、基于节点嵌入向量、连接嵌入向量和性能参数输入至全连接层,将节点嵌入向量、连接嵌入向量和性能参数在全连接层融合之后形成特征整合信息,将特征整合信息输入预设的生成网络,以生成 异构多节点互联拓扑结构。Step S2, based on the node embedding vector, connection embedding vector and performance parameters input to the fully connected layer, the node embedding vector, connection embedding vector and performance parameters are fused in the fully connected layer to form feature integration information, and the feature integration information is input into the preset Generate networks to generate heterogeneous multi-node interconnection topologies.
在其中一个实施方式中,生成网络包括但是不限于上采样层和卷积层,如图10所示即为本申请的生成网络的示意图。其中,空间上采样层使用由图A ω定义的聚合函数进行操作,该图由一个包含V个顶点和E个边的图S(V,E)映射到一个更大的图S'(V',E'),通过给新顶点集分配不同的重要度,网络可以学习到A ω的最佳值,从而对图进行良好的上采样。上采样层的上采样过程具体包括:将图S(V,E)映射到一个包含N*n个顶点和E*m个邻边的图S'(V',E');基于图S'(V',E')生成第一邻接矩阵,并获得第一邻接矩阵的初始值;基于第一邻接矩阵的初始值,对第一邻接矩阵进行训练,以获得第一邻接矩阵的最佳值A ω。基于下式获得图S'(V',E')顶点特征: In one embodiment, the generation network includes, but is not limited to, an upsampling layer and a convolutional layer, as shown in FIG. 10 , which is a schematic diagram of the generation network of this application. where the spatial upsampling layer operates using an aggregation function defined by a graph A ω that maps a graph S(V,E) with V vertices and E edges to a larger graph S'(V' , E'), by assigning different importances to new sets of vertices, the network can learn the optimal value of A ω for good upsampling of the graph. The upsampling process of the upsampling layer specifically includes: mapping the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V',E') Generate the first adjacency matrix and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the optimal value of the first adjacency matrix A ω . The graph S'(V', E') vertex features are obtained based on the following formula:
Figure PCTCN2022096236-appb-000019
Figure PCTCN2022096236-appb-000019
其中,f in为图S'(V',E')的顶点特征,k ij表示图S'(V',E')的顶点j和顶点i之间的测地距离A ω为计算N*n个顶点中任一;顶点的权重后得到的图S'(V',E')的最佳值;f j表示图S(V,E)的顶点特征。 Among them, f in is the vertex feature of the graph S'(V', E'), kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E'), and ω is the calculation N* Any of the n vertices; the optimal value of the graph S'(V', E') obtained after the weight of the vertex; f j represents the vertex feature of the graph S(V, E).
在其中一个实施例中,在上采样层进行上采样处理之后,卷积层基于全局图和独立图进行卷积运算。卷积层的运算过程如图11所示,具体地,初始化独立图为图S(V,E);基于下式生成独立图:In one embodiment, after the upsampling layer performs upsampling processing, the convolution layer performs convolution operations based on the global graph and the independent graph. The operation process of the convolutional layer is shown in Figure 11. Specifically, the independent graph is initialized as graph S(V,E); the independent graph is generated based on the following formula:
Figure PCTCN2022096236-appb-000020
Figure PCTCN2022096236-appb-000020
其中,C k表示独立图,f in示图S'(V',E')的顶点特征W θk
Figure PCTCN2022096236-appb-000021
分别为嵌入函数θ和
Figure PCTCN2022096236-appb-000022
的参数,SoftMax为归一化函数;其中,归一化函数为;其中,归一化函数为:
Among them, C k represents the independent graph, fin represents the vertex features W θk of the graph S'(V', E') and
Figure PCTCN2022096236-appb-000021
are embedding functions θ and
Figure PCTCN2022096236-appb-000022
Parameters, SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
Figure PCTCN2022096236-appb-000023
Figure PCTCN2022096236-appb-000023
其中,C k表示独立图,f in示图S'(V',E')的顶点特征W θk
Figure PCTCN2022096236-appb-000024
分别为嵌入函数θ和
Figure PCTCN2022096236-appb-000025
的参数,SoftMax为归一化函数;其中,归一化函数为;其中,归一化函数为:
Among them, C k represents the independent graph, fin represents the vertex features W θk of the graph S'(V', E') and
Figure PCTCN2022096236-appb-000024
are embedding functions θ and
Figure PCTCN2022096236-appb-000025
Parameters, SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
Figure PCTCN2022096236-appb-000026
Figure PCTCN2022096236-appb-000026
其中,B k表示全局图,C k表示独立图α表示调整独立图权重的参数,f in表示图S'(V',E')的顶点特征,K v表示空间维的核大小,W k表示1x1卷积操作的权重向量。 Among them, B k represents the global graph, C k represents the independent graph α represents the parameter to adjust the weight of the independent graph, fin represents the vertex features of the graph S'(V', E'), K v represents the kernel size of the spatial dimension, W k A vector of weights representing the 1x1 convolution operation.
步骤S3、基于预构建的判别网络,确保异构多节点互联拓扑结构满足预设的准确性要求。其中,判别网络与生成网络的运算结构存在相似性,如图12所示,即为本申请的判别网络的内部结构。具体地,判别网络使用的是与生成网络学习的权值不同、具有可训练权值φ的聚合矩阵B φ,由于该聚合是从一个较大 的图S'(V',E')映射到一个较小的图S 1(V 1,E 1),基于下式,获得图S 1(V 1,E 1)的顶点特征: Step S3. Based on the pre-built discriminant network, ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements. Among them, there is a similarity between the operation structure of the discrimination network and the generation network, as shown in FIG. 12 , which is the internal structure of the discrimination network of the present application. Specifically, the discriminative network uses an aggregation matrix B φ with trainable weights φ that is different from the weights learned by the generator network, since the aggregation is mapped from a larger graph S'(V', E') to A smaller graph S 1 (V 1 , E 1 ), based on the following formula, obtains the vertex features of graph S 1 (V 1 , E 1 ):
Figure PCTCN2022096236-appb-000027
Figure PCTCN2022096236-appb-000027
其中,f i为图S 1(V 1,E 1)的顶点特征;f j'表示图S’(V‘,E’)的顶点特征;k表示图S’(V‘,E’)的顶点j和顶点i之间的测地距离B φ为具有可训练权值φ的聚合矩阵B φAmong them, f i is the vertex feature of graph S 1 (V 1 , E 1 ); f j ' is the vertex feature of graph S'(V', E'); k is the graph S'(V', E') The geodesic distance B φ between vertex j and vertex i is an aggregation matrix B φ with trainable weights φ.
在其中一个实施例中,确保异构多节点互联拓扑结构满足预设的准确性要求具体包括:将异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构基于下式进行交叉熵损失运算;In one embodiment, ensuring that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements specifically includes: intersecting the heterogeneous multi-node interconnection topology with the preset real heterogeneous multi-node interconnection topology based on the following formula entropy loss operation;
Figure PCTCN2022096236-appb-000028
Figure PCTCN2022096236-appb-000028
其中,E表示分布函数的期望值,P data代表实际拓扑样本的分布,x是P data中的真实样本,P z表示输入噪音的分布,D(x)表示判别样本为正确的概率,G(z)表示异构多节点互联拓扑图;z表示输入噪音;基于下式,获取拓扑重建损失结果: Among them, E represents the expected value of the distribution function, P data represents the distribution of actual topological samples, x is the real sample in P data , P z represents the distribution of input noise, D(x) represents the probability of judging the sample as correct, G(z ) represents the heterogeneous multi-node interconnection topology; z represents the input noise; based on the following formula, the topology reconstruction loss result is obtained:
Figure PCTCN2022096236-appb-000029
Figure PCTCN2022096236-appb-000029
其中,P t表示异构多节点互联拓扑结构P t'表示真实异构多节点互联拓扑结构,L topo表示异构多节点互联拓扑结构与真实异构多节点互联拓扑结构对应节点的拓扑距离;根据拓扑重建损失与交叉熵损失运算结果,基于下式获取异构多节点互联拓扑结构的最终损失: Among them, P t represents the heterogeneous multi-node interconnection topology P t 'represents the real heterogeneous multi-node interconnection topology, L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology; According to the calculation results of topology reconstruction loss and cross-entropy loss, the final loss of the heterogeneous multi-node interconnection topology is obtained based on the following formula:
L=L cGAN+λL rec L=L cGAN +λL rec
其中,λ为重建项加权L rec为拓扑重建损失L cGAN为交叉熵损失;将异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若异构多节点互联拓扑结构的最终损失大于预设的异构多节点互联拓扑结构的损失,则重复执行步骤S2直至异构多节点互联拓扑结构的最终损失不大于预设的异构多节点互联拓扑结构的损失,确保异构多节点互联拓扑结构满足预设的准确性要求。 Among them, λ is the weight of the reconstruction item L rec is the topology reconstruction loss L cGAN is the cross-entropy loss; compare the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the heterogeneous The final loss of the multi-node interconnection topology is greater than the preset loss of the heterogeneous multi-node interconnection topology, then repeat step S2 until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset heterogeneous multi-node interconnection topology loss, to ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements.
实施例二:Embodiment two:
本实施例的方法,包括以下步骤:基于图卷积网络模型,特征提取节点信息和拓扑结构,以获取节点信息和拓扑结构的低维向量表示;输入性能参数和低维向量表示至全连接层,以生成特征整合信息;将特征整合信息输入预设的生成网络,以生成异构多节点互联拓扑结构;基于预构建的判别网络,确保异构多节点互联拓扑结构满足预设的准确性要求。The method of this embodiment includes the following steps: based on the graph convolutional network model, feature extraction of node information and topology to obtain low-dimensional vector representations of node information and topology; input performance parameters and low-dimensional vector representations to the fully connected layer , to generate feature integration information; input feature integration information into the preset generation network to generate heterogeneous multi-node interconnection topology; based on the pre-built discriminant network, ensure that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements .
在其中一个实施方式中,生成网络具体包括:上采样层和卷积层;卷积层基于上采样层的上采样处理结果,以生成异构多节点互联拓扑结构。In one embodiment, the generating network specifically includes: an upsampling layer and a convolutional layer; the convolutional layer generates a heterogeneous multi-node interconnection topology based on the upsampling processing result of the upsampling layer.
在其中一个实施方式中,低维向量表示包括节点嵌入向量和连接嵌入向量;基于下式,获取得到图卷积网络模型的节点嵌入向量:In one of the implementations, the low-dimensional vector representation includes a node embedding vector and a connection embedding vector; based on the following formula, the node embedding vector of the graph convolutional network model is obtained:
Figure PCTCN2022096236-appb-000030
Figure PCTCN2022096236-appb-000030
其中,v i表示节点
Figure PCTCN2022096236-appb-000031
表示节点v i的邻居节点,e ij表示节点v i和v j的连接,mean表示平均值函数;基于下式,获取得到图卷积网络模型的连接嵌入向量:
Among them, v i represents the node
Figure PCTCN2022096236-appb-000031
Represents the neighbor nodes of node v i , e ij represents the connection between nodes v i and v j , mean represents the average function; based on the following formula, the connection embedding vector of the graph convolutional network model is obtained:
Figure PCTCN2022096236-appb-000032
Figure PCTCN2022096236-appb-000032
其中,
Figure PCTCN2022096236-appb-000033
Figure PCTCN2022096236-appb-000034
表示两个不同大小的前馈网络,
Figure PCTCN2022096236-appb-000035
是对应于邻边的可学习的1x1权重,concat表示拼接函数,基于节点特征创建节点向量,v i和v j均表示节点。
in,
Figure PCTCN2022096236-appb-000033
and
Figure PCTCN2022096236-appb-000034
Denotes two feed-forward networks of different sizes,
Figure PCTCN2022096236-appb-000035
is the learnable 1x1 weight corresponding to the adjacent edge, concat represents the concatenation function, and creates a node vector based on node features, and v i and v j both represent nodes.
在其中一个实施方式中,上采样层的上采样过程具体包括:假设特征整合信息为一个包含V个顶点和E个邻边的图S(V,E);基于图S(V,E)依次进行如下操作:将图S(V,E)映射到一个包含N*n个顶点和E*m个邻边的图S'(V',E');基于图S'(V',E'),生成第一邻接矩阵,并获得第一邻接矩阵的初始值;基于第一邻接矩阵的初始值,对第一邻接矩阵进行训练,以获得第一邻接矩阵的最佳值A ωIn one of the implementations, the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges; based on the graph S(V, E), sequentially Perform the following operations: map the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V', E' ), generate the first adjacency matrix, and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the optimal value A ω of the first adjacency matrix.
在其中一个实施方式中,基于下式获得图S'(V',E')顶点特征:In one of the implementations, the graph S'(V', E') vertex features are obtained based on the following formula:
Figure PCTCN2022096236-appb-000036
Figure PCTCN2022096236-appb-000036
其中,f in为图S'(V',E')的顶点特征,k ij表示图S'(V',E')的顶点j和顶点i之间的测地距离A ω为计算N*n个顶点中任一;顶点的权重后得到的图S'(V',E')的最佳值;f j表示图S(V,E)的顶点特征。 Among them, f in is the vertex feature of the graph S'(V', E'), kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E'), and ω is the calculation N* Any of the n vertices; the optimal value of the graph S'(V', E') obtained after the weight of the vertex; f j represents the vertex feature of the graph S(V, E).
在其中一个实施方式中,卷积层基于上采样层的上采样结果生成全局图和独立图,并基于全局图和独立图进行卷积运算;卷积运算具体包括以下步骤:初始化独立图为图S(V,E),基于下式生成独立图:In one of the implementations, the convolutional layer generates a global graph and an independent graph based on the upsampling result of the upsampling layer, and performs a convolution operation based on the global graph and the independent graph; the convolution operation specifically includes the following steps: initializing the independent graph as a graph S(V,E), which generates an independent graph based on:
Figure PCTCN2022096236-appb-000037
Figure PCTCN2022096236-appb-000037
其中,C k表示独立图,f in示图S'(V',E')的顶点特征W θk
Figure PCTCN2022096236-appb-000038
分别为嵌入函数θ和
Figure PCTCN2022096236-appb-000039
的参数,SoftMax为归一化函数;其中,归一化函数为;其中,归一化函数为:
Among them, C k represents the independent graph, fin represents the vertex features W θk of the graph S'(V', E') and
Figure PCTCN2022096236-appb-000038
are embedding functions θ and
Figure PCTCN2022096236-appb-000039
Parameters, SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
Figure PCTCN2022096236-appb-000040
Figure PCTCN2022096236-appb-000040
其中,N表示图S'(V',E')的顶点数量θ(v i)
Figure PCTCN2022096236-appb-000041
分别表示两个初始值不等的1x1卷积层。
Among them, N represents the number of vertices θ(v i ) of the graph S'(V', E')
Figure PCTCN2022096236-appb-000041
Respectively represent two 1x1 convolutional layers with different initial values.
在其中一个实施方式中,基于下式获取异构多节点互联拓扑结构的特征值:In one of the implementation manners, the eigenvalues of the heterogeneous multi-node interconnection topology are obtained based on the following formula:
Figure PCTCN2022096236-appb-000042
Figure PCTCN2022096236-appb-000042
其中,B k表示全局图,C k表示独立图α表示调整独立图权重的参数,f in表示图S'(V',E')的顶点特征,K v表示空间维的核大小,W k表示1x1卷积操作的权重向量。卷积层的内部的具体地运算过程如图11所示,图中,B k是全局图,对每一层是唯一的。B k是独立图,用于学习每个样本特定的拓扑结构。 Among them, B k represents the global graph, C k represents the independent graph α represents the parameter to adjust the weight of the independent graph, fin represents the vertex features of the graph S'(V', E'), K v represents the kernel size of the spatial dimension, W k A vector of weights representing the 1x1 convolution operation. The specific operation process inside the convolutional layer is shown in Figure 11. In the figure, B k is a global map, which is unique to each layer. Bk is an independent graph for learning a per-sample specific topology.
θ和
Figure PCTCN2022096236-appb-000043
是两个嵌入函数,在这里为1x1的卷积层。K v表示子图的数量,
Figure PCTCN2022096236-appb-000044
表示残差操作
Figure PCTCN2022096236-appb-000045
表示矩阵乘操作
Figure PCTCN2022096236-appb-000046
是控制两种图重要性权重的门。通过门控机制来调整不同层中独立图的重要性,对每层使用一个不同的α该值通过训练来学习和更新。
θ and
Figure PCTCN2022096236-appb-000043
are two embedding functions, here a 1x1 convolutional layer. Kv denotes the number of subgraphs,
Figure PCTCN2022096236-appb-000044
represents the residual operation
Figure PCTCN2022096236-appb-000045
Represents the matrix multiplication operation
Figure PCTCN2022096236-appb-000046
is the gate that controls the importance weights of the two graphs. The importance of independent graphs in different layers is adjusted through a gating mechanism, using a different α value for each layer that is learned and updated through training.
在其中一个实施方式中,确保异构多节点互联拓扑结构满足预设的准确性要求具体包括:In one of the implementation manners, ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements specifically includes:
将异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构基于下式进行交叉熵损失运算;The heterogeneous multi-node interconnection topology and the preset real heterogeneous multi-node interconnection topology are calculated based on the following formula for cross-entropy loss;
Figure PCTCN2022096236-appb-000047
Figure PCTCN2022096236-appb-000047
其中,E表示分布函数的期望值,P data代表实际拓扑样本的分布,x是P data中的真实样本,P z表示输入噪音的分布,D(x)表示判别样本为正确的概率,G(z)表示异构多节点互联拓扑图;z表示输入噪音;基于下式,获取拓扑重建损失结果: Among them, E represents the expected value of the distribution function, P data represents the distribution of actual topological samples, x is the real sample in P data , P z represents the distribution of input noise, D(x) represents the probability of judging the sample as correct, G(z ) represents the heterogeneous multi-node interconnection topology; z represents the input noise; based on the following formula, the topology reconstruction loss result is obtained:
Figure PCTCN2022096236-appb-000048
Figure PCTCN2022096236-appb-000048
其中,P t表示异构多节点互联拓扑结构P t'表示真实异构多节点互联拓扑结构,L topo表示异构多节点互联拓扑结构与真实异构多节点互联拓扑结构对应节点的拓扑距离;根据拓扑重建损失与交叉熵损失运算结果,基于下式获取异构多节点互联拓扑结构的最终损失: Among them, P t represents the heterogeneous multi-node interconnection topology P t 'represents the real heterogeneous multi-node interconnection topology, L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology; According to the calculation results of topology reconstruction loss and cross-entropy loss, the final loss of the heterogeneous multi-node interconnection topology is obtained based on the following formula:
L=L cGAN+λL rec L=L cGAN +λL rec
其中,λ为重建项加权L rec为拓扑重建损失L cGAN为交叉熵损失;将异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若异构多节点互联拓扑结构的最终损失大于预设的异构多节点互联拓扑结构的损失,则重复执行上采样操作和图卷积操作直至异构多节点互联拓扑结构的最终损失不大于预设的异构多节点互联拓扑结构的损失,确保异构多节点互联拓扑结构满足预设的准确性要求。其中,上采样操作包括:上采样层的上采样过程具体包括:假设特征整合信息为一个包含V个顶点和E个邻边的图S(V,E);基于图S(V,E)依次进行如下操作:将图S(V,E)映射到一个包含N*n个顶点和E*m个邻边的图S'(V',E');基于图S'(V',E'),生成第一邻接矩阵,并获得第一邻接矩阵的初始值;基于第一邻接矩阵的初始值,对第一邻接矩阵进行训练,以获得第一邻接矩阵的最佳值A ω;基 于最佳值A ω,获得图S'(V',E')的顶点特征。基于下式获得图S'(V',E')顶点特征: Among them, λ is the weight of the reconstruction item L rec is the topology reconstruction loss L cGAN is the cross-entropy loss; compare the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the heterogeneous If the final loss of the multi-node interconnection topology is greater than the preset loss of the heterogeneous multi-node interconnection topology, the upsampling operation and graph convolution operation are repeated until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset heterogeneous multi-node interconnection topology. The loss of the heterogeneous multi-node interconnection topology ensures that the heterogeneous multi-node interconnection topology meets the preset accuracy requirements. Among them, the upsampling operation includes: the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S(V, E) containing V vertices and E adjacent edges; Perform the following operations: map the graph S(V, E) to a graph S'(V', E') containing N*n vertices and E*m adjacent edges; based on the graph S'(V', E' ), generate the first adjacency matrix, and obtain the initial value of the first adjacency matrix; based on the initial value of the first adjacency matrix, train the first adjacency matrix to obtain the best value A ω of the first adjacency matrix; Get the best value A ω , get the vertex features of the graph S'(V', E'). The graph S'(V', E') vertex features are obtained based on the following formula:
Figure PCTCN2022096236-appb-000049
Figure PCTCN2022096236-appb-000049
其中,f in为图S'(V',E')的顶点特征,k ij表示图S'(V',E')的顶点j和顶点i之间的测地距离A ω为计算N*n个顶点中任一顶点的权重后得到的图S'(V',E')的最佳值;f j表示图S(V,E)的顶点特征。 Among them, f in is the vertex feature of the graph S'(V', E'), kij represents the geodesic distance A between the vertex j and the vertex i of the graph S'(V', E'), and ω is the calculation N* The optimal value of the graph S'(V', E') obtained after the weight of any vertex in the n vertices; f j represents the vertex feature of the graph S(V, E).
图卷积操作包括卷积层基于上采样层的上采样结果生成全局图和独立图,并基于全局图和独立图进行卷积运算;卷积运算具体包括以下步骤:初始化独立图为图S(V,E),基于下式生成独立图:The graph convolution operation includes the convolution layer generating a global graph and an independent graph based on the upsampling results of the upsampling layer, and performing convolution operations based on the global graph and the independent graph; the convolution operation specifically includes the following steps: initializing the independent graph as a graph S( V,E), generating an independent graph based on:
Figure PCTCN2022096236-appb-000050
Figure PCTCN2022096236-appb-000050
其中,C k表示独立图,f in示图S'(V',E')的顶点特征W θk
Figure PCTCN2022096236-appb-000051
分别为嵌入函数θ和
Figure PCTCN2022096236-appb-000052
的参数,SoftMax为归一化函数;其中,归一化函数为;其中,归一化函数为:
Among them, C k represents the independent graph, fin represents the vertex features W θk of the graph S'(V', E') and
Figure PCTCN2022096236-appb-000051
are embedding functions θ and
Figure PCTCN2022096236-appb-000052
Parameters, SoftMax is a normalization function; where, the normalization function is; where, the normalization function is:
Figure PCTCN2022096236-appb-000053
Figure PCTCN2022096236-appb-000053
N表示图S'(V',E')的顶点数量θ(v i)、
Figure PCTCN2022096236-appb-000054
分别表示两个初始值不等的1x1卷积层。
N represents the number of vertices θ(v i ) of the graph S'(V', E'),
Figure PCTCN2022096236-appb-000054
Respectively represent two 1x1 convolutional layers with different initial values.
基于下式获取异构多节点互联拓扑结构的特征值:基于下式获取异构多节点互联拓扑结构的特征值:Obtain the eigenvalues of the heterogeneous multi-node interconnection topology based on the following formula: Obtain the eigenvalues of the heterogeneous multi-node interconnection topology based on the following formula:
Figure PCTCN2022096236-appb-000055
Figure PCTCN2022096236-appb-000055
其中,B k表示全局图,C k表示独立图α表示调整独立图权重的参数,f in表示图S'(V',E')的顶点特征K v表示空间维的核大小W k表示1x1卷积操作的权重向量。 Among them, B k represents the global graph, C k represents the independent graph α represents the parameter to adjust the weight of the independent graph, fin represents the vertex features of the graph S'(V', E'), K v represents the kernel size of the spatial dimension W k represents 1x1 A vector of weights for the convolution operation.
在其中一个实施方式中,性能参数包括噪音信号、性能要求、功耗要求和成本要求。In one embodiment, the performance parameters include noise signal, performance requirements, power consumption requirements and cost requirements.
实施例三:Embodiment three:
本实施例提供一种存储有计算机可读指令的非易失性计算机可读存储介质,当程序被一个或多个处理器执行时,使得一个或多个处理器执行上述任一实施例中提供的异构多节互联拓扑生成方法的步骤。This embodiment provides a non-volatile computer-readable storage medium storing computer-readable instructions. When the program is executed by one or more processors, one or more processors execute the program provided in any of the above-mentioned embodiments. The steps of the heterogeneous multi-node interconnect topology generation method.
本领域内的技术人员应明白,本申请实施例中的实施例可提供为方法、系统、或计算机程序产品。因此,本申请实施例中可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本申请实施例中可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art should understand that the embodiments in the embodiments of the present application may be provided as methods, systems, or computer program products. Therefore, the embodiment of the present application may be in the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Moreover, the embodiments of the present application may take the form of computer program products implemented on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) containing computer-usable program codes therein. .
本申请实施例中是参照根据本申请实施例中实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌 入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。Embodiments of the present application are described with reference to flowcharts and/or block diagrams of methods, devices (systems), and computer program products according to embodiments of the present application. It should be understood that each procedure and/or block in the flowchart and/or block diagram, and a combination of procedures and/or blocks in the flowchart and/or block diagram can be realized by computer program instructions. These computer program instructions may be provided to a general purpose computer, special purpose computer, embedded processor, or processor of other programmable data processing equipment to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing equipment produce a An apparatus for realizing the functions specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to operate in a specific manner, such that the instructions stored in the computer-readable memory produce an article of manufacture comprising instruction means, the instructions The device realizes the function specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded onto a computer or other programmable data processing device, causing a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process, thereby The instructions provide steps for implementing the functions specified in the flow chart or blocks of the flowchart and/or the block or blocks of the block diagrams.
注意,上述仅为本申请的较佳实施例及所运用技术原理。本领域技术人员会理解,本申请不限于这里的特定实施例,对本领域技术人员来说能够进行各种明显的变化、重新调整和替代而不会脱离本申请的保护范围。因此,虽然通过以上实施例对本申请进行了较为详细的说明,但是本申请不仅仅限于以上实施例,在不脱离本申请构思的情况下,还可以包括更多其它等效实施例,而本申请的范围由所附的权利要求范围决定。Note that the above are only preferred embodiments and technical principles used in this application. Those skilled in the art will understand that the present application is not limited to the specific embodiments here, and various obvious changes, readjustments and substitutions can be made by those skilled in the art without departing from the protection scope of the present application. Therefore, although the present application has been described in detail through the above embodiments, the present application is not limited to the above embodiments, and can also include more other equivalent embodiments without departing from the concept of the present application. The scope is determined by the scope of the appended claims.

Claims (11)

  1. 一种异构多节点互联拓扑生成方法,其特征在于:所述方法包括以下步骤:A heterogeneous multi-node interconnect topology generation method, characterized in that: the method includes the following steps:
    基于图卷积网络模型,特征提取节点信息和拓扑结构,以获取节点信息和拓扑结构的低维向量表示;Based on the graph convolutional network model, feature extraction node information and topology to obtain low-dimensional vector representation of node information and topology;
    输入性能参数和所述低维向量表示至全连接层,以生成特征整合信息;Input performance parameters and the low-dimensional vector representation to the fully connected layer to generate feature integration information;
    将所述特征整合信息输入预设的生成网络,以生成异构多节点互联拓扑结构;及Inputting the feature integration information into a preset generation network to generate a heterogeneous multi-node interconnection topology; and
    获取所述异构多节点互联拓扑结构的特征值,确保所述异构多节点互联拓扑结构满足预设的准确性要求。The characteristic value of the heterogeneous multi-node interconnection topology is acquired to ensure that the heterogeneous multi-node interconnection topology meets a preset accuracy requirement.
  2. 根据权利要求1所述的异构多节点互联拓扑生成方法,其特征在于:所述生成网络具体包括:上采样层、卷积层、全连接层、批归一化层、修正线性单元和S型函数。The heterogeneous multi-node interconnection topology generation method according to claim 1, wherein: the generation network specifically includes: an upsampling layer, a convolutional layer, a fully connected layer, a batch normalization layer, a modified linear unit, and a S type function.
  3. 根据权利要求1所述的异构多节点互联拓扑生成方法,其特征在于:所述低维向量表示包括节点嵌入向量和连接嵌入向量;The heterogeneous multi-node interconnection topology generation method according to claim 1, wherein the low-dimensional vector representation includes node embedding vectors and connection embedding vectors;
    基于下式,获取得到所述图卷积网络模型的节点嵌入向量:Based on the following formula, the node embedding vector of the graph convolutional network model is obtained:
    Figure PCTCN2022096236-appb-100001
    Figure PCTCN2022096236-appb-100001
    其中ν i表示节点,
    Figure PCTCN2022096236-appb-100002
    表示节点ν i的邻居节点,e ij表示节点ν i和ν j的连接,mean表示平均值函数;及
    where ν i represents a node,
    Figure PCTCN2022096236-appb-100002
    Represents the neighbor nodes of node ν i , e ij represents the connection between nodes ν i and ν j , mean represents the average function; and
    基于下式,获取得到所述图卷积网络模型的连接嵌入向量:Based on the following formula, the connection embedding vector of the graph convolutional network model is obtained:
    Figure PCTCN2022096236-appb-100003
    Figure PCTCN2022096236-appb-100003
    其中,f c0和f c1表示两个不同大小的前馈网络,w ij e表示对应邻边的可学习的1x1权重,concat表示拼接函数,基于节点特征创建节点向量,v i和v j均表示节点。 Among them, f c0 and f c1 represent two feed-forward networks of different sizes, w ij e represents the learnable 1x1 weight of the corresponding adjacent edge, concat represents the concatenation function, and creates a node vector based on node features, and v i and v j both represent node.
  4. 根据权利要求2所述的异构多节点互联拓扑生成方法,其特征在于:所述上采样层的上采样过程具体包括:假设特征整合信息为一个包含V个顶点和E个邻边的图S(V,E);基于所述图S(V,E)依次进行如下操作:The heterogeneous multi-node interconnection topology generation method according to claim 2, characterized in that: the upsampling process of the upsampling layer specifically includes: assuming that the feature integration information is a graph S containing V vertices and E adjacent edges (V, E); Carry out the following operations sequentially based on the graph S (V, E):
    将所述图S(V,E)映射到一个包含N*n个顶点和E*m个邻边的图S'(V',E');Mapping said graph S(V,E) to a graph S'(V',E') containing N*n vertices and E*m adjacent edges;
    基于所述图S'(V',E'),生成第一邻接矩阵,并获得所述第一邻接矩阵的初始值;及Based on the graph S'(V', E'), generate a first adjacency matrix, and obtain an initial value of the first adjacency matrix; and
    基于所述第一邻接矩阵的初始值,对所述第一邻接矩阵进行训练,以获得所述第一邻接矩阵的最佳值A ωBased on the initial value of the first adjacency matrix, the first adjacency matrix is trained to obtain an optimal value A ω of the first adjacency matrix.
  5. 根据权利要求4所述的异构多节点互联拓扑生成方法,其特征在于:基下式获得所述图S'(V',E')的顶点特征:The heterogeneous multi-node interconnection topology generation method according to claim 4, characterized in that: the following formula obtains the vertex characteristics of the graph S' (V', E'):
    Figure PCTCN2022096236-appb-100004
    Figure PCTCN2022096236-appb-100004
    其中,f in为图S'(V',E')的顶点特征k ij表示图S'(V',E')的顶点j和顶点i之间的测地距离,A ω为 计算所述N*n个顶点中任一顶点的权重后得到的图S'(V',E')的最佳值,f j表示图S(V,E)的顶点特征。 Among them, f in is the vertex feature k ij of the graph S'(V', E'), which means the geodesic distance between the vertex j and the vertex i of the graph S'(V', E'), A ω is the calculation described The optimal value of the graph S'(V', E') obtained after the weight of any vertex in the N*n vertices, and f j represents the vertex feature of the graph S(V, E).
  6. 根据权利要求2所述的异构多节点互联拓扑生成方法,其特征在于:所述卷积层基于所述上采样层的上采样结果生成全局图和独立图,并基于所述全局图和所述独立图进行卷积运算;所述卷积运算具体包括以下步骤:The heterogeneous multi-node interconnection topology generation method according to claim 2, characterized in that: the convolution layer generates a global graph and an independent graph based on the upsampling result of the upsampling layer, and based on the global graph and the The independent graph is carried out convolution operation; The convolution operation specifically includes the following steps:
    初始化所述独立图为图S(V,E),并基于下式生成所述独立图::The independent graph is initialized as a graph S(V, E), and the independent graph is generated based on the following formula::
    Figure PCTCN2022096236-appb-100005
    Figure PCTCN2022096236-appb-100005
    其中,C k表示独立图,f in表示图S'(V',E')的顶点特征W θk
    Figure PCTCN2022096236-appb-100006
    分别为嵌入函数θ和
    Figure PCTCN2022096236-appb-100007
    的参数,SoftMax为归一化函数;及
    where C k represents the independent graph, fin represents the vertex features W θk of the graph S'(V', E') and
    Figure PCTCN2022096236-appb-100006
    are embedding functions θ and
    Figure PCTCN2022096236-appb-100007
    The parameters of , SoftMax is a normalization function; and
    其中,所述归一化函数为:Wherein, the normalization function is:
    Figure PCTCN2022096236-appb-100008
    Figure PCTCN2022096236-appb-100008
    其中,N表示图S'(V',E')的顶点数量,θ(v i)、
    Figure PCTCN2022096236-appb-100009
    分别表示两个初始值不等的1x1卷积层。
    Among them, N represents the number of vertices in the graph S'(V', E'), θ(v i ),
    Figure PCTCN2022096236-appb-100009
    Respectively represent two 1x1 convolutional layers with different initial values.
  7. 根据权利要求6所述的异构多节点互联拓扑生成方法,其特征在于:基于下式获取所述异构多节点互联拓扑结构的特征值:The heterogeneous multi-node interconnection topology generation method according to claim 6, characterized in that: the characteristic value of the heterogeneous multi-node interconnection topology is obtained based on the following formula:
    Figure PCTCN2022096236-appb-100010
    Figure PCTCN2022096236-appb-100010
    其中B k表示全局图,C k表示独立图α表示调整独立图权重的参数。f in表示图S'(V',E')的顶点特征K v表示空间维的核大小W K表示1x1卷积操作的权重向量。 where B k represents the global graph, and C k represents the independent graph α represents the parameter to adjust the weight of the independent graph. f in represents the vertex feature K of the graph S'(V', E') , v represents the kernel size W of the spatial dimension, and K represents the weight vector of the 1x1 convolution operation.
  8. 根据权利要求7所述的异构多节点互联拓扑生成方法,其特征在于:确保所述异构多节点互联拓扑结构满足预设的准确性要求具体包括:The method for generating a heterogeneous multi-node interconnection topology according to claim 7, wherein ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements specifically includes:
    将所述异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构,基于下式进行交叉熵损失运算:The heterogeneous multi-node interconnection topology and the preset real heterogeneous multi-node interconnection topology are used to perform cross-entropy loss calculation based on the following formula:
    Figure PCTCN2022096236-appb-100011
    Figure PCTCN2022096236-appb-100011
    其中,E表示分布函数的期望值,P data代表实际拓扑样本的分布,x是P data中的真实样本,P z表示输入噪音的分布,D(x)表示判别样本为正确的概率,G(z)表示异构多节点互联拓扑图;z表示输入噪音; Among them, E represents the expected value of the distribution function, P data represents the distribution of actual topological samples, x is the real sample in P data , P z represents the distribution of input noise, D(x) represents the probability of judging the sample as correct, G(z ) represents the heterogeneous multi-node interconnection topology; z represents the input noise;
    基于下式,获取拓扑重建损失结果:Based on the following formula, the topology reconstruction loss result is obtained:
    Figure PCTCN2022096236-appb-100012
    Figure PCTCN2022096236-appb-100012
    其中,P t表示异构多节点互联拓扑结构P t’表示真实异构多节点互联拓扑结构,L topo表示异构多节点 互联拓扑结构与真实异构多节点互联拓扑结构对应节点的拓扑距离; Among them, P t represents the heterogeneous multi-node interconnection topology P t 'represents the real heterogeneous multi-node interconnection topology, L topo represents the topological distance between the heterogeneous multi-node interconnection topology and the corresponding node of the real heterogeneous multi-node interconnection topology;
    根据所述拓扑重建损失结果与所述交叉熵损失运算结果,基于下式获取所述异构多节点互联拓扑结构的最终损失:According to the topology reconstruction loss result and the cross-entropy loss calculation result, the final loss of the heterogeneous multi-node interconnection topology is obtained based on the following formula:
    L=L cGAN+λL rec L=L cGAN +λL rec
    其中,λ为重建项加权,L rec为拓扑重建损失,L cGAN为交叉熵损失;及 Among them, λ is the weight of the reconstruction item, L rec is the topology reconstruction loss, and L cGAN is the cross-entropy loss; and
    将所述异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若所述异构多节点互联拓扑结构的最终损失大于所述预设的异构多节点互联拓扑结构的损失,则重复执行权利要求4~7直至所述异构多节点互联拓扑结构的最终损失不大于所述预设的异构多节点互联拓扑结构的损失,确保所述异构多节点互联拓扑结构满足预设的准确性要求。Comparing the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the final loss of the heterogeneous multi-node interconnection topology is greater than the preset heterogeneous multi-node interconnection topology the loss of the multi-node interconnection topology, repeat claims 4 to 7 until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset loss of the heterogeneous multi-node interconnection topology, ensuring that the heterogeneous multi-node interconnection topology The multi-node interconnection topology meets the preset accuracy requirements.
  9. 根据权利要求1所述的异构多节点互联拓扑生成方法,其特征在于:所述性能参数包括噪音信号、性能要求、功耗要求和成本要求。The method for generating a heterogeneous multi-node interconnection topology according to claim 1, wherein the performance parameters include noise signals, performance requirements, power consumption requirements and cost requirements.
  10. 根据权利要求1所述的异构多节点互联拓扑生成方法,其特征在于,所述确保所述异构多节点互联拓扑结构满足预设的准确性要求,包括:The heterogeneous multi-node interconnection topology generation method according to claim 1, wherein the ensuring that the heterogeneous multi-node interconnection topology meets preset accuracy requirements includes:
    基于所述异构多节点互联拓扑结构的特征值,获得所述异构多节点互联拓扑结构与预设的真实异构多节点互联拓扑结构的交叉熵损失结果及拓扑重建损失结果;Obtaining a cross-entropy loss result and a topology reconstruction loss result between the heterogeneous multi-node interconnection topology and a preset real heterogeneous multi-node interconnection topology based on the eigenvalues of the heterogeneous multi-node interconnection topology;
    基于所述交叉熵损失结果及所述拓扑重建损失结果,获得所述异构多节点互联拓扑结构的最终损失;及Obtaining a final loss of the heterogeneous multi-node interconnection topology based on the cross-entropy loss result and the topology reconstruction loss result; and
    将所述异构多节点互联拓扑结构的最终损失与预设的异构多节点互联拓扑结构的损失进行比较,若所述异构多节点互联拓扑结构的最终损失大于所述预设的异构多节点互联拓扑结构的损失,则重复执行所述上采样层的上采样过程和所述卷积层的卷积运算直至所述异构多节点互联拓扑结构的最终损失不大于所述预设的异构多节点互联拓扑结构的损失。Comparing the final loss of the heterogeneous multi-node interconnection topology with the preset loss of the heterogeneous multi-node interconnection topology, if the final loss of the heterogeneous multi-node interconnection topology is greater than the preset heterogeneous multi-node interconnection topology the loss of the multi-node interconnection topology, repeat the upsampling process of the upsampling layer and the convolution operation of the convolution layer until the final loss of the heterogeneous multi-node interconnection topology is not greater than the preset Loss of heterogeneous multi-node interconnect topology.
  11. 一个或多个存储有计算机可读指令的非易失性计算机可读存储介质,其特征在于,所述计算机可读指令被一个或多个处理器执行时,使得所述一个或多个处理器执行如权利要求1至10中任一项所述的方法的步骤。One or more non-transitory computer-readable storage media storing computer-readable instructions, wherein when the computer-readable instructions are executed by one or more processors, the one or more processors Carrying out the steps of the method as claimed in any one of claims 1 to 10.
PCT/CN2022/096236 2022-01-10 2022-05-31 Method for generating heterougeneous multi-node interconnection topology, and storage medium WO2023130656A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202210024578.5 2022-01-10
CN202210024578.5A CN114050975B (en) 2022-01-10 2022-01-10 Heterogeneous multi-node interconnection topology generation method and storage medium

Publications (1)

Publication Number Publication Date
WO2023130656A1 true WO2023130656A1 (en) 2023-07-13

Family

ID=80196189

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/096236 WO2023130656A1 (en) 2022-01-10 2022-05-31 Method for generating heterougeneous multi-node interconnection topology, and storage medium

Country Status (2)

Country Link
CN (1) CN114050975B (en)
WO (1) WO2023130656A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114050975B (en) * 2022-01-10 2022-04-19 苏州浪潮智能科技有限公司 Heterogeneous multi-node interconnection topology generation method and storage medium
CN114726739B (en) * 2022-04-18 2024-04-09 深圳市智象科技有限公司 Topology data processing method, device, equipment and storage medium
CN114884908B (en) * 2022-04-29 2024-02-13 浪潮电子信息产业股份有限公司 Data synchronization method, device, equipment and storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210034737A1 (en) * 2019-07-30 2021-02-04 Sakif Hossain Khan Detection of adverserial attacks on graphs and graph subsets
CN112417219A (en) * 2020-11-16 2021-02-26 吉林大学 Hyper-graph convolution-based hyper-edge link prediction method
CN112925989A (en) * 2021-01-29 2021-06-08 中国计量大学 Group discovery method and system of attribute network
CN114050975A (en) * 2022-01-10 2022-02-15 苏州浪潮智能科技有限公司 Heterogeneous multi-node interconnection topology generation method and storage medium

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111914484B (en) * 2020-08-07 2024-03-26 中国南方电网有限责任公司 Recursive graph convolution network system for power grid transient stability evaluation
CN112163219A (en) * 2020-08-27 2021-01-01 北京航空航天大学 Malicious program identification and classification method based on word embedding and GCN
CN112651492B (en) * 2020-12-30 2023-10-13 广州大学华软软件学院 Self-connection width graph convolution neural network model system and training method
CN112800903B (en) * 2021-01-19 2022-08-26 南京邮电大学 Dynamic expression recognition method and system based on space-time diagram convolutional neural network
CN113222328B (en) * 2021-03-25 2022-02-25 中国科学技术大学先进技术研究院 Air quality monitoring equipment point arrangement and site selection method based on road section pollution similarity
CN113240187B (en) * 2021-05-26 2022-10-11 合肥工业大学 Prediction model generation method, system, device, storage medium and prediction method
CN113904786B (en) * 2021-06-29 2023-05-30 重庆大学 False data injection attack identification method based on line topology analysis and tide characteristics
CN113780470B (en) * 2021-09-28 2024-03-08 西安聚全网络科技有限公司 Graph classification method based on self-adaptive multichannel cross graph convolutional network

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20210034737A1 (en) * 2019-07-30 2021-02-04 Sakif Hossain Khan Detection of adverserial attacks on graphs and graph subsets
CN112417219A (en) * 2020-11-16 2021-02-26 吉林大学 Hyper-graph convolution-based hyper-edge link prediction method
CN112925989A (en) * 2021-01-29 2021-06-08 中国计量大学 Group discovery method and system of attribute network
CN114050975A (en) * 2022-01-10 2022-02-15 苏州浪潮智能科技有限公司 Heterogeneous multi-node interconnection topology generation method and storage medium

Also Published As

Publication number Publication date
CN114050975A (en) 2022-02-15
CN114050975B (en) 2022-04-19

Similar Documents

Publication Publication Date Title
WO2023130656A1 (en) Method for generating heterougeneous multi-node interconnection topology, and storage medium
JP7233656B2 (en) Task Activation for Accelerated Deep Learning
US20220374288A1 (en) Distributed placement of linear operators for accelerated deep learning
Chen et al. NoC-based DNN accelerator: A future design paradigm
US20220129302A1 (en) Data processing system and method for heterogeneous architecture
US20220197714A1 (en) Training a neural network using a non-homogenous set of reconfigurable processors
US11392740B2 (en) Dataflow function offload to reconfigurable processors
WO2015196833A1 (en) Many-core processor system integrated with network router, and integration method and implementation method thereof
US11463272B2 (en) Scalable in-network computation for massively-parallel shared-memory processors
CN113469355B (en) Multi-model training pipeline in distributed system
US11789733B2 (en) Instruction processing apparatus, acceleration unit, and server
Parthasarathy et al. DEFER: distributed edge inference for deep neural networks
CN113452655A (en) Distributed training method, gradient communication device and computing equipment
Ma et al. FPGA-based AI smart NICs for scalable distributed AI training systems
CN111191784A (en) Transposed sparse matrix multiplied by dense matrix for neural network training
Truong et al. Hybrid electrical/optical switch architectures for training distributed deep learning in large-scale
CN115222014A (en) Acceleration unit and server for neural network model execution
Tariq et al. A novel meta-heuristic for green computing on vfi-noc-hmpsocs
US20240177034A1 (en) Simulating quantum computing circuits using kronecker factorization
JP7389176B2 (en) acceleration system
Denholm Improving low latency applications for reconfigurable devices
Zhu et al. Research on A Chiplet-based DSA (Domain-Specific Architectures) Scalable Convolutional Acceleration Architecture
Chen et al. A Survey on Graph Neural Network Acceleration: A Hardware Perspective
Yang et al. Versa-DNN: A Versatile Architecture Enabling High-Performance and Energy-Efficient Multi-DNN Acceleration
Ortega-Cisneros Design and Implementation of a NoC-Based Convolution Architecture With GEMM and Systolic Arrays

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22918114

Country of ref document: EP

Kind code of ref document: A1