CN110837891B - Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture - Google Patents

Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture Download PDF

Info

Publication number
CN110837891B
CN110837891B CN201911014330.5A CN201911014330A CN110837891B CN 110837891 B CN110837891 B CN 110837891B CN 201911014330 A CN201911014330 A CN 201911014330A CN 110837891 B CN110837891 B CN 110837891B
Authority
CN
China
Prior art keywords
neurons
neuron
winning
weight
input
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911014330.5A
Other languages
Chinese (zh)
Other versions
CN110837891A (en
Inventor
李丽
张衡
傅玉祥
黄延
何国强
何书专
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University
Original Assignee
Nanjing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University filed Critical Nanjing University
Priority to CN201911014330.5A priority Critical patent/CN110837891B/en
Publication of CN110837891A publication Critical patent/CN110837891A/en
Application granted granted Critical
Publication of CN110837891B publication Critical patent/CN110837891B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention relates to a self-organizing mapping method and a self-organizing mapping system based on a SIMD architecture, wherein the method comprises the following steps: the distance calculation module is used for finding the best matching competition layer neuron with the minimum distance by calculating the Manhattan distance between the input vector and the weight vector of the corresponding competition layer neuron; the learning rate and field radius calculation module is used for updating the learning rate and the field radius through shift operation; a cooperation module in which neurons winning in competition are not individually excited but are excited collectively by determining a field in which the neurons winning in competition are centered; and the weight updating module is used for updating the corresponding weights of the neurons in the inner part of the radius of the field of the winning neuron and storing the calculated weights back to the SRAM on the chip. The method supports the arbitrary classification of samples and the operation of arbitrary characteristics, and can meet the requirements of low complexity and high precision.

Description

Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture
Technical Field
The invention relates to a hardware implementation of an unsupervised neural network algorithm, in particular to a hardware implementation of a self-organizing mapping algorithm based on a SIMD architecture, which can represent input data in a low-dimensional space and can also be used in applications such as data visualization, clustering and the like.
Background
The self-organizing map (SOM) algorithm is an unsupervised learning algorithm for clustering and high-dimensional visualization, and is an artificial neural network developed by simulating the characteristics of human brain on signal processing. The SOM algorithm was proposed by Teuvo Kohonen in 1981, and has now become the most widely applied self-organizing neural network method, in which the wta (winner Takes all) competition mechanism reflects the most fundamental feature of the self-organizing map learning algorithm.
The SOM algorithm has good clustering effect due to its characteristics such as topology structure maintenance, probability distribution maintenance, instructor-free learning and visualization, and thus receives wide attention, and various results of application and research on the SOM algorithm are emerging, and are widely applied to a plurality of information processing fields such as voice recognition, image processing, classification and clustering, combinatorial optimization (such as TSP problem), data analysis and prediction.
Compared with other neural networks, the structure of the self-organizing neural network and the learning rule thereof have own characteristics. In the network structure, the self-organizing neural network is a two-layer network formed by an input layer and a competition layer; the neurons between the two layers realize bidirectional connection, and the self-organizing neural network has no hidden layer. The competition layer can be a one-dimensional, two-dimensional or three-dimensional array, and the input layer is the sample characteristic quantity.
In the SOM network model, each weight coefficient vector can be regarded as an internal representation of the input vector in the neural network, or it is a mapping of the input vector, and the purpose of the self-organizing function of the SOM network model is to make the neural network converge on a representation form by adjusting the weight coefficients. In this representation, each neuron is only specifically matched or sensitive to a certain input pattern. In other words, the purpose of the SOM is to allow the morphological representation of the weight coefficients of the neurons to indirectly mimic the input pattern.
Disclosure of Invention
The purpose of the invention is as follows: the self-organizing mapping method based on the SIMD architecture is provided, so that a good clustering effect of a self-organizing mapping algorithm is realized under limited resources, and the training speed and the decision speed of the algorithm are improved.
The technical scheme is as follows: a self-organizing mapping method based on SIMD architecture includes the following steps:
s1: storing the initialization weight W and the input neurons into a storage unit according to a rule, and initializing a field h, a learning rate and the maximum training times of the network;
s2: taking one sample from the total samples as network input, and taking corresponding input features and weights from a storage unit;
s3: calculating Manhattan distances from the input neurons to all the neurons in the competition layer, determining position coordinates of the winning neurons according to the minimum distance, calculating the distance from the input neurons to the neurons in the competition layer with coordinates (i, j), finding out the minimum distance value, and determining a calculation formula of the winning neurons
Figure BDA0002245202000000021
S4: taking out the input and weight from the memory unit, and adjusting the network weight of the winning neuron and the neuron in the field, wherein the formula of the weight is Wi(n+1)=Wi(n)+η(n)hj,i(n)(X-Wi(n)), restoring the updated weight to the storage unit, and overwriting the weight before updating;
s5: when all learning samples are not calculated, repeating the steps of S2-S5, otherwise executing S6;
s6: adjusting learning rate function and domain function, learning rate
Figure BDA0002245202000000022
Wherein eta0Is the initial learning rate; radius of area of
Figure BDA0002245202000000023
Wherein
Figure BDA0002245202000000024
S7: if the current training times do not reach the set maximum training times, returning to S2, otherwise, terminating the loop and ending the learning process.
In a further embodiment, the method comprises the following steps: the data storage unit comprises 8N subunits, wherein 4N subunits store input neuron data, randomly generated weights are stored in the other 2N subunits in a ping-pong manner, winning neurons obtained through training are stored in the other 2N subunits, and the winning neurons obtained through next training cover the last winning neuron data.
In a further embodiment, the method comprises the following steps: the whole input sample can be traversed by D batches, the total training frequency is K, and the frequency of ping-pong operation required by the weight is DxK.
In a further embodiment, the method comprises the following steps: under the condition that computing resources and storage resources are fixed, the same set of resources are used, one controller is adopted to control a plurality of processors, and the spatial parallel operation is realized by executing the operation of the same operation respectively.
A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the steps of the method as described above are implemented when the computer program is executed by the processor.
A computer-readable storage medium, on which a computer program is stored which, when being executed by a processor, carries out the steps of the method described above.
A SIMD architecture based self-organizing map system comprising:
the data storage module comprises a plurality of storage units with preset depths, and each storage unit comprises a plurality of subunits which are respectively used for storing input neuron data, randomly generated weights and trained winning neuron data;
the distance calculation module is used for respectively taking out the data of each input neuron and the corresponding weight from the storage unit, calculating the Manhattan distance sum of the input neuron and the weight corresponding to each neuron of the competition layer, calculating the minimum distance by a bubble sorting method, obtaining the coordinate of a winning neuron, and finally storing the coordinate into the storage module;
the learning rate and field radius calculation module is used for calculating the learning rate and the field radius by using shift operation and adopting two registers for storage, and the calculation result is written back to the original register again each time;
the weight updating module and the cooperation module are used for taking out the neuron data and the weights from the storage unit and carrying out updating operation, and the updated weights are stored in the storage unit according to ping-pong storage.
In a further embodiment, the data storage unit comprises 8N subunits, wherein 4N subunits store input neuron data, randomly generated weights are ping-pong stored in another 2N subunits, the winning neuron obtained from the training is stored in another 2N subunits, and the winning neuron obtained from the next training overwrites the last winning neuron data.
Has the advantages that:
first, the learning rate and the field radius calculation of the invention both use shift operation, thereby greatly reducing the operation resources and the operation time.
Secondly, the learning rate and the field radius are gradually converged along with the training times, so that the clustering effect can be better improved.
Thirdly, calculating the Manhattan distance sum of the input neuron and each neuron of the competition layer, and selecting the winning neuron by a bubble sorting method, thereby saving the operation resource and improving the calculation speed.
Fourthly, the same set of resources are adopted in the reasoning link and the training process, and the utilization rate of the operation resources and the storage resources can be improved by adopting multi-path parallel computation.
In conclusion, the invention can effectively save operation resources and improve the calculation speed, has wide application prospect and has good application value for different occasions.
Drawings
Fig. 1 is a hardware implementation architecture diagram of the self-organizing map algorithm in the present invention.
Fig. 2 is a network architecture diagram of the self-organizing map algorithm.
FIG. 3 is a schematic diagram of data arrangement in the present invention.
Fig. 4 is a schematic diagram of hardware modules in the present invention.
FIG. 5 is a schematic diagram of a computing unit design according to the present invention.
FIG. 6 is a flowchart illustrating the detailed process of updating weights in the present invention.
Fig. 7 is a flow chart of a hardware implementation of the self-organizing map algorithm of the present invention.
Detailed Description
SOMs typically include the following three links:
and (3) competition links: i.e. the selection of the winning neuron. According to the principle of WTA (Winner Takes all), the discrimination function value of each input vector is calculated, the minimum distance is compared, and the neuron used as a competitive winner is obtained according to the minimum distance.
The cooperation link, also called the coordination link: the neurons that win the competition are not individually stimulated, but a neighborhood centered on the winning neuron is first determined, and all neurons in the neighborhood are stimulated together according to a certain rule within the radius of the domain of the winning neuron.
And (3) self-adaptive link: the weights of all neurons in the neighborhood of the winning neuron are adjusted appropriately to increase their discrimination function values for the input pattern. Even if the weight vectors of the output layer neurons change with the input vectors.
The SOM network can map any dimension input mode into a one-dimensional, two-dimensional or three-dimensional graph on an output layer, and keep the topological structure of the graph unchanged; the network can make the weight vector space and the probability distribution of the input mode tend to be consistent through repeated learning of the input mode, namely probability retentivity. The neurons of the competition layer of the network compete for the response opportunity to the input mode, and the weights related to the winning neuron are adjusted towards the direction more beneficial to the competition of the winning neuron, namely, the winning neuron is used as the center of a circle, excitatory side feedback is shown for the adjacent neurons, inhibitory side feedback is shown for the far adjacent neurons, the neighbors mutually stimulate, and the far adjacent neurons mutually inhibit. In general, neighbor is a competition layer neuron within a continuously decreasing field radius with increasing training times, from a neuron emitting a signal as a center of a circle; distal neighbors refer to competing layer neurons with radii outside the radius of the field. Neurons further than the far neighbours exhibit weak excitatory effects, and this interaction is also known as the "mexican hat" because the curve of the interaction resembles the hat worn by mexicans.
The learning steps of the SOM algorithm are as follows:
1. and initializing, assigning a network weight to a random value between [0 and 1], and setting an initial value of a learning rate, a neighborhood radius and a total learning frequency T.
2. Sampling, randomly selecting an input mode from the data set, normalizing and inputting the input mode into the neural network.
3. And competing, calculating the distances between all the neurons and the input mode, and finding out the winning neuron corresponding to the input mode.
4. And cooperatively determining a neighborhood range according to a neighborhood function of the neural network.
5. And self-adapting, and updating the weight of the neuron in the neighborhood.
6. And judging whether all samples are input into the neural network, if so, jumping to 7, and otherwise, jumping to 2.
7. The learning rate and neighborhood functions are updated.
8. And stopping the condition, judging whether the iteration number n exceeds T, if so, finishing the algorithm, and otherwise, jumping to 2.
A hardware implementation of a self-organizing map algorithm based on SIMD architecture,
the acceleration core comprises a distance calculation module, a learning rate and field radius calculation module, a cooperation module, a weight value updating module and a data storage module.
In the design framework, under the condition that computing resources and storage resources are fixed, the same set of resources are used in the reasoning link and the training link of the algorithm, and the computing resources are utilized as much as possible by adopting multi-path parallel computing. In the hardware implementation design process, the reasoning process and the training process both adopt a full-flow design mode, and the neuron of the competition layer is supported to be a one-dimensional or two-dimensional graph.
The on-chip SRAM is designed to have storage resources divided into 8N banks, the depth of each bank is M, 4N banks are allocated to store data of all input neurons, and ping-pong access is carried out on randomly generated weights in the other 2N banks; and storing the winning neurons obtained by the last sequential training to the other 2N banks, wherein the winning neurons obtained by the next training cover the data of the last winning neurons. And storing the weights in 2N banks, wherein the whole input sample can be traversed only by D times of calculation, and the total training time is K, and the time of ping-pong operation required by the weights is DxK. The distance calculation module respectively takes out the data of each input neuron and the corresponding weight from the bank, then calculates the Manhattan distance sum of the input neuron and the weight corresponding to each neuron of the competition layer, calculates the minimum distance by a bubble sorting method, obtains the coordinate of the winning neuron, and finally stores the coordinate into the bank.
The learning rate and the field radius calculation module are reduced along with the change of the training times, the learning rate and the field radius calculation both use shift operation, the learning rate and the field radius are stored by two registers, and the calculation result is written back to the original register every time.
The weight updating module respectively takes the input neurons and the weights from the bank, and the weight is Wi(n+1)=Wi(n)+η(n)hj,i(n)(X-Wi(n)), h within the radius of the domain according to the cooperative modular approachj,i(n) is 1, h outside the radius of the domainj,iAnd (n) is 0, and the updated weight is stored in the bank according to ping-pong storage.
And the cooperation module calculates Euclidean distances between the winning neuron and other competition layer neurons, and the competition layer neurons within the radius of the field of the winning neuron are jointly excited according to a certain rule.
According to one aspect of the invention, a hardware implementation of a self-organizing map algorithm based on a SIMD architecture is provided, the acceleration core of the invention comprising: the device comprises a data storage module, a distance calculation module, a learning rate and field radius calculation module, a weight value updating module and a cooperation module.
Under the condition that computing resources and storage resources are fixed, the same set of resources are used in the reasoning link and the training link of the algorithm, the SIMD architecture can adopt one controller to control a plurality of processors, and spatial parallel operation can be realized for the operation of respectively executing the same operation.
Specifically, the on-chip data storage module sets storage resources to be 32N banks, the depth of each bank is M, 16N banks are allocated to store data of all input neurons, and ping-pong access is performed on randomly generated weights in the other 8N banks; and storing the winning neurons obtained by the last sequential training into the other 8N banks, covering the data of the last winning neuron by the winning neurons obtained by the next training, storing the weights in the 8N banks, traversing the whole input sample by D batches (banks), wherein the total training frequency is K, and the frequency of ping-pong operation required by the weights is D multiplied by K.
The distance calculation module is obtained by judging the Manhattan distance between the input and each competitive neuron in the competitive layer, the competitive neuron with the minimum Manhattan distance is the winning neuron, the Manhattan distance is solved by obtaining the distance difference only by using an adder, then the size is judged by a comparator, 8N paths of parallel calculation can be carried out under the condition that the storage resources are set to be 32N banks, and the winning neuron is obtained by a bubble sorting method.
The learning rate and the field radius calculation module are reduced along with the change of the training times, and the learning rate
Figure BDA0002245202000000061
Wherein eta0Is the initial learning rate; radius of area of
Figure BDA0002245202000000062
Wherein
Figure BDA0002245202000000063
The learning rate and the field radius are calculated by using shift operation, the learning rate and the field radius are stored by using two registers after right shift is performed once every iteration, and the calculation result is written back to the original register again every time.
The weight updating module respectively takes the input neurons and the weights from the bank, and the weight is Wi(n+1)=Wi(n)+η(n)hj,i(n)(X-WiAnd (N)), according to the cooperation module flow, the updated weight is stored in the bank according to a ping-pong storage mode, and under the condition that the storage resources are set to 32N banks, 16N paths of parallel calculation can be realized.
The cooperation module calculates the Euclidean distance between the winning neuron and other competition layer neurons, can remove the root of the Euclidean distance, and performs square calculation on the field radius, so that the logic resource and the operation time are reduced, under the condition that the storage resource is set to be 32N banks, 8N paths of parallel calculation are realized, and the competition layer neurons within the field radius of the winning neuron are jointly excited according to a certain rule.
The hardware architecture shown in fig. 1 is an example of the present invention, and the example implements the construction of a self-organizing map algorithm, including a control module, a distance calculation module, a learning rate and domain radius calculation module, a cooperation module, a weight value update module, and the transportation and storage of data.
Fig. 2 is a network structure diagram of the self-organizing map algorithm, wherein each circle of the input layer represents a feature, each circle of the competition layer represents a classification, but the actual classification situation is less than the number of circles of the competition layer. Each input has a weight for the competition layer neurons.
Fig. 3 is a schematic diagram of data arrangement, where the storage resource is set to 32N banks, each bank has a depth of M, 16N banks are allocated to store data of all input neurons, and a randomly generated weight is used for ping-pong access in another 8N banks.
Fig. 4 is a schematic diagram of hardware modules, and the process is as follows:
s1: after receiving the starting signal, the address generation module generates a source data address, and the controller reads the source data according to the signal;
s2: the training unit is used for training, iteration is carried out for multiple times to obtain updated weights, and the trained weights are stored on the SRAM by the controller;
s3: the reasoning unit classifies the source data according to the trained weight;
the process of the design schematic diagram of the computing unit of the collaboration module shown in FIG. 5 is as follows:
s1: calculating the domain radius of the winning neuron, and updating the domain radius through a shifting operation;
s2: giving the fixed coordinate position of the neuron of the competition layer, and calculating the square of the Euclidean distance between the winning neuron and the neuron of the competition layer through 8 paths in parallel;
s3: comparing the square of Euclidean distance between the winning neuron and the neuron in the competition layer with the square of the radius of the domain of the winning neuron, wherein the neuron updates the weight according to the rule within the radius of the domain;
the specific process of updating the weights shown in fig. 6 is as follows:
s1: taking out input neurons and weights from the bank for weight updating;
s2: obtaining a competition layer neuron needing to update the weight value through a cooperation module according to Wi(n+1)=Wi(n)+η(n)hj,i(n)(X-Wi(n)) updating the weights, wherein the weight updating is performed through 16 paths of parallel calculation;
s3: the updated weight is stored in the bank again, and the weight before updating is covered;
example 1
Through the above description of the modules, the hardware implementation process of the self-organizing map algorithm shown in fig. 7 is as follows:
s1: storing the initialization weight W and the input neurons into an SRAM according to a rule, and initializing a field h, a learning rate and the maximum training times of a network;
s2: firstly, taking a sample from the total sample as network input, and taking corresponding input characteristics and weights from an SRAM;
s3: and calculating Manhattan distances from the input neuron to all the competitive layer neurons, and determining the position coordinates of the winning neuron according to the minimum distance. Calculating the distance between the neuron in competition layer with coordinates (i, j), finding the minimum distance value, and determining the calculation formula of winning neuron
Figure BDA0002245202000000081
S4: taking input and weight from SRAM, and adjusting network weight of winning neuron and neuron in field, wherein weight formula is Wi(n+1)=Wi(n)+η(n)hj,i(n)(X-Wi(n)), restoring the updated weight toIn the bank, covering the weight before updating;
s5: when all learning samples are not calculated, repeating the steps of S2-S5, otherwise executing S6;
s6: adjusting learning rate function and domain function, learning rate
Figure BDA0002245202000000082
Wherein eta0Is the initial learning rate; radius of area of
Figure BDA0002245202000000083
Wherein
Figure BDA0002245202000000084
S7: if the current training times do not reach the set maximum training times, returning to S2, otherwise, terminating the loop and ending the learning process.
The self-organizing mapping algorithm based on the SIMD architecture supports the operation of any sample classification and any sample size, reduces the source data calculation amount of the traditional hardware implementation mode, balances calculation and storage resources to realize the maximized multi-path parallel, and can effectively save operation resources and improve the calculation speed. As a typical clustering algorithm in the field of machine learning, the method has wide application prospect and good application value for different occasions.
Although the preferred embodiments of the present invention have been described in detail, the present invention is not limited to the details of the embodiments, and various equivalent modifications can be made within the technical spirit of the present invention, and the scope of the present invention is also within the scope of the present invention.

Claims (5)

1. The self-organizing mapping method based on the SIMD architecture is characterized in that under the condition that computing resources and storage resources are fixed, the same set of resources are used, one controller is adopted to control a plurality of processors, and the operations of the same operation are respectively executed, so that the parallel operation on the space is realized; the method comprises the following steps:
s1: storing the initialization weight W and the input neurons into a storage unit according to a rule, and initializing a field h, a learning rate and the maximum training times of the network; the storage unit comprises 8N subunits, wherein 4N subunits store input neuron data, randomly generated weights are stored in the other 2N subunits in a ping-pong manner, winning neurons obtained by training are stored in the other 2N subunits, and the winning neurons obtained by the next training cover the last winning neuron data;
s2: taking one sample from the total samples as network input, and taking corresponding input features and weights from a storage unit;
s3: calculating Manhattan distances from the input neurons to all the neurons in the competition layer, determining position coordinates of the winning neurons according to the minimum distance, calculating the distance from the input neurons to the neurons in the competition layer with coordinates (i, j), finding out the minimum distance value, and determining a calculation formula of the winning neurons
Figure DEST_PATH_IMAGE001
S4: taking out the input and weight from the storage unit, and adjusting the network weight of the winning neuron and the neuron in the field, wherein the weight is expressed by
Figure DEST_PATH_IMAGE002
The updated weight is stored in the storage unit again to cover the weight before updating;
s5: when all learning samples are not calculated, repeating the steps of S2-S5, otherwise executing S6;
s6: adjusting learning rate function and domain function, learning rate
Figure DEST_PATH_IMAGE003
Wherein
Figure DEST_PATH_IMAGE004
Is the initial learning rate; radius of area of
Figure DEST_PATH_IMAGE005
Wherein
Figure DEST_PATH_IMAGE006
S7: if the current training times do not reach the set maximum training times, returning to the step S2, otherwise, terminating the circulation and ending the learning process.
2. A SIMD architecture based self-organizing map method according to claim 1, comprising the steps of: the whole input sample can be traversed by D batches, the total training frequency is K, and the frequency of ping-pong operation required by the weight is DxK.
3. A computer device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the steps of the method of any of claims 1 to 2 are implemented by the processor when executing the computer program.
4. A computer-readable storage medium, on which a computer program is stored, which, when being executed by a processor, carries out the steps of the method of any one of claims 1 to 2.
5. A SIMD architecture based self-organizing map system comprising:
the data storage module comprises a plurality of storage units with preset depths, each storage unit comprises 8N subunits, wherein 4N subunits store input neuron data, randomly generated weights are stored in the other 2N subunits in a ping-pong manner, the winning neurons obtained by training are stored in the other 2N subunits, and the winning neurons obtained by the next training cover the last winning neuron data;
the distance calculation module is used for respectively taking out the data of each input neuron and the corresponding weight from the storage unit, calculating the Manhattan distance sum of the input neuron and the weight corresponding to each neuron of the competition layer, calculating the minimum distance by a bubble sorting method, obtaining the coordinate of a winning neuron, and finally storing the coordinate into the storage module;
the learning rate and field radius calculation module is used for calculating the learning rate and the field radius by using shift operation and adopting two registers for storage, and the calculation result is written back to the original register again each time;
the weight updating module and the cooperation module are used for taking out the neuron data and the weights from the storage unit and carrying out updating operation, and the updated weights are stored in the storage unit according to ping-pong storage.
CN201911014330.5A 2019-10-23 2019-10-23 Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture Active CN110837891B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911014330.5A CN110837891B (en) 2019-10-23 2019-10-23 Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911014330.5A CN110837891B (en) 2019-10-23 2019-10-23 Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture

Publications (2)

Publication Number Publication Date
CN110837891A CN110837891A (en) 2020-02-25
CN110837891B true CN110837891B (en) 2022-05-17

Family

ID=69575598

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911014330.5A Active CN110837891B (en) 2019-10-23 2019-10-23 Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture

Country Status (1)

Country Link
CN (1) CN110837891B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112015854B (en) * 2020-07-17 2023-07-18 河海大学常州校区 Heterogeneous data attribute association method based on self-organizing mapping neural network
CN113705858B (en) * 2021-08-02 2023-07-11 西安交通大学 Shortest path planning method, system, equipment and storage medium for multiple target areas
CN114637720B (en) * 2021-12-13 2024-04-30 西安电子科技大学 On-chip optical network mapping method based on growth type annular SOM neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009099008A (en) * 2007-10-18 2009-05-07 Seiko Epson Corp Parallel arithmetic unit and parallel arithmetic method
CN103019656A (en) * 2012-12-04 2013-04-03 中国科学院半导体研究所 Dynamically reconfigurable multi-stage parallel single instruction multiple data array processing system
CN104823482A (en) * 2013-09-27 2015-08-05 华为技术有限公司 Switching method for terminal, base station and system
CN108734274A (en) * 2017-04-24 2018-11-02 英特尔公司 Calculation optimization mechanism for deep neural network
CN108805798A (en) * 2017-05-05 2018-11-13 英特尔公司 Fine granularity for deep learning frame calculates communication and executes

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9449257B2 (en) * 2012-12-04 2016-09-20 Institute Of Semiconductors, Chinese Academy Of Sciences Dynamically reconstructable multistage parallel single instruction multiple data array processing system
US10732621B2 (en) * 2016-05-09 2020-08-04 Strong Force Iot Portfolio 2016, Llc Methods and systems for process adaptation in an internet of things downstream oil and gas environment
CN106529060B (en) * 2016-11-15 2021-09-03 中国电力科学研究院有限公司 Load sequence modeling method and system
CN108170147B (en) * 2017-12-31 2020-10-16 南京邮电大学 Unmanned aerial vehicle task planning method based on self-organizing neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009099008A (en) * 2007-10-18 2009-05-07 Seiko Epson Corp Parallel arithmetic unit and parallel arithmetic method
CN103019656A (en) * 2012-12-04 2013-04-03 中国科学院半导体研究所 Dynamically reconfigurable multi-stage parallel single instruction multiple data array processing system
CN104823482A (en) * 2013-09-27 2015-08-05 华为技术有限公司 Switching method for terminal, base station and system
CN108734274A (en) * 2017-04-24 2018-11-02 英特尔公司 Calculation optimization mechanism for deep neural network
CN108805798A (en) * 2017-05-05 2018-11-13 英特尔公司 Fine granularity for deep learning frame calculates communication and executes

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
A Pipelined Speculative SIMD Architecture for SOM ANN;0.Hammami 等;《Proceedings of International Conference on Neural Networks (ICNN"97)》;20020806;985-990 *
Multi-port SRAM with Multi-bank for Self-organizing Maps Neural Network;FengweiAn;《 14th IEEE International Conference on Solid-State and Integrated Circuit Technology (ICSICT)》;20181206;1-3 *
Parallel FPGA Implementation of Self-Organizing Maps;K. Ben Khalifa 等;《The 16th International Conference on Microelectronics, 2004. ICM 2004》;20050531;709-712 *
基于SOM算法的高维数据可视化;石丽红;《中国优秀博硕士学位论文全文数据库(硕士) 信息科技辑》;20140215(第(2014)02期);I138-606 *
用Norm Matrix实现自组织映射网络的可视化;郭景峰 等;《小型微型计算机系统》;20131115;第34卷(第11期);2630-2634 *

Also Published As

Publication number Publication date
CN110837891A (en) 2020-02-25

Similar Documents

Publication Publication Date Title
Ojha et al. Metaheuristic design of feedforward neural networks: A review of two decades of research
CN110837891B (en) Self-organizing mapping method and system based on SIMD (Single instruction multiple data) architecture
Liao et al. On the importance of normalisation layers in deep learning with piecewise linear activation units
Abraham Adaptation of fuzzy inference system using neural learning
Smith Neural networks for combinatorial optimization: a review of more than a decade of research
Juang et al. A locally recurrent fuzzy neural network with support vector regression for dynamic-system modeling
Yu et al. Evolutionary fuzzy neural networks for hybrid financial prediction
Khan et al. Artificial neural network (ANNs)
Dozono et al. Convolutional self organizing map
CN114511021A (en) Extreme learning machine classification algorithm based on improved crow search algorithm
Ajibade et al. A hybrid chaotic particle swarm optimization with differential evolution for feature selection
CN108983180A (en) A kind of high-precision radar sea clutter forecast system of colony intelligence
Salimi et al. Extended mixture of MLP experts by hybrid of conjugate gradient method and modified cuckoo search
Amari Mathematical theory of neural learning
Fortier et al. DOSI: Training artificial neural networks using overlapping swarm intelligence with local credit assignment
Espinós Longa et al. Swarm Intelligence in Cooperative Environments: Introducing the N-Step Dynamic Tree Search Algorithm
Szepesvári et al. Topology learning solved by extended objects: a neural network model
de Lima Mendes et al. gaCNN: Composing cnns and gas to build an optimized hybrid classification architecture
Wong et al. R-POPTVR: a novel reinforcement-based POPTVR fuzzy neural network for pattern classification
Weng et al. Optimal in-place self-organization for cortical development: Limited cells, sparse coding and cortical topography
Maksim et al. Development of a software library for game artificial intelligence
Zarth et al. Optimization of neural networks weights and architecture: A multimodal methodology
Li et al. An efficient feature selection algorithm for computer-aided polyp detection
CN113705858B (en) Shortest path planning method, system, equipment and storage medium for multiple target areas
Prieto et al. Simulation and hardware implementation of competitive learning neural networks

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant