CN105095967B - A kind of multi-modal neuromorphic network core - Google Patents

A kind of multi-modal neuromorphic network core Download PDF

Info

Publication number
CN105095967B
CN105095967B CN201510419465.5A CN201510419465A CN105095967B CN 105095967 B CN105095967 B CN 105095967B CN 201510419465 A CN201510419465 A CN 201510419465A CN 105095967 B CN105095967 B CN 105095967B
Authority
CN
China
Prior art keywords
dendron
unit
modal
network core
aixs cylinder
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201510419465.5A
Other languages
Chinese (zh)
Other versions
CN105095967A (en
Inventor
裴京
施路平
王栋
邓磊
徐海峥
李国齐
马骋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Ling Xi Technology Co. Ltd.
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201510419465.5A priority Critical patent/CN105095967B/en
Publication of CN105095967A publication Critical patent/CN105095967A/en
Application granted granted Critical
Publication of CN105095967B publication Critical patent/CN105095967B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Feedback Control In General (AREA)

Abstract

The present invention provides a kind of multi-modal neuromorphic network core, including:Mode register, aixs cylinder input block, synapse weight memory cell, dendron unit and neuron computing unit.The multi-modal neuromorphic network core can both carry out artificial neural network computing, can also carry out impulsive neural networks computing, and can switch as required between artificial neural network operational mode and impulsive neural networks operational mode.

Description

A kind of multi-modal neuromorphic network core
Technical field
The present invention relates to a kind of neuromorphic network core.
Background technology
Neutral net is the computing system that a kind of mimic biology brain cynapse-neuronal structure carries out data processing, by dividing The connection composition of calculate node and interlayer for multilayer.Each node simulates a neuron, performs some certain operations, such as Activation primitive, the connection analog neuron cynapse between node, connects and is represented for the weighted value inputted from last layer node Synapse weight.Neutral net has powerful non-linear, adaptive information disposal ability.
Neuron in artificial neural network is used as itself after the accumulated value from connection input is handled by the use of activation primitive Output.Corresponding to different network topology structures, neuron models and learning rules, artificial neural network includes perceiving again The tens of kinds of network models such as device, Hopfield networks, Boltzmann machine, it is possible to achieve diversified function, pattern-recognition, Complex control, signal transacting and optimization etc. have application.Traditional artificial neural network data, which may be considered, to be passed through The frequency information coding of neuron pulse, each layer neuron are serially run successively.The nerve of artificial Neural Network Simulation biology System hlerarchy, but influence of the information processing architecture such as time serieses of matching cortex completely to study is the failure to, and make For real biological cortex in processing information for, the study to information data is not independent static state, but over time There is the contact of context.Impulsive neural networks are the new neural networks occurred in recent ten years, the third generation of being known as Neutral net.Data in impulsive neural networks are encoded with the space time information of neuron pulse signal, the input and output of network with And the information transmission between neuron shows as the pulse of neuron transmission and sends the temporal information of pulse, neuron needs same Shi Binghang is run.Compared with traditional artificial neural network, impulsive neural networks are in information processing manner, neuron models, simultaneously Row etc. has a relatively big difference, and the method for operation is closer to real biosystem.Impulsive neural networks application accurate timing Pulse train nerve information is encoded and handled, this computation model comprising time calculating elements more biology explain Property, it is the effective tool for carrying out complicated space time information processing, multi-modal information can be handled and information processing is more real-time. But the discontinuity of the neuron models of impulsive neural networks, the complexity of space-time code, the uncertainty of network structure cause It is difficult to mathematically complete to the overall description of network, therefore, it is difficult to build effective and general supervised learning algorithm, limitation Its calculation scale and accuracy.
The content of the invention
In view of this, it is necessory to provide it is a kind of can carry out artificial neural networks or enter horizontal pulse nerve The neuromorphic network core of network calculations.
A kind of multi-modal neuromorphic network core, including:Mode register, aixs cylinder input block, synapse weight storage are single Member, dendron unit and neuron computing unit;
The mode register is connected with the aixs cylinder input block, dendron unit and neuron computing unit, control Said units operate in artificial neural network pattern or impulsive neural networks pattern;
The aixs cylinder input block is connected with the dendron unit, is received and is stored aixs cylinder input;
The synapse weight memory cell is connected with the dendron unit, stores synapse weight matrix;
The dendron unit is connected with the neuron computing unit, including dendron multiplicaton addition unit and dendron summing elements, When operating in artificial neural network pattern, the aixs cylinder input vector is sent into the dendron multiplicaton addition unit with synapse weight matrix and entered Row multiply-add operation, when operating in impulsive neural networks pattern, the aixs cylinder input vector is sent into the tree with synapse weight matrix Prominent summing elements carry out accumulating operation;
The neuron computing unit includes the first computing unit and the second computing unit, operates in artificial neural network mould During formula, the multiply-add operation result that the dendron multiplicaton addition unit is sent is sent into the first computing unit progress artificial neural network fortune Calculate, when operating in impulsive neural networks pattern, the accumulating operation result that the dendron summing elements are sent is sent into second meter Calculate unit and carry out impulsive neural networks calculating.
Compared with prior art, multi-modal neuromorphic network core provided by the invention can both carry out artificial neural network Computing, impulsive neural networks computing can also be carried out, and can be as required in artificial neural network operational mode and pulse god Through switching between Network operation mode, real-time, multi-modal or complicated space-time signal can be carried out and calculate and can guarantee that calculating Accuracy.
Brief description of the drawings
Fig. 1 is in the hybrid system of the artificial neural network that first embodiment of the invention provides and impulsive neural networks Basic computational ele- ment structure chart.
Fig. 2 is the cascaded structure schematic diagram of the present invention
Fig. 3 is the parallel-connection structure schematic diagram of the present invention.
Fig. 4 is the parallel organization schematic diagram of the present invention.
Fig. 5 is the learning structure schematic diagram of the present invention.
Fig. 6 is the feedback arrangement schematic diagram of the present invention.
Fig. 7 is that elementary layer level structure schematic diagram is calculated in hybrid system provided by the invention.
Fig. 8 is the hybrid system of artificial neural network provided by the invention and impulsive neural networks.
Fig. 9 is the signal that the numerical quantities of artificial neural network output are converted to pulse train in second embodiment of the invention Figure.
Figure 10 is that the frequency coding pulse train in second embodiment of the invention by impulsive neural networks output is converted to number Value amount schematic diagram.
Figure 11 is that the Population Coding pulse train in second embodiment of the invention by impulsive neural networks output is converted to number Value amount schematic diagram.
Figure 12 is that the time encoding pulse train in second embodiment of the invention by impulsive neural networks output is converted to number Value amount schematic diagram.
Figure 13 is that the binary-coding pulse train in second embodiment of the invention by impulsive neural networks output is converted to number Value amount schematic diagram.
Figure 14 is the multi-modal neuromorphic network nuclear structure block diagram that third embodiment of the invention provides.
When Figure 15 is that the multi-modal neuromorphic network core that third embodiment of the invention provides operates in artificial neural network Structured flowchart.
Figure 16 is the operational flow diagram of one time step of multi-modal neuromorphic network core under artificial neural network pattern.
When Figure 17 is that the multi-modal neuromorphic network core that third embodiment of the invention provides operates in impulsive neural networks Structured flowchart.
Figure 18 is the operational flow diagram of one time step of multi-modal neuromorphic network core under impulsive neural networks pattern.
Figure 19 is the hybrid system of the artificial neural network that fourth embodiment of the invention provides and impulsive neural networks.
Figure 20 is fourth embodiment of the invention routing nodes structured flowchart.
Figure 21 is fourth embodiment of the invention routing data inclusion composition.
Figure 22 is fourth embodiment of the invention routing nodes workflow diagram.
Main element symbol description
Hybrid system 100 Mode register 211
Basic computational ele- ment 110 Aixs cylinder input block 212
First basic computational ele- ment 110a Synapse weight memory cell 213
Second basic computational ele- ment 110b Dendron unit 214
Unit 111 Dendron multiplicaton addition unit 214a
Neuron 115 Dendron summing elements 214b
Cynapse 116 Neuron computing unit 215
Composite computing unit 120 First computing unit 215a
Series connection recombiner unit 120a Second computing unit 215b
Recombiner unit in parallel 120b Dendron expands memory cell 2151
Parallel composition unit 120c Parameter storage unit 2152
Learn recombiner unit 120d Integrate leak calculation unit 2153
Feedback complex unit 120e Trigger signal counter 216
Hybrid system 200 Controller 217
Neuromorphic network core 210 Routing node 220
Multi-modal neuromorphic network core 210a
Following embodiment will combine above-mentioned accompanying drawing and further illustrate the present invention.
Embodiment
Multi-modal neuromorphic network core provided by the invention is made below in conjunction with the accompanying drawings and the specific embodiments further Detailed description.
First embodiment of the invention provides a kind of hybrid system 100 of artificial neural network and impulsive neural networks, Including at least two basic computational ele- ments 110, at least two basic computational ele- ment 110, at least one is ANN Network computing unit, artificial neural networks are undertaken, at least one is impulsive neural networks computing unit, undertakes pulse nerve net Network is calculated, and at least two basic computational ele- ment 110 is connected with each other according to topological structure, realizes neural computing work(jointly Energy.
Fig. 1 is referred to, an at least artificial neural networks unit calculates with an at least impulsive neural networks Unit is considered as an independent neutral net respectively, and the neutral net includes multiple neurons 115, the plurality of neuron Connected between 115 by cynapse 116, form single or multiple lift structure.Synapse weight represents postsynaptic neuron and receives the presynaptic The weighted value of neuron output.
An at least impulsive neural networks computing unit is used to perform impulsive neural networks calculating to the data received. The data at least transmitted between the input data of an impulsive neural networks computing unit, output data and neuron 115 are Spike sequence, the model of the neuron 115 at least described in an impulsive neural networks computing unit is based on spike arteries and veins Rush calculate neuron models, can be but be not limited to leakage-integration-fire model, spike response model and At least one of Hodgkin-Huxley models.
An at least artificial neural networks unit is used to perform artificial neural networks to the data received. The data at least transmitted between the input data of an artificial neural networks unit, output data and neuron 115 are Numerical quantities.At least artificial neural networks unit is further according to neuron models, network structure, learning algorithm Difference, can be perceptron neural network computing unit, BP neural network computing unit, Hopfield neural computing lists Member, adaptive resonance theory neural computing unit, depth conviction neural computing unit and convolutional neural networks calculate At least one of unit.
An at least artificial neural networks unit and at least an impulsive neural networks computing unit Topology connection To form a complex neural network computing unit.
The topological structure of the Topology connection includes cascaded structure, parallel-connection structure, parallel organization, learning structure and feedback At least one of structure.
Fig. 2 is referred to, described two basic computational ele- ments 110 are connected in series to form a series connection composite computing unit 120a.Described two basic computational ele- ments 110 are respectively the first basic computational ele- ment 110a and the second basic computational ele- ment 110b, The output end of the first basic computational ele- ment 110a connects the second basic computational ele- ment 110b input.Described first is basic In computing unit 110a and the second basic computational ele- ment 110b, one is artificial neural network computing unit, and another is pulse Neural computing unit.System input first passes around the first basic computational ele- ment 110a processing, and the result after processing is used as the Two basic computational ele- ment 110b input, the result after the second basic computational ele- ment 110b processing export for system.
Fig. 3 is referred to, described two basic basic computational ele- ments 110 are connected in parallel to form a recombiner unit in parallel 120b.Described two basic computational ele- ments 110 are respectively the first basic computational ele- ment 110a and the second basic computational ele- ment 110b, Input connection the second basic computational ele- ment 110b of first basic computational ele- ment 110a input, described first Basic computational ele- ment 110a output end connects the output end of the second basic computational ele- ment 110b.First basic calculating In unit 110a and the second basic computational ele- ment 110b, one is artificial neural network computing unit, and another is pulse nerve Network calculations unit.System input is input to the first basic computational ele- ment 110a and second basic computational ele- ment simultaneously 110b carries out parallel processing, the place that the first basic computational ele- ment 110a and the second basic computational ele- ment 110b are each obtained Reason result is collected, and is exported as system.
Fig. 4 is referred to, described two basic basic computational ele- ments 110 are connected in parallel to form a parallel recombiner unit 120c.Described two basic computational ele- ments 110 are respectively the first basic computational ele- ment 110a and the second basic computational ele- ment 110b, The input of the first basic computational ele- ment 110a and the second basic computational ele- ment 110b input are each independent, the first base This computing unit 110a output end connects the second basic computational ele- ment 110b output end.First basic computational ele- ment In 110a and the second basic computational ele- ment 110b, one is artificial neural network computing unit, and another is impulsive neural networks Computing unit.System input is divided into 2 two parts of input 1 and input, wherein the first basic computational ele- ment 110a is sent into input 1, By the first basic computational ele- ment 110a processing, the second basic computational ele- ment 110b is sent into input 2, by the second basic calculating list First 110b processing, make after the first basic computational ele- ment 110a and the second basic computational ele- ment 110b result is collected Exported for system.
Fig. 5 is referred to, it is multiple that described two basic basic computational ele- ments 110 and a unit 111 form a study Close unit 120d.Described two basic computational ele- ments 110 are respectively the first basic computational ele- ment 110a and the second basic calculating list First 110b.In the first basic computational ele- ment 110a and the second basic computational ele- ment 110b, one is artificial neural network meter Unit is calculated, another is impulsive neural networks computing unit.System input obtains after the first basic computational ele- ment 110a processing Reality output, the difference of the reality output and target output is input to unit 111, the unit 111 is according to this Difference adjusts the parameters such as the network structure of the second basic computational ele- ment 110b, synapse weight, in the unit Practising algorithm can be for Delta rules, BP algorithm, simulated annealing, genetic algorithm etc., the learning algorithm used in the present embodiment For BP algorithm.The second basic computational ele- ment 110b output can as the first basic computational ele- ment 110a network structure, The parameters such as synapse weight, or according to the second basic computational ele- ment 110b the first basic computational ele- ment of output adjustment 110a's The parameters such as network structure, synapse weight.
Refer to Fig. 6, in a certain embodiment of the present invention, described two basic computational ele- ments 110 form a feedback complex list First 120e.Described two basic computational ele- ments 110 are respectively the first basic computational ele- ment 110a and the second basic computational ele- ment 110b.First basic computational ele- ment 110a output end is connected with the second basic computational ele- ment 110b input, and second is basic Computing unit 110b result of calculation is output to the first basic computational ele- ment 110a as feedback.First basic computational ele- ment In 110a and the second basic computational ele- ment 110b, one be artificial neural network computing unit another be impulsive neural networks meter Calculate unit.System input exports after the first basic computational ele- ment 110a processing, and output result is as the second basic computational ele- ment 110b input, the second basic computational ele- ment 110b output are input to the first basic computational ele- ment 110a as value of feedback.
Each example is to be combined two basic computational ele- ments 110 under certain topological structure above, is formed various Composite computing unit, further, greater number of basic computational ele- ment 110 can also be carried out under certain topological structure Combination, forms various composite computing units, and various composite computing units are carried out into group again under certain topological structure Close, form increasingly complex mixing and calculate structure, structure is calculated so as to produce rich and varied mixing.Refer to Fig. 7, first layer Composite computing unit 120 is that a serial mixing calculates structure, and the first layer composite computing unit 120 is decomposed into two in the second layer The series connection of individual composite computing unit 120, further, the second layer composite computing unit 120 are broken down into two in third layer The parallel connection of composite computing unit 120, above-mentioned decomposable process can go on always, until being broken down into substantially in last layer Computing unit 110, basic computational ele- ment 110 are minimum computing unit structure.Fig. 8 is referred to, the figure is by above-mentioned level The mixing of specific an artificial neural network and impulsive neural networks that design method obtains calculates structure, including series connection Structure, parallel-connection structure and feedback arrangement.
The hybrid system 100 of the artificial neural network and impulsive neural networks further comprises at least one form Converting unit, it is arranged between the artificial neural networks unit and pulse nerve net computing unit, form conversion Unit is used to realize the data transfer between different types of neural computing unit.The format conversion unit can be by people The numerical quantities of artificial neural networks computing unit output are converted to pulse train, or impulsive neural networks computing unit is exported Pulse train is converted to numerical quantities and realizes that form is changed, to ensure to carry out data between different types of basic computational ele- ment 110 Transmission.
Further, the hybrid system 100 of the artificial neural network and impulsive neural networks may include multiple institutes State artificial neural network unit and multiple impulsive neural networks units, the plurality of artificial neural network unit and multiple pulses god Topology connection is realized with above-described topological structure through NE.
The artificial neural network and the hybrid system 100 of impulsive neural networks that first embodiment of the invention provides combine The computation schemas of two kinds of neutral nets of artificial neural network and impulsive neural networks, artificial neural network is used to needing accurate Data processing either need complete mathematical describe computing unit, by impulsive neural networks be used for need snap information handle or Person is complicated space-time signal processing or needs to handle the computing unit of multi-modal signal (such as audio visual signal) simultaneously, is formed Real-time, multi-modal or complicated space-time signal and the system that can guarantee that counting accuracy can be carried out.Such as done by the system Audio visual Integrated Real-time Processing, it can will include the multi-modal complicated space-time signal input pulse nerve of image (video) and sound Network calculations unit is pre-processed, quick to reduce space-time characteristic needed for signal Space-time Complexity or extract real-time, pretreatment Data afterwards continue to input artificial neural networks unit, can be by building Complete mathematic model or having supervision in this step The mode of learning algorithm builds the artificial neural network that can be realized compared with precise information processing, ensure that the accuracy of output.
Second embodiment of the invention provides a kind of mixed communication method of artificial neural network and impulsive neural networks nerve, Including:Whether the data type of sender and recipient are consistent in the neutral net basic computational ele- ment 110 that judgement is communicated, Carry out data transmission if consistent, if inconsistent, perform Data Format Transform, be by the data type conversion that sender sends With recipient's identical data type, and carry out data transmission.The data type of artificial neural network is numerical value, pulse nerve net The data type of network is pulse train.
Specifically, during the Data Format Transform is performed, if sender is the artificial neural networks Unit, recipient are impulsive neural networks computing unit, then the numeric format output of artificial neural network are converted into pulse sequence Row input pulse neutral net;If sender is impulsive neural networks computing unit, recipient is artificial neural networks list Member, then the pulse train form output of impulsive neural networks is converted into numerical value input artificial neural network.
The form of the different types of data input of basic computational ele- ment 110 output difference.Artificial neural network is based on Numerical operation, input and output are all numerical value, and impulsive neural networks are based on pulse train computing, and input and output are all pulse train. Different types of basic computational ele- ment 110 is integrated into same mixing to calculate in structure, it is necessary to solve the problems, such as to communicate each other. Need to provide a kind of artificial neural network the mixed communication method neural with impulsive neural networks so that artificial neural network meter Calculating the numeric format output of unit can be received by impulsive neural networks computing unit, and impulsive neural networks computing unit The output of pulse train form can be received by artificial neural networks unit.
Refer to Fig. 9, in a certain embodiment of the present invention, the impulsive neural networks unit is data receiver, the people Artificial neural networks unit is data sender, and the communication process is:The numerical quantities that artificial neural networks unit is exported The pulse train of respective frequencies, and the input using the pulse train as impulsive neural networks computing unit are converted to, it is described right Frequency is answered to refer to that the frequency of the pulse train after conversion is directly proportional to the size of numerical quantities.
In a certain embodiment of the present invention, the artificial neural network unit is data receiver, the impulsive neural networks Unit is data sender, and the communication process is:The pulse train form output of impulsive neural networks is converted to corresponding big Small numerical quantities, and using the numerical quantities as the input as artificial neural network.Compiled according to impulsive neural networks pulse train The difference of code mode, can be divided into following 4 kinds of situations again:
Figure 10 is referred to, pulse train uses frequency coding, and network output effective information is only represented by pulse frequency.Work as arteries and veins When rushing neutral net and using this coded system, the pulse train of frequency coding is converted to the method for numerical quantities is:By pulse Sequence is converted to the numerical quantities of corresponding numerical value, and the corresponding numerical value refers to the size of numerical quantities and the frequency of pulse train into just Than, that is, the inverse process of the above-mentioned communication means from artificial neural network to impulsive neural networks.
Figure 11 is referred to, pulse train is Population Coding, the characterized same information of output of multiple neurons, effectively letter Cease the neuron number that spike is sent for synchronization.It is corresponding logical when impulsive neural networks use this coded system Letter method is:The neuron number of the transmission spike of synchronization is converted into corresponding numerical value, the size of numerical value and granting The neuron number of spike is directly proportional.
Figure 12 is referred to, pulse train is time encoding, and effective information is the time that neuron sends spike.Work as arteries and veins When rushing neutral net and using this coded system, corresponding communication means is:The time that neuron is sent to spike changes For corresponding numerical value, size and the spike Time Of Release exponentially functional relation of numerical value.
Figure 13 is referred to, pulse train is binary-coding, and whether effective information provides point for neuron within a certain period of time Peak pulse.When impulsive neural networks use this coded system, corresponding communication means is:If point is provided in limiting time Peak pulse, then numerical value is 1, and otherwise numerical value is 0.
The mixed communication method of the artificial neural network that second embodiment of the invention provides and impulsive neural networks nerve is real The direct communication of artificial neural network and impulsive neural networks is showed, above two neutral net can be realized on this basis Hybrid operation.
Figure 14 is referred to, third embodiment of the invention provides a kind of multi-modal neuromorphic network core 210a, including:Pattern Register 211, aixs cylinder input block 212, synapse weight memory cell 213, dendron unit 214 and neuron computing unit 215.
In the present embodiment, the multi-modal neuromorphic network core 210a has a bars input aixs cylinder, b bars dendron and b nerve First cell space, every aixs cylinder are connected with a bar dendrons respectively, are a cynapse at each tie point, share a × b cynapse, connection weight It is synapse weight again.The corresponding pericaryon of every dendron.Therefore the multi-modal neuromorphic network core 210a is maximum The network size that can be carried is a inputs × b neurons.Wherein a, b are the integer more than 0, and a, b value are typically according to reality Depending on.A minimums can be taken as 1, and maximum can be realized to be limited to hardware resource, including storage resource, logical resource etc. Single neuromorphic network core maximum aixs cylinder inputs number;B minimum desirable 1, maximum while hardware resource is limited to, also by It is limited to the neural n ary operation total degree that each neuromorphic network core can perform in the single trigger signal cycle.A in the present embodiment Value with b is 256.It is appreciated that in actual applications, the specific number of above-mentioned input aixs cylinder, dendron and pericaryon Mesh can be adjusted according to concrete condition.
The mode register 211 controls the operational mode of the multi-modal neuromorphic network core 210a.The pattern Register 211 is connected with the aixs cylinder input block 212, dendron unit 214 and neuron computing unit 215, and control is above-mentioned Unit operates in artificial neural network pattern or impulsive neural networks pattern.The width of the mode register 211 is 1, Its register value can pass through user configuration.
The aixs cylinder input block 212 is connected with the dendron unit 214.The aixs cylinder input block 212 receives a bar axles It is prominent to input and store.In the present embodiment, every aixs cylinder has the memory cell of 16.When operating in impulsive neural networks pattern, The input of every aixs cylinder is all the spike of 1, and the memory cell stores the aixs cylinder input of 16 time steps.Work as operation In artificial neural network pattern, the input of every aixs cylinder is all the signed number of 8, and the memory cell stores a time 8 aixs cylinders input of step.
The synapse weight memory cell 213 is connected with the dendron unit 214, the synapse weight memory cell 213 A × b synapse weight is stored, in the present embodiment, the synapse weight is the signed number of 8.
The input of the dendron unit 214 connects the aixs cylinder input block 212 and synapse weight memory cell 213, Output end connects the neuron computing unit 215.The dendron unit 214 performs a aixs cylinder input vector and a × b cynapses The vector-matrix multiplication computing of weight matrix, a result of calculation that computing obtains are that the aixs cylinder of a neuron is inputted.Institute Stating dendron unit 214 includes dendron multiplicaton addition unit 214a and dendron summing elements 214b.When operating in artificial neural network pattern When, the aixs cylinder input vector is sent into the dendron multiplicaton addition unit 214a with synapse weight matrix and carries out multiply-add operation, moment of a vector Battle array multiplication is realized by multiplier and accumulator.When operating in impulsive neural networks pattern, the aixs cylinder input vector is with dashing forward Touch weight matrix and be sent into the dendron summing elements progress accumulating operation, now, aixs cylinder input is 1, and vector-matrix multiplication is led to Cross data selector and accumulator is realized.
The neuron computing unit 215 is used to perform neuron calculating, including the meters of the first computing unit 215a and second Calculate unit 215b.When operating in artificial neural network pattern, multiply-add operation result that the dendron multiplicaton addition unit 214a is sent It is sent into the first computing unit 215a and carries out artificial neural network computing.Neuron passes through one 1024 × 8 in the present embodiment The look-up tables'implementation Any Nonlinear Function function of position, is exported as 8 bit values.When operating in impulsive neural networks pattern, institute State the accumulating operation result that dendron summing elements 214b is sent and be sent into the second computing unit 215b progress impulsive neural networks Calculate.Neuron in the present embodiment is integration-leakage-igniting (LIF) model, is exported for spike signal and when cephacoria electricity Position.
The multi-modal neuromorphic network core 210a further comprises trigger signal counter 216, the trigger signal Counter 216 receives trigger signal and records trigger signal number, i.e., current time walks.The trigger signal for the fixed cycle when Clock signal, the cycle is 1ms in the present embodiment.
The multi-modal neuromorphic network core 210a further comprises controller 217.The controller 217 and the axle Prominent input block 212, synapse weight memory cell 213, dendron unit 214, neuron computing unit 215 connect, the control Device 217 controls above-mentioned aixs cylinder input block 212, synapse weight memory cell 213, dendron unit 214, neuron computing unit 215 operation sequential.In addition the controller 217 be also responsible for controlling the multi-modal neuromorphic network core 210a startup and Terminate, the controller 217 starts the multi-modal neuromorphic network core 210a computings simultaneously in the trigger signal rising edge Control computing flow.
Impulsive neural networks operational mode and the people of the multi-modal neuromorphic network core 210a will be introduced respectively below Artificial neural networks operational mode.
Figure 15 is referred to, the multi-modal neuromorphic network core 210a operates in artificial neural network pattern, the pattern Under the aixs cylinder input that receives of the aixs cylinder input block 212 be the packet of 16, including 8 target aixs cylinder sequence numbers (0~ 255) and 8 input (signed number -128~127).The now memory of the aixs cylinder input block 212 256 × 8, this is deposited Reservoir records 8 inputs in 256 aixs cylinders.
The synapse weight memory cell 213 stores 256 × 256 synaptic weight values, and each synaptic weight value is 8 Signed number.
The dendron multiplicaton addition unit 214a includes multiplier and adder, is calculated per time step by neuron on its dendron The product of 256 dimension synapse weight vectors and 256 dimension aixs cylinder input vectors, and using result of calculation as first computing unit 215a input.
The first computing unit 215a performs non-linear/linear activation primitive meter of neuron in artificial neural network Calculate.Before use, first according to used in the neuromorphic network activation primitive, calculate for value be 0~1023 input Corresponding 8 output, and above-mentioned result of calculation is stored in the look-up table storage of neuron computing unit.Operationally, in the future Inputted from the multiply-add result of dendron of the dendron multiplicaton addition unit 214a as the address of look-up table, the look-up table address is stored 8 outputs are neuron output.
Refer to Figure 16, under artificial neural network pattern, described multi-modal mono- time step of neuromorphic network core 210a Operational process comprise the following steps:
S11, after detecting trigger signal, from adding 1, the controller 217 starts for the circulation of trigger signal counter 216 Multi-modal neuromorphic network core 210a operations;
S12, the aixs cylinder input block 212 read current time step axle according to the value of the trigger signal counter 216 Prominent input vector, and the dendron multiplicaton addition unit 214a is sent to, the synapse weight memory cell 213 sequential reads out 1~256 Number neuron dendron synapse weight vector, and it is sent to the dendron multiplicaton addition unit 214a;
S13, the dendron multiplicaton addition unit 214a calculated successively according to input sequence the aixs cylinder input vector with described 1~ No. 256 neuron dendron synapse weight vector products, and it is sent to the first computing unit 215a;
S14, the first computing unit 215a obtain using the output of the dendron multiplicaton addition unit 214a as table address is searched Go out neuron output;
S15, the controller 217 stop the multi-modal neuromorphic network core 210a operations, and return to step S11.
Figure 17 is referred to, the multi-modal neuromorphic network core 210a operates in impulsive neural networks pattern, the pattern Under the aixs cylinder input that receives of the aixs cylinder input block 212 be the packet of 12, including 8 target aixs cylinder sequence numbers (0~ And 4 delayed datas (0~15) 255).The delayed data represents the difference that the time step that the input comes into force walks with current time Value.Now the memory of the aixs cylinder input block 212 is 256 × 16, the memory record it is current in 256 aixs cylinders and it The input of totally 16 time steps afterwards.If certain position is 1, corresponding aixs cylinder activates on corresponding time step, i.e. input is 1;If certain position For 0, then correspond to aixs cylinder and do not activated on corresponding time step, be i.e. input is 0.
The synapse weight memory cell 213 stores 256 × 256 synaptic weight values, and each synaptic weight value is 8 Signed number.
The dendron summing elements 214b includes data selector and adder, calculates current time on every dendron successively Walk all activated cynapse weight and, and using result of calculation as the second computing unit in the neuron computing unit 215 215b input.
The second computing unit 215b is used to carry out impulsive neural networks calculating, and the neuron in the present embodiment is product Point-leakage-igniting (LIF) model.The second computing unit 215b further comprises:Dendron expands memory cell 2151, ginseng Number memory cell 2152 and integration leak calculation unit 2153.In each time step, the second computing unit 215b is transported successively Row 256 times, the computing of 256 neurons is realized by time-multiplexed mode.Wherein described dendron expands memory cell 2151 Comprising 256 memory cell, input data bag is expanded for receiving the extraneous dendron sent, the dendron expands input data bag Including transmitting terminal membrane potential of neurons value and target nerve member sequence number, and according to neuron sequence number by transmitting terminal membrane potential of neurons Corresponding memory cell is arrived in value storage.The parameter storage unit 2152 includes 256 memory cell, stores 256 neurons Film potential, threshold value and leakage value.The integration leak calculation unit 2153 each neuron is performed successively integration-leakage- Ignition operation, when film potential exceedes positive threshold value, output spike pulse signal and current film potential value.It is wherein described to integrate-let out Leakage-ignition operation is as follows:
Film potential=former film potential+dendron input+dendron expands input-leakage value
If film potential is more than positive threshold value, spike signal and film potential value are sent, and film potential is reset.If Film potential is less than negative threshold value, then does not send signal and reset film potential.
Figure 18 is referred to, one time step of impulsive neural networks pattern of the multi-modal neuromorphic network core 210a Operational process comprises the following steps:
S21, after detecting trigger signal, from adding 1, the controller 217 starts for the circulation of trigger signal counter 216 Multi-modal neuromorphic network core 210a operations;
S22, the aixs cylinder input block 212 read current time step axle according to the value of the trigger signal counter 216 Prominent input vector, and the dendron summing elements 214b is sent to, the synapse weight memory cell 213 sequential reads out 1~256 The dendron synapse weight vector of number neuron, and it is sent to the dendron summing elements 214b;
S23, the dendron summing elements 214b calculated successively according to input sequence the aixs cylinder input vector with described 1~ No. 256 neuron dendron synapse weight vector products, and it is sent to the second computing unit 215b;
S24, the second computing unit 215b read 1~No. No. 256 god from the parameter storage unit 2152 successively Calculated through first parameter, and with the input value from the dendron summing elements 214b, output spike pulse signal and current Film potential;
S25, after neuron calculating terminates, the controller 217 stops multi-modal neuromorphic network core 210a operations, And return to step S21.
It is possible to further obtain two kinds on the basis of the multi-modal neuromorphic network core 210a that provides the present embodiment Single mode neuromorphic network core.A kind of single mode neuromorphic network core operated under artificial neural network pattern, including: Aixs cylinder input block 212, synapse weight memory cell 213, dendron multiplicaton addition unit 214a and the first computing unit 215a.It is a kind of The single mode neuromorphic network core operated under impulsive neural networks pattern, including:Aixs cylinder input block 212, synapse weight Memory cell 213, dendron summing elements 214b and the second computing unit 215b.Above two single mode neuromorphic network core It is that corresponding simplification has been done on the basis of 3rd embodiment, the specific attachment structure and function of internal each unit are referred to Third embodiment of the invention.
It is appreciated that multi-modal neuromorphic network core 210a and two kinds of single modes that third embodiment of the invention is provided State neuromorphic network core can calculate as the mixing of first embodiment of the invention artificial neural network and impulsive neural networks Basic computational ele- ment 110 in system 100.
The multi-modal neuromorphic network core 210a that third embodiment of the invention provides can both carry out artificial neural network Computing, impulsive neural networks computing can also be carried out, and can be as required in artificial neural network operational mode and pulse god Through switching between Network operation mode, real-time, multi-modal or complicated space-time signal can be carried out and calculate and can guarantee that calculating Accuracy.
Figure 19 is referred to, fourth embodiment of the invention provides a kind of mixed stocker of artificial neural network and impulsive neural networks System 200, including:Multiple neuromorphic network cores 210 and with the plurality of neuromorphic network core multiple routes correspondingly Node 220, the multiple routing node 220 form the route network of m × n network structures, wherein, m, n are whole more than 0 Number.Line direction in the m n array is defined as X-direction, column direction is defined as Y-direction, each pair neuromorphic network core 210 with Routing node 220 has an only local XY coordinate.
The neuromorphic network core 210 corresponds with routing node 220, refers to each neuromorphic network core 210 Corresponding to a routing node 220, each routing node 220 corresponds to a neuromorphic network core 210.One corresponding to mutually To in neuromorphic network core 210 and routing node 220, neuromorphic network core 210 is referred to as the local nerve of routing node 220 Form network core, routing node 220 are referred to as the local routing node of neuromorphic network core 210.The neuromorphic network core 210 input both is from its local routing node, and neuron result of calculation can also be sent to the local routing node, and then pass through Route network exports or is sent to target nerve form network core.
In the present embodiment, the quantity of the neuromorphic network core 210 and routing node 220 is 9,9 routes Node 220 forms the route network of 3 × 3 network structures.
The neuromorphic network core 210 can be single mode neuromorphic network core, or multi-modal neural shape State network core, such as the multi-modal neuromorphic network core 210a that third embodiment of the invention provides.So-called single mode nerve shape State network core refers to that the neuromorphic network core can only operate in artificial neural network pattern or impulsive neural networks pattern;It is so-called Multi-modal neuromorphic network core refers to that the neuromorphic network core has two kinds of operational modes:Artificial neural network pattern and pulse Network mode, it can be switched by configuring inherent parameters and realizing between above two operational mode, each multi-modal god Can single configuration operation pattern through form network core.
The neuromorphic network core 210 has multiple neuron elements and the input of multiple aixs cylinders.In the present embodiment, the god There are 256 neuron elements and 256 aixs cylinder inputs through form network core 210, maximum can undertake the god for including 256 neurons Through network operations.If the scale of neural network to be realized is more than 256 neurons, need to realize will be more by route network Individual neuromorphic network core 210 connects together on the topology, and each neuromorphic network core 210 undertakes a part of nerve Network operations, collectively constitute a big neutral net.The hybrid system 200 that the present embodiment is provided has 9 neuromorphic nets Network core 210, maximum can undertake the neural computing for including 2304 neurons, and the neural computing can be artificial god Hybrid network computing through network operations, impulsive neural networks computing or artificial neural network and impulsive neural networks.
When the neuromorphic network core 210 is operated under artificial neural network pattern, the neuromorphic network core 210 In the result of calculation of 256 neuron elements be all the numerical value of 8;When neuromorphic network core 210 operates in pulse nerve When under network mode, the result of calculation of the unit of 256 neurons in the neuromorphic network core 210 is all the spike of 1 Pulse.Which kind of no matter operate under pattern, the result of calculation of the unit of neuron can all be sent directly to the neuromorphic network The local routing node of core 210.
The multiple routing node 220 forms the route network of network structure, and the route network undertakes data-transformation facility, The data transfer includes:Transmission between system input, system output and neuromorphic network core 210, wherein neuromorphic Between network core 210 transmission again be divided into the internuclear data transfer of model identical neuromorphic network and different mode neuromorphic The internuclear data transfer of network.Any routing node 220 in the route network can use an XY that is unique and determining Plane coordinates represents.In the present embodiment, 9 routing nodes 220 form 3 × 3 array, and line direction in array is defined as into X side To column direction is defined as Y-direction, and each routing node 220 can be born with it in X positive directions, X negative directions, Y positive directions, Y The adjacent routing node in direction directly carries out data transmission, and forms mesh topology.It is appreciated that the mesh topology In addition to above-mentioned network structure, or other common structures such as hub-and-spoke configuration, bus type structure.
The data of the route network transmission include:System input data, system output data and neuromorphic network The data transmitted between core 210.Above-mentioned data are transmitted according to routing rule set in advance in route network.The present embodiment Middle routing rule is:Data are transmitted in X direction first, are exported again along Y-direction after reaching target X-coordinate routing node, directly To the routing node for reaching target XY coordinates.If representing the coordinate of starting routing node with (x0, y0), (x1, y1) represents mesh The coordinate of routing node is marked, above-mentioned routing rule is:(x0,y0)→(x1,y0)→(x1,y1).It is appreciated that actually should In, different routing rules can also be set according to specific circumstances.
The implementation process transmitted individually below between introducing system input, system output and neuromorphic network core 210.
The transmitting procedure of the system input data is:System input data is first inputted to appointing for route network outermost One routing node, target nerve form network core can be sent to according to above-mentioned routing rule by route network afterwards.
The transmitting procedure of the system output data is:The result of calculation of the neuromorphic network core 210 is sent first To local routing node, any route of route network outermost is sent to according to above-mentioned routing rule by route network afterwards Node, then be sent to by the routing node outside system, complete system output.
The process transmitted between the neuromorphic network core 210 is:The result of calculation of the neuromorphic network core 210 is first Local routing node is first sent to, target routing node is sent to according to above-mentioned routing rule by route network afterwards, then by The target routing node is sent to its local neuromorphic network core, completes the data transfer between neuromorphic network core 210.
Refer to Figure 20, the routing node 220 includes the routing table with multiple memory cell, the routing table it is each Memory cell corresponds to a neuron elements of local neuromorphic network core.The memory cell stores corresponding neuron list XY coordinate address, target aixs cylinder input sequence number and the delayed data of the purpose neuromorphic network core of member output.In the present embodiment Neuromorphic network core 210 includes 256 neuron elements, and the number of memory cell is also 256 in the routing table.
In the present embodiment, communicated between the system input data, system output data and neuromorphic network core 210 Data are transmitted in the form of 32 route data bags between routing node 220.Figure 21 show the lattice of above-mentioned route data bag Formula, 6 target nerve form network core X-direction addresses, 6 target nerve form network core Y sides are included in the route data bag Sequence numbers and 8 data are inputted to address, 4 aixs cylinder delays, 8 target aixs cylinders, altogether 32 data.Wherein, the aixs cylinder of 4 is prolonged When operate under impulsive neural networks pattern effectively when target nerve form network core, 8 data are artificial neural network pattern The output of lower neuron.
The packet that the routing node 220 is dealt into local neuromorphic network core is from above-mentioned 32 route data bags 4 aixs cylinders delay of interception, 8 target aixs cylinders input sequence numbers and 8 data, altogether 20 data.For operating in artificial god Through the neuromorphic network core 210 under network mode, after receiving the packet from local routing node, 8 data are made For the input of corresponding sequence number aixs cylinder.For operating in the neuromorphic network core 210 under impulsive neural networks pattern, receive and From after the packet of local routing node, input of the corresponding sequence number aixs cylinder after corresponding be delayed is put 1.
Figure 22 is referred to, the workflow of the routing node 220 includes:
S1, the routing node 220 are received come the neuron result of calculation for local neuromorphic network core of hanging oneself;
S2, the routing node 220 read the routing iinformation of corresponding neuron from routing table, and by the routing iinformation with The neuron result of calculation is combined as route data bag;
S3, the routing node 220 judge the sending direction of the route data bag, and the route is sent according to judged result Packet.
In step S1, when operating under artificial neural network pattern, the neuron result of calculation is output data;Work as fortune For row under artificial neural network pattern, the neuron result of calculation is spike.
In step S3, the route data bag include target nerve form network core in the relative address x of X-direction and Relative address y in the Y direction.The routing node 220 judges the sending direction of the route data bag according to x and y numerical value, Specially:Work as x>When 0, the route data bag is sent to the adjacent routing node of X positive directions, works as x<When 0, by the route Packet is sent to the adjacent routing node of X negative directions, works as y>When 0, it is adjacent that the route data bag is sent to Y positive directions Routing node, work as y<When 0, the route data bag is sent to the adjacent routing node of Y negative directions, as x=y=0, by institute State the local neuromorphic network core that route data bag directly transmits back the routing node 220.Next routing node receives After the route data bag that one routing node is sent, the relative address in the route data bag is modified, is specially:If x<0, Revised relative address x '=x+1 is then made, if x>0, then revised relative address x '=x-1 is made, if y<0, then order amendment Relative address y '=y+1 afterwards, if y>0, then make revised relative address y '=y-1.The relative address often passes through one Routing node 220 is once corrected, until x=y=0.
In the present embodiment, the judgement and transmission of X-direction are carried out first, after X-direction, which is sent, to be terminated, i.e. x '=0, then carry out The judgement and transmission of other direction, sent until by route data bag to target nerve form network core.
The artificial neural network and the hybrid system 200 of impulsive neural networks that fourth embodiment of the invention provides combine people The computation schema of two kinds of neutral nets of artificial neural networks and impulsive neural networks, both with impulsive neural networks complicated space-time letter Number disposal ability, and can makes full use of artificial neural network abundant and powerful computing capability, can carry out real-time, multi-modal Or complicated space-time signal calculates and can guarantee that the accuracy of calculating.
In addition, those skilled in the art can also do other changes in spirit of the invention, certainly, these are according to present invention essence The change that god is done, it should all be included within scope of the present invention.

Claims (11)

  1. A kind of 1. multi-modal neuromorphic network core, it is characterised in that including:Mode register, aixs cylinder input block, cynapse power Weight memory cell, dendron unit and neuron computing unit;
    The mode register is connected with the aixs cylinder input block, dendron unit and neuron computing unit, and control is above-mentioned Unit operates in artificial neural network pattern or impulsive neural networks pattern;
    The aixs cylinder input block is connected with the dendron unit, is received and is stored aixs cylinder input;
    The synapse weight memory cell is connected with the dendron unit, stores synapse weight matrix;
    The dendron unit is connected with the neuron computing unit, including dendron multiplicaton addition unit and dendron summing elements, operation In artificial neural network pattern, the aixs cylinder input vector is sent into the dendron multiplicaton addition unit with synapse weight matrix and multiplied Add computing, when operating in impulsive neural networks pattern, the aixs cylinder input vector is sent into the dendron with synapse weight matrix and tired out Unit is added to carry out accumulating operation;
    The neuron computing unit includes the first computing unit and the second computing unit, operates in artificial neural network pattern When, the multiply-add operation result that the dendron multiplicaton addition unit is sent is sent into the first computing unit progress artificial neural network fortune Calculate, when operating in impulsive neural networks pattern, the accumulating operation result that the dendron summing elements are sent is sent into second meter Calculate unit and carry out impulsive neural networks calculating.
  2. 2. multi-modal neuromorphic network core as claimed in claim 1, it is characterised in that the multi-modal neuromorphic network Core further comprises trigger signal counter, and the trigger signal counter is connected with the aixs cylinder input block, receives triggering Signal simultaneously records trigger signal number, i.e., current time is walked, and the trigger signal number is sent into the aixs cylinder input block.
  3. 3. multi-modal neuromorphic network core as claimed in claim 1, it is characterised in that the multi-modal neuromorphic network Core further comprises controller, the controller and the aixs cylinder input block, synapse weight memory cell, dendron unit, god Connected through first computing unit, control the operation sequential of connected said units, the controller controls the multi-modal god Startup and termination through form network core.
  4. 4. multi-modal neuromorphic network core as claimed in claim 1, it is characterised in that the dendron multiplicaton addition unit includes multiplying Musical instruments used in a Buddhist or Taoist mass and adder, for calculating the product of synapse weight vector and aixs cylinder input vector on its dendron.
  5. 5. multi-modal neuromorphic network core as claimed in claim 3, it is characterised in that first computing unit includes one Individual look-up table, look-up table is inputted using the multiply-add result as the address of look-up table, the output that the address is stored is nerve Member output.
  6. 6. multi-modal neuromorphic network core as claimed in claim 5, it is characterised in that described under artificial neural network pattern The operational process of one time step of multi-modal neuromorphic network core comprises the following steps:
    S11, after detecting trigger signal, for the trigger signal counter cycle from adding 1, the controller starts multi-modal nerve Form network core is run;
    S12, the aixs cylinder input block read current time step aixs cylinder input vector according to the value of the trigger signal counter, And the dendron multiplicaton addition unit is sent to, the synapse weight memory cell sequential reads out 1~n neuron dendron synapse weights Vector, and it is sent to the dendron multiplicaton addition unit;
    S13, the dendron multiplicaton addition unit calculate the aixs cylinder input vector and 1~n nerves according to input sequence successively First dendron synapse weight vector product, and it is sent to first computing unit;
    S14, first computing unit show that neuron is defeated using the output of the dendron multiplicaton addition unit as table address is searched Go out;
    S15, the controller stop the multi-modal neuromorphic network core operation, and return to step S11.
  7. 7. multi-modal neuromorphic network core as claimed in claim 1, it is characterised in that the dendron summing elements include number According to selector and adder, calculate successively on every dendron the weight of all activated cynapse of current time step and.
  8. 8. multi-modal neuromorphic network core as claimed in claim 3, it is characterised in that neural in second computing unit Member is integration-leakage-fire model, and second computing unit includes:Dendron expands memory cell, parameter storage unit and product Divide leak calculation unit, the expansion memory cell is connected with the integration leak calculation unit, stores transmitting terminal neuron membrane Potential value, the parameter storage unit are connected with the integration leak calculation unit, store the film potential of neuron, threshold value and let out Leakage value, the integration leak calculation unit perform integration-leakage-ignition operation.
  9. 9. multi-modal neuromorphic network core as claimed in claim 8, it is characterised in that described under impulsive neural networks pattern The operational process of one time step of multi-modal neuromorphic network core comprises the following steps:
    S21, after detecting trigger signal, for the trigger signal counter cycle from adding 1, the controller starts multi-modal nerve Form network core is run;
    S22, the aixs cylinder input block read current time step aixs cylinder input vector according to the value of the trigger signal counter, And the dendron summing elements are sent to, the synapse weight memory cell sequential reads out the dendron cynapse power of 1~n neurons Weight vector, and it is sent to the dendron summing elements;
    S23, the dendron summing elements calculate the aixs cylinder input vector and 1~n nerves according to input sequence successively First dendron synapse weight vector product, and it is sent to second computing unit;
    S24, second computing unit read 1~No. No. n neuron parameter from the parameter storage unit successively, and with coming Calculated from the input value of the dendron summing elements, output spike pulse signal and current film potential;
    S25, the controller stop the multi-modal neuromorphic network core operation, and return to step S21.
  10. 10. multi-modal neuromorphic network core as claimed in claim 1, it is characterised in that the multi-modal neuromorphic net There is network core a bars input aixs cylinder, b bars dendron and b pericaryon, every aixs cylinder to be connected respectively with b bar dendrons, each connection It is a cynapse at point, shares a × b cynapse, the corresponding pericaryon of every dendron, wherein a, b are to be whole more than 0 Number.
  11. 11. multi-modal neuromorphic network core as claimed in claim 10, it is characterised in that the a=b=256.
CN201510419465.5A 2015-07-16 2015-07-16 A kind of multi-modal neuromorphic network core Active CN105095967B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510419465.5A CN105095967B (en) 2015-07-16 2015-07-16 A kind of multi-modal neuromorphic network core

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510419465.5A CN105095967B (en) 2015-07-16 2015-07-16 A kind of multi-modal neuromorphic network core

Publications (2)

Publication Number Publication Date
CN105095967A CN105095967A (en) 2015-11-25
CN105095967B true CN105095967B (en) 2018-02-16

Family

ID=54576341

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510419465.5A Active CN105095967B (en) 2015-07-16 2015-07-16 A kind of multi-modal neuromorphic network core

Country Status (1)

Country Link
CN (1) CN105095967B (en)

Families Citing this family (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109242094B (en) * 2016-01-20 2020-05-08 中科寒武纪科技股份有限公司 Apparatus and method for performing artificial neural network forward operations
CN105719000B (en) * 2016-01-21 2018-02-16 广西师范大学 A kind of neuron hardware unit and the method with this unit simulation impulsive neural networks
US10878318B2 (en) * 2016-03-28 2020-12-29 Google Llc Adaptive artificial neural network selection techniques
US10990872B2 (en) * 2016-03-31 2021-04-27 International Business Machines Corporation Energy-efficient time-multiplexed neurosynaptic core for implementing neural networks spanning power- and area-efficiency
CN108510064B (en) * 2016-04-18 2021-12-10 中国科学院计算技术研究所 Processing system and method for artificial neural network comprising multiple core processing modules
US20170330069A1 (en) * 2016-05-11 2017-11-16 Kneron Inc. Multi-layer artificial neural network and controlling method thereof
CN106022468B (en) * 2016-05-17 2018-06-01 成都启英泰伦科技有限公司 the design method of artificial neural network processor integrated circuit and the integrated circuit
CN106056211B (en) * 2016-05-25 2018-11-23 清华大学 Neuron computing unit, neuron computing module and artificial neural networks core
CN106201651A (en) * 2016-06-27 2016-12-07 鄞州浙江清华长三角研究院创新中心 The simulator of neuromorphic chip
DE102016216944A1 (en) * 2016-09-07 2018-03-08 Robert Bosch Gmbh Method for calculating a neuron layer of a multilayer perceptron model with a simplified activation function
US11270193B2 (en) * 2016-09-30 2022-03-08 International Business Machines Corporation Scalable stream synaptic supercomputer for extreme throughput neural networks
CN106875005B (en) * 2017-01-20 2019-09-20 清华大学 Adaptive threshold neuronal messages processing method and system
CN106845633B (en) * 2017-01-25 2021-07-09 北京灵汐科技有限公司 Neural network information conversion method and system
CN106909969B (en) * 2017-01-25 2020-02-21 清华大学 Neural network information receiving method and system
CN106845632B (en) * 2017-01-25 2020-10-16 清华大学 Method and system for converting impulse neural network information into artificial neural network information
CN106971229B (en) * 2017-02-17 2020-04-21 清华大学 Neural network computing core information processing method and system
CN106971228B (en) * 2017-02-17 2020-04-07 北京灵汐科技有限公司 Method and system for sending neuron information
US10909449B2 (en) 2017-04-14 2021-02-02 Samsung Electronics Co., Ltd. Monolithic multi-bit weight cell for neuromorphic computing
CN107657312B (en) * 2017-09-18 2021-06-11 东南大学 Binary network implementation system for speech common word recognition
US20190332924A1 (en) * 2018-04-27 2019-10-31 International Business Machines Corporation Central scheduler and instruction dispatcher for a neural inference processor
KR102650660B1 (en) * 2018-09-03 2024-03-25 삼성전자주식회사 Neuromorphic apparatus and method for processing multi-bits neuromorphic operation thereof
CN109858620B (en) * 2018-12-29 2021-08-20 北京灵汐科技有限公司 Brain-like computing system
CN109901878B (en) * 2019-02-25 2021-07-23 北京灵汐科技有限公司 Brain-like computing chip and computing equipment
CN110263926B (en) * 2019-05-18 2023-03-24 南京惟心光电系统有限公司 Pulse neural network based on photoelectric computing unit, system and operation method thereof
CN110322010B (en) * 2019-07-02 2021-06-25 深圳忆海原识科技有限公司 Pulse neural network operation system and method for brain-like intelligence and cognitive computation
KR20210012839A (en) * 2019-07-26 2021-02-03 에스케이하이닉스 주식회사 Semiconductor device for performing an operation
CN111565152B (en) * 2020-03-27 2022-04-29 中国人民解放军国防科技大学 Brain-like chip routing system data communication method based on routing domain division
CN111667064B (en) * 2020-04-22 2023-10-13 南京惟心光电系统有限公司 Hybrid neural network based on photoelectric computing unit and operation method thereof
CN112966814B (en) * 2021-03-17 2023-05-05 上海新氦类脑智能科技有限公司 Information processing method of fusion impulse neural network and fusion impulse neural network
CN113554162B (en) * 2021-07-23 2022-12-20 上海新氦类脑智能科技有限公司 Axon input extension method, device, equipment and storage medium
CN113313240B (en) * 2021-08-02 2021-10-15 成都时识科技有限公司 Computing device and electronic device
CN114861864A (en) * 2022-02-24 2022-08-05 天津大学 Neuron network modeling method and device with dendritic morphology
CN114611686B (en) * 2022-05-12 2022-08-30 之江实验室 Synapse delay implementation system and method based on programmable neural mimicry core

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6269352B1 (en) * 1995-10-13 2001-07-31 Stmicroelectronics S.R.L. Low-voltage, very-low-power conductance mode neuron
CN1516070A (en) * 2003-01-08 2004-07-28 剑 王 Associative memory neural network
CN102457419A (en) * 2010-10-22 2012-05-16 中国移动通信集团广东有限公司 Method and device for optimizing transmission network route based on neural network model

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB0811057D0 (en) * 2008-06-17 2008-07-23 Univ Ulster Artificial neural network architecture

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6269352B1 (en) * 1995-10-13 2001-07-31 Stmicroelectronics S.R.L. Low-voltage, very-low-power conductance mode neuron
CN1516070A (en) * 2003-01-08 2004-07-28 剑 王 Associative memory neural network
CN102457419A (en) * 2010-10-22 2012-05-16 中国移动通信集团广东有限公司 Method and device for optimizing transmission network route based on neural network model

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"A Spike-Timing-Based Integrated Model";Hu Jun etal.;《Neural Computation》;20121130;第451-472页 *
"Model-Based Online Learning With Kernels";Li Guoqi etal.;《IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS》;20130331;第356-369页 *

Also Published As

Publication number Publication date
CN105095967A (en) 2015-11-25

Similar Documents

Publication Publication Date Title
CN105095967B (en) A kind of multi-modal neuromorphic network core
CN105095961B (en) A kind of hybrid system of artificial neural network and impulsive neural networks
CN105095965B (en) The mixed communication method of artificial neural network and impulsive neural networks nerve
CN105095966B (en) The hybrid system of artificial neural network and impulsive neural networks
Ferreira et al. An approach to reservoir computing design and training
Ishibuchi et al. Fuzzy neural networks with fuzzy weights and fuzzy biases
CN107423839A (en) A kind of method of the intelligent building microgrid load prediction based on deep learning
Ernoult et al. Updates of equilibrium prop match gradients of backprop through time in an RNN with static input
CN106570597A (en) Content popularity prediction method based on depth learning under SDN architecture
CN107909206A (en) A kind of PM2.5 Forecasting Methodologies based on deep structure Recognition with Recurrent Neural Network
CN104077438B (en) Power network massive topologies structure construction method and system
Igiri et al. Effect of learning rate on artificial neural network in machine learning
Velasco et al. Back propagation artificial neural networks for the analysis of quality control charts
Zong et al. Price forecasting for agricultural products based on BP and RBF Neural Network
CN104915714A (en) Predication method and device based on echo state network (ESN)
CN105260556B (en) The overhead crane modeling method of hair clip mutation operation RNA genetic algorithm
Prado et al. FPGA based implementation of a Fuzzy Neural Network modular architecture for embedded systems
Jing et al. Study on characteristic of fractional master-slave neural network
Wang et al. Convolution neural network based load model parameter selection considering short-term voltage stability
Qu et al. An Optimization Method of SNNs for Shortest Path Problem
Mayrhofer et al. Devs simulation of spiking neural networks
CN106570586A (en) Field detection vehicle path planning method for electric energy metering device
Kwon et al. Nonlinear mapping of interval vectors by neural networks
Ann et al. Calculation of hybrid multi-layered perceptron neural network output using matrix multiplication
Sambatti et al. Self-configured neural network for data assimilation using FPGA for ocean circulation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20180212

Address after: 100036 Beijing city Haidian District West Sanhuan Road No. 10 wanghailou B block two layer 200-30

Patentee after: Beijing Ling Xi Technology Co. Ltd.

Address before: 100084 Beijing Beijing 100084-82 mailbox

Patentee before: Tsinghua University

TR01 Transfer of patent right