CN106056212B - A kind of artificial neural networks core - Google Patents

A kind of artificial neural networks core Download PDF

Info

Publication number
CN106056212B
CN106056212B CN201610354206.3A CN201610354206A CN106056212B CN 106056212 B CN106056212 B CN 106056212B CN 201610354206 A CN201610354206 A CN 201610354206A CN 106056212 B CN106056212 B CN 106056212B
Authority
CN
China
Prior art keywords
neuron
address
module
router
interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610354206.3A
Other languages
Chinese (zh)
Other versions
CN106056212A (en
Inventor
马骋
张震
王世凯
李晶
熊剑平
朱荣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tsinghua University
Original Assignee
Tsinghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tsinghua University filed Critical Tsinghua University
Priority to CN201610354206.3A priority Critical patent/CN106056212B/en
Publication of CN106056212A publication Critical patent/CN106056212A/en
Application granted granted Critical
Publication of CN106056212B publication Critical patent/CN106056212B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means

Abstract

The present invention relates to a kind of artificial neural networks cores, including:One router-module, at least a neuron computing module, at least a nuclear control device, the neuron computing module and the nuclear control device correspond.The router-module receives and parses through outer input data, and by after parsing address information and aixs cylinder value information be sent into corresponding neuron computing module;And the neuron calculated result that the neuron computing module exports is sent to destination address, and send this frame data processing complement mark to the corresponding nuclear control device of the neuron computing module.Neuron calculated result is sent to the router-module for carrying out neuron calculating by the neuron computing module.The nuclear control device enters next neuron period after receiving described frame data processing complement mark.The networking of multiple artificial neural networks cores may be implemented using artificial neural networks core provided by the invention.

Description

A kind of artificial neural networks core
Technical field
The invention belongs to artificial neural networks field, in particular to a kind of neuron computing unit.
Background technique
Artificial neural network is to use for reference computation model made of biological brain cynapse-neuronal structure develops, can be concurrently Large-scale complex operation is carried out, and has the characteristics that nonlinearity, adaptivity.With biological brain neural network structure class Seemingly, artificial neural network may be defined as by basic structures such as neuron computing unit, aixs cylinder unit, dendron unit, cynapse units Composition.Neuron computing unit is most basic computing unit, can be carried out simple mathematical operation;Aixs cylinder unit is responsible for output mind Through first calculated result, a neuron has an aixs cylinder;Dendron unit is the input that neuron calculates, and a neuron can have more A dendron;Cynapse unit indicates the weight of a upper neuron axon and the connection of next neuron dendron;Between neuron and neuron By aixs cylinder, cynapse, dendricity layering connection to constitute neural network.
Artificial neural network is coupled by multilayer neuron computing unit, passes through a large amount of distributed simple operation units Interaction realizes that parallel complex nonlinear calculates, and has powerful information processing capability.In research artificial neural networks When model, the common equipment such as general purpose computer that calculates often will appear the disadvantages of calculating efficiency is low, and time-consuming, and energy consumption is high.Closely The characteristics of Nian Lai, part research institution is directed to artificial neural network, develops general neural computing chip, such as IBM TrueNorth has tentatively achieved the purpose that carry out artificial neural network operation in chip.However, current majority neural network meters Chip is calculated using by computing unit(Neuron)And connection unit(Cynapse matrix)Isolated mentality of designing, due to using The layout type that fixed scale couples entirely, to limit the connection number of neuron in monokaryon, while in non-full connection nerve net In the application of network, the waste for not coupling cynapse point storage resource will also result in.
Summary of the invention
In view of this, it is necessory to provide a kind of neural computing chip that flexible networking may be implemented.
A kind of artificial neural networks core, including:One router-module, at least a neuron computing module, at least one Nuclear control device, the neuron computing module and the nuclear control device correspond;The router-module, for receiving and solving Analyse outer input data, and by after parsing address information and aixs cylinder value information be sent into corresponding neuron computing module;It is described Neuron computing module is sent to the router-module for carrying out neuron calculating, and by neuron calculated result;It is described Router-module is also used to the neuron calculated result that the neuron computing module exports being sent to destination address, and to The corresponding nuclear control device of the neuron computing module sends this frame data processing complement mark;The nuclear control device, for connecing Enter next neuron period after receiving described frame data processing complement mark.
Compared with prior art, multiple artificial minds may be implemented using artificial neural networks core provided by the invention Networking through network query function core realizes the flexible expansion and the effective use of storage resource for calculating core, enhances and calculate core Adaptability avoids additional resource overhead.
Detailed description of the invention
Fig. 1 is the neuron computing unit schematic diagram that first embodiment of the invention provides.
Fig. 2 is the neuron computing module schematic diagram that second embodiment of the invention provides.
Fig. 3 is the artificial neural networks core schematic diagram that third embodiment of the invention provides.
Fig. 4 is the artificial neural networks core schematic diagram that fourth embodiment of the invention provides.
Fig. 5 is the internuclear transmitting data frame format of artificial neural networks core that fourth embodiment of the invention provides.
Fig. 6 is that artificial neural networks core router data provided by the invention receives and dispatches flow chart.
Fig. 7 is artificial neural networks nucleus neuron computing module calculated result transmission flow figure provided by the invention.
Fig. 8 is the artificial neural networks core schematic diagram that fifth embodiment of the invention provides.
Fig. 9 is artificial neural networks core networking schematic diagram provided by the invention.
Main element symbol description
Artificial neural networks core 10、20、30、40
Router-module 100、400
Path control deivce 110、410
Upper transmitting and receiving caching and interface 120
Left transmitting and receiving caching and interface 130
Lower transmitting and receiving caching and interface 140
Right transmitting and receiving caching and interface 150
Route information table 160
Neuron-router interface 170
Peripheral sensory neuron-router interface 471
Nervus opticus member-router interface 472
Third nerve member-router interface 473
Fourth nerve member-router interface 474
Neuron computing module 200
Peripheral sensory neuron computing module 200a
Nervus opticus member computing module 200b
Third nerve member computing module 200c
Fourth nerve member computing module 200d
Neuron computing unit 210
Decoder module 211
Address weight module 212
Multiplier 213
Accumulator 214
Send caching and interface 220
Function information table 230
Nuclear control device 300
First nuclear control device 300a
Second nuclear control device 300b
Third nuclear control device 300c
4th nuclear control device 300d
The present invention that the following detailed description will be further explained with reference to the above drawings.
Specific embodiment
Below in conjunction with the accompanying drawings and the specific embodiments, artificial neural networks core provided by the invention is made further It is described in detail.
Referring to Figure 1, first embodiment of the invention provides a kind of neuron computing unit 210, including decoder module 211, Address weight module 212, multiplier 213, accumulator 214.
The decoder module 211 is for receiving neural network information and parsing.The neural network information includes address letter Breath and aixs cylinder value information, address information therein is sent into the address weight module 212 by the decoder module 211, by aixs cylinder Value information is sent into the multiplier 213.
Storage address weight is to list in the address weight module 212, realizes to being input to the address weight module 212 Address information matched.If input address is matched with stored address information in address weight module 212, address power Molality block 212 exports corresponding weighted value to multiplier 213, and the weighted value of output is the numerical value changed in a certain range;If Input address and stored address information mismatch, then address weight module 212 exports zero to multiplier 213.In addition, institute Address effective marker state can also be arranged according to matching judgment result by stating address weight module 212.Specifically, if input address It being matched with address information stored in address weight module 212, the address weight module 212 issues address effective marker, If stored address information mismatches in input address and address weight module 212, the address weight module 212 does not change Address effective marker state.
The multiplier 213 receives the weighted value that the address weight module 212 exports and the decoder module 211 is defeated The weighted value is multiplied by aixs cylinder value out with aixs cylinder value, and product is sent into accumulator 214.
The calculated result that multiplier 213 exports is added up and is exported by the realization of accumulator 214.
It is integrated with calculating that the neuron computing unit 210 that first embodiment of the invention provides realizes addressing, by adopting With addressing with calculate integrated design, breach layout type that fixed scale couples entirely to the limit of neuron computing unit number System, simultaneously because taking the connecting mode of address matching, avoids the wasting of resources caused by useless connection, to realize flexibility Expansion and the effective use of storage resource, enhance neural computing efficiency.
Second embodiment of the invention further provides for a kind of neuron computing module, for establishing connecting relation between neuron And carry out neuron calculating.Fig. 2 is referred to, the neuron computing module 200 includes:Multiple neuron computing units 210, hair Send caching and interface 220, function information table 230.
The multiple neuron computing unit 210 is received from external address information and aixs cylinder value information, and to described Address information is judged, if the address information received is matched with the address information that itself is stored, to the letter received Breath carries out neural computing, and calculated result and address effective marker are sent to the transmission caching and interface 220, if connecing The address information received and the address information of itself storage mismatch, then do not carry out neural network meter to the information received It calculates.
Each neuron computing unit further comprises in multiple neuron computing units described in the present embodiment:Decode mould Block 211, address weight module 212, multiplier 213, accumulator 214.Neuron computing unit and the present invention in the present embodiment 210 structure of neuron computing unit and function that first embodiment provides are essentially identical, the mould of address weight described in the present embodiment Block 212 is also used to match when input address with stored address information in address weight module 212, then address weight module 212 export corresponding weighted value to multiplier 213, and issue address effective marker to transmission caching and interface 220;If input Address and stored address information mismatch, then address weight module 212 exports zero to multiplier 213, and unchanged Location effective marker state.
The transmission caching and interface 220 are used to calculate the multiple neuron in external global clock gclk triggering The calculated result and address effective marker of unit 210 are latched, and there is criterion in the address for resetting each neuron computing unit 210 Will, and specified neuron calculated result is exported to the function information table 230 and obtains transformation result.
The neuron calculated result is converted to corresponding functional value by the function information table 230.
Single neuron computing unit will be addressed and be calculated in the neuron computing module that second embodiment of the invention provides It is combined into one, neuron addressing can be carried out simultaneously and calculates, by using addressing and calculate integrated design, breach fixed gauge Limitation of the layout type that mould couples entirely to neuron computing unit number, simultaneously because taking the connection side of address matching Formula avoids the wasting of resources caused by useless connection, to realize flexible expansion and the effective use of storage resource, enhances Neural computing efficiency.
Third embodiment of the invention provides a kind of artificial neural networks core, including:Neuron computing module, nuclear control Device.The neuron computing module enters different nerves for carrying out neuron calculating under the control of the nuclear control device The transmission flow of first calculated result.The neuron computing module can be existing various neuron computing modules, this implementation It is illustrated by taking the neuron computing module 200 that second embodiment of the invention provides as an example in example.Fig. 3 specifically is referred to, manually Neural computing core 10 includes neuron computing module 200, nuclear control device 300.
The nuclear control device 300 is artificial by issuing neuron period N_peri and comparing pulse N_data_comp control Neural computing core 10 enters the transmission flow of different neuron calculated results.When entering the new neuron period, pass through The comparison of pulse N_data_comp triggering neuron calculated result, output are compared in output, and wait data processed result mark R_ N_data_done or N_data_null is effectively to enter the new neuron period.The nuclear control device 300 is in global clock The lower sending cycle into new round neuron calculated result of gclk triggering.
The neuron computing module 200 is used to establish connecting relation and progress neuron calculating between neuron, with this hair The neuron computing module structure and function that bright second embodiment provides are essentially identical, and transmission described in the present embodiment is cached and connect Mouth 220 locks the calculated result of each neuron computing unit 210 and address effective marker when global clock gclk is triggered It deposits, and resets the address effective marker of each neuron computing unit 210, hereafter pulse N_ is compared in the sending of nuclear control device 300 Data_comp reads serial number and Current neural after neuron computing module 200 receives the relatively pulse N_data_comp The neuron calculated result is then sent into function letter if effective by the identical neuron address effective marker of first period N_peri Breath table 230 simultaneously obtains transformation result, if invalid, then issues data invalid mark N_data_null to nuclear control device 300, drives Kinetonucleus controller 300 enters next neuron period.The transmission caching and interface 220, which are worked as, judges that address effective marker is effective When, the transformation result of function information table 230 is exported by SPI interface, while resetting the tired of corresponding neuron computing unit 210 Add device 214, when judge address effective marker for it is invalid when, accumulator 214 is not handled, is simultaneously emitted by data invalid mark N_data_null, driving nuclear control device 300 enter next neuron period.
The artificial neural networks core 10 that third embodiment of the invention provides is set by using addressing is integrated with calculating Thinking is counted, the connection number of single neuron computing unit also can be changed in neuron computing unit number in dynamically configurable core.
Fourth embodiment of the invention provides a kind of artificial neural networks core, including router-module, neuron calculate Module, nuclear control device.Neuron calculated result is sent to institute for carrying out neuron calculating by the neuron computing module State router-module.The neuron computing module can be existing various neuron computing modules, with this in the present embodiment It is illustrated for the neuron computing module 200 that invention second embodiment provides.Specifically refer to Fig. 4, artificial neural network Calculating core 20 includes router-module 100, neuron computing module 200, nuclear control device 300.
The present embodiment and 3rd embodiment difference is to increase router-module 100, now to the router-module 100 into Row is described in detail.
The router-module 100 is for receiving and parsing through outer input data, and by the address information and axis after parsing Prominent value information is sent into corresponding neuron computing module 200;And the neuron meter for exporting the neuron computing module 200 It calculates result and is sent to destination address, and send this frame data to nuclear control device 300 and handle complement mark.
The router-module 100 further comprises:Path control deivce 110, internuclear interaction caching and interface, routing iinformation Table 160, neuron-router interface 170.Wherein the internuclear interaction caching and interface further comprise right four hairs in upper lower-left Send reception caching and interface, i.e., it is upper to transmit and receive caching and interface 120, left transmitting and receiving caching and interface 130, lower transmitting and receiving Caching and interface 140, right transmitting and receiving caching and interface 150.The number of caching and its data-interface is transmitted and received in practical application Mesh can do variation appropriate according to specific requirements.The router-module 100 is interacted with external data using serial peripheral Interface (Serial Peripheral Interface, SPI).The right four transmitting and receivings caching in upper lower-left is internuclear interaction caching Area.Neuron-the router interface 170 is used to receive the neuron calculated result of the output of neuron computing module 200, this reality It applies in example using SPI interface, is only capable of one frame calculated result of storage.
When the router-module 100 receives data, external interface(Internuclear interaction caching including neighboring router module And interface and local neuron interface)Caching non-empty mark is sent to the router-module 100, prompts local router module 100 carry out data receiver.External interface works in slave pattern at this time, and local interface works in holotype.If local router mould It is less than to receive caching for corresponding direction in block 100, then starts a frame data to receive according to first in, first out rule, and to path control deivce 110 issue reception caching non-empty mark, and data frame in 110 pairs of receptions cachings of path control deivce is prompted to handle;If local road Receiving caching by corresponding direction in device module 100 has expired, then without data receiver, outside sends the pending datas such as caching and sends.
When the router-module 100 sends data, path control deivce 110, which is first read, sends buffer status.If sending slow Deposit less than, then data frame to be sent be written according to first in, first out rule and send caching, at the same send out to external interface send it is slow Non-empty mark is deposited, external receive is waited;If sending caching has expired, path control deivce 110 skips the transmission of this frame data, under reading One frame data simultaneously parse.
The path control deivce 110 is the core for routing data parsing and transmission control.It is cached according to upper transmitting and receiving And interface 120, left transmitting and receiving caching and interface 130, lower transmitting and receiving caching and interface 140, right transmitting and receiving are cached and are connect Mouth 150, the sequence of neuron-router interface 170 are successively read whether each reception caching is empty.If non-empty, then one is read Frame data are parsed;If it is empty, then it skips to next direction and receives caching.
Fig. 5 is referred to, which is the internuclear transmitting data frame format of artificial neural networks core, i.e., in internuclear interaction caching One frame data format of storage, the data frame include:Neuron computing module address field, axis in target core address field, core Prominent address field, neuron output field.Specifically in the present embodiment, the target core address field further comprises:Left and right Direction target core address field, up and down direction target core address field.Wherein, left and right directions target core address field, upper and lower It is respectively 8 bit signed numbers to target core address field, highest order is 0 to represent to the left or upwards, and highest order is 1 to represent to the right Or downwards;Neuron computing module address field is 4 bit unsigned numbers in core, represents neuron computing module in some core; Aixs cylinder address field is 8 bit unsigned numbers, carries out address matching for neuron computing unit in neuron computing module;Mind It is divided into 8 bit unsigned numbers through first output field.The data frame format that the neuron-router interface 170 stores only includes Neuron output field in Fig. 5 data frame.Single artificial neural networks core issues the network position that data frame can reach It sets, neuron computing module number and single neuron computing unit can couple in core the aixs cylinder address number upper limit is by the core Between transmitting data frame determine.It should be noted that the present embodiment has been merely given as a kind of specific form of data frames, practical application The sequence of each field and bit number can make the appropriate adjustments in middle data frame.
The path control deivce 110 carries out data parsing after reading one frame data of internuclear interaction caching.If being sent to this The data frame on ground, i.e. the left and right directions target core address, up and down direction target core address of data frame are 0, then remove data frame In target core address field, while neuron computing module address and aixs cylinder address and neuron are exported into two words in analytic kernel Section is sent into target nerve member computing module, is then removed in reception caching and is resolved data frame;If not being sent to local Data frame, then according to first left and right, it is rear above and below sequence by target core address decrement, then detect target send caching whether be it is full, If not dealing with while skipping to the next direction reception caching of reading then completely, if not full, then by treated, data frame feeding is sent out It send caching medium to be sent, while removing in reception caching and being resolved data frame.
After the path control deivce 110 reads reception one frame data of caching of neuron-router interface 170, with nuclear control The Current neural member period N_peri that device 300 exports makees the input of route information table 160, extracts and stores in route information table 160 Different neuronal target addresses, and by the neuron of address and reading export carry out framing, while detect target send caching Whether be it is full, if it is full do not deal with then while skipping to read next direction and receive caching, if not full then will treated counts It is medium to be sent that transmission caching is sent into according to frame, while being removed in the reception caching of neuron-router interface 170 and being resolved data frame And send this frame data to nuclear control device 300 and handle complement mark R_N_data_done, driving nuclear control device 300 enters next The neuron period.Neuron-the router interface 170 issues data effective marker N_data_ in neuron computing module 200 After en, start if reception caching is less than and receive a frame data, keeps reception to wait if receiving caching and having expired.Neuron calculates Module 200 empties data effective marker N_data_en after being successfully transmitted a frame data to neuron-router interface 170.
What is stored in the route information table 160 is the destination address of different neurons, is inputted as the Current neural member period, Output is corresponding destination address.
20 operational process of artificial neural networks core includes router data transmitting-receiving and neuron computing module meter It calculates result and sends two parts, refer to Fig. 6, the router data transmitting-receiving comprises the steps of:
S101, the Neural Network Data packet to be received such as router-module 100;
S102, router-module 100 receive Neural Network Data packet and parse;
S103, router-module 100 judge whether the Neural Network Data packet received is sent to local, if so then execute S105, if otherwise executing S104;
S104 will be sent into transmission by the sequence above and below behind first left and right after target core address decrement in the Neural Network Data packet Caching returns to S101;
S105, by the Neural Network Data packet aixs cylinder address and aixs cylinder value be sent into neuron computing module 200, the mind Each neuron computing unit is passed to through network packet;
S106, each neuron computing unit judge whether the aixs cylinder address matches with itself storage address, if then S107 is executed, if otherwise returning to S101;
S107, neuron computing unit exports weighted value corresponding with the aixs cylinder address, by the weighted value and the axis Prominent value is sent into accumulator after being multiplied, and returns to S101.
Fig. 7 is referred to, 200 calculated result of the neuron computing module transmission comprises the steps of:
S201 waits global clock triggering to come into force;
S202, judges whether global clock triggering comes into force, if so then execute S203, if otherwise returning to S201;
Each neuron computing module accumulator result is latching to corresponding transmission caching and interface by S203;
S204, judges whether each transmission caching and interface receive the comparison pulse that corresponding nuclear control device issues, if S205 is then executed, if otherwise returning to S204;
S205, reading address value neuron computing module identical with the Current neural member period is sent in caching and interface In data;
S206 judges whether the data are effective, if so then execute S207, if otherwise executing S215;
S207 by the data input function look-up table and exports result;
S208 judges whether the output result is effective, if so then execute S209, if otherwise executing S215;
The output result is sent to caching and router is waited to receive by S209, resets corresponding neuron accumulator;
S210 waits router idle, S211 is executed if the router free time, if router is not idle to return to S210;
S211, router-module start a data cached reception of neuron computing module;
S212 judges that neuron computing module issues whether information handles completion, if so then execute S213, if otherwise returning S212;
S213, router-module complete pulse to the processing of nuclear control device photos and sending messages;
S214, nuclear control device judges whether the last one neuron period, if then returning to S201, if otherwise executing S216;
S215, to nuclear control device photos and sending messages data invalid pulse;
S216, nuclear control device drive the neuron period to substitute, and pulse is compared in sending, return to S204.
The artificial neural networks core 20 that fourth embodiment of the invention provides calculates mould using single-router single neuron The working method of block realizes the networking of multiple artificial neural networks cores 20 by router-module 100.
Fifth embodiment of the invention provides a kind of artificial neural networks core, including a router-module, multiple nerves First computing module, multiple nuclear control devices, the multiple neuron computing module and the multiple nuclear control device correspond.This reality It applies example to be with 3rd embodiment difference, 3rd embodiment is configuration of the single channel by single neuron computing module, and the present embodiment It is single channel by the configuration of multi-neuron computing module, i.e., single router-module can be established with multiple neuron computing modules to be joined It connects.
The present embodiment is specifically illustrated so that single channel is by four neuron computing modules as an example, calculates mould through member in practical application The quantity of block can do variation appropriate according to specific requirements.Refer to Fig. 8, a kind of artificial neural networks core 30, packet Include router-module 400, peripheral sensory neuron computing module 200a, the first nuclear control device 300a, nervus opticus member computing module 200b, the second nuclear control device 300b, third nerve member computing module 200c, third nuclear control device 300c, fourth nerve member calculate Module 200d, the 4th nuclear control device 300d.
The router-module 400 is distinguished with the router-module 100 in 3rd embodiment to be, router-module 400 It include multiple neuron-router interfaces, i.e. peripheral sensory neuron-router interface 471, nervus opticus member-router interface 472, third nerve member-router interface 473, fourth nerve member-router interface 474.Every group of nuclear control device, neuron calculate Module, neuron-router interface respectively with path control deivce 410 by single channel in embodiment three by single neuron computing module Configuration mode establishes connection.
A router-module corresponds to multiple minds in the artificial neural networks core 30 that fourth embodiment of the invention provides Through first computing module, if multiple neuron computing module has the function of same or similar, cluster calculation may be implemented, can drop Low router number of nodes improves transmission bandwidth.The flexible expansion and the effective use of storage resource for calculating core are realized, with enhancing The adaptability for calculating core, avoids additional resource overhead.
Fig. 9 is referred to, the present invention further provides a kind of artificial neural networks systems, including multiple artificial neural network Network calculates core 40, and multiple artificial neural networks core 40 realizes being coupled to each other for multiple directions by router-module 100. Only gived in embodiment include 9 artificial neural networks core 40 situation, and each artificial neural networks core 40 include 4 transmitting and receivings caching and interface, realizes being coupled to each other for 4 directions up and down.It is appreciated that practical application The transmitting and receiving caching and interface for including in the number of middle artificial neural networks core and each artificial neural networks core Number can change according to actual application scenarios.
Present embodiments provide a kind of artificial neural network by forming multiple 40 networkings of artificial neural networks core Network computing system realizes the network connection of more artificial neural networks cores 40.
In addition, those skilled in the art can also do other variations in spirit of that invention, certainly, these are according to the present invention The variation that spirit is done, all should be comprising within scope of the present invention.

Claims (11)

1. a kind of artificial neural networks system, which is characterized in that including:One router-module, an at least neuron calculate Module, at least a nuclear control device, the neuron computing module and the nuclear control device correspond;
The router-module, for receiving and parsing through outer input data, and by after parsing address information and aixs cylinder value believe Breath is sent into corresponding neuron computing module;
The neuron computing module is sent to the router for carrying out neuron calculating, and by neuron calculated result Module;
The router-module, the neuron calculated result for being also used to export the neuron computing module is with being sent to target Location, and this frame data processing complement mark is sent to the corresponding nuclear control device of the neuron computing module;
The nuclear control device, for entering next neuron period after receiving described frame data processing complement mark.
2. artificial neural networks system as described in claim 1, which is characterized in that the artificial neural networks core system System includes a router-module, N number of neuron computing module, N number of nuclear control device, N number of neuron computing module and the N A nuclear control device corresponds, and N is the integer greater than 1.
3. artificial neural networks system as described in claim 1, which is characterized in that
The nuclear control device is also used to issue the neuron period to itself corresponding neuron computing module and compares pulse;
The neuron computing module is also used to after receiving the comparison pulse that itself corresponding nuclear control device issues, reads Address value neuron identical with Current neural member period address effective marker, if effectively then by the neuron calculated result It is converted, if in vain then to itself corresponding nuclear control device sending data invalid mark;
The neuron computing module, be also used to judge conversion after neuron calculated result it is whether effective, if effectively then will The neuron calculated result after changing is sent to the router-module, if then issuing in vain to itself corresponding nuclear control device Data invalid mark;
The nuclear control device is also used to laggard in the data invalid mark for receiving itself corresponding neuron computing module transmission Enter next neuron period.
4. artificial neural networks system as claimed in claim 3, which is characterized in that the neuron computing module includes Multiple neuron computing units send caching and interface, function information table;
The multiple neuron computing unit receives and parses through address information and aixs cylinder value letter that the router-module is sent Breath, if the address information received is matched with the address information that neuron computing unit itself stores, which calculates single Member carries out neural computing to the information received, and calculated result and address effective marker are sent to described send and delayed It deposits and interface, if the address information received and the address information of neuron computing unit itself storage mismatch, not to this The information received carries out neural computing;
Transmission caching and interface, in the triggering of external global clock by the calculating of the multiple neuron computing unit As a result and address effective marker is latched, and resets the address effective marker of the multiple neuron computing unit;And After receiving the comparison pulse that itself corresponding nuclear control device issues, serial number neuron identical with the Current neural member period is read Address effective marker, if effective, which is sent into the function information table by the transmission caching and interface And transformation result is obtained, if invalid, the transmission caching and interface issue data invalid mark to itself corresponding nuclear control device Will;
The neuron calculated result is converted to corresponding functional value by the function information table;
Transmission caching and interface, are also used to judge whether the transformation result effective, if effective, transmissions cache and Interface exports the transformation result, and if invalid, the transmission caching and interface issue data to itself corresponding nuclear control device Invalid flag.
5. artificial neural networks system as claimed in claim 4, which is characterized in that the router-module further wraps It includes:Neuron-router interface, path control deivce, route information table, internuclear interaction caching and interface;
Neuron-the router interface, for receiving the neuron calculated result of the neuron computing module output;
The path control deivce, the neuron calculated result exported for reading the neuron-router interface, and will The Current neural member period of nuclear control device output corresponding with the neuron computing module is sent to the route information table;
The route information table, for storing the destination address of different neurons, current mind that the path control deivce is inputted It is corresponding destination address through first periodic conversion;
The path control deivce, the destination address for being also used to export the route information table and the neuron calculated result group The internuclear interaction caching and interface are sent at the first data frame, and by first data frame;
The internuclear interaction caching and interface, for sending first data frame.
6. artificial neural networks system as claimed in claim 5, which is characterized in that first data frame includes:
Target core address field stores target artificial neural network computing system address;
Neuron computing module address field in core stores target nerve member in target artificial neural network computing system and calculates mould Block address;
Aixs cylinder address field stores target nerve member computing unit address in target nerve member computing module;
Neuron output field stores neuron calculated result.
7. artificial neural networks system as claimed in claim 6, which is characterized in that
The internuclear interaction caching and interface, are also used to receive extrinsic neural network packet;
The path control deivce is also used to parse the extrinsic neural network packet, if the extrinsic neural network packet To be sent to local data packet, then by the extrinsic neural network packet aixs cylinder address and aixs cylinder value be sent into target nerve First computing module, if the extrinsic neural network packet is not to be sent to local data packet, to the extrinsic neural network Target core address in data packet is updated, and the updated extrinsic neural network packet is sent into the internuclear interaction Caching and interface.
8. artificial neural networks system as claimed in claim 7, which is characterized in that the neuron-router interface Using Serial Peripheral Interface (SPI), slave pattern is worked in when the Serial Peripheral Interface (SPI) is as sender, works in master when as recipient Mode.
9. artificial neural networks system as claimed in claim 8, which is characterized in that the internuclear interaction caching and interface Using Serial Peripheral Interface (SPI), slave pattern is worked in when the Serial Peripheral Interface (SPI) is as sender, works in master when as recipient Mode.
10. a kind of data receiver method of artificial neural networks system as claimed in claim 9, which is characterized in that packet Include following steps:
S101, the Neural Network Data packet to be received such as described router-module;
S102, the router-module receive Neural Network Data packet and parse;
S103, whether the Neural Network Data packet that the router mould judgement receives is sent to local, if so then execute S105, if Otherwise S104 is executed;
S104 will be sent into transmission caching by the sequence above and below behind first left and right after target core address decrement in the Neural Network Data packet, Return to S101;
S105, by the Neural Network Data packet aixs cylinder address and aixs cylinder value be sent into neuron computing module, the neural network Data packet is passed to each neuron computing unit;
S106, each neuron computing unit judge whether the aixs cylinder address matches with itself storage address, if so then execute S107, if otherwise returning to S101;
S107, neuron computing unit exports weighted value corresponding with the aixs cylinder address, by the weighted value and the aixs cylinder value It is sent into accumulator after multiplication, returns to S101.
11. a kind of data transmission method for uplink of artificial neural networks system as claimed in claim 9, which is characterized in that packet Include following steps:
S201 waits global clock triggering to come into force;
S202, judges whether global clock triggering comes into force, if so then execute S203, if otherwise returning to S201;
Each neuron computing module accumulator result is latching to corresponding transmission caching and interface by S203;
S204, judges whether each transmission caching and interface receive the comparison pulse that corresponding nuclear control device issues, if then holding Row S205, if otherwise returning to S204;
S205 reads address value neuron computing module identical with the Current neural member period and sends in caching and in interface Data;
S206 judges whether the data are effective, if so then execute S207, if otherwise executing S215;
S207 by the data input function look-up table and exports result;
S208 judges whether the output result is effective, if so then execute S209, if otherwise executing S215;
The output result is sent to caching and router is waited to receive by S209, resets corresponding neuron accumulator;
S210 waits router idle, S211 is executed if the router free time, if router is not idle to return to S210;
S211, router-module start a data cached reception of neuron computing module;
S212 judges that neuron computing module issues whether information handles completion, if so then execute S213, if otherwise returning S212;
S213, router-module complete pulse to the processing of nuclear control device photos and sending messages;
S214, nuclear control device judges whether the last one neuron period, if then returning to S201, if otherwise executing S216;
S215, to nuclear control device photos and sending messages data invalid pulse;
S216, nuclear control device drive the neuron period to substitute, and pulse is compared in sending, return to S204.
CN201610354206.3A 2016-05-25 2016-05-25 A kind of artificial neural networks core Active CN106056212B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610354206.3A CN106056212B (en) 2016-05-25 2016-05-25 A kind of artificial neural networks core

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610354206.3A CN106056212B (en) 2016-05-25 2016-05-25 A kind of artificial neural networks core

Publications (2)

Publication Number Publication Date
CN106056212A CN106056212A (en) 2016-10-26
CN106056212B true CN106056212B (en) 2018-11-23

Family

ID=57175911

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610354206.3A Active CN106056212B (en) 2016-05-25 2016-05-25 A kind of artificial neural networks core

Country Status (1)

Country Link
CN (1) CN106056212B (en)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108256641A (en) * 2016-12-28 2018-07-06 上海磁宇信息科技有限公司 For the cellular array internal network communication method of cellular array computing system
WO2018137412A1 (en) * 2017-01-25 2018-08-02 清华大学 Neural network information reception method, sending method, system, apparatus and readable storage medium
CN106971228B (en) * 2017-02-17 2020-04-07 北京灵汐科技有限公司 Method and system for sending neuron information
CN106971229B (en) * 2017-02-17 2020-04-21 清华大学 Neural network computing core information processing method and system
CN110413551B (en) 2018-04-28 2021-12-10 上海寒武纪信息科技有限公司 Information processing apparatus, method and device
WO2019001418A1 (en) 2017-06-26 2019-01-03 上海寒武纪信息科技有限公司 Data sharing system and data sharing method therefor
CN109426553A (en) 2017-08-21 2019-03-05 上海寒武纪信息科技有限公司 Task cutting device and method, Task Processing Unit and method, multi-core processor
CN109214616B (en) * 2017-06-29 2023-04-07 上海寒武纪信息科技有限公司 Information processing device, system and method
CN109214502B (en) 2017-07-03 2021-02-26 清华大学 Neural network weight discretization method and system
CN108334942B (en) * 2017-12-22 2020-08-04 清华大学 Data processing method, device, chip and storage medium of neural network
CN108345936B (en) * 2018-01-31 2020-12-04 清华大学 Neuromorphic chip, system and method based on internal state mark
CN110399977A (en) * 2018-04-25 2019-11-01 华为技术有限公司 Pond arithmetic unit
CN110188872B (en) * 2019-06-05 2021-04-13 北京灵汐科技有限公司 Heterogeneous cooperative system and communication method thereof
CN112163673B (en) * 2020-09-28 2023-04-07 复旦大学 Population routing method for large-scale brain-like computing network
CN112242963B (en) * 2020-10-14 2022-06-24 广东工业大学 Rapid high-concurrency neural pulse data packet distribution and transmission method and system
CN112650705A (en) * 2020-12-31 2021-04-13 清华大学 Routing control method and artificial intelligence processor
CN112817898A (en) * 2021-02-08 2021-05-18 清华大学 Data transmission method, processor, chip and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0664516A2 (en) * 1994-01-19 1995-07-26 Nippon Telegraph And Telephone Corporation Neural network with reduced calculation amount
CN101681449A (en) * 2007-06-15 2010-03-24 佳能株式会社 Calculation processing apparatus and method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0664516A2 (en) * 1994-01-19 1995-07-26 Nippon Telegraph And Telephone Corporation Neural network with reduced calculation amount
US5630024A (en) * 1994-01-19 1997-05-13 Nippon Telegraph And Telephone Corporation Method and apparatus for processing using neural network with reduced calculation amount
CN101681449A (en) * 2007-06-15 2010-03-24 佳能株式会社 Calculation processing apparatus and method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
数字神经元芯片的设计与应用;罗莉;《计算机研究与发展》;19980930;全文 *

Also Published As

Publication number Publication date
CN106056212A (en) 2016-10-26

Similar Documents

Publication Publication Date Title
CN106056212B (en) A kind of artificial neural networks core
CN106056211B (en) Neuron computing unit, neuron computing module and artificial neural networks core
CN105989539B (en) A kind of financial transaction market obtain system and acquisition methods
CN107766148A (en) A kind of isomeric group and task processing method and device
CN106951926A (en) The deep learning systems approach and device of a kind of mixed architecture
CN107710237A (en) Deep neural network divides on server
CN106294233B (en) A kind of transfer control method and device of direct memory access
CN106815329A (en) A kind of data cached update method and device
CN107430704A (en) Neural network algorithm is realized in nerve synapse substrate based on the metadata associated with neural network algorithm
CN107688853A (en) A kind of device and method for being used to perform neural network computing
CN104461943B (en) Method for reading data, device and system
CN112468401B (en) Network-on-chip routing communication method for brain-like processor and network-on-chip
CN109246027B (en) Network maintenance method and device and terminal equipment
Goh et al. Nonlocal evolution of weighted scale-free networks
CN108334942A (en) Data processing method, device, chip and the storage medium of neural network
CN107241305A (en) A kind of network protocol analysis system and its analysis method based on polycaryon processor
CN110188872A (en) A kind of isomery cooperative system and its communication means
CN108365996B (en) FPGA + ARM architecture-based on-chip network simulation platform
CN105430060B (en) Constraint publish-subscribe model space flight measurement and control data exchange system
CN104717272A (en) Event stream processing system and method thereof
CN109327511A (en) A kind of data request method and server based on http protocol
CN108595379A (en) A kind of parallelization convolution algorithm method and system based on multi-level buffer
CN113312177A (en) Wireless edge computing system and optimization method based on federal learning
CN110086731A (en) A kind of cloud framework lower network data stabilization acquisition method
CN113642700A (en) Cross-platform multi-modal public opinion analysis method based on federal learning and edge calculation

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant