CN105844330A - Data processing method of neural network processor and neural network processor - Google Patents

Data processing method of neural network processor and neural network processor Download PDF

Info

Publication number
CN105844330A
CN105844330A CN201610165618.2A CN201610165618A CN105844330A CN 105844330 A CN105844330 A CN 105844330A CN 201610165618 A CN201610165618 A CN 201610165618A CN 105844330 A CN105844330 A CN 105844330A
Authority
CN
China
Prior art keywords
data
absolute value
input data
nonlinear mapping
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610165618.2A
Other languages
Chinese (zh)
Other versions
CN105844330B (en
Inventor
费旭东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN201610165618.2A priority Critical patent/CN105844330B/en
Publication of CN105844330A publication Critical patent/CN105844330A/en
Application granted granted Critical
Publication of CN105844330B publication Critical patent/CN105844330B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means

Abstract

The invention provides a data processing method of a neural network processor and a neural network processor. The method includes the following steps that: input data and corresponding weight absolute values are added together through an adder, wherein the input data are data of the output of a previous stage, and the input data and the weight absolute values are n-element vectors; and n-term data obtained after adding the input data and the corresponding weight absolute values together are subjected to n times of first nonlinear mapping; results obtained after the first nonlinear mapping are subjected to n times of accumulation operation through an accumulator, wherein the accumulation operation includes weight sign bit-controlled adding operation and subtraction operation; and a result obtained after the n times of accumulation operation is subjected to second nonlinear mapping, so that a processing result can be obtained, and data output is carried out, wherein the second nonlinear mapping is formulated according to the rules of neural network nonlinear mapping and the inverse mapping of the first nonlinear mapping. With the method adopted, quantization efficiency can be improved, and storage requirements and bandwidth requirements of data can be decreased.

Description

The data processing method of neural network processor and neural network processor
Technical field
The present embodiments relate to electronic chip technology field, particularly relate to a kind of neural network processor Data processing method and neural network processor.
Background technology
Neutral net and degree of depth learning algorithm have been obtained for extremely successful application, and are in and develop rapidly During, industry generally expects that new calculation contributes to realizing the most universal, complicated intelligence and answers With.Neutral net and degree of depth learning algorithm achieve very prominent in recent years in image recognition application Achievement, therefore neutral net and the optimization of degree of depth learning algorithm and high efficiency are realized beginning to focus on also by industry Paying attention to, such as facebook, it is excellent that the company such as Qualcomm, baidu, google has all put into neutral net Change the research of algorithm.Qualcomm company has issued integrated neural network processing module in chip of future generation Plan, improve neural network algorithm treatment effeciency, the improvement of related algorithm, and chip realize effect Rate is that it is paid close attention to and the key problem of research.
Fig. 1 is the calculating model of neural networks schematic diagram of a n level (layer), and Processing with Neural Network is therein The calculating shape of one neuron is such as: y=f (x1*w1+x2*w2+ ...+xn*wn+b), calculating is that classification is carried out , the output of previous stage is the input of rear stage.Fig. 2 is traditional computational methods flow chart, and prime exports Inputting (x1, x2 ... xn) as data, x1, x2 ... xn are multiplied with corresponding weight parameter respectively, Then completed x1*w1+x2*w2+ by accumulator ... the accumulation operations of+xn*wn+b, then reflect through non-linear Penetrate y=f (result after Lei Jia) and obtain result of calculation, finally complete data output.
It can be seen that in above-mentioned data processing method, the computation complexity ratio of the multiplication owing to relating to Higher, under certain calculation accuracy requires, storage demand and the bandwidth demand of corresponding data are the highest, Computational efficiency is the highest.
Summary of the invention
The embodiment of the present invention provides data processing method and the Processing with Neural Network of a kind of neural network processor Device, to solve the storage demand of data present in existing processing method and bandwidth demand is higher, calculate Inefficient problem.
First aspect, the embodiment of the present invention provides the data processing method of a kind of neural network processor, should Method includes: first passes through adder and will input data and the addition of corresponding weight absolute value, inputs data For the data of previous stage output, input data and weight absolute value are n unit vector.The most successively will input N item data after data and corresponding weight absolute value are added carries out n the first nonlinear mapping.Then By accumulator, the result after the first nonlinear mapping is carried out n accumulation operations, and accumulation operations includes power The add operation of weight sign bit control and subtraction.Finally the result after n accumulation operations is carried out Two nonlinear mapping obtain result and carry out data output, and the second nonlinear mapping is according to neutral net The rule of nonlinear mapping and the inverse mapping of the first nonlinear mapping are formulated.It is achieved thereby that complicated is taken advantage of Method calculates and is changed into additional calculation, improves quantitative efficiency, and the capacity of storage and bandwidth can be compressed, because of This storage demand reducing data and bandwidth demand, improve computational efficiency.And input data and do not limit Quantify in 0/1 binary so that computational accuracy meets the demand of actual application network, can be suitable for except nerve net The application target of wider scope beyond network calculating.
In a possible design, the first nonlinear mapping is the M power conversion of 2, and M is input number According to corresponding weight absolute value be added after n item data in every.Use 2 M power conversion by Simple in mapping relations, realization price of hardware is low.
In a possible design, when the first nonlinear mapping is the M power conversion of m, and m is not equal to When 2, for making circuit realiration simple, the M power of m is converted the M power conversion being converted to 2, logical Before crossing adder general's input data and the addition of corresponding weight absolute value, also include: input data are taken advantage of With Proportional coefficient K1, and/or, weight absolute value is multiplied by Proportional coefficient K2, K1With K2Equal or different; Or, after data and the addition of corresponding weight absolute value being inputted by adder, also include: to phase Every in n item data after adding is multiplied by Proportional coefficient K3.Wherein, K1、K2、K3It is not equal to 0.
In a possible design, K1、K2、K3For 1+1/2NOr 1-1/2N
In a possible design, when input data or weight absolute value are equal to 0, accumulation operations is for working as Front cumulative item remains unchanged.Weight sign bit is for time negative, and accumulation operations is subtraction.Accumulator is in Maintenance state, owing to, during actual neural computing, no matter counting input data or weight, Have substantial amounts of 0, the most so can simplify process, reduce power consumption.
In a possible design, realize the first nonlinear mapping or second non-linear by analog circuit Map or accumulation operations.Simulation non-linear conversion, the realization of addition all instantaneous can complete, disobeys It is disinclined to the speed of digital dock.
Second aspect, the embodiment of the present invention provides a kind of neural network processor, including: add circuit, Being added for inputting data and corresponding weight absolute value, input data are the data of previous stage output, Input data and weight absolute value are n unit vector.First nonlinear mapping circuit, for inputting successively N item data after data and corresponding weight absolute value are added carries out n the first nonlinear mapping.Cumulative Circuit, for the result after the first nonlinear mapping is carried out n accumulation operations, accumulation operations includes power The add operation of weight sign bit control and subtraction.Second nonlinear mapping circuit, for tiring out n time Result after add operation carries out the second nonlinear mapping and obtains result and carry out data output, and second is non- Rule and the inverse mapping of the first nonlinear mapping that Linear Mapping maps according to Neural Network Based Nonlinear are formulated.
In a possible design, the first nonlinear mapping circuit is the M power translation circuit of 2, M For input data and corresponding weight absolute value be added after n item data in every.
In a possible design, also include: the first mlultiplying circuit, for inputting at add circuit Before data and corresponding weight absolute value are added, input data are multiplied by Proportional coefficient K1;And/or, the Square law circuit, for before add circuit will input data and corresponding weight absolute value is added, right Weight absolute value is multiplied by Proportional coefficient K2, K1With K2Equal or different;Or, the 3rd mlultiplying circuit, For after add circuit will input data and corresponding weight absolute value is added, to the n item after being added Every in data is multiplied by Proportional coefficient K3;Wherein, K1、K2、K3It is not equal to 0.
In a possible design, K1、K2、K3For 1+1/2NOr 1-1/2N
In a possible design, when input data or weight absolute value are equal to 0, accumulation operations is for working as Front cumulative item remains unchanged;Weight sign bit is for time negative, and accumulation operations is subtraction.
At neutral net provided in each possible design of above-mentioned second aspect and above-mentioned second aspect Reason device, its beneficial effect may refer to be carried in each possible design of above-mentioned first aspect and first aspect The beneficial effect come, does not repeats them here.
The data processing method of the neural network processor that the embodiment of the present invention provides and Processing with Neural Network Device, will input data by adder and corresponding weight absolute value will be added, and input data are that previous stage is defeated The data gone out, the n item data after input data and corresponding weight absolute value being added the most successively is carried out N the first nonlinear mapping, carries out n time by the result after the first nonlinear mapping by accumulator and adds up Operation, accumulation operations includes add operation and the subtraction that weight sign bit controls, finally tired by n time Result after add operation carries out the second nonlinear mapping and obtains result and carry out data output, and second is non- Rule and the inverse mapping of the first nonlinear mapping that Linear Mapping maps according to Neural Network Based Nonlinear are formulated. It is changed into additional calculation it is achieved thereby that calculated by complicated multiplication, improves quantitative efficiency, the appearance of storage Amount and bandwidth can be compressed, and therefore reduce storage demand and the bandwidth demand of data, improve calculating effect Rate.And input data and be not limited to 0/1 binary quantization so that computational accuracy meets actual application network Demand, can be suitable for the application target of wider scope in addition to neural computing.
Accompanying drawing explanation
In order to be illustrated more clearly that the embodiment of the present invention or technical scheme of the prior art, below will be to reality Execute the required accompanying drawing used in example or description of the prior art to be briefly described, it should be apparent that under, Accompanying drawing in the description of face is some embodiments of the embodiment of the present invention, comes for those of ordinary skill in the art Say, on the premise of not paying creative work, it is also possible to obtain other accompanying drawing according to these accompanying drawings.
Fig. 1 is the calculating model of neural networks schematic diagram of a n level (layer);
Fig. 2 is traditional computational methods flow chart;
Fig. 3 is the flow chart of the data processing method embodiment one of neural network processor of the present invention;
Fig. 4 is the computing block diagram of the data processing method embodiment two of neural network processor of the present invention;
Fig. 5 is the computing block diagram of the data processing method embodiment three of neural network processor of the present invention;
Fig. 6 be neural network processor of the present invention data processing method embodiment three in 2 M power become Change schematic diagram;
Fig. 7 is the structural representation of neural network processor embodiment one of the present invention;
Fig. 8 is the structural representation of neural network processor embodiment two of the present invention;
Fig. 9 is the structural representation of neural network processor embodiment three of the present invention.
Detailed description of the invention
For making the purpose of the embodiment of the present invention, technical scheme and advantage clearer, below in conjunction with this Accompanying drawing in bright embodiment, is clearly and completely described the technical scheme in the embodiment of the present invention, Obviously, described embodiment is a part of embodiment of the embodiment of the present invention rather than whole embodiments. Based on the embodiment in the embodiment of the present invention, those of ordinary skill in the art are not making creative work The every other embodiment obtained under premise, broadly falls into the scope of embodiment of the present invention protection.
The embodiment of the present invention provides data processing method and the Processing with Neural Network of a kind of neural network processor Device, can be applicable to the such as data such as view data, speech data, video data needs to carry out neutral net The scene calculated, the data received carry out neural computing (single-stage neutral net as input data Calculate or Multilevel ANN calculate), the following data processing method provided by the embodiment of the present invention, Complicated multiplication calculates and is changed into additional calculation, improves quantitative efficiency, and capacity and the bandwidth of storage are permissible Compression, therefore reduces storage demand and the bandwidth demand of data, improves computational efficiency.
The neural network processor that the embodiment of the present invention provides may have physical entity form.Such as, at cloud End server application scenario, can be independent process chip, in terminal (such as mobile phone) application, and Ke Yishi A module in terminal handler chip.The input of information needs from voice, image, natural language etc. The various information wanting Intelligent treatment input, and through necessary pretreatment, (such as sampling, analog digital conversion, feature carries Take) form the data of pending neural network computing.The output of information deliver to other subsequent treatment modules or Software, such as figure or other be appreciated that available manifestation mode.Wherein, beyond the clouds under application form, Before and after neural network processor, the processing unit of level such as can be undertaken by other server operation unit, Under terminal applies environment, before and after neural network processor, level processing unit can be by its of terminal software and hardware He completes part (such as including sensor, interface circuit etc.).
Fig. 3 is the flow chart of the data processing method embodiment one of neural network processor of the present invention, such as figure Shown in 1, the method for the present embodiment may include that
S101, input data and corresponding weight absolute value being added by adder, input data are front The data of one-level output, input data and weight absolute value are n unit vector.
For Multilevel ANN, the output of previous stage is as the input of rear stage.Weight therein includes Weight absolute value and weight sign bit.
S102, successively will input data and corresponding weight absolute value be added after n item data carry out n time First nonlinear mapping.
Wherein, the first nonlinear mapping can be any truth of a matter m M power conversion, it is also possible to be shape such as Y=A*mB, the nonlinear transformation of y=B*B.Preferably, the first nonlinear mapping is the M power change of 2 Change, M be input data and corresponding weight absolute value be added after data in every.Use the M of 2 Power conversion is simple due to mapping relations, and realization price of hardware is low.
S103, the result after the first nonlinear mapping is carried out n accumulation operations by accumulator.
Accumulation operations includes add operation and the subtraction that weight sign bit controls, i.e. that is, weight Sign bit is negative, then accumulator does subtraction, and weight sign bit is that just accumulator does additive operation.
S104, the result after n accumulation operations is carried out the second nonlinear mapping obtain result and go forward side by side Row data export, rule that the second nonlinear mapping maps according to Neural Network Based Nonlinear and first non-linear The inverse mapping mapped is formulated.
Wherein, when the first nonlinear mapping is the M power conversion of m, when m is not equal to 2, for making circuit Realize simple, the M power of m is converted the M power conversion being converted to 2, especially by: in S101 Before inputting data and the addition of corresponding weight absolute value by adder, also include:
Input data are multiplied by Proportional coefficient K1, and/or, weight absolute value is multiplied by Proportional coefficient K2, K1With K2Equal or different;Or, will input data and the addition of corresponding weight absolute value by adder Afterwards, also include: to the n item data after being added every is multiplied by Proportional coefficient K3.Wherein, K1、 K2、K3It is not equal to 0, alternatively, K1、K2、K3Can be 1+1/2NOr 1-1/2N
By aforesaid operations, can be by M time that M power shift conversion is 2 of m when circuit realiration Power conversion realizes, because the mapping relations of the M power conversion of 2 are simple, realization price of hardware is low.
Further, there is the situation equal to 0 in input data or weight absolute value, and weight there is also as negative The situation of number, therefore, when input data or weight absolute value are equal to 0, accumulation operations is current cumulative item Remain unchanged.Accumulator is in maintenance state, due to during actual neural computing, no matter Number input data or weight, exist substantial amounts of 0, the most so can simplify process, reduces power consumption.
Weight sign bit is for time negative, and accumulation operations is subtraction, i.e. accumulator does subtraction.Do and subtract Method is exactly to negate original number to add 1.
It is alternatively possible to realize the first nonlinear mapping or the second nonlinear mapping or tired by analog circuit Add operation.Simulation non-linear conversion, the realization of addition all can instantaneous complete, and disobeys and is disinclined to numeral The speed of clock.
It should be noted that adder mentioned above, accumulator and hereafter carry in the embodiment of the present invention To multiplier be discrete, the circuit of physics, not by universal cpu realize software module.
The data processing method of the neural network processor that the present embodiment provides, will input number by adder Being added according to corresponding weight absolute value, input data are the data of previous stage output, the most successively by defeated Enter the n item data after data and the addition of corresponding weight absolute value and carry out n the first nonlinear mapping, will Result after first nonlinear mapping carries out n accumulation operations by accumulator, and accumulation operations includes weight The add operation of sign bit control and subtraction, finally carry out second by the result after n accumulation operations Nonlinear mapping obtains result and carries out data output, and the second nonlinear mapping is non-according to neutral net The rule of Linear Mapping and the inverse mapping of the first nonlinear mapping are formulated.It is achieved thereby that by complicated multiplication Calculating is changed into additional calculation, improves quantitative efficiency, and the capacity of storage and bandwidth can be compressed, therefore Reduce storage demand and the bandwidth demand of data, improve computational efficiency.And input data and be not limited to 0/1 binary quantifies so that computational accuracy meets the demand of actual application network, can be suitable for except neutral net meter The application target of wider scope beyond calculation.
Describe the embodiment of the present invention in detail provide below in conjunction with formulation process and a specific embodiment The processing procedure of data processing method of neural network processor.
Specifically, complicated multiplication is calculated and is changed into additional calculation, in detail by the M power conversion using m Formula proving process as follows:
As a example by background technology to be mentioned the calculating of shape such as y=f (x1*w1+x2*w2+ ...+xn*wn+b),
First assume that x > 0, s (w) are the sign bit of W.
Y=f (s (w1) eln(x1)+ln(|w1|)+s(w2)eln(x2)+ln(|w2|)+…s(wn)eln(xn)+ln(|wn|)+b)
Ln (y)=ln (f (s (w1) eln(x1)+ln(|w1|)+s(w2)eln(x2)+ln(|w2|)+…s(wn)eln(xn)+ln(|wn|)+b))
Assuming u=ln (y), v=ln (x), c=ln (| w |), this formula is expressed as again:
U=ln (f (s (w1) ev1+c1+s(w2)ev2+c2+…s(wn)evn+cn+b))
If exponential relationship is the index of 2, logarithmic relationship is the logarithm with 2 as the end, and computing formula is written as:
Y=f (s (w1) 2log 2 (x1)+log 2 (|w1|)+s(w2)2log 2 (x2)+log 2 (|w2|)+…+s(wn)2log 2 (xn)+log 2 (|wn|)…+b)
log2Y=log2f(s(w1)2log 2 (x1)+log 2 (|w1|)+s(w2)2log 2 (x2)+log 2 (|w2|)+…+s(wn)2log 2 (xn)+log 2 (|wn| )…+b)
Assume u=log2Y, v=log2X, c=log2| w |, this formula is expressed as again:
U=log2f(s(w1)2v1+c1+s(w2)2v2+c2+…s(wn)2vn+cn+b)
Make g (x)=2(x), ff (x)=log2F (x), then computing formula is abbreviated as:
U=ff (s (w1) g (v1+c1)+s (w2) g (v2+c2)+... s (wn) g (vn+cn)+b)
Wherein, corresponding first nonlinear mapping of g (x), corresponding second nonlinear mapping of ff (x).Fig. 4 is this The computing block diagram of the data processing method embodiment two of bright neural network processor, as shown in Figure 4, prime is defeated Going out as input data, input data and corresponding weight absolute value are n unit vector, first pass through adder Input data and corresponding weight absolute value are added, n item data v1+c1 after being added, v2+c2,, Vn+cn, n item data v1+c1 after will add up successively, v2+c2,, vn+cn carry out n the first non-thread Property map (g (x)=2(x)), obtain the result 2 after the first nonlinear mappingv1+c1、2v2+c2、、、2vn+cn, Then carry out n accumulation operations and obtain 2v1+c1+2v2+c2+…2vn+cn+ b, if input data or weight are 0, The item that the most currently adds up remains unchanged.If weight sign bit is negative, then accumulator does subtraction.Finally Result after n accumulation operations carries out the second nonlinear mapping (ff (x)) obtain result and go forward side by side line number According to output.
If exponential relationship is the index with m as the end, logarithmic relationship is the logarithm with m as the end, computing formula It is written as:
Y=f (s (w1) mlog m (x1)+log m (|w1|)+s(w2)mlog m (x2)+log m (|w2|)+…+s(wn)mlog m (xn)+log m (|wn|) …+b)
logmY=logmf(s(w1)mlog m (x1)+log m (|w1|)+s(w2)mlog m (x2)+log m (|w2|)+…+s(wn)mlog m (xn)+l og m (|wn|)…+b)
Assume u=logmY, v=logmX, c=logm| w |, this formula is expressed as again:
U=logmf(s(w1)mv1+c1+s(w2)mv2+c2+…+s(wn)mvn+cn+b)
U=logmf(s(w1)2(v1+c1)*log 2 m+s(w2)2(v2+c2)*log 2 m+…+s(wn)2(vn+cn)*log 2 m+b)
Make g (x)=2(x), ff (x)=logmF (x), k=log2M, then computing formula is abbreviated as:
U=ff (s (w1) g (c1*k+v1*k)+s (w2) g (c2*k+v2*k)+... s (wn) g (cn*k+vn*k)+b)
Wherein corresponding first nonlinear mapping of g (x), corresponding second nonlinear mapping of ff (x).As it is shown in figure 5, Fig. 5 is the computing block diagram of the data processing method embodiment three of neural network processor of the present invention.Above-mentioned calculating K in formula is the proportionality coefficient taken advantage of, by inputting data and corresponding weight absolute value in adder It is multiplied by proportionality coefficient k before or after addition, the M power of m can be converted the M power conversion being converted to 2, Make mapping relations simple, also simple on circuit realiration, and then realization price of hardware is low.In conjunction with Fig. 5, this The method implemented includes:
S201, from the data (i.e. prime output) of n-1 DBMS output unit output as the input of n level Data.
Input data and the corresponding weight absolute value of this grade are done phase add operation by the adder of S202, n level.
S203, on the passage that input data and weight absolute value are added, or in input data and corresponding Weight absolute value be added after result on, be multiplied by proportionality coefficient k, conventional proportionality coefficient k such as type is such as 1+1/2N, 1-1/2NProportionality coefficient.
S204, successively will input data and corresponding weight absolute value be added after n item data carry out n order One nonlinear mapping.
Wherein, the first nonlinear mapping e.g. can simplify design 2 M power conversion.Fig. 6 is this The M power conversion schematic diagram of 2 in the data processing method embodiment three of invention neural network processor, such as figure Shown in 6, specifically include:
In data after S2041, addition, the floating-point part of data, through decoder circuit, be converted to anti- Reflect the binary number of floating-point power actual size.
S2042, the fractional part of data, through the combinational circuit tabled look-up or simplify, complete the index of data Transformed mappings.
Data after S2043, exponential transform, and the power of floating-point, combined the simplest by combinational circuit Single 1, to n position multiplication, forms last result.
S205, accumulator carry out n accumulation operations to the cumulative item of output after the first nonlinear mapping.If Input data or weight are 0, and the item that the most currently adds up remains unchanged.If weight sign bit is negative, then add up Device does subtraction.
S206, the result after n accumulation operations is carried out the second nonlinear mapping.Usually, this is second non- Rule and the inverse mapping of the first nonlinear mapping that Linear Mapping maps according to Neural Network Based Nonlinear are formulated. One special case of this conversion is to include the cascades of 2 kinds of conversion, and it 1 is the non-linear change that neutral net is original Change, such as Sigmoid, ReLU etc., its 2 be adjust the data regularity of distribution logarithmic transformation.
S207, result is carried out data output.
Such as, a corresponding shape is such as: the calculating of y=f (x1*w1+x2*w2+ ...+xn*wn+b), x, w, The initial data precision of y needs the floating-point (32bit) of single precision, is at least also required to the fixed-point number of 16bit, right The data bandwidth answered is by the proportional calculating of 16bit, simultaneously need to the multiplier of 16bit.
First can be to x, w, y carry out logarithmic transformation, A=logmB.Experiment shows, the number after conversion According to, the quantization of 8 bits can meet the requirement of practicality, even can also accomplish 6 bits, 4 bits further.
Data after conversion are designated as P.Q, the part before wherein P is arithmetic point, and Q is arithmetic point portion below Point, P has p position, and Q has q position, and p+q is total bit number of data.
The present embodiment represents correspondence with Pv.Qv, the x after conversion, w, y data.
Therefore, the additional calculation after logarithmic transformation is just changed into:
There is no the situation of carry: Px.Qx+Pw.Qw=(Px+Pw). (Qx+Qw) or
There is the situation of carry: Px.Qx+Pw.Qw=(Px+Pw+1). (Qx+Qw-1)
It is denoted as: Pt.Qt
Pt.Qt does a simple multiplication, and the coefficient taken advantage of is: Log2M, by this multiplication, permissible The m that follow-up needs are donePt.Qt, be converted to 2Ps.QsDo.And Ps.Qs=Pt.Qt*Log2m。
As a rule, the logarithm or the index that take 2 can meet application requirement.In particular cases, if The logarithmic relationship suitably adjusted can preferably meet the requirement of application, can fit by taking advantage of this coefficient Work as regulation.By a sub-addition, satisfiable adjusting range be type such as: 1+1/2N, 1-1/2NAny Coefficient.Assuming that Pt is 4bit, Qt is 4bit, follow-up calculating, and Pt is directly sent to 4 lines and turns the decoder of 16 lines, And Qt directly send the look-up table of 4bit a to 4bit, or the simple combination electricity of reflection index mapping data rule Road.
Fig. 7 is the structural representation of neural network processor embodiment one of the present invention, as it is shown in fig. 7, this The neural network processor of embodiment may include that add circuit the 11, first nonlinear mapping circuit 12, Summation circuit 13 and the second nonlinear mapping circuit 14.Wherein, add circuit 11 will be for inputting data Being added with corresponding weight absolute value, input data are the data of previous stage output, input data and weight Absolute value is n unit vector.First nonlinear mapping circuit 12 will be for inputting data and corresponding power successively N item data after weight absolute value is added carries out n the first nonlinear mapping.Summation circuit 13 is for by the Result after one nonlinear mapping carries out n accumulation operations, and accumulation operations includes what weight sign bit controlled Add operation and subtraction.Second nonlinear mapping circuit 14 is for by the result after n accumulation operations Carrying out the second nonlinear mapping obtain result and carry out data output, the second nonlinear mapping is according to god Formulate through the rule of network nonlinear mapping and the inverse mapping of the first nonlinear mapping.
Wherein, the first nonlinear mapping circuit 12 can be the M power translation circuit of any truth of a matter m, Can also is that shape such as y=A*mB, the non-linear transform circuit of y=B*B.Preferably, first non-linear reflects Radio road 12 is the M power translation circuit of 2, and M is added for input data and corresponding weight absolute value After n item data in every.
Further, when the first nonlinear mapping circuit 12 is the M power translation circuit of m, m In 2 time, for making circuit realiration simple, the M power of m is converted the M power conversion being converted to 2, Specifically, Fig. 8 is the structural representation of neural network processor embodiment two of the present invention, as shown in Figure 8, On the basis of neural network processor shown in Fig. 7, further, also include: the first mlultiplying circuit 15, For, before add circuit 11 will input data and corresponding weight absolute value is added, input data being taken advantage of With Proportional coefficient K1;And/or, the second mlultiplying circuit 16, for add circuit 11 will input data and Before corresponding weight absolute value is added, weight absolute value is multiplied by Proportional coefficient K2, K1With K2Equal Or.
Fig. 9 is the structural representation of neural network processor embodiment three of the present invention, as it is shown in figure 9, On the basis of neural network processor shown in Fig. 7, further, also include: the 3rd mlultiplying circuit 17, For after add circuit 11 will input data and corresponding weight absolute value is added, to the n after being added Every in item data is multiplied by Proportional coefficient K3
Wherein, K1、K2、K3It is not equal to 0.Optionally, K1、K2、K3Can be 1+1/2NOr 1-1/2N
Further, when input data or weight absolute value are equal to 0, accumulation operations is current cumulative Xiang Wei Hold constant;Weight sign bit is for time negative, and accumulation operations is subtraction.Accumulator is in maintenance state, Owing to, during actual neural computing, no matter counting input data or weight, existing substantial amounts of 0, the most so can simplify process, reduce power consumption.
The device of the present embodiment, may be used for performing the technical scheme of embodiment of the method shown in Fig. 3, in fact Existing principle is similar to, and here is omitted.
The neural network processor that the present embodiment provides, will input data and corresponding power by add circuit Weight absolute value is added, and input data are the data of previous stage output, and then the first nonlinear mapping circuit depends on Secondary will input data and corresponding weight absolute value be added after n item data carry out n time first and non-linear reflect Penetrating, the result after the first nonlinear mapping is carried out n accumulation operations by accumulator by summation circuit, tired Add operation includes add operation and the subtraction that weight sign bit controls, the second last nonlinear mapping electricity Result after n accumulation operations is carried out the second nonlinear mapping and obtains result and to carry out data defeated by road Go out, rule that the second nonlinear mapping maps according to Neural Network Based Nonlinear and the first nonlinear mapping inverse Map and formulate.It is changed into additional calculation it is achieved thereby that calculated by complicated multiplication, improves quantitative efficiency, Capacity and the bandwidth of storage can be compressed, and therefore reduce storage demand and the bandwidth demand of data, improve Computational efficiency.And input data and be not limited to 0/1 binary quantization so that computational accuracy meets reality should By the demand of network, the application target of wider scope in addition to neural computing can be suitable for.
One of ordinary skill in the art will appreciate that: realize all or part of step of above-mentioned each method embodiment Suddenly can be completed by the hardware that programmed instruction is relevant.Aforesaid program can be stored in a computer can Read in storage medium.This program upon execution, performs to include the step of above-mentioned each method embodiment;And Aforesaid storage medium includes: ROM, RAM, magnetic disc or CD etc. are various can store program code Medium.
It is last it is noted that various embodiments above is only in order to illustrate the technical scheme of the embodiment of the present invention, It is not intended to limit;Although the embodiment of the present invention being described in detail with reference to foregoing embodiments, It will be understood by those within the art that: it still can be to the technology described in foregoing embodiments Scheme is modified, or the most some or all of technical characteristic is carried out equivalent;And these are repaiied Change or replace, not making the essence of appropriate technical solution depart from the embodiment of the present invention each embodiment technical side The scope of case.

Claims (11)

1. the data processing method of a neural network processor, it is characterised in that including:
To input data by adder and corresponding weight absolute value is added, described input data are previous The data of level output, described input data and described weight absolute value are n unit vector;
N item data after described input data and corresponding weight absolute value being added successively carries out n order One nonlinear mapping;
Result after first nonlinear mapping is carried out n accumulation operations, described cumulative behaviour by accumulator Make to include add operation and the subtraction that weight sign bit controls;
Result after n accumulation operations is carried out the second nonlinear mapping obtain result and carry out data Output, rule that described second nonlinear mapping maps according to Neural Network Based Nonlinear and described first non-thread Property map inverse mapping formulate.
Method the most according to claim 1, it is characterised in that described first nonlinear mapping is 2 The conversion of M power, M is in the n item data after described input data and corresponding weight absolute value are added Every.
Method the most according to claim 1 and 2, it is characterised in that described by adder by defeated Before entering data and the addition of corresponding weight absolute value, also include:
Described input data are multiplied by Proportional coefficient K1, and/or, described weight absolute value is multiplied by ratio system Number K2, K1With K2Equal or different;Or,
Described by adder will input data and corresponding weight absolute value be added after, also include:
To the n item data after being added every is multiplied by Proportional coefficient K3
Wherein, K1、K2、K3It is not equal to 0.
Method the most according to claim 3, it is characterised in that K1、K2、K3For 1+1/2NOr 1-1/2N
5. according to the method described in any one of claim 1-4, it is characterised in that
When described input data or described weight absolute value are equal to 0, described accumulation operations is current cumulative item Remain unchanged;
Described weight sign bit is for time negative, and described accumulation operations is subtraction.
6. according to the method described in any one of claim 1-5, it is characterised in that real by analog circuit Show described first nonlinear mapping or described second nonlinear mapping or described accumulation operations.
7. a neural network processor, it is characterised in that including:
Add circuit, is added for inputting data and corresponding weight absolute value, and described input data are The data of previous stage output, described input data and described weight absolute value are n unit vector;
First nonlinear mapping circuit, for successively by described input data and corresponding weight absolute value phase N item data after adding carries out n the first nonlinear mapping;
Summation circuit, for the result after the first nonlinear mapping being carried out n accumulation operations, described tired Add operation includes add operation and the subtraction that weight sign bit controls.
Second nonlinear mapping circuit, non-linear reflects for the result after n accumulation operations carries out second Penetrating and obtain result and carry out data output, described second nonlinear mapping is according to Neural Network Based Nonlinear The rule mapped and the inverse mapping of described first nonlinear mapping are formulated.
Neural network processor the most according to claim 7, it is characterised in that described first non-thread Property mapping circuit is the M power translation circuit of 2, and M is described input data and corresponding weight absolute value Every in n item data after addition.
Neural network processor the most according to claim 7, it is characterised in that also include:
First mlultiplying circuit, for inputting data and corresponding weight absolute value phase at described add circuit Before adding, described input data are multiplied by Proportional coefficient K1;And/or,
Second mlultiplying circuit, for inputting data and corresponding weight absolute value phase at described add circuit Before adding, described weight absolute value is multiplied by Proportional coefficient K2, K1With K2Equal or different;
Or,
3rd mlultiplying circuit, for inputting data and corresponding weight absolute value phase at described add circuit After adding, to the n item data after being added every is multiplied by Proportional coefficient K3
Wherein, K1、K2、K3It is not equal to 0.
Neural network processor the most according to claim 9, it is characterised in that K1、K2、K3 For 1+1/2NOr 1-1/2N
11. according to the neural network processor described in any one of claim 7-10, it is characterised in that institute When stating input data or described weight absolute value equal to 0, described accumulation operations is that current cumulative item maintains not Become;
Described weight sign bit is for time negative, and described accumulation operations is subtraction.
CN201610165618.2A 2016-03-22 2016-03-22 The data processing method and neural network processor of neural network processor Active CN105844330B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610165618.2A CN105844330B (en) 2016-03-22 2016-03-22 The data processing method and neural network processor of neural network processor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610165618.2A CN105844330B (en) 2016-03-22 2016-03-22 The data processing method and neural network processor of neural network processor

Publications (2)

Publication Number Publication Date
CN105844330A true CN105844330A (en) 2016-08-10
CN105844330B CN105844330B (en) 2019-06-28

Family

ID=56587746

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610165618.2A Active CN105844330B (en) 2016-03-22 2016-03-22 The data processing method and neural network processor of neural network processor

Country Status (1)

Country Link
CN (1) CN105844330B (en)

Cited By (35)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107169563A (en) * 2017-05-08 2017-09-15 中国科学院计算技术研究所 Processing system and method applied to two-value weight convolutional network
CN107256424A (en) * 2017-05-08 2017-10-17 中国科学院计算技术研究所 Three value weight convolutional network processing systems and method
CN107292458A (en) * 2017-08-07 2017-10-24 北京中星微电子有限公司 A kind of Forecasting Methodology and prediction meanss applied to neural network chip
CN107766935A (en) * 2016-08-22 2018-03-06 耐能有限公司 Multilayer artificial neural networks
WO2018120019A1 (en) * 2016-12-30 2018-07-05 上海寒武纪信息科技有限公司 Compression/decompression apparatus and system for use with neural network data
CN108510065A (en) * 2018-03-30 2018-09-07 中国科学院计算技术研究所 Computing device and computational methods applied to long Memory Neural Networks in short-term
CN108734280A (en) * 2017-04-21 2018-11-02 上海寒武纪信息科技有限公司 A kind of arithmetic unit and method
WO2019001463A1 (en) * 2017-06-30 2019-01-03 华为技术有限公司 Data processing method and apparatus
WO2019001418A1 (en) * 2017-06-26 2019-01-03 上海寒武纪信息科技有限公司 Data sharing system and data sharing method therefor
CN109214509A (en) * 2017-07-05 2019-01-15 中国科学院沈阳自动化研究所 One kind being used for deep neural network high speed real-time quantization structure and operation implementation method
CN109242091A (en) * 2018-09-03 2019-01-18 郑州云海信息技术有限公司 Image-recognizing method, device, equipment and readable storage medium storing program for executing
CN109284822A (en) * 2017-07-20 2019-01-29 上海寒武纪信息科技有限公司 A kind of neural network computing device and method
CN109376854A (en) * 2018-11-02 2019-02-22 矽魅信息科技(上海)有限公司 More truth of a matter logarithmic quantization method and devices for deep neural network
CN109409515A (en) * 2017-04-06 2019-03-01 上海寒武纪信息科技有限公司 Arithmetic unit and method
TWI657381B (en) * 2017-03-03 2019-04-21 美商慧與發展有限責任合夥企業 Electronic device and related method
CN109726805A (en) * 2017-10-30 2019-05-07 上海寒武纪信息科技有限公司 The method for carrying out neural network processor design using black box simulator
US10387298B2 (en) 2017-04-04 2019-08-20 Hailo Technologies Ltd Artificial neural network incorporating emphasis and focus techniques
CN110163231A (en) * 2018-02-14 2019-08-23 倍加科技股份有限公司 Establish the method, computer system and recording medium of data identification model
CN110413255A (en) * 2018-04-28 2019-11-05 北京深鉴智能科技有限公司 Artificial neural network method of adjustment and device
CN110533174A (en) * 2018-05-24 2019-12-03 华为技术有限公司 The circuit and method of data processing in nerve network system
CN110689125A (en) * 2017-10-30 2020-01-14 上海寒武纪信息科技有限公司 Computing device
CN111105019A (en) * 2018-10-25 2020-05-05 上海登临科技有限公司 Neural network operation device and operation method
CN111126590A (en) * 2016-12-23 2020-05-08 中科寒武纪科技股份有限公司 Artificial neural network operation device and method
WO2020093654A1 (en) * 2018-11-06 2020-05-14 Genesys Logic, Inc. Multichip system and data processing method adapted to the same for implementing neural network application
CN111767998A (en) * 2018-02-27 2020-10-13 上海寒武纪信息科技有限公司 Integrated circuit chip device and related product
CN112153081A (en) * 2020-11-24 2020-12-29 浙江齐安信息科技有限公司 Method for detecting abnormal state of industrial network
US11010338B2 (en) 2017-04-06 2021-05-18 Shanghai Cambricon Information Technology Co., Ltd Data screening device and method
US11221929B1 (en) 2020-09-29 2022-01-11 Hailo Technologies Ltd. Data stream fault detection mechanism in an artificial neural network processor
US11238334B2 (en) 2017-04-04 2022-02-01 Hailo Technologies Ltd. System and method of input alignment for efficient vector operations in an artificial neural network
US11237894B1 (en) 2020-09-29 2022-02-01 Hailo Technologies Ltd. Layer control unit instruction addressing safety mechanism in an artificial neural network processor
US11263077B1 (en) 2020-09-29 2022-03-01 Hailo Technologies Ltd. Neural network intermediate results safety mechanism in an artificial neural network processor
US11544545B2 (en) 2017-04-04 2023-01-03 Hailo Technologies Ltd. Structured activation based sparsity in an artificial neural network
US11551028B2 (en) 2017-04-04 2023-01-10 Hailo Technologies Ltd. Structured weight based sparsity in an artificial neural network
US11615297B2 (en) 2017-04-04 2023-03-28 Hailo Technologies Ltd. Structured weight based sparsity in an artificial neural network compiler
US11811421B2 (en) 2020-09-29 2023-11-07 Hailo Technologies Ltd. Weights safety mechanism in an artificial neural network processor

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102523453A (en) * 2011-12-29 2012-06-27 西安空间无线电技术研究所 Super large compression method and transmission system for images
CN103019656A (en) * 2012-12-04 2013-04-03 中国科学院半导体研究所 Dynamically reconfigurable multi-stage parallel single instruction multiple data array processing system
CN103501224A (en) * 2013-09-23 2014-01-08 长春理工大学 Asymmetric image encryption and decryption method based on quantum cell neural network system
CN105260773A (en) * 2015-09-18 2016-01-20 华为技术有限公司 Image processing device and image processing method
CN105389592A (en) * 2015-11-13 2016-03-09 华为技术有限公司 Method and apparatus for identifying image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102523453A (en) * 2011-12-29 2012-06-27 西安空间无线电技术研究所 Super large compression method and transmission system for images
CN103019656A (en) * 2012-12-04 2013-04-03 中国科学院半导体研究所 Dynamically reconfigurable multi-stage parallel single instruction multiple data array processing system
CN103501224A (en) * 2013-09-23 2014-01-08 长春理工大学 Asymmetric image encryption and decryption method based on quantum cell neural network system
CN105260773A (en) * 2015-09-18 2016-01-20 华为技术有限公司 Image processing device and image processing method
CN105389592A (en) * 2015-11-13 2016-03-09 华为技术有限公司 Method and apparatus for identifying image

Cited By (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107766935B (en) * 2016-08-22 2021-07-02 耐能有限公司 Multilayer artificial neural network
CN107766935A (en) * 2016-08-22 2018-03-06 耐能有限公司 Multilayer artificial neural networks
CN111126590A (en) * 2016-12-23 2020-05-08 中科寒武纪科技股份有限公司 Artificial neural network operation device and method
CN111126590B (en) * 2016-12-23 2023-09-29 中科寒武纪科技股份有限公司 Device and method for artificial neural network operation
WO2018120019A1 (en) * 2016-12-30 2018-07-05 上海寒武纪信息科技有限公司 Compression/decompression apparatus and system for use with neural network data
TWI657381B (en) * 2017-03-03 2019-04-21 美商慧與發展有限責任合夥企業 Electronic device and related method
US11315009B2 (en) 2017-03-03 2022-04-26 Hewlett Packard Enterprise Development Lp Analog multiplier-accumulators
US11461615B2 (en) 2017-04-04 2022-10-04 Hailo Technologies Ltd. System and method of memory access of multi-dimensional data
US11514291B2 (en) 2017-04-04 2022-11-29 Hailo Technologies Ltd. Neural network processing element incorporating compute and local memory elements
US11461614B2 (en) 2017-04-04 2022-10-04 Hailo Technologies Ltd. Data driven quantization optimization of weights and input data in an artificial neural network
US11544545B2 (en) 2017-04-04 2023-01-03 Hailo Technologies Ltd. Structured activation based sparsity in an artificial neural network
US11354563B2 (en) 2017-04-04 2022-06-07 Hallo Technologies Ltd. Configurable and programmable sliding window based memory access in a neural network processor
US11551028B2 (en) 2017-04-04 2023-01-10 Hailo Technologies Ltd. Structured weight based sparsity in an artificial neural network
US11263512B2 (en) 2017-04-04 2022-03-01 Hailo Technologies Ltd. Neural network processor incorporating separate control and data fabric
US11615297B2 (en) 2017-04-04 2023-03-28 Hailo Technologies Ltd. Structured weight based sparsity in an artificial neural network compiler
US11238331B2 (en) 2017-04-04 2022-02-01 Hailo Technologies Ltd. System and method for augmenting an existing artificial neural network
US10387298B2 (en) 2017-04-04 2019-08-20 Hailo Technologies Ltd Artificial neural network incorporating emphasis and focus techniques
US11238334B2 (en) 2017-04-04 2022-02-01 Hailo Technologies Ltd. System and method of input alignment for efficient vector operations in an artificial neural network
US11216717B2 (en) 2017-04-04 2022-01-04 Hailo Technologies Ltd. Neural network processor incorporating multi-level hierarchical aggregated computing and memory elements
US11675693B2 (en) 2017-04-04 2023-06-13 Hailo Technologies Ltd. Neural network processor incorporating inter-device connectivity
CN109409515B (en) * 2017-04-06 2022-08-09 上海寒武纪信息科技有限公司 Arithmetic device and method
CN109409515A (en) * 2017-04-06 2019-03-01 上海寒武纪信息科技有限公司 Arithmetic unit and method
US10896369B2 (en) 2017-04-06 2021-01-19 Cambricon Technologies Corporation Limited Power conversion in neural networks
US11010338B2 (en) 2017-04-06 2021-05-18 Shanghai Cambricon Information Technology Co., Ltd Data screening device and method
CN108734280A (en) * 2017-04-21 2018-11-02 上海寒武纪信息科技有限公司 A kind of arithmetic unit and method
CN107256424B (en) * 2017-05-08 2020-03-31 中国科学院计算技术研究所 Three-value weight convolution network processing system and method
CN107256424A (en) * 2017-05-08 2017-10-17 中国科学院计算技术研究所 Three value weight convolutional network processing systems and method
US11551068B2 (en) 2017-05-08 2023-01-10 Institute Of Computing Technology, Chinese Academy Of Sciences Processing system and method for binary weight convolutional neural network
CN107169563A (en) * 2017-05-08 2017-09-15 中国科学院计算技术研究所 Processing system and method applied to two-value weight convolutional network
WO2019001418A1 (en) * 2017-06-26 2019-01-03 上海寒武纪信息科技有限公司 Data sharing system and data sharing method therefor
US11748452B2 (en) 2017-06-30 2023-09-05 Huawei Technologies Co., Ltd. Method for data processing by performing different non-linear combination processing
WO2019001463A1 (en) * 2017-06-30 2019-01-03 华为技术有限公司 Data processing method and apparatus
US11334758B2 (en) 2017-06-30 2022-05-17 Huawei Technologies Co., Ltd. Method and apparatus of data processing using multiple types of non-linear combination processing
CN109214509A (en) * 2017-07-05 2019-01-15 中国科学院沈阳自动化研究所 One kind being used for deep neural network high speed real-time quantization structure and operation implementation method
CN109214509B (en) * 2017-07-05 2021-07-06 中国科学院沈阳自动化研究所 High-speed real-time quantization structure and operation implementation method for deep neural network
CN109284822B (en) * 2017-07-20 2021-09-21 上海寒武纪信息科技有限公司 Neural network operation device and method
CN110597558B (en) * 2017-07-20 2021-11-12 上海寒武纪信息科技有限公司 Neural network task processing system
CN110597558A (en) * 2017-07-20 2019-12-20 上海寒武纪信息科技有限公司 Neural network task processing system
US11481215B2 (en) 2017-07-20 2022-10-25 Cambricon (Xi'an) Semiconductor Co., Ltd. Calculation method and related product
CN109284822A (en) * 2017-07-20 2019-01-29 上海寒武纪信息科技有限公司 A kind of neural network computing device and method
CN107292458A (en) * 2017-08-07 2017-10-24 北京中星微电子有限公司 A kind of Forecasting Methodology and prediction meanss applied to neural network chip
CN107292458B (en) * 2017-08-07 2021-09-10 北京中星微人工智能芯片技术有限公司 Prediction method and prediction device applied to neural network chip
CN109726805B (en) * 2017-10-30 2021-02-09 上海寒武纪信息科技有限公司 Method for designing neural network processor by using black box simulator
CN110689125A (en) * 2017-10-30 2020-01-14 上海寒武纪信息科技有限公司 Computing device
CN109726805A (en) * 2017-10-30 2019-05-07 上海寒武纪信息科技有限公司 The method for carrying out neural network processor design using black box simulator
CN110163231A (en) * 2018-02-14 2019-08-23 倍加科技股份有限公司 Establish the method, computer system and recording medium of data identification model
CN111767998A (en) * 2018-02-27 2020-10-13 上海寒武纪信息科技有限公司 Integrated circuit chip device and related product
CN108510065A (en) * 2018-03-30 2018-09-07 中国科学院计算技术研究所 Computing device and computational methods applied to long Memory Neural Networks in short-term
CN110413255A (en) * 2018-04-28 2019-11-05 北京深鉴智能科技有限公司 Artificial neural network method of adjustment and device
CN110533174A (en) * 2018-05-24 2019-12-03 华为技术有限公司 The circuit and method of data processing in nerve network system
CN110533174B (en) * 2018-05-24 2023-05-12 华为技术有限公司 Circuit and method for data processing in neural network system
CN109242091A (en) * 2018-09-03 2019-01-18 郑州云海信息技术有限公司 Image-recognizing method, device, equipment and readable storage medium storing program for executing
CN109242091B (en) * 2018-09-03 2022-03-22 郑州云海信息技术有限公司 Image recognition method, device, equipment and readable storage medium
CN111105019B (en) * 2018-10-25 2023-11-10 上海登临科技有限公司 Neural network operation device and operation method
CN111105019A (en) * 2018-10-25 2020-05-05 上海登临科技有限公司 Neural network operation device and operation method
CN109376854B (en) * 2018-11-02 2022-08-16 矽魅信息科技(上海)有限公司 Multi-base logarithm quantization device for deep neural network
CN109376854A (en) * 2018-11-02 2019-02-22 矽魅信息科技(上海)有限公司 More truth of a matter logarithmic quantization method and devices for deep neural network
TWI715281B (en) * 2018-11-06 2021-01-01 創惟科技股份有限公司 Multichip system, data processing method adapted to the same, and non-transitory computer-readable medium for implementing neural network application
WO2020093654A1 (en) * 2018-11-06 2020-05-14 Genesys Logic, Inc. Multichip system and data processing method adapted to the same for implementing neural network application
US11221929B1 (en) 2020-09-29 2022-01-11 Hailo Technologies Ltd. Data stream fault detection mechanism in an artificial neural network processor
US11263077B1 (en) 2020-09-29 2022-03-01 Hailo Technologies Ltd. Neural network intermediate results safety mechanism in an artificial neural network processor
US11811421B2 (en) 2020-09-29 2023-11-07 Hailo Technologies Ltd. Weights safety mechanism in an artificial neural network processor
US11237894B1 (en) 2020-09-29 2022-02-01 Hailo Technologies Ltd. Layer control unit instruction addressing safety mechanism in an artificial neural network processor
CN112153081A (en) * 2020-11-24 2020-12-29 浙江齐安信息科技有限公司 Method for detecting abnormal state of industrial network

Also Published As

Publication number Publication date
CN105844330B (en) 2019-06-28

Similar Documents

Publication Publication Date Title
CN105844330B (en) The data processing method and neural network processor of neural network processor
KR102476343B1 (en) Apparatus and method for supporting neural network calculation of fixed-point numbers with relatively few digits
US20200117981A1 (en) Data representation for dynamic precision in neural network cores
CN107506828A (en) Computing device and method
WO2019238029A1 (en) Convolutional neural network system, and method for quantifying convolutional neural network
US11537879B2 (en) Neural network weight discretizing method, system, device, and readable storage medium
CN112508125A (en) Efficient full-integer quantization method of image detection model
CN108510065A (en) Computing device and computational methods applied to long Memory Neural Networks in short-term
CN110929865A (en) Network quantification method, service processing method and related product
WO2022111002A1 (en) Method and apparatus for training neural network, and computer readable storage medium
WO2020098368A1 (en) Adaptive quantization method and apparatus, device and medium
CN110175641A (en) Image-recognizing method, device, equipment and storage medium
CN114792359A (en) Rendering network training and virtual object rendering method, device, equipment and medium
US11651198B2 (en) Data processing method and apparatus for neural network
US20230143985A1 (en) Data feature extraction method and related apparatus
CN112085154A (en) Asymmetric quantization for compression and inference acceleration of neural networks
CN111930670B (en) Heterogeneous intelligent processing quantization device, quantization method, electronic device and storage medium
CN114444688A (en) Neural network quantization method, apparatus, device, storage medium, and program product
KR20220042315A (en) Method and apparatus for predicting traffic data and electronic device
CN114970357A (en) Energy-saving effect evaluation method, system, device and storage medium
CN111382854B (en) Convolutional neural network processing method, device, equipment and storage medium
JP2023505237A (en) Flexible Precision Neural Inference Processing Unit
CN110751274A (en) Neural network compression method and system based on random projection hash
CN110782003A (en) Neural network compression method and system based on Hash learning
US20220180177A1 (en) An efficient method for vlsi implementation of useful neural network activation functions

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant