WO1992001257A1 - Virtual neurocomputer architectures for neural networks - Google Patents

Virtual neurocomputer architectures for neural networks Download PDF

Info

Publication number
WO1992001257A1
WO1992001257A1 PCT/US1991/003318 US9103318W WO9201257A1 WO 1992001257 A1 WO1992001257 A1 WO 1992001257A1 US 9103318 W US9103318 W US 9103318W WO 9201257 A1 WO9201257 A1 WO 9201257A1
Authority
WO
WIPO (PCT)
Prior art keywords
neuron
values
units
weight
adder
Prior art date
Application number
PCT/US1991/003318
Other languages
English (en)
French (fr)
Inventor
Gerald George Pechanek
Stamatis Vassiliadis
Jose Guadalupe Delgado-Frias
Original Assignee
International Business Machines Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US07/526,866 external-priority patent/US5065339A/en
Application filed by International Business Machines Corporation filed Critical International Business Machines Corporation
Publication of WO1992001257A1 publication Critical patent/WO1992001257A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/10Interfaces, programming languages or software development kits, e.g. for simulating neural networks

Definitions

  • These inventions relate to computers and computer systems and particularly to virtual neurocomputer architectures for neural networks.
  • LEARNING MACHINE SCALABLE FLOW VIRTUAL LEARNING NEUROCOMPUTER G. G. Pechanek,
  • PLAN PYRAMID LEARNING ARCHITECTURE NEUROCOMPUTER, G. G.
  • the neurons modeled on the neural processor are simulated in either a "direct” or a "virtual” implementation.
  • each neuron would have a physical processing element (PE) available which could operate simultaneously in parallel with the other neuron PE's active in the system.
  • PE physical processing element
  • multiple neurons are assigned to individual hardware processing elements (PEs), requiring that the PE's processing be shared across its “virtual” neurons.
  • the performance of the network will be greater under the "direct” approach, but due to the large number of neurons utilized in many network simulations and technology constraints limiting the number of "direct” neuron PEs which can be implemented, many neurocomputer designs utilize the "virtual" neurons concept to simulate more neurons than are available directly.
  • the basic concept involved in a virtual neurocomputer is to provide some degree of parallelism, if possible, and then to divide the total number of neurons to be implemented in a network among the parallel processing elements available and to use time division multiplexing per physical processing element. This naturally spreads a single time unit out to cover the processing required for the number of virtual neural nodes assigned to a single physical PE.
  • a virtual neural node represents one neuron in the network being simulated.
  • a uni-processor can be used but it must handle all processing for all neural nodes in a sequential manner. Because of this, many virtual neurocomputers use a parallel arrangement of microprocessors. Parallel arrangements for neurocomputing may be reviewed. See Hecht-Nielsen 90: Savely IEEE 87 and Treleaven 89.
  • the structure used usually allows for floating point hardware accelerators to be added for improved performance of each neural node calculation.
  • An efficient communications network between the physical PE's is also required among the parallel processing elements to improve performance.
  • For these virtual neurocomputers to function there must be local memory for the physical processors containing the network interconnection structure, weight matrix, and virtual PE activation state memory.
  • our virtual system is scalable and applicable to networks and permits our improved hardware, illustrated in other applications, to be used where the physical hardware does not have enough neurons to achieve the results of the desired application.
  • the new apparatus which acts as a virtual scalable neural array processor apparatus has an N neuron structure having weight multiplication units containing a multiplier, weight storage, and neuron output value storage for V neurons where V > N and V is the number of neurons in the neural network to be emulated on the N neuron structure.
  • the first architectural preferred embodiment is the approach to be discussed. It modifies the SNAP structure to allow the sharing of the physical synapse processing elements such that a neuron input is not represented by a column or row of N synapse processors but by multiple rows or columns of synapse processors.
  • a second architectural preferred embodiment to be discussed is the approach for TSNAP which replicates groups of synapse processing elements until sufficient synapse processing elements are available to cover the V neurons to be modeled while keeping N activation function generators (sigmoid generators).
  • a third architectural preferred embodiment to be discussed is the general approach, representing a common approach for either SNAP or TSNAP, which adds an iterative adder with a partial summation storage element to each of the inputs of the N sigmoid generators and increases the weight storage appropriately at each synapse processor.
  • FIGURE 1 illustrates the Virtual Neurocomputer.
  • FIGURE 2 shows a SNAP-V1 with eight Virtual Neurons.
  • FIGURE 3 illustrates a SNAP-V1 Bypass Adder.
  • FIGURE 4 illustrates a SNAP tag matching multiplier
  • FIGURE 5 shows SNAP-shots 1-4.
  • FIGURE 6 shows SNAP-shot 5.
  • FIGURE 7 shows SNAP-shots 6 through 9.
  • FIGURE 8 shows SNAP-shot 10.
  • FIGURE 9 illustrates SNAP-V2 with eight Virtual Neurons.
  • FIGURE 10 illustrates a four Neuron T-SNAP.
  • FIGURE 11 shows an 9-Neuron T-SNAP weight matrix with Neuron Y values.
  • FIGURE 12 shows an eight Neuron Virtual T-SNAP with 4 physical Neurons.
  • FIGURE 13 shows a 16-Neuron T-SNAP weight matrix with Neuron Y values.
  • FIGURE 14 shows a 16-Neuron Virtual T-SNAP, and the Step 1 calculation of Y1 ', Y2', Y3', and Y4'.
  • FIGURE 15 shows a 16-Neuron Virtual T-SNAP, and the Step 2 calculation of Y5', Y6', Y7', and Y8'.
  • FIGURE 16 shows a 16-Neuron Virtual T-SNAP, and the Step 3 calculation of Y9', YA', YB', and YC.
  • FIGURE 17 shows a 16-Neuron Virtual TSNAP, with four triangular sections and a Step 4 calculation of UD', YE', YF' and YG'.
  • FIGURE 18 shows a 4-Neuron TSNAP-V2 with 16 Neuron weights
  • FIGURE 19 shows a Virtual Architecture Comparison Summary.
  • FIGURES may be separated in parts and as a convention we place the top of the FIGURE as the first sheet, with subsequent sheets proceeding down and across when viewing the FIGURE, in the event that multiple sheets are used.
  • FIG. 1 The characteristics of a virtual neurocomputer are shown in Figure 1. Reference here may be had to a related system from Hecht-Nielsen illustrated by Sou'cek 88.
  • Sou'cek 88 There are up to R physical processing elements depicted in Figure 1.
  • each physical processor would be assigned the neural node calculation task for K neurons. This task is the calculation of the sum of products (weight ⁇ connected neuron output), the activation function, and possibly an output function.
  • the parallel broadcast bus is used to communicate network interconnection information, neuron activation state values, etc. between the physical processors. In the structure of Figure 1 , the parallel bus is controlled by the system controller to ensure no bus conflicts. After the virtual neuron activation output values are calculated, the local system copies for these values must be updated in each neural processing element.
  • the number V will be used to denote the number of neurons contained in the network to be modeled on the neurocomputer; the number N will denote the number physical neurons available in the physical implementation.
  • V For virtual processing V > N.
  • N and V are powers of 2.
  • the computational tasks to be implemented by the neurocomputer architectures described herein are given by equation 1 and 2 which are based on a subset of the full Parallel Distributed Processing model. See here Rumelhart 86 and the Hopfield network illustrated by Hopfield 84.
  • V is the number of neurons in the neural network.
  • Y 1 is the j th neuron output value connected to the i th neuron input through a connection weight of W ij .
  • F(z) is the neuron activation function which many times is set equal to a sigmoid activation function whose form, for example is:
  • T is a global control parameter used to modify the slope of the sigmoid function for a given set of z values.
  • Equations 1 and 2 for a completely connected V neuron network, contain four basic operations:
  • the HOST computer assumes the responsibilities of initializing the network architectures.
  • the HOST computer will be responsible for the loading of the number of neurons in the network to be simulated by the architecture, all the connection weights, the initial neuron values, the number of network update cycles to be run, as well as starting the model into execution.
  • the HOST computer is also provided with the ability to read the neuron values at the completion of network execution.
  • the initialization time and the Host processing time are considered as a separate issue. Only the performance during execution will be considered.
  • the first architectural preferred embodiment is the approach to be discussed. It modifies the SNAP structure (see Vassiiiadis SNAP 90) to allow the sharing of the physical synapse processing elements such that a neuron input is not represented by a column or row of N synapse processors but rather by multiple rows or columns of synapse processors.
  • a second architectural preferred embodiment to be discussed is the approach for TSNAP (see Pechanek T-SNAP) which replicates groups of synapse processing elements until sufficient synapse processing elements are available to cover the V neurons to be modeled while keeping N activation function generators (sigmoid generators)
  • a third architectural preferred embodiment to be discussed is the general approach, representing a common approach for either SNAP or TSNAP, which adds an iterative adder with a partial summation storage element to each of the inputs of the N sigmoid generators and increases the weight storage appropriately at each synapse processor.
  • the performance of each virtual neurocomputer is discussed.
  • a summary compares each virtual architecture approach in terms of implementation costs and performance. The "best" architecture for a given application will depend upon these costs and will be chosen for the particular application.
  • the system defined clock period is C, with all delays specified as multiples of C.
  • the number of stages in the communicating adder tree is log 2 N. where N is the total number of physical neurons.
  • This sequence of events requires a simple control mechanism such as the use of a counter whose output value is compared against delay values representing the listed events, namely: the multiplier delay, the log 2 N communicating adder tree - add mode delay, the sigmoid delay, and the log 2 N communicating adder tree - communications mode delay.
  • SNAP-V1 SNAP VIRTUAL ARCHITECTURE 1
  • SNAP-V1 The first approach to be described, termed SNAP-V1 , restricts V to a maximum of N 2 . Greater than N 2 neurons can be simulated with this architecture but a more involved control process would be required and this expanded capability will not be discussed in this description.
  • SNAP-V1 the SNAP neuron definition (Vassiiiadis SNAP 90) is changed to accommodate the greater number of weighted inputs required per neuron.
  • Figure 2 shows a virtual eight neuron network modeled on a 4 neuron SNAP. Sufficient internal storage for the weights and Y values is assumed to be available within the multiplier cell structures.
  • a variation of the communicating adder to be used in the adder tree is required.
  • This variation shown in Figure 3 SNAP-V1 Bypass Adder, allows the adder to be bypassed in a forward direction as well as the reverse communication mode direction.
  • the various bypass paths are indicated by an arrow in Figure 3, and this symbology is also utilized in Figure 2 to indicate where the bypass adders are used.
  • Table 1 depicts the driver controls, DR1 , DR2, DR3, DR4, and DR5, which are driven from a central common tree controller point, encoded from D3, D2, and D1 control signals.
  • the sigmoid generators In order to ensure the neuron values are reverse communicated to the proper neuron input, the sigmoid generators must also tag the generated value. This tag accompanies the neuron value in its return path through the adder trees.
  • the multiplier cell Y value input register will then require a neuron Y value tag matching compare function. This is depicted in Figure 4 where the loading of the Y value register is under control of the tag matching function. Two Y value registers are required, an old value and a new value.
  • SNAP-V2 SNAP VIRTUAL ARCHITECTURE 2
  • the total number of weight registers in terms of K is:
  • TSNAP-V2 period K ⁇ K ⁇ M + (K + 1 )( log 2 N)C + ⁇ 1A + ⁇ s ) (4) TSNAP-V1 : TSNAP VIRTUAL ARCHITECTURE 1
  • a virtual implementation on the TSNAP is our normally preferred embodiment, implementing the functions achieved by T-SNAP. This will be demonstrated utilizing a 4, 8, and a 16 neuron network modeled on a 4 neuron TSNAP.
  • Figure 10 depicts the 4 neuron T-SNAP.
  • the function F for all equations is assumed to be the sigmoid function.
  • the equation for Neuron 3 is printed here for easy reference with Figure 10.
  • Y 3 F(W 31 Y 1 + W 32 Y 2 + W 33 Y 3 + W 34 Y 4 )
  • S represents the sigmoid generators.
  • Figure 11 represents the 8 neuron weights and Y values required when modeled on an 8 neuron TSNAP.
  • Figure 12 modifies the TSNAP structure of Figure 10 by replicating the diagonal cells, G-Cells, and addition trees and producing the partial summation on the diagonal of the structure. The summation is shown on the diagonal to demonstrate that the summation tree structure can be placed as required.
  • the triangular sections replicated exclude the sigmoid generator though the size of the section is based on the number of sigmoid generators which is equal to the number of physical neurons N of the TSNAP section. For TSNAP-V1 it will be assumed that the triangular sections are replicated enough times to handle the largest network to be modeled on the system. The number of replicated triangular sections is given by:
  • Y 3 F(W 31 Y 1 + W 32 Y 2 + W 33 Y 3 + W 34 Y 4 + W 35 Y 5 + W 36 Y 6 + W 37 Y 7 + W 38 Y 8 )
  • ⁇ 7 F(W 71 Y 1 + W 72 Y 2 + W 73 Y 3 + W 74 Y 4 + W 75 Y 5 + W 76 Y 6 + W 77 Y 7 + W 78 Y 8 )
  • the neuron values would be tagged such that only the correct neuron values are loaded into the proper triangular section.
  • the compare of the tag can be done prior to sending the neuron value into a triangular section or at each Y value storage register.
  • Figure 12 One of the advantages of these illustrated embodiments which we prefer is that the basic concept shown in Figure 12 can be continued for larger networks by replication of the multipliers and communicating adder trees while keeping the same number of neuron activation function generators.
  • This approach for a 16 neuron network is shown in Figure 14, 15, 16, and 17.
  • Figures 14, 15, 16, and 17 For reference all the weights and neuron values for the 16 neurons, as would be used in a 16 neuron T-SNAP, are shown in Figure 13.
  • Figures 14, 15, 16, and 17 contain 4 replicated T-SNAP sections.
  • the weights and output values for neurons 1 to 4 are depicted in Figure 14 STEP 1.
  • Figure 15 STEP 2 shows the weights and structure required for neurons 5 to 8.
  • Figure 16 STEP 3 shows neurons 9 to 12
  • Figure 17 STEP 4 shows neurons 13 to 16.
  • Y 3 F(W 3,1 Y 1 + W 3,2 Y 2 + W 3,3 Y 3 + W 3,4 Y 4 + W 3,5 Y 5 + W 3,6 Y 6 + W 3 ,7 Y 7 + W 3,8 Y 8 ) +
  • Y 7 F(W 7,1 Y 1 + W 7,2 Y 2 + W 7,3 Y 3 + W 7,4 Y 4 + W 7,5 Y 5 + W 7,6 Y 6 + W 7,7 Y 7 + W 7,8 Y 8 ) + (W 7,9 Y 9 + W 7,10 Y 10 + W 7,11 Y 11 + W 7, 12 Y 12 + W 7,13 Y 13 + W 7,14 Y 14 + W 7,15 Y 15 + W 7,16 Y 16 )
  • Y 12 F(W 12, 1 Y 1 + W 12,2 Y 2 + W 12,3 Y 3 + W 12,4 Y 4 + W 12,5 Y 5 + W 12,6 Y 6 + W 12,7 Y 7 + W 12,8 Y 8 ) + (W 12,9 Y 9 + W 12,10 Y 10 + W 12,11 Y 11 + W 12, 12 Y 12 + W 12,13 Y 13 + W 12,14 Y 14 + W 12,15 Y 15 + W 12,16 Y 16 )
  • Y 14 F(W 14,1 Y 1 + W 14.2 Y 2 + W 14,3 Y 3 + W 14,4 Y 4 + W 14.5 Y 5 + W 14,6 Y 6 + W 14,7 Y 7 + W 14,8 Y 8 ) +
  • S represents the sigmoid generators.
  • TSNAP - V1 period K( ⁇ M + 2( log 2 N + log 2 K) ⁇ A + ⁇ s )
  • TSNAP-V2 TSNAP VIRTUAL ARCHITECTURE 2
  • the TSNAP-V2 performance is:
  • TSNAP - V2 period K(K ⁇ M + 2( log 2 N ⁇ A + ⁇ IA ) + ⁇ s )

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • Evolutionary Computation (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Neurology (AREA)
  • Feedback Control In General (AREA)
  • Multi Processors (AREA)
  • Devices For Executing Special Programs (AREA)
  • Complex Calculations (AREA)
  • Executing Machine-Instructions (AREA)
  • Advance Control (AREA)
PCT/US1991/003318 1990-05-22 1991-05-17 Virtual neurocomputer architectures for neural networks WO1992001257A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US526,866 1990-05-22
US07/526,866 US5065339A (en) 1990-05-22 1990-05-22 Orthogonal row-column neural processor
US68278691A 1991-04-08 1991-04-08
US682,786 1991-04-08

Publications (1)

Publication Number Publication Date
WO1992001257A1 true WO1992001257A1 (en) 1992-01-23

Family

ID=27062243

Family Applications (4)

Application Number Title Priority Date Filing Date
PCT/US1991/003315 WO1991018349A1 (en) 1990-05-22 1991-05-17 Scalable flow virtual learning neurocomputer
PCT/US1991/003316 WO1991018350A1 (en) 1990-05-22 1991-05-17 A learning machine synapse processor system apparatus
PCT/US1991/003318 WO1992001257A1 (en) 1990-05-22 1991-05-17 Virtual neurocomputer architectures for neural networks
PCT/US1991/003317 WO1991018351A1 (en) 1990-05-22 1991-05-17 Pyramid learning architecture neurocomputer

Family Applications Before (2)

Application Number Title Priority Date Filing Date
PCT/US1991/003315 WO1991018349A1 (en) 1990-05-22 1991-05-17 Scalable flow virtual learning neurocomputer
PCT/US1991/003316 WO1991018350A1 (en) 1990-05-22 1991-05-17 A learning machine synapse processor system apparatus

Family Applications After (1)

Application Number Title Priority Date Filing Date
PCT/US1991/003317 WO1991018351A1 (en) 1990-05-22 1991-05-17 Pyramid learning architecture neurocomputer

Country Status (4)

Country Link
US (3) US5509106A (US06174465-20010116-C00003.png)
EP (4) EP0484506A1 (US06174465-20010116-C00003.png)
JP (4) JP2746350B2 (US06174465-20010116-C00003.png)
WO (4) WO1991018349A1 (US06174465-20010116-C00003.png)

Families Citing this family (36)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2647327B2 (ja) * 1992-04-06 1997-08-27 インターナショナル・ビジネス・マシーンズ・コーポレイション 大規模並列コンピューティング・システム装置
JP2572522B2 (ja) * 1992-05-12 1997-01-16 インターナショナル・ビジネス・マシーンズ・コーポレイション コンピューティング装置
US5517667A (en) * 1993-06-14 1996-05-14 Motorola, Inc. Neural network that does not require repetitive training
BR9508898A (pt) * 1994-09-07 1997-11-25 Motorola Inc Sistema para reconhecer sons falados
US6128720A (en) * 1994-12-29 2000-10-03 International Business Machines Corporation Distributed processing array with component processors performing customized interpretation of instructions
US5659785A (en) * 1995-02-10 1997-08-19 International Business Machines Corporation Array processor communication architecture with broadcast processor instructions
US5799134A (en) * 1995-03-13 1998-08-25 Industrial Technology Research Institute One dimensional systolic array architecture for neural network
US6023753A (en) * 1997-06-30 2000-02-08 Billion Of Operations Per Second, Inc. Manifold array processor
US6167502A (en) * 1997-10-10 2000-12-26 Billions Of Operations Per Second, Inc. Method and apparatus for manifold array processing
JP3413213B2 (ja) * 1997-12-19 2003-06-03 ビーエイイー システムズ パブリック リミテッド カンパニー バイナリーコードコンバーター及びコンパレーター
US7254565B2 (en) * 2001-07-26 2007-08-07 International Business Machines Corporation Method and circuits to virtually increase the number of prototypes in artificial neural networks
JP3987782B2 (ja) * 2002-10-11 2007-10-10 Necエレクトロニクス株式会社 アレイ型プロセッサ
GB2400201A (en) * 2003-04-05 2004-10-06 Hewlett Packard Development Co Network modelling its own response to a requested action
US8443169B2 (en) * 2005-03-28 2013-05-14 Gerald George Pechanek Interconnection network connecting operation-configurable nodes according to one or more levels of adjacency in multiple dimensions of communication in a multi-processor and a neural processor
CN111291873A (zh) * 2014-07-21 2020-06-16 徐志强 预制性突触的模拟方法及装置
US9747546B2 (en) 2015-05-21 2017-08-29 Google Inc. Neural network processor
CN105512724B (zh) * 2015-12-01 2017-05-10 中国科学院计算技术研究所 加法器装置、数据累加方法及数据处理装置
US11308383B2 (en) 2016-05-17 2022-04-19 Silicon Storage Technology, Inc. Deep learning neural network classifier using non-volatile memory array
JP6890615B2 (ja) * 2016-05-26 2021-06-18 タータン エーアイ リミテッド ディープニューラルネットワークについての加速器
CN111310893B (zh) * 2016-08-05 2023-11-21 中科寒武纪科技股份有限公司 一种用于执行神经网络运算的装置及方法
US9946539B1 (en) 2017-05-23 2018-04-17 Google Llc Accessing data in multi-dimensional tensors using adders
US10534607B2 (en) 2017-05-23 2020-01-14 Google Llc Accessing data in multi-dimensional tensors using adders
US11551064B2 (en) 2018-02-08 2023-01-10 Western Digital Technologies, Inc. Systolic neural network engine capable of forward propagation
US11461579B2 (en) 2018-02-08 2022-10-04 Western Digital Technologies, Inc. Configurable neural network engine for convolutional filter sizes
US10853034B2 (en) 2018-03-30 2020-12-01 Intel Corporation Common factor mass multiplication circuitry
JP6902000B2 (ja) 2018-07-10 2021-07-14 株式会社東芝 演算装置
EP4009183A1 (en) 2018-10-18 2022-06-08 Shanghai Cambricon Information Technology Co., Ltd Network-on-chip data processing method and device
CN109614876B (zh) * 2018-11-16 2021-07-27 北京市商汤科技开发有限公司 关键点检测方法及装置、电子设备和存储介质
CN109657788A (zh) * 2018-12-18 2019-04-19 北京中科寒武纪科技有限公司 数据处理方法、装置及相关产品
US11270763B2 (en) 2019-01-18 2022-03-08 Silicon Storage Technology, Inc. Neural network classifier using array of three-gate non-volatile memory cells
US11409352B2 (en) 2019-01-18 2022-08-09 Silicon Storage Technology, Inc. Power management for an analog neural memory in a deep learning artificial neural network
US11023559B2 (en) 2019-01-25 2021-06-01 Microsemi Soc Corp. Apparatus and method for combining analog neural net with FPGA routing in a monolithic integrated circuit
US11270771B2 (en) 2019-01-29 2022-03-08 Silicon Storage Technology, Inc. Neural network classifier using array of stacked gate non-volatile memory cells
US10929058B2 (en) 2019-03-25 2021-02-23 Western Digital Technologies, Inc. Enhanced memory device architecture for machine learning
US11783176B2 (en) 2019-03-25 2023-10-10 Western Digital Technologies, Inc. Enhanced storage device memory architecture for machine learning
US11423979B2 (en) 2019-04-29 2022-08-23 Silicon Storage Technology, Inc. Decoding system and physical layout for analog neural memory in deep learning artificial neural network

Family Cites Families (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4796199A (en) * 1987-02-24 1989-01-03 Oregon Graduate Center Neural-model, information-handling architecture and method
US4858147A (en) * 1987-06-15 1989-08-15 Unisys Corporation Special purpose neurocomputer system for solving optimization problems
US5014235A (en) * 1987-12-15 1991-05-07 Steven G. Morton Convolution memory
FR2625347B1 (fr) * 1987-12-23 1990-05-04 Labo Electronique Physique Structure de reseau de neurones et circuit et arrangement de reseaux de neurones
US4953099A (en) * 1988-06-07 1990-08-28 Massachusetts Institute Of Technology Information discrimination cell
EP0349819B1 (de) * 1988-07-05 1993-12-22 Siemens Aktiengesellschaft In integrierter Schaltungstechnik ausgeführtes digitales neuronales Netz
GB2224139A (en) * 1988-10-24 1990-04-25 Philips Electronic Associated Digital data processing apparatus
FR2639461A1 (fr) * 1988-11-18 1990-05-25 Labo Electronique Physique Arrangement bidimensionnel de points memoire et structure de reseaux de neurones utilisant un tel arrangement
DE68927474T2 (de) * 1988-12-29 1997-05-22 Sharp Kk Neuro-Rechner
JPH02287670A (ja) * 1989-04-27 1990-11-27 Mitsubishi Electric Corp 半導体神経回路網
JP2517410B2 (ja) * 1989-05-15 1996-07-24 三菱電機株式会社 学習機能付集積回路装置
US5148514A (en) * 1989-05-15 1992-09-15 Mitsubishi Denki Kabushiki Kaisha Neural network integrated circuit device having self-organizing function
US5148515A (en) * 1990-05-22 1992-09-15 International Business Machines Corp. Scalable neural array processor and method
US5243688A (en) * 1990-05-22 1993-09-07 International Business Machines Corporation Virtual neurocomputer architectures for neural networks

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Design of Parallel Hardware Neural Network Systems from Custom Analog VSLI 'Building Block' Chips; EBERHARDT et al; IJCNN; Vol. 2; June 18-22, 1989 pp. II-183 to II-190. *
Implementing Neural Nets with Programmable Logic; JACQUES J. VIDAL; IEEE Trans. on Acoustics, Speech, and Signal Processing; Vol. 36, No. 7; July 1988; pp 1180-1190. *

Also Published As

Publication number Publication date
US5509106A (en) 1996-04-16
EP0486684A1 (en) 1992-05-27
JPH04505824A (ja) 1992-10-08
JP2663996B2 (ja) 1997-10-15
JP2746350B2 (ja) 1998-05-06
WO1991018350A1 (en) 1991-11-28
EP0484522A4 (US06174465-20010116-C00003.png) 1994-03-23
EP0486635A1 (en) 1992-05-27
EP0484506A4 (US06174465-20010116-C00003.png) 1994-03-23
EP0484522A1 (en) 1992-05-13
JP2502867B2 (ja) 1996-05-29
WO1991018349A1 (en) 1991-11-28
EP0486684A4 (US06174465-20010116-C00003.png) 1994-03-23
JP2663995B2 (ja) 1997-10-15
JPH04507027A (ja) 1992-12-03
EP0486635A4 (US06174465-20010116-C00003.png) 1994-03-23
WO1991018351A1 (en) 1991-11-28
US5617512A (en) 1997-04-01
US5542026A (en) 1996-07-30
EP0484506A1 (en) 1992-05-13
JPH04507026A (ja) 1992-12-03
JPH05500429A (ja) 1993-01-28

Similar Documents

Publication Publication Date Title
WO1992001257A1 (en) Virtual neurocomputer architectures for neural networks
US5483620A (en) Learning machine synapse processor system apparatus
US5613044A (en) Learning machine synapse processor system apparatus
US5243688A (en) Virtual neurocomputer architectures for neural networks
RU2131145C1 (ru) Нейропроцессор, устройство для вычисления функций насыщения, вычислительное устройство и сумматор
US5146543A (en) Scalable neural array processor
US5337395A (en) SPIN: a sequential pipeline neurocomputer
US5065339A (en) Orthogonal row-column neural processor
US5148515A (en) Scalable neural array processor and method
Lehmann et al. A generic systolic array building block for neural networks with on-chip learning
US5640586A (en) Scalable parallel group partitioned diagonal-fold switching tree computing apparatus
US5146420A (en) Communicating adder tree system for neural array processor
US5251287A (en) Apparatus and method for neural processing
CN220569161U (zh) 可调适存储器内运算电路
Krikelis et al. Implementing neural networks with the associative string processor
Delgado-Frias et al. A VLSI pipelined neuroemulator
Misra et al. Neural network simulation on a reduced-mesh-of-trees organization
Wherrett et al. Digital optical circuits for 2D data processing
Aikens II et al. A neuro-emulator with embedded capabilities for generalized learning
Lundheim et al. A Neural Network for Global Second Level Trigger-A Real-time Implementation on DecPeRLe-1
Faure et al. A cellular architecture dedicated to neural net emulation
Krikelis A novel massively associative processing architecture for the implementation of artificial neural networks
JPH02287862A (ja) ニューラルネットワーク演算装置
Malhotra et al. Evaluation of electronic artificial neural network implementations
Vassiliadis et al. SPIN: The sequential pipelined neuroemulator

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 1991918394

Country of ref document: EP

AK Designated states

Kind code of ref document: A1

Designated state(s): JP

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB GR IT LU NL SE

WWP Wipo information: published in national office

Ref document number: 1991918394

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 1991918394

Country of ref document: EP