WO2019165679A1 - 一种包括比特转换装置的神经网络处理器及其方法 - Google Patents
一种包括比特转换装置的神经网络处理器及其方法 Download PDFInfo
- Publication number
- WO2019165679A1 WO2019165679A1 PCT/CN2018/082179 CN2018082179W WO2019165679A1 WO 2019165679 A1 WO2019165679 A1 WO 2019165679A1 CN 2018082179 W CN2018082179 W CN 2018082179W WO 2019165679 A1 WO2019165679 A1 WO 2019165679A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- bit
- data
- neural network
- bit conversion
- conversion
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
- G06N3/063—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/048—Activation functions
Definitions
- the present invention relates to the field of artificial intelligence, and more particularly to improvements in neural network processors.
- the deep learning technology of artificial intelligence has been rapidly developed in recent years, and it has been solved in solving high-level abstract cognitive problems such as image recognition, speech recognition, natural language understanding, weather prediction, gene expression, content recommendation and intelligent robots. It has been widely used and proven to have excellent performance. This has made the development and improvement of artificial intelligence technology a research hotspot in academia and industry.
- Deep neural network is one of the highest level of perceptual models in the field of artificial intelligence. This kind of network simulates the neural connection structure of the human brain by establishing a model, and uses multiple transformation stages to describe the data features for images, videos and Large-scale data processing tasks such as audio have brought about breakthroughs.
- the model of deep neural network belongs to an operation model, which contains a large number of nodes. These nodes adopt a mesh interconnection structure, which is called a neuron of deep neural network.
- the strength of the connection between the two nodes represents the weighted value of the signal between the two nodes, ie the weight, to correspond to the memory in the biological neural network.
- a dedicated processor for neural network computing has also been developed accordingly.
- a large amount of data needs to be repeatedly convolved, activated, pooled, etc., which requires a large amount of computation time, which seriously affects the user experience. This makes how to reduce the computation time of the neural network becomes an improved strategy for the neural network processor.
- a neural network processor including a bit conversion device including the following:
- An input interface a control unit, a data conversion unit, and an output interface
- the control unit is configured to generate a control signal for the data conversion unit
- the input interface is configured to receive original data
- the data conversion unit is configured to perform bit conversion on the original data according to the control signal to convert the original data into a bit conversion result expressed by using fewer bits;
- the output interface is configured to output the bit conversion result to the bit conversion device.
- control unit is configured to determine a rule for performing bit conversion according to the set parameter or the input parameter to generate the control signal
- the parameter includes information related to the number of bits of the original data and the number of bits of the bit conversion result.
- the data conversion unit is configured to determine a reserved bit and a truncation bit in the original data according to the control signal, and according to the reserved bit of the original data and the The highest bit of the truncated bits of the original data determines the bit conversion result.
- the data conversion unit is configured to determine a reserved bit and a truncation bit in the original data according to the control signal, and use a reserved bit in the original data as a Describe the bit conversion result.
- the data conversion unit is configured to perform bit conversion on the original data according to the control signal, and convert the original data into a bit conversion expressed by using half of the original number of bits result.
- control unit generates a control signal for the data conversion unit
- the input interface receives raw data from outside the bit conversion device that needs to perform bit conversion
- the data conversion unit performs bit conversion on the original data according to the control signal to convert the original data into a bit conversion result expressed by using fewer bits;
- the output interface outputs the bit conversion result to the bit conversion device.
- step 1) comprises:
- control unit determines a rule for performing bit conversion according to the set parameter or the input parameter
- control unit generates a control signal corresponding to the rule
- the parameter includes information related to the number of bits of the original data and the number of bits of the bit conversion result.
- step 3 comprises:
- the data conversion unit determines the bit conversion result based on the reserved bit of the original data and the highest bit of the truncated bits of the original data according to the control signal.
- step 3 comprises:
- the data conversion unit uses a reserved bit in the original data as the bit conversion result according to the control signal.
- the buffered neural network data is input to the bit conversion device to perform steps 1)-4), or
- the result of the convolution operation is input to the bit conversion device to perform steps 1)-4).
- a computer readable storage medium having stored therein a computer program, when executed, for implementing the method of any of the above.
- the present invention provides a bit conversion apparatus for a neural network processor that can be used to adjust the number of bits used to express data in various computational processes of a neural network. By reducing the number of bits used to express the data, the hardware cost required for the calculation can be reduced, the computational speed can be increased, the need for the data storage space of the neural network processor can be reduced, and the energy consumption for performing neural network calculations can be reduced.
- FIG. 1 shows a block diagram of a bit conversion device in accordance with one embodiment of the present invention
- FIG. 2 is a connection diagram of respective units in a bit conversion device according to an embodiment of the present invention.
- FIG. 3 is a flow chart of a method of bit-transforming neural network data using a bit conversion device as shown in FIG. 1 according to an embodiment of the present invention
- 4a is a hardware configuration diagram for performing bit conversion in a "rounding mode" in a data conversion unit of a bit conversion device according to an embodiment of the present invention
- 4b is a hardware configuration diagram for performing bit conversion in a "direct truncation mode" in a data conversion unit of a bit conversion device according to an embodiment of the present invention.
- the inventors believe that by appropriately reducing the number of bits of data involved in the calculation of the neural network, for example, using fewer bits to represent data that would otherwise require more bits, reducing the amount of computation to reduce the neural network. Calculation time. This is because the inventors have found in the prior art that the neural network algorithm has relatively high fault tolerance for the intermediate result of the calculation, even if fewer bits are used to represent that more bits are needed to represent The practice of the data changes the accuracy of the data involved in the calculation and thus the accuracy of the intermediate results obtained, but this does not have a large impact on the final output of the neural network.
- bit conversion the manner in which the bit of the data used for the calculation is reduced.
- bit conversion the process of adjusting the number of binary bits required to express a value.
- bit conversion For example, for a decimal value of 0.5, the result of using Q7 fixed-point data is 01000000 (where Q7 uses the leftmost first bit of the 8 bits as the sign bit, and the remaining 7 bits represent the fractional part, thereby To represent the decimal between -1 and 1 with a precision of 7), when performing bit conversion, the result originally expressed by Q7 can be modified to be represented by Q3, and the result is 0100 (similar to Q7, Q3 is also used).
- the first leftmost bit is used as a sign bit, except that it uses 3 bits to represent the fractional part, which can represent a fractional precision of -1 between 1 and 1.
- the present invention proposes a bit conversion device for a neural network processor.
- a rule for performing bit conversion can be determined by the bit conversion device according to a parameter set or based on a user input to perform bit conversion on the data.
- the neural network processor can process a relatively small amount of data, thereby increasing the processing speed and reducing the energy consumption of the neural network processor.
- the inventor believes that in the logic combination circuit, the speed of the data operation is inversely proportional to the number of bits of the numerical expression; the energy consumption of the data operation is proportional to the bit of the numerical expression; therefore, after the bit conversion of the data, the accelerated calculation can be achieved. With the effect of reducing power consumption.
- FIG. 1 shows a bit conversion device 101 according to an embodiment of the present invention, comprising: an input bus unit 102 as an input interface, a data conversion unit 103, an output bus unit 104 as an output interface, and a control unit 105.
- the input bus unit 102 is configured to acquire neural network data that needs to be bit-converted to be provided to the data conversion unit 103.
- input bus unit 102 can receive and/or transmit a plurality of data to be converted in parallel.
- the data conversion unit 103 is configured to perform bit conversion on the neural network data from the input bus unit 102 according to a rule for performing bit conversion determined, for example, or based on a parameter input by the user.
- the output bus unit 104 is configured to output the result of the bit conversion obtained by the processing by the data conversion unit 103 from the bit conversion device 101 to provide means for performing subsequent processing in the neural network processing.
- the control unit 105 is configured to determine a rule of bit conversion, and select a corresponding bit conversion mode to control the operation of the data conversion unit 103 to perform bit conversion.
- the control unit 105 may determine a rule for performing bit conversion by analyzing the set parameters or parameters input by the user to select from among various conversion modes set in advance.
- the parameter may include the number of bits of the data to be converted and the number of bits of the data to be converted, or the binary representation of the data to be converted and the binary representation of the converted data, such as Q7. Q3 and so on. For example, based on the parameters entered by the user, it is determined that the neural network data represented by Q7 is converted to be represented by Q3.
- input bus unit 102 and/or output bus unit 104 can receive and/or transmit a plurality of data to be converted in parallel.
- the input bus unit has a bit number of 128 bits
- the output bus has a bit number of 64 bits.
- the control unit receives parameters input by the user from outside the bit conversion device for generating a mode switching signal for the data conversion unit according to the determined bit conversion rule, so that the data conversion unit can know which manner is needed in the current situation To perform bit conversion. And, the control unit may further generate an input control signal for controlling the input bus unit to start receiving data or suspending receiving data, and an output control signal for controlling the output bus unit to start outputting or suspending the output bit conversion result.
- a method of performing bit conversion of neural network data using the bit conversion apparatus as shown in FIG. 1 will be described below by way of an embodiment. Referring to Figure 3, the method includes:
- Step 1 The control unit 105 in the bit conversion device 101 determines the rule of the bit conversion used based on the set conversion demand parameter or the parameter input by the user.
- the set conversion requirement parameter, the parameter input by the user includes information related to the number of bits of the neural network data to be converted and the number of converted data bits.
- the set conversion requirement parameter, the parameter input by the user may further include a cut rule when performing bit conversion, such as a rule of "rounding off” or "direct truncation".
- the control unit 105 selects from a preset bit conversion mode.
- the bit conversion mode comprises a "rounding mode” and a "direct truncation mode", and the processing for the two different modes will be described in a subsequent step.
- Step 2 The input bus unit 102 in the bit conversion device 101 supplies the neural network data obtained by the bit conversion device 101 to the data conversion unit 103.
- the input bus unit 102 herein may include a plurality of interfaces capable of receiving data in parallel to receive neural network data from outside the bit conversion device 101 that needs to perform bit conversion in parallel. Similarly, the input bus unit 102 may also include a plurality of interfaces capable of outputting data in parallel, thereby providing data to the data conversion unit 103 in parallel for bit conversion processing.
- Step 3 The data conversion unit 103 performs bit conversion on the neural network data that needs to perform bit conversion in accordance with the rule of bit conversion determined by the control unit 105.
- a control signal from the control unit 105 can be received by the data conversion unit 103 to perform bit conversion in accordance with the rule.
- the inventors have found that, when reducing the number of bits of the data used for the calculation, if the reduced number of bits is greater than or equal to half the number of bits of the original data, the neural network processor can be made in hardware cost, processing speed, and A compromise between accuracy rates. Therefore, in the present invention, it is preferable to reduce the number of bits of the neural network data that needs to perform bit conversion to half of the original, for example, to perform bit conversion using a fixed hardware structure to convert 32-bit data into 16 bits, and 16 bits.
- the data is converted into 8 bits, 8 bits of data are converted into 4 bits, 4 bits of data are converted into 2 bits, and 2 bits of data are converted into 1 bit.
- each bit of the neural network data that needs to perform bit conversion may be divided into reserved bits and truncated bits according to the rule, wherein the reserved bits are among the bits of the neural network data.
- a high one or more bits, the truncated bits being the remaining bits in the various bits of the neural network data. For example, for the 8-bit data 10101111, if the number of bits is reduced to half of the original number, the reserved bit is 1010 and the truncation bit is 1111.
- 4a shows a hardware device structure for performing bit conversion in the "rounding mode" in the data conversion unit 103, in which 16 8-bit neural network data requiring bit conversion is performed in parallel, according to an embodiment of the present invention.
- Input into the data conversion unit 103 among the 4 bit reserved bits in each 8-bit neural network data, bits other than the sign bit (for example, a 1 , a 2 , a 3 ) and the highest bit of the corresponding truncation bit are removed ( For example, a 4 ) is used as two inputs of the adder, respectively, and the output of the adder and the sign bit in the neural network data are used together as a result of performing bit conversion on the 8-bit neural network data.
- the neural network data input to the conversion unit 103 is 10101111 (inverse), indicating that it represents -0.6328125 in decimal, and its truncation bit is 1111, which will be truncated.
- the highest bit 1 is added to the three bits 010 other than the sign bit in the reserved bit, and the result of bit conversion based on the sign bit in the neural network data and the result of the adder is 1011 (reverse code), indicating Decimal -0.625.
- 4b shows a hardware device structure for performing bit conversion in a "direct truncation mode" in the data conversion unit 103, in which 16 8-bit neural network data requiring bit conversion is paralleled, according to an embodiment of the present invention.
- Input into the data conversion unit 103 4 bit reserved bits (for example, a 0 , a 1 , a 2 , a 3 ) in each 8-bit neural network data are directly used as execution bits for the 8 bit neural network data. The result of the conversion.
- Step 4 The result of the bit conversion obtained by the processing of the data conversion unit 103 by the output bus unit 104 is output from the bit conversion device 101 to provide means for performing subsequent processing in the neural network processing.
- bit conversion apparatus provided by the above-described embodiments of the present invention can be used as part of a neural network processor in various calculation processes for neural networks.
- the buffered neural network data may be bit-converted using a bit conversion device when the buffering of the neural network data has been completed and the convolution operation has not been completed.
- the buffer may be used by the bit conversion device.
- the network data is bit-converted, and the result obtained by the bit conversion is supplied to a unit for performing a convolution operation to perform a convolution operation.
- the bit conversion means may be used to perform bit conversion on the result of the convolution operation when the convolution operation on the data has been completed and the activation operation has not been completed.
- the accumulation operation of the convolution operation unit tends to increase the number of bits of the result of the obtained convolution operation, in order to adapt to the bit number requirements of subsequent operations (for example, for some hardware-implemented activation).
- the number of bits used is often fixed, and the result of the convolution operation needs to be bit-converted.
- the present invention provides a bit conversion apparatus for a neural network processor that can be used to adjust the number of bits used to express data in various calculation processes of a neural network.
- a bit conversion apparatus for a neural network processor that can be used to adjust the number of bits used to express data in various calculation processes of a neural network.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- General Physics & Mathematics (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Neurology (AREA)
- Image Processing (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (11)
- 一种神经网络处理器,该神经网络处理器中包括比特转换装置,该比特转换装置包括:输入接口、控制单元、数据转换单元、和输出接口;其中,所述控制单元用于产生针对所述数据转换单元的控制信号;所述输入接口用于接收原始数据;所述数据转换单元用于根据所述控制信号对所述原始数据进行比特转换,以将所述原始数据转换为采用更少的比特位数进行表达的比特转换结果;所述输出接口用于将所述比特转换结果输出所述比特转换装置。
- 根据权利要求1所述的神经网络处理器,其中所述控制单元用于根据设置的参数或者输入的参数确定执行比特转换的规则,以产生所述控制信号;其中,所述参数包括与所述原始数据的比特位数以及所述比特转换结果的比特位数相关的信息。
- 根据权利要求2所述的神经网络处理器,其中所述数据转换单元用于根据所述控制信号,确定所述原始数据中的保留位以及截断位,并且根据所述原始数据的保留位以及所述原始数据的截断位中的最高位确定所述比特转换结果。
- 根据权利要求2所述的神经网络处理器,其中所述数据转换单元用于根据所述控制信号,确定所述原始数据中的保留位以及截断位,并且将所述原始数据中的保留位作为所述比特转换结果。
- 根据权利要求1所述的神经网络处理器,其中所述数据转换单元用于根据所述控制信号对所述原始数据进行比特转换,以原始数据转化为采用原本一半的比特位数进行表达的比特转换结果。
- 一种采用如权利要求1-5中任意一项所述的神经网络处理器对神经网络的数据进行比特转换的方法,包括:1)所述控制单元产生针对数据转换单元的控制信号;2)所述输入接口接收来自所述比特转换装置外部的需要执行比特转换的原始数据;3)所述数据转换单元根据所述控制信号对所述原始数据进行比特转换,以将所述原始数据转换为采用更少的比特位数进行表达的比特转换结果;4)所述输出接口将所述比特转换结果输出所述比特转换装置。
- 根据权利要求6所述的方法,其中步骤1)包括:1-1)所述控制单元根据设置的参数或者输入的参数确定执行比特转换的规则;1-2)所述控制单元产生与所述规则对应的控制信号;其中,所述参数包括与所述原始数据的比特位数以及所述比特转换结果的比特位数相关的信息。
- 根据权利要求7所述的方法,其中步骤3)包括:所述数据转换单元根据所述控制信号,基于所述原始数据的保留位以及所述原始数据的截断位中的最高位确定所述比特转换结果。
- 根据权利要求7所述的方法,其中步骤3)包括:所述数据转换单元根据所述控制信号,将所述原始数据中的保留位作为所述比特转换结果。
- 根据权利要求6-9中任意一项所述的方法,在已完成对神经网络数据的缓存、并且尚未完成卷积运算时,将缓存的神经网络数据输入所述比特转换装置以执行步骤1)-4),或者在已完成对数据的卷积运算、并且尚未完成激活运算时,将卷积运算的结果输入所述比特转换装置以执行步骤1)-4)。
- 一种计算机可读存储介质,其中存储有计算机程序,所述计算机程序在被执行时用于实现如权利要求6-10中任意一项所述的方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810170612.3A CN108345938A (zh) | 2018-03-01 | 2018-03-01 | 一种包括比特转换装置的神经网络处理器及其方法 |
CN201810170612.3 | 2018-03-01 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2019165679A1 true WO2019165679A1 (zh) | 2019-09-06 |
Family
ID=62959552
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2018/082179 WO2019165679A1 (zh) | 2018-03-01 | 2018-04-08 | 一种包括比特转换装置的神经网络处理器及其方法 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN108345938A (zh) |
WO (1) | WO2019165679A1 (zh) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113392954B (zh) * | 2020-03-13 | 2023-01-24 | 华为技术有限公司 | 终端网络模型的数据处理方法、装置、终端以及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106447034A (zh) * | 2016-10-27 | 2017-02-22 | 中国科学院计算技术研究所 | 一种基于数据压缩的神经网络处理器、设计方法、芯片 |
CN107145939A (zh) * | 2017-06-21 | 2017-09-08 | 北京图森未来科技有限公司 | 一种神经网络优化方法及装置 |
CN107330515A (zh) * | 2016-04-29 | 2017-11-07 | 北京中科寒武纪科技有限公司 | 一种用于执行人工神经网络正向运算的装置和方法 |
CN107423816A (zh) * | 2017-03-24 | 2017-12-01 | 中国科学院计算技术研究所 | 一种多计算精度神经网络处理方法和系统 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106796668B (zh) * | 2016-03-16 | 2019-06-14 | 香港应用科技研究院有限公司 | 用于人工神经网络中比特深度减少的方法和系统 |
CN107340993B (zh) * | 2016-04-28 | 2021-07-16 | 中科寒武纪科技股份有限公司 | 运算装置和方法 |
CN107203808B (zh) * | 2017-05-08 | 2018-06-01 | 中国科学院计算技术研究所 | 一种二值卷积装置及相应的二值卷积神经网络处理器 |
CN107292458B (zh) * | 2017-08-07 | 2021-09-10 | 北京中星微人工智能芯片技术有限公司 | 一种应用于神经网络芯片的预测方法和预测装置 |
-
2018
- 2018-03-01 CN CN201810170612.3A patent/CN108345938A/zh active Pending
- 2018-04-08 WO PCT/CN2018/082179 patent/WO2019165679A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107330515A (zh) * | 2016-04-29 | 2017-11-07 | 北京中科寒武纪科技有限公司 | 一种用于执行人工神经网络正向运算的装置和方法 |
CN106447034A (zh) * | 2016-10-27 | 2017-02-22 | 中国科学院计算技术研究所 | 一种基于数据压缩的神经网络处理器、设计方法、芯片 |
CN107423816A (zh) * | 2017-03-24 | 2017-12-01 | 中国科学院计算技术研究所 | 一种多计算精度神经网络处理方法和系统 |
CN107145939A (zh) * | 2017-06-21 | 2017-09-08 | 北京图森未来科技有限公司 | 一种神经网络优化方法及装置 |
Also Published As
Publication number | Publication date |
---|---|
CN108345938A (zh) | 2018-07-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP7273108B2 (ja) | モデルトレーニング方法、装置、電子デバイス、記憶媒体、プログラム | |
CN107451658B (zh) | 浮点运算定点化方法及系统 | |
US20230048031A1 (en) | Data processing method and apparatus | |
CN109478144A (zh) | 一种数据处理装置和方法 | |
US20220335304A1 (en) | System and Method for Automated Design Space Determination for Deep Neural Networks | |
KR102655950B1 (ko) | 뉴럴 네트워크의 고속 처리 방법 및 그 방법을 이용한 장치 | |
CN108345934B (zh) | 一种用于神经网络处理器的激活装置及方法 | |
JP2022118263A (ja) | 自然言語処理モデルの訓練方法、自然言語処理方法、装置、電子機器、記憶媒体及びプログラム | |
CN111047045B (zh) | 机器学习运算的分配系统及方法 | |
JP2022173453A (ja) | ディープラーニングモデルのトレーニング方法、自然言語処理方法及び装置、電子機器、記憶媒体及びコンピュータプログラム | |
CN110781686A (zh) | 一种语句相似度计算方法、装置及计算机设备 | |
CN115017178A (zh) | 数据到文本生成模型的训练方法和装置 | |
CN112884146A (zh) | 一种训练基于数据量化与硬件加速的模型的方法及系统 | |
WO2020093885A1 (zh) | 一种异构协同计算系统 | |
EP3444758B1 (en) | Discrete data representation-supporting apparatus and method for back-training of artificial neural network | |
WO2019165679A1 (zh) | 一种包括比特转换装置的神经网络处理器及其方法 | |
CN116762080A (zh) | 神经网络生成装置、神经网络运算装置、边缘设备、神经网络控制方法以及软件生成程序 | |
CN117151178A (zh) | 一种面向fpga的cnn定制网络量化加速方法 | |
CN112183744A (zh) | 一种神经网络剪枝方法及装置 | |
CN116976461A (zh) | 联邦学习方法、装置、设备及介质 | |
JP2022078286A (ja) | データ処理モデルのトレーニング方法、装置、電子機器及び記憶媒体 | |
CN114462592A (zh) | 模型训练方法、装置、电子设备及计算机可读存储介质 | |
CN113935456A (zh) | 脉冲神经网络层内数据处理方法及设备、处理芯片 | |
CN114692865A (zh) | 一种神经网络量化训练方法、装置及相关产品 | |
CN108416435B (zh) | 一种具有低带宽激活装置的神经网络处理器及其方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 18908091 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18908091 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18908091 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 280921) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 18908091 Country of ref document: EP Kind code of ref document: A1 |