CN110298442A - A kind of multidomain treat-ment method of intelligent weight information - Google Patents

A kind of multidomain treat-ment method of intelligent weight information Download PDF

Info

Publication number
CN110298442A
CN110298442A CN201910483424.0A CN201910483424A CN110298442A CN 110298442 A CN110298442 A CN 110298442A CN 201910483424 A CN201910483424 A CN 201910483424A CN 110298442 A CN110298442 A CN 110298442A
Authority
CN
China
Prior art keywords
weight information
memory
stored
multidomain treat
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910483424.0A
Other languages
Chinese (zh)
Other versions
CN110298442B (en
Inventor
景蔚亮
杜源
陈邦明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Xinchu Integrated Circuit Co Ltd
Original Assignee
Shanghai Xinchu Integrated Circuit Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Xinchu Integrated Circuit Co Ltd filed Critical Shanghai Xinchu Integrated Circuit Co Ltd
Priority to CN201910483424.0A priority Critical patent/CN110298442B/en
Publication of CN110298442A publication Critical patent/CN110298442A/en
Application granted granted Critical
Publication of CN110298442B publication Critical patent/CN110298442B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/06Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
    • G06N3/063Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons using electronic means

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Computational Linguistics (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Neurology (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Image Analysis (AREA)

Abstract

The present invention relates to information technology field more particularly to a kind of multidomain treat-ment methods of intelligent weight information, applied to the neural network with weight information;Wherein, during constantly train to neural network, the weight information that change frequency is greater than the first preset threshold is stored in a first kind memory;And change frequency is stored in one second class memory less than the weight information of the first preset threshold;It can be by constantly learning, weight information is stored in different positions, so that training speed is getting faster, and the type of storaging medium increases, the storaging medium that can be adapted for the feature setting of the weight information of storage, evidence of fetching from other memory again is not needed, the price of trained neural network model equipment is reduced yet.

Description

A kind of multidomain treat-ment method of intelligent weight information
Technical field
The present invention relates to information technology field more particularly to a kind of multidomain treat-ment methods of intelligent weight information.
Background technique
Important branch of the deep learning as artificial intelligence, performer key player, software-hardware synergism innovation are depth It practises and injects new power.The optimization acceleration of hardware is then set about from two stages of deep learning, respectively training stage and reasoning rank Section.There is tensor processor (tensor processor unit, the TPU) generation in special useful reasoning stage, also takes into account training CPU (Central Processing Unit central processing unit, abbreviation CPU), the GPU (Graphics in stage and reasoning stage Processing Unit graphics processor, abbreviation GPU), NPU (neural-networkprocess units network processing unit, Abbreviation NPU) etc..The weight information of the neural network in these frameworks requires to access data by memory at present, because refreshing Quantity through weight information in network is very big, so needing the memory of large capacity, and DRAM (Dynamic Random Access Memory dynamic random access memory, abbreviation DRAM) be volatile memory, need constantly to refresh, thus its power consumption compared with Greatly, simultaneously as DRAM power failure data will be lost, it is therefore necessary to by the weight information storage of trained model to large capacity In nonvolatile memory, such as HDD (Hard Disk Drive hard disk drive, abbreviation HDD), SSD (Solid State Disk solid state hard disk, abbreviation SSD) in, when powering on next time, need fetch from HDD or SSD evidence, in this way causing property It can be with the waste of power consumption.In addition, the relevance of most weight information is smaller in neural network, and most of weight information Variation it is relatively slow or be basically unchanged, then, when storing these weights, can't be distinguished for current neural network These weight informations, the weight information relatively slow for those constant or variations, DRAM equally will do it refreshing, update weight Information then causes the waste of unnecessary time and power consumption in this way.
Summary of the invention
In view of the above-mentioned problems, the invention proposes a kind of multidomain treat-ment method of intelligent weight information, applied to having The neural network of weight information;
Wherein, during constantly train to the neural network, change frequency is greater than the first preset threshold The weight information be stored in a first kind memory;And
The weight information that change frequency is less than first preset threshold is stored in one second class memory.
Above-mentioned multidomain treat-ment method, wherein judge the change frequency of the weight information in institute by a processing system State the size between the first preset threshold.
Above-mentioned multidomain treat-ment method, wherein the processing system is the computer with an operating system.
Above-mentioned multidomain treat-ment method, wherein there is relevance between the weight information;
Also the weight information that relevance is greater than the second preset threshold is stored in the first kind memory;And
Relevance is stored in the second class memory less than the weight information of the second preset threshold.
Above-mentioned multidomain treat-ment method, wherein the first kind memory is dynamic RAM.
Above-mentioned multidomain treat-ment method, wherein the second class memory is that phase transition storage or flash memory or resistance are dynamic State random access memory or ferroelectric dynamic random access memory, or magnetic dynamic RAM.
Above-mentioned multidomain treat-ment method, wherein the Application of Neural Network is in a learning model.
The utility model has the advantages that a kind of multidomain treat-ment method of intelligent weight information proposed by the present invention, it can be by constantly Weight information, is stored in different positions by study, so that training speed is getting faster, and the type of storaging medium increases, The storaging medium that can be adapted for the feature setting of the weight information of storage, does not need to take from other memory again yet Data reduce the price of trained neural network model equipment.
Detailed description of the invention
Fig. 1 is that each unit connects to be formed in the multidomain treat-ment method of weight information intelligent in one embodiment of the invention Structure principle chart.
Specific embodiment
Invention is further explained with reference to the accompanying drawings and examples.
In a preferred embodiment, a kind of multidomain treat-ment method of intelligent weight information is proposed, can be applied In the neural network with weight information;
Wherein, during constantly train to neural network, change frequency is greater than to the power of the first preset threshold Weight information is stored in a first kind memory;And
Change frequency is stored in one second class memory less than the weight information of the first preset threshold.
The multidomain treat-ment method of the weight information of above-mentioned intelligence can be applied to the study stage of neural network;Due to power The weight information that weight information is being gradually changed with trained change, therefore stored in first kind memory generally can be with training Process constantly tail off;First kind memory is generally memory, completes the learning process of neural network;At the beginning of learning process, All weight informations, which can be, to be stored entirely in first kind memory, when weight information is stored into the second class memory Afterwards, corresponding weight information can be deleted in first kind memory.
In a preferred embodiment, judge the change frequency of weight information in the first default threshold by a processing system Size between value.
In above-mentioned technical proposal, as shown in Figure 1, the connection of processing system and first kind memory and the second class memory can With as shown in the figure.
In above-described embodiment, it is preferable that processing system can be the computer with an operating system.
In above-described embodiment, it is preferable that there is relevance between weight information;
Also the weight information that relevance is greater than the second preset threshold is stored in first kind memory;And
Relevance is stored in the second class memory less than the weight information of the second preset threshold.
In a preferred embodiment, first kind memory can be dynamic RAM.
In a preferred embodiment, the second class memory can for phase transition storage or flash memory or resistance dynamic with Machine memory or ferroelectric dynamic random access memory, or magnetic dynamic RAM etc..
Above-mentioned multidomain treat-ment method, wherein Application of Neural Network is in a learning model.
Specifically, the variation speed of weight information and the relevance of weight are determined by operating system.Carrying out learning tasks Before, by the way that one second preset threshold m is arranged, when operating system calculates the relevance between the weight information in neural network When coefficient n is greater than m, i.e. the relevance of judgement weight information is larger, otherwise, when m is greater than n, that is, determines the association of weight information Property is smaller.Meanwhile one first preset threshold x can be set, when operating system calculates the change of the weight information in neural network When changing property coefficient y greater than x, i.e. the variation of judgement weight information is very fast, otherwise, when x is greater than y, that is, determines the change of weight information Change slower.The big and/or biggish weight information of relevance will be changed to be put into first kind memory, such as memory dram, and that Small and/or relevance is lesser is put into the second class memory for a little weight informations variations, such as nonvolatile memory, is equivalent to Have compressed the weight information in DRAM, in this way can so that the process of training is accelerated, meanwhile, memory can be allowed to give full play to its length Place, while avoiding consuming a large amount of power consumption.
Meanwhile in training process over time, it has more and more weight informations to tend towards stability, operating system The lesser weight information of partially tend towards stability in first kind memory and/or relevance is judged, and is stored in In two class memories, it is always maintained at the high quality of weight information in first kind memory, i.e. first kind memory so always Store that relevance is big and/or the weight information that changes greatly is continuously updated and eliminates the first kind by constantly learning and deposit Weight information in reservoir, and the information that these are eliminated is transferred in the second class memory, gets over trained process Come it is faster, meanwhile, but also first kind memory can vacate the weight letter that space is gone storage relevance big and/or changed greatly Breath.
In conclusion a kind of multidomain treat-ment method of intelligent weight information proposed by the present invention, is applied to have weight The neural network of information;Wherein, during constantly train to neural network, change frequency is greater than the first default threshold The weight information of value is stored in a first kind memory;And the weight information by change frequency less than the first preset threshold is deposited In Chu Yi the second class memory;Weight information can be stored in different positions by constantly learning, so that training speed Degree is getting faster, and the type of storaging medium increases, and can be deposited for what the feature setting of the weight information of storage was adapted Medium is stored up, evidence of fetching from other memory again is not needed yet, reduces the price of trained neural network model equipment.
By description and accompanying drawings, the exemplary embodiments of the specific structure of specific embodiment are given, based on present invention essence Mind can also make other conversions.Although foregoing invention proposes existing preferred embodiment, however, these contents are not intended as Limitation.
For a person skilled in the art, after reading above description, various changes and modifications undoubtedly be will be evident. Therefore, appended claims should regard the whole variations and modifications for covering true intention and range of the invention as.It is weighing The range and content of any and all equivalences, are all considered as still belonging to the intent and scope of the invention within the scope of sharp claim.

Claims (7)

1. a kind of multidomain treat-ment method of intelligent weight information, applied to the neural network with weight information;
It is characterized in that, change frequency is greater than the first default threshold during constantly train to the neural network The weight information of value is stored in a first kind memory;And
The weight information that change frequency is less than first preset threshold is stored in one second class memory.
2. multidomain treat-ment method according to claim 1, which is characterized in that judge that the weight is believed by a processing system Size of the change frequency of breath between first preset threshold.
3. multidomain treat-ment method according to claim 2, which is characterized in that the processing system is with an operating system Computer.
4. multidomain treat-ment method according to claim 1, which is characterized in that have relevance between the weight information;
Also the weight information that relevance is greater than the second preset threshold is stored in the first kind memory;And
Relevance is stored in the second class memory less than the weight information of the second preset threshold.
5. multidomain treat-ment method according to claim 1, which is characterized in that the first kind memory is deposited for dynamic random Reservoir.
6. multidomain treat-ment method according to claim 1, which is characterized in that the second class memory is phase change memory Device or flash memory or resistance dynamic RAM or ferroelectric dynamic random access memory, or magnetic dynamic RAM.
7. multidomain treat-ment method according to claim 1, which is characterized in that the Application of Neural Network is in a learning model In.
CN201910483424.0A 2019-06-04 2019-06-04 Intelligent partition processing method for weight information Active CN110298442B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910483424.0A CN110298442B (en) 2019-06-04 2019-06-04 Intelligent partition processing method for weight information

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910483424.0A CN110298442B (en) 2019-06-04 2019-06-04 Intelligent partition processing method for weight information

Publications (2)

Publication Number Publication Date
CN110298442A true CN110298442A (en) 2019-10-01
CN110298442B CN110298442B (en) 2024-03-22

Family

ID=68027591

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910483424.0A Active CN110298442B (en) 2019-06-04 2019-06-04 Intelligent partition processing method for weight information

Country Status (1)

Country Link
CN (1) CN110298442B (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150317556A1 (en) * 2014-04-30 2015-11-05 Prophetstor Data Services, Inc. Adaptive quick response controlling system for software defined storage system for improving performance parameter
CN107909142A (en) * 2017-11-14 2018-04-13 深圳先进技术研究院 A kind of parameter optimization method of neutral net, system and electronic equipment
CN108122031A (en) * 2017-12-20 2018-06-05 杭州国芯科技股份有限公司 A kind of neutral net accelerator architecture of low-power consumption

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150317556A1 (en) * 2014-04-30 2015-11-05 Prophetstor Data Services, Inc. Adaptive quick response controlling system for software defined storage system for improving performance parameter
CN107909142A (en) * 2017-11-14 2018-04-13 深圳先进技术研究院 A kind of parameter optimization method of neutral net, system and electronic equipment
CN108122031A (en) * 2017-12-20 2018-06-05 杭州国芯科技股份有限公司 A kind of neutral net accelerator architecture of low-power consumption

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
李丽霞: "BP神经网络设计探讨", 《现代预防医学》 *

Also Published As

Publication number Publication date
CN110298442B (en) 2024-03-22

Similar Documents

Publication Publication Date Title
CN110276367B (en) Learning classification device and learning classification method
US11663125B2 (en) Cache configuration performance estimation
CN108122031B (en) Low-power consumption neural network accelerator device
CN112329948B (en) Multi-agent strategy prediction method and device
EP3973401B1 (en) Interleaving memory requests to accelerate memory accesses
US7934058B2 (en) Predictive caching of assets to improve level load time on a game console
CN105242871A (en) Data writing method and apparatus
JP2005322215A5 (en)
CN112051969A (en) Hierarchical storage data self-adaptive migration method based on deep reinforcement learning
CN109240280A (en) Anchoring auxiliary power positioning system control method based on intensified learning
CN114692829A (en) DNN model-based checkpoint selection method, equipment and storage medium
CN110298442A (en) A kind of multidomain treat-ment method of intelligent weight information
KR102192325B1 (en) Data management device supporting high speed artificial neural network operation with caching data based on data locality of artificial neural networks
KR20200139620A (en) Data management device supporting high speed artificial neural network operation with caching data based on data locality of artificial neural networks
Ali Cache replacement algorithm
CN109685208A (en) A kind of method and device accelerated for the dilute combization of neural network processor data
JP2019164629A (en) Model generation device, information processing device, model generation method and program
CN115148248B (en) Deep learning-based DRAM (dynamic random Access memory) refreshing method and device
Gulden et al. Beyond Zipf: An agent-based understanding of city size distributions
US20210200584A1 (en) Multi-processor system, multi-core processing device, and method of operating the same
CN109857742A (en) A kind of big data virtual tables fast display method and system
KR101351648B1 (en) Multi-media server for managing contents using prediction of user's requests, and management method thereof
JP2019160256A (en) Learning discrimination device and method for learning discrimination
Walker Rentier state theory 50 Years on: New developments
Min et al. Learning to play visual doom using model-free episodic control

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant