WO2020259543A1 - 一种基于联合优化回声状态网络的热数据预测方法 - Google Patents
一种基于联合优化回声状态网络的热数据预测方法 Download PDFInfo
- Publication number
- WO2020259543A1 WO2020259543A1 PCT/CN2020/097950 CN2020097950W WO2020259543A1 WO 2020259543 A1 WO2020259543 A1 WO 2020259543A1 CN 2020097950 W CN2020097950 W CN 2020097950W WO 2020259543 A1 WO2020259543 A1 WO 2020259543A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- particle
- storage layer
- echo state
- state network
- fitness value
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0602—Interfaces specially adapted for storage systems specifically adapted to achieve a particular effect
- G06F3/0614—Improving the reliability of storage systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0628—Interfaces specially adapted for storage systems making use of a particular technique
- G06F3/0638—Organizing or formatting or addressing of data
- G06F3/064—Management of blocks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/06—Digital input from, or digital output to, record carriers, e.g. RAID, emulated record carriers or networked record carriers
- G06F3/0601—Interfaces specially adapted for storage systems
- G06F3/0668—Interfaces specially adapted for storage systems adopting a particular infrastructure
- G06F3/0671—In-line storage system
- G06F3/0673—Single storage device
- G06F3/0679—Non-volatile semiconductor memory device, e.g. flash memory, one time programmable memory [OTP]
Definitions
- the invention belongs to the technical field of chaotic time series prediction, and specifically relates to a thermal data prediction method based on a joint optimization echo state network.
- NAND flash memory As a non-volatile storage technology, NAND flash memory is widely used in communication systems and consumer electronic products. Compared with hard disk drives, NAND flash memory has higher access speed and power efficiency. In consumer electronic devices based on NAND flash memory, a large number of applications rely on NAND flash memory for data exchange, file storage and video storage. NAND flash memory is mainly used to store large-capacity data. The NAND structure can provide extremely high cell density, which can achieve high storage density, high writing and high erasing speed; therefore, NAND flash memory is mostly used for large-capacity data storage, such as solid state drives. The demand for NAND flash memory will continue to grow in the future, mainly in cloud computing, Internet of Things and data centers.
- NAND flash memory faces at least two challenges, namely, remote update and limited durability, which limits its large-scale application.
- NAND flash memory has the defect that the operation cannot be overwritten, that is, a new write operation cannot be performed on a page before the flash memory is erased. Therefore, improper updates will generate many invalid pages and dead pages, which will reduce efficiency and performance.
- NAND flash memory has a limited lifespan because the flash memory block can only withstand a limited number of erasing. If the block's erasing number is greater than the block's maximum erasable number, it will be unusable.
- Garbage Collection (GC) and Wear Leveling (Wear Leveling, WL) allocate frequently written data (i.e.
- HDI hot data identification
- the essence of HDI is to try to understand the access behavior of hot data well so as to intelligently allocate different data to the appropriate blocks.
- traditional HDI has the following two problems. One is that the memory overhead is large. At present, most of the hot data identification mechanisms are based on the idea of identifying hot data pages in NAND flash memory.
- the core principle of these mechanisms is the page counter, which records the number of read and write operations of the logical page corresponding to the NAND flash memory page within a certain period of time. If the number of reads and writes is greater than the set threshold, the requested page is judged as a hot page, otherwise it is a cold page. Another serious problem is that the recognition accuracy is not high.
- the hot data recognition mechanism based on the Bloom filter is widely used in the recognition of the hot and cold data of the SSD, but the inherent defect of the Bloom filter is the false positive, which means that the data does not belong to the set. Wrongly judged to be in the set.
- the hot data identification method based on load request size and load access mode has a single consideration factor, and the local characteristics of the load are not fully considered, and the accuracy of hot data identification is not high.
- the purpose of the present invention is to provide a thermal data prediction method based on a joint optimization echo state network, which creatively proposes to replace the traditional thermal data recognition with thermal data prediction, and constructs a joint optimization echo state network, so that the predicted thermal data More real-time and reliable.
- the present invention provides a hot data prediction method based on a joint optimization echo state network, which includes the following steps:
- the position information of the particle includes the initial position and the position range of the particle, and the position of each particle is represented by a storage layer parameter in the echo state network;
- the quantum particle swarm algorithm is used to update the particle position based on the position range of each particle.
- the L2+ adaptive L1/2 regularization constrained echo state network is used to calculate the output weight and calculate the global optimal fitness value.
- the particle position corresponding to the global best fitness value at the end of the iteration is used as the best storage layer parameter;
- y represents the obtained predicted logical block address
- the data on the predicted logical block address is the thermal data
- x is the logical block address where the input historical thermal data is located
- W out represents the output weight.
- the historical thermal data The address of the logical block is used in the echo state network training process in step S2 and step S3.
- the present invention innovatively replaces the thermal data recognition module in the flash memory conversion layer with thermal data prediction while continuing to use the solid state hard disk structure framework, and uses the joint optimization echo state network to predict thermal data.
- the joint optimization includes two parts. The first part uses quantum The particle swarm algorithm iterative optimization determines the optimal storage layer parameters of the echo state network. The second part uses the L2+adaptive L1/2 regularization constrained ESN to obtain high sparsity output weights.
- the present invention iteratively searches the quantum particle swarm algorithm Optimal and L2+adaptive L1/2 regularization constraints are combined to obtain the best storage layer parameters, and the joint optimized echo state network used for prediction is more real-time and reliable.
- the present invention uses the logical block address where the historical hot data is located to train the echo state network to obtain the final output weight, and then uses it to predict the logical block address where the hot data is located.
- step S2 the execution process of iterative optimization and determination of optimal storage layer parameters in step S2 is as follows:
- the current position of each particle is sequentially used as the storage layer parameter in the echo state network and the output weight is calculated;
- S23 Based on the principle of minimum fitness value, select the individual best fitness value, individual best parameter, and global best fitness value and global best parameter of each particle according to the fitness value of each particle;
- the particle position selected as the global best fitness value is the global best parameter
- step S25 Determine whether the number of iterations reaches the maximum number of iterations, if not, return to step S24 for the next iteration calculation; otherwise, use the current global optimal parameter as the optimal storage layer parameter.
- any particle j is updated according to the following formula:
- P j (t+1) and P j (t) respectively represent the position of particle j after and before update
- u j are random numbers
- sbest j and sbest i represent the best individual parameters of the jth and ith particles
- mbest is the average value of the current individual best parameters of all particles
- iter and iter max are the current iteration times respectively
- the maximum number of iterations ⁇ max and ⁇ min are the inertia factors respectively
- N is the total number of particles.
- the calculation formula for the fitness value of any particle j is:
- Fitness represents the fitness value of the current particle j, ⁇ 1 and ⁇ 2 are both regularization coefficients, and W out is the output weight corresponding to the current particle j;
- Y represents the logical block address of the historical thermal data used for network training
- X represents the state information of the storage layer updated based on the previous segment of the logical block address of the historical hot data based on network training, and
- X*W out represents the prediction result corresponding to the latter segment of the logical block address of the historical thermal data.
- the process of calculating the output weights using the L2+adaptive L1/2 regularization constraint echo state network is as follows:
- the input layer-storage layer weight matrix and the storage layer internal connection weight matrix are related to the storage layer parameters in the echo state network;
- the state information X is composed of state node information X(t);
- U(t) represents the t-th data in the input variable U
- X(t) and X(t-1) respectively represent the t-th and t-1th state node information
- T of t is determined by the input
- the data length of the variable U is determined
- W in and W x respectively represent the input layer-storage layer weight matrix in the echo state network, the storage layer internal connection weight matrix, logsig( ⁇ ) represents the activation function
- E represents the loss function
- ⁇ 1 and ⁇ 2 are both regularization coefficients.
- step U403 is: simplify the loss function, and then calculate the output weight by using the coordinate descent algorithm;
- I is the identity matrix
- the method of solving matrix W′ out is to calculate each element separately, the value of the k-th element in the m-th row of W′ out is as follows:
- Y′ k (t) represents the t-th element in the k-th row of Y′
- X′ j (t) represents the t-th element in the j-th row of X′
- L is the number of output layer nodes
- n is the number of storage layer nodes.
- step U403 it also includes adaptive optimization of the output weight obtained in step U403, and the optimization process is as follows:
- the converted loss function is:
- K is the number of nodes in the input layer.
- the storage layer parameters in the echo state network include four key parameters: internal connection spectrum radius, storage layer scale, input layer scale factor and storage layer sparsity.
- the parameters required to initialize the quantum particle swarm algorithm in step S1 include the particle swarm size N, the maximum number of iterations iter max , and the inertia factors ⁇ max and ⁇ min .
- the particle position parameter is set to a boundary value corresponding to the exceeding position range.
- the present invention innovatively proposes to replace traditional thermal data recognition with thermal data prediction.
- the public thermal data prediction technology can predict the nature of the next data one or even several beats in advance based on historical access behavior, and actively allocate storage to
- the solid state drive (SSD) corresponding block (hot/cold data block) is more active than the traditional hot data recognition, and the implementation process of the present invention uses joint optimization to improve the accuracy of network prediction, thereby obtaining better Accurate thermal data recognition effect and efficient thermal data prediction will better serve garbage collection and wear leveling technologies, and ultimately improve the life of solid state drives.
- the neural network method retains more characteristic information for the input, and more comprehensively classifies the thermal data.
- the present invention performs joint optimization on the echo state network.
- the L2 regularization constraint obtains a good generalization ability through the trade-off between the model deviation and the prediction variance so as to obtain the weight of continuous shrinkage, but it cannot produce sparse solution; L1/2 regularization can generate very sparse solutions, but when there is a high degree of correlation between the predictor variables, L1/2 cannot play a good regulatory role.
- the present invention uses L2+adaptive L1/2 regularization to train the least square Multiplication can obtain the advantages of two kinds of regularization, thereby improving the prediction accuracy of thermal data.
- optimizing the parameters of the echo state network storage layer based on the QPSO algorithm can solve the problem that the storage layer parameters cannot be determined when building the model.
- this algorithm removes the velocity information of the particles based on the wave-particle duality and only retains the position information, which can effectively reduce the complexity of the calculation, and at the same time obtain the storage layer parameters of the adaptive model, thereby further improving Prediction accuracy; further, the present invention combines L2+adaptive L1/2 regularization and QPSO algorithm to obtain the best storage layer parameters and improve prediction accuracy.
- Figure 1 is a typical architecture of a NAND flash memory system
- FIG. 2 is a flowchart of a method for predicting hot data based on a joint optimized echo state network according to an embodiment of the present invention
- Fig. 3 is a specific algorithm flow chart of the iterative optimization of the quantum particle swarm algorithm of the present invention; wherein, after the execution of step U304 in Fig. 3A is completed, it turns to step U305 in Fig. 3B.
- Fig. 4 is a specific algorithm flow chart of the present invention using L2+adaptive L1/2 constrained echo state network to calculate output weights.
- Fig. 5 is a performance comparison diagram of four actual workloads according to an embodiment of the present invention.
- the present invention provides a hot data prediction method based on a joint optimization echo state network, which is mainly applied to a NAND flash memory system.
- the typical architecture of a NAND flash memory system includes module B101 (user operation) and module B102. (File system) and module B103 (solid state drive). The actual operation of the user will affect the solid state drive through the file system.
- the solid state drive also includes a flash memory conversion layer, a flash controller, and a NAND flash array.
- the flash memory conversion layer includes an address allocation unit, a garbage collection unit, a wear leveling unit, and a thermal data prediction unit.
- the present invention innovatively proposes to use the thermal data prediction unit
- the traditional hot data identification method usually passively analyzes user access behavior, and allocates and stores the corresponding data to the corresponding block (hot/cold data area) of the solid state drive (SSD) through the Flash Transport Layer Protocol (FTL) Block), this method has high hot data missed detection or false alarms when responding to requests with complex access behaviors.
- the hot data prediction technology disclosed in the present invention can predict the nature of the next data one or even a few beats in advance based on historical access behavior, and actively allocate and store it to the corresponding block (hot/cold data block) of the solid state drive (SSD).
- the thermal data prediction method proposed by the present invention is essentially "predictive thermal data identification".
- the predicted logical block address information finally obtained by the present invention is used for garbage collection and wear leveling processing.
- the traditional thermal data recognition is to accurately and efficiently distinguish which data is valid data.
- the present invention provides a thermal data prediction method based on a joint optimization echo state network, which replaces thermal data identification with thermal data prediction, and has high-precision prediction, which specifically includes the following steps:
- the position information of the particle includes the initial position and the position range of the particle.
- the position of each particle is represented by the storage layer parameters in the echo state network (ESN).
- the storage layer parameters in the echo state network include the internal connection spectrum radius, the storage layer scale, and the input.
- the layer scale factor and the storage layer sparsity, the dimension of each particle is initialized to 4 in this example, that is, each particle is a 1*4 matrix, which represents the 4 parameters of the ESN storage layer.
- Determine the range of the 4 parameters set the parameter range as the position range of all particles, and randomly assign a value to each particle within the position range during initialization. In the subsequent update process, it can be regarded as the particle continuously moving towards the maximum within the specified range. If the particle moves beyond the specified range, the particle position information is updated to the boundary value.
- Each particle position represents a specific value of ESN storage layer parameters.
- the parameters required by the quantum particle swarm algorithm include the particle swarm size N, the maximum number of iterations Itermax, the inertia factors ⁇ max and ⁇ min (used to update the particle position information later).
- the quantum particle swarm algorithm is used to update the particle position based on the position range of each particle, and the echo state network with L2+adaptive L1/2 regularization constraint is used to calculate the output weight during each update process to obtain the global optimal fitness value.
- the particle position corresponding to the global optimal fitness value is used as the optimal storage layer parameter.
- the current position of each particle is sequentially used as the storage layer parameter in the echo state network and the output weight is calculated;
- S23 Based on the principle of minimum fitness value, select the individual best fitness value, individual best parameter, and global best fitness value and global best parameter of each particle according to the fitness value of each particle;
- the particle position selected as the global best fitness value is the global best parameter
- step S25 Determine whether the number of iterations reaches the maximum number of iterations, if not, return to step S24 for the next iteration calculation; otherwise, use the current global optimal parameter as the optimal storage layer parameter.
- FIG. 3 provides an example flowchart as shown in FIG. 3, which includes the following steps:
- U302 Set the location of the j-th particle as the ESN storage layer parameter, and use L2+adaptive L1/2 regularization to constrain the least square calculation that appears in the training to obtain a higher sparsity output weight Wout.
- the detailed steps of calculating the output weight Wout with the ESN of the L2+adaptive L1/2 regularization constraint are shown in Figure 4, which will be described in detail below.
- U303 Calculate the fitness value corresponding to the j-th particle based on the output weight Wout corresponding to the j-th particle.
- the calculation formula is as follows:
- ⁇ 1 and ⁇ 2 are regularization coefficients, and W out is the output weight corresponding to the current particle j; Y represents the last part of the logical block address of the historical hot data used for network training, and X represents network-based training The state information of the storage layer updated before the logical block address where the historical thermal data is located, X*W out represents the prediction result corresponding to the latter section of the logical block address where the historical thermal data is located.
- D301 Determine whether all particles have completed the fitness value calculation, if not, add 1 to j, and return to step U302 to calculate the fitness value of the next particle. If all particles have completed the fitness value calculation, proceed to step U304.
- the individual best fitness value, individual best parameter, and global best fitness value and global best parameter of each particle are selected according to the fitness value of each particle. After all particles have calculated their fitness values, compare and judge, record the fitness value of each particle as the individual best fitness value fsbest, and the position of each particle as the individual best parameter sbest; record the smallest particle fitness value among all particles The fitness value is the global best fitness value fgbest, and its corresponding position is the global best parameter gbest. These obtained parameters will be used in subsequent iterations to optimize.
- sbest i represents the individual best parameter of the i-th particle
- mbest is the average value of the current individual best parameters of all particles, that is, the average value of each dimension parameter of all particles is used to update the particle position information.
- P j (t+1) and P j (t) respectively represent the position of the particle j after and before the update
- u j are random numbers between (0,1), where ⁇ is calculated as:
- the parameter ⁇ representing the step length of the particle movement is larger, and the particles can move to the optimal position faster; while the smaller ⁇ in the later stage of the iteration means that the parameter is in the optimal position.
- the nearby particles decrease the step size, and move closer to the best position more accurately each time.
- D302, D303, U308, U309 are: update the individual according to the newly calculated fitness value Best and global best. If the newly calculated fitness value is less than the individual best fitness value of the particle, the individual best fitness value of the particle is updated to the newly calculated fitness value, and the individual best fitness value of the particle is updated at the same time.
- the best parameter is the parameter of the current particle; if the newly calculated fitness value is less than the global best fitness value, the fitness value is updated to the global best fitness value, and the global best parameter is the parameter of the particle at the same time.
- D304 Judge whether all particles have been updated, if not, j+1 and return to U306, use the updated particle parameters to recalculate mbest, and update the position information of the next particle. If all particles have been updated, proceed to D305 .
- D305 Judge whether the number of iterations has reached the maximum number of iterations, if not, add 1 to iter and return to U305 for the next iteration. If the maximum number of iterations has been reached, the final global optimal parameters are derived for subsequent training to jointly optimize the echo state network. Predict the logical block address.
- y represents the obtained predicted logical block address
- the data on the predicted logical block address is the thermal data
- x is the logical block address where the input historical thermal data is located
- W out represents the output weight.
- y is the predicted access address. It is worth noting that both x and Wout can be multi-dimensional variables, and the obtained y is a one-dimensional variable.
- the data on the obtained logical block address will be classified as hot data for Garbage collection and wear leveling treatment.
- a set of storage layer parameters are determined, namely, the internal connection spectrum radius, storage layer scale, input layer scale factor and storage layer sparsity.
- the process of calculating output weights using the echo state network with L2+adaptive L1/2 regularization constraints in the present invention is as follows:
- U401 Obtain the input layer-storage layer weight matrix in the echo state network, the internal connection weight matrix of the storage layer, and use the previous segment of the logical block address where the historical thermal data is located as the input variable U, and the latter segment as the actual result Y.
- the Echo State Network is a low-complexity and fast-converging calculation scheme, which is suitable for temporal data classification and prediction tasks.
- the ESN network architecture includes three layers: input layer, storage layer and output layer.
- the weight of the input layer-storage layer is Win
- the internal connection weight of the storage layer is Wx
- the storage layer-output layer is Wout.
- the input layer-storage layer weight W in ⁇ R n ⁇ K is determined based on the storage layer parameters, and the storage layer internal connection weight W x ⁇ R n ⁇ n .
- the logical block address of historical thermal data selected in the embodiment of the present invention is the logical block address of historical thermal data recorded by the user. Other feasible embodiments
- the selected length can be other, and the present invention does not specifically limit it.
- the general idea is to use the previous address to predict the latter address, and then compare the predicted latter address with the actual address to adjust the network. Part of it is the original feature of the echo state network, which is not described in detail in the present invention.
- the state information X is composed of state node information X(t);
- U(t) represents the t-th data in the input variable U
- X(t) and X(t-1) represent the t-th and t-1 state node information respectively
- the number of nodes is determined by the input variable U Determined by the data length
- W in and W x respectively represent the input layer-storage layer weight matrix in the echo state network, and the internal connection weight matrix of the storage layer.
- Logsig( ⁇ ) represents the activation function, which can approximate any non- Linear function, and then the neural network can be applied to the nonlinear model.
- the activation function we directly multiply the input amount by the input layer scale coefficient and transform it into the corresponding range of the activation function. Since the input is sequentially calculated, t can be understood as the time.
- E represents the loss function
- ⁇ 1 and ⁇ 2 are both regularization coefficients.
- the present invention simplifies the loss function E, and then uses the coordinate descent algorithm to calculate the output weight;
- the method of solving matrix W′ out is to calculate each element separately, the value of the k-th element in the m-th row of W′ out is as follows:
- Y′ k (t) represents the t-th element in the k-th row of Y′
- X′ j (t) represents the t-th element in the j-th row of X′
- the output weight W out is calculated using the relationship between the matrix W'out and the output weight W out .
- This embodiment also includes adaptive optimization of the output weight obtained in step U403, and the optimization is U404:
- the converted loss function is:
- n is the number of storage layer nodes
- K is the number of output layer nodes.
- the present invention innovatively uses thermal data prediction instead of thermal data identification, which improves the accuracy of thermal data discrimination.
- Financial1 is a write-intensive tracking file.
- MSR is a common workload for large-scale enterprise servers. Distilled represents a typical use mode of personal computers.
- MillSSD is collected from industrial automatic optical inspection instruments and has Runcore RCS hardware configuration -V- T25SSD (512GB, SATA2), Intel X2 7400 and 2G DDR3. MillSSD is also a write-intensive tracking file because it has the effect of substantial image backup. The performance comparison result of this embodiment is shown in Figure 5.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
Description
Claims (10)
- 一种基于联合优化回声状态网络的热数据预测方法,其特征在于:包括如下步骤:S1:初始化量子粒子群算法所需参数以及每个粒子的位置信息;其中,粒子的位置信息包括粒子初始位置以及位置范围,每个粒子的位置以回声状态网络中存储层参数表示;S2:利用量子粒子群算法迭代寻优确定最佳存储层参数;其中,基于每个粒子的位置范围采用量子粒子群算法更新粒子位置,每次更新过程中利用L2+自适应L1/2正则化约束的回声状态网络计算输出权值并计算出全局最佳适应值,迭代结束时全局最佳适应值对应的粒子位置作为最佳存储层参数;S3:基于回声状态网络中最佳存储层参数采用L2+自适应L1/2正则化约束的回声状态网络计算最终输出权值;S4:利用最终输出权值和输入的历史热数据所在逻辑区块地址预测出热数据,预测公式如下:y=x*W out式中,y表示获取的预测逻辑区块地址,预测逻辑区块地址上的数据为热数据,x为输入的历史热数据所在逻辑区块地址,W out表示输出权值,所述历史热数据所在的逻辑区块地址用于步骤S2和步骤S3中回声状态网络训练过程。
- 根据权利要求1所述的方法,其特征在于:步骤S2中迭代寻优确定最佳存储层参数的执行过程如下:S21:将每个粒子的位置依次作为回声状态网络中存储层参数,并分别采用L2+自适应L1/2正则化约束的回声状态网络计算出每个粒子对应输出权值;其中,将每个粒子的当前位置依次作为回声状态网络中的存储层参数并计算输出权值;S22:利用每个粒子对应的输出权值计算出每个粒子的适应值;S23:基于最小适应值原则根据每个粒子的适应值选择出每个粒子的个 体最佳适应值、个体最佳参数,以及全局最佳适应值以及全局最佳参数;其中,选为全局最佳适应值的粒子位置为全局最佳参数;S24:在粒子的位置范围内更新每个粒子的位置,并基于每个粒子的更新位置重新计算每个粒子的适应值,并基于最小适应值原则更新每个粒子的个体最佳适应值、个体最佳参数,以及全局最佳适应值以及全局最佳参数;S25:判断迭代次数是否达到最大迭代次数,若未达到,返回步骤S24进行下一次迭代计算;否则,将当前的全局最佳参数作为最佳存储层参数。
- 根据权利要求1所述的方法,其特征在于:采用L2+自适应L1/2正则化约束的回声状态网络计算输出权值的过程如下:U401:获取回声状态网络中输入层-存储层权值矩阵,存储层内部连接 权值矩阵,以及利用历史热数据所在逻辑区块地址的前段作为输入变量U,后段作为实际结果Y;其中,输入层-存储层权值矩阵以及存储层内部连接权值矩阵与回声状态网络中存储层参数相关;U402:基于输入变量U更新存储层的状态信息X,状态信息X由状态节点信息X(t)构成;X(t)=log sig(U(t)W in+X(t-1)W x)式中,U(t)表示输入变量U中第t个数据,X(t)、X(t-1)分别表示第t个、第t-1个状态节点信息,t的最大值T由输入变量U的数据长度决定,Win、Wx分别表示回声状态网络中输入层-存储层权值矩阵,存储层内部连接权值矩阵,logsig(·)表示激活函数;U403:基于L2+自适应L1/2正则化约束下的损失函数获取损失函数最小值下的输出权值;式中,E表示损失函数,λ 1、λ 2均为正则化系数。
- 根据权利要求1所述的方法,其特征在于:回声状态网络中存储层参数包括内部连接谱半径,存储层规模,输入层比例系数和存储层稀疏度四个关键参数。
- 根据权利要求1所述的方法,其特征在于:步骤S1中初始化量子粒子群算法所需参数包括粒子群规模N,最大迭代次数iter max、以及惯性因子ω max和ω min。
- 根据权利要求1所述的方法,其特征在于:更新粒子位置时,若粒子移动距离超过了粒子对应的位置范围,则将粒子位置参数设置为超过位置范围对应的边界值。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910566123.4 | 2019-06-27 | ||
CN201910566123.4A CN110554838B (zh) | 2019-06-27 | 2019-06-27 | 一种基于联合优化回声状态网络的热数据预测方法 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020259543A1 true WO2020259543A1 (zh) | 2020-12-30 |
Family
ID=68735438
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/097950 WO2020259543A1 (zh) | 2019-06-27 | 2020-06-24 | 一种基于联合优化回声状态网络的热数据预测方法 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110554838B (zh) |
WO (1) | WO2020259543A1 (zh) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110554838B (zh) * | 2019-06-27 | 2020-08-14 | 中南大学 | 一种基于联合优化回声状态网络的热数据预测方法 |
CN112448697A (zh) * | 2020-10-30 | 2021-03-05 | 合肥工业大学 | 一种基于量子粒子群算法的有源滤波器优化方法及系统 |
CN112731019B (zh) * | 2020-12-21 | 2022-10-14 | 合肥工业大学 | 一种anpc三电平逆变器故障诊断方法 |
CN115841067A (zh) * | 2022-10-12 | 2023-03-24 | 大连理工大学 | 针对航空发动机故障预警的量子回声状态网络模型构建方法 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103020434A (zh) * | 2012-11-30 | 2013-04-03 | 南京航空航天大学 | 一种基于粒子群优化的最小二乘支持向量机组合预测方法 |
US20170046079A1 (en) * | 2015-08-11 | 2017-02-16 | International Business Machines Corporation | Read distribution in a three-dimensional stacked memory based on thermal profiles |
CN109901800A (zh) * | 2019-03-14 | 2019-06-18 | 重庆大学 | 一种混合内存系统及其操作方法 |
CN110554838A (zh) * | 2019-06-27 | 2019-12-10 | 中南大学 | 一种基于联合优化回声状态网络的热数据预测方法 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10307109B2 (en) * | 2011-04-20 | 2019-06-04 | Novo Nordisk A/S | Glucose predictor based on regularization networks with adaptively chosen kernels and regularization parameters |
US10014026B1 (en) * | 2017-06-20 | 2018-07-03 | Seagate Technology Llc | Head delay calibration and tracking in MSMR systems |
CN109726858A (zh) * | 2018-12-21 | 2019-05-07 | 新奥数能科技有限公司 | 基于动态时间规整的热负荷预测方法及装置 |
CN109656485A (zh) * | 2018-12-24 | 2019-04-19 | 合肥兆芯电子有限公司 | 区分热数据和冷数据的方法 |
-
2019
- 2019-06-27 CN CN201910566123.4A patent/CN110554838B/zh active Active
-
2020
- 2020-06-24 WO PCT/CN2020/097950 patent/WO2020259543A1/zh active Application Filing
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103020434A (zh) * | 2012-11-30 | 2013-04-03 | 南京航空航天大学 | 一种基于粒子群优化的最小二乘支持向量机组合预测方法 |
US20170046079A1 (en) * | 2015-08-11 | 2017-02-16 | International Business Machines Corporation | Read distribution in a three-dimensional stacked memory based on thermal profiles |
CN109901800A (zh) * | 2019-03-14 | 2019-06-18 | 重庆大学 | 一种混合内存系统及其操作方法 |
CN110554838A (zh) * | 2019-06-27 | 2019-12-10 | 中南大学 | 一种基于联合优化回声状态网络的热数据预测方法 |
Non-Patent Citations (1)
Title |
---|
ZHU, HONGQIU ET AL.: "Hybrid-QPSO-based Parameters Optimization of LS-SVM and its Application", JOURNAL OF CENTRAL SOUTH UNIVERSITY (SCIENCE AND TECHNOLOGY), vol. 42, no. 4, 30 April 2011 (2011-04-30), DOI: 20200901141252A * |
Also Published As
Publication number | Publication date |
---|---|
CN110554838A (zh) | 2019-12-10 |
CN110554838B (zh) | 2020-08-14 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020259543A1 (zh) | 一种基于联合优化回声状态网络的热数据预测方法 | |
US10896126B2 (en) | Storage device, method and non-volatile memory device performing garbage collection using estimated number of valid pages | |
Yang et al. | Reducing garbage collection overhead in {SSD} based on workload prediction | |
WO2021143883A1 (zh) | 神经网络的自适应搜索方法及装置 | |
CN110968272B (zh) | 基于时间序列预测的海量小文件存储性能优化方法及系统 | |
CN107025071A (zh) | 非易失性存储器装置及其垃圾收集方法 | |
US11734103B2 (en) | Behavior-driven die management on solid-state drives | |
CN103150245A (zh) | 确定数据实体的访问特性的方法和存储控制器 | |
CN106569962A (zh) | 一种基于增强时间局部性的热点数据识别方法 | |
Wang et al. | A novel buffer management scheme based on particle swarm optimization for SSD | |
EP3651024B1 (en) | Method of operating storage device, storage device performing the same and storage system including the same | |
Gupta et al. | Relevance feedback based online learning model for resource bottleneck prediction in cloud servers | |
CN108829343B (zh) | 一种基于人工智能的缓存优化方法 | |
Lima et al. | Evaluation of recurrent neural networks for hard disk drives failure prediction | |
Poczeta et al. | Analysis of fuzzy cognitive maps with multi-step learning algorithms in valuation of owner-occupied homes | |
WO2023202484A1 (zh) | 神经网络模型的修复方法和相关设备 | |
CN115827651B (zh) | 一种低能耗的机载嵌入式数据库内存事务管理方法及系统 | |
Cao et al. | A loan default discrimination model using cost-sensitive support vector machine improved by PSO | |
Liu et al. | Semi‐supervised breast histopathological image classification with self‐training based on non‐linear distance metric | |
Ha et al. | Dynamic hot data identification using a stack distance approximation | |
Ushasukhanya et al. | Development and optimization of deep convolutional neural network using taguchi method for real-time electricity conservation system | |
Reddy et al. | Analysis of Stock Market Value Prediction using Simple Novel Long Short Term Memory Algorithm in Comparison with Back Propagation Algorithm for Increased Accuracy Rate | |
Lee et al. | File Access Characteristics of Deep Learning Workloads and Cache-Friendly Data Management | |
Chen et al. | Pruning deep feature networks using channel importance propagation | |
Kargar et al. | E2-NVM: A Memory-Aware Write Scheme to Improve Energy Efficiency and Write Endurance of NVMs using Variational Autoencoders. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20833324 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20833324 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205 DATED 25/05/2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20833324 Country of ref document: EP Kind code of ref document: A1 |