WO2022110861A1 - Method and apparatus for data set caching in network training, device, and storage medium - Google Patents

Method and apparatus for data set caching in network training, device, and storage medium Download PDF

Info

Publication number
WO2022110861A1
WO2022110861A1 PCT/CN2021/109237 CN2021109237W WO2022110861A1 WO 2022110861 A1 WO2022110861 A1 WO 2022110861A1 CN 2021109237 W CN2021109237 W CN 2021109237W WO 2022110861 A1 WO2022110861 A1 WO 2022110861A1
Authority
WO
WIPO (PCT)
Prior art keywords
data set
trained
node
training
cache
Prior art date
Application number
PCT/CN2021/109237
Other languages
French (fr)
Chinese (zh)
Inventor
赵仁明
陈培
Original Assignee
苏州浪潮智能科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 苏州浪潮智能科技有限公司 filed Critical 苏州浪潮智能科技有限公司
Publication of WO2022110861A1 publication Critical patent/WO2022110861A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/082Learning methods modifying the architecture, e.g. adding, deleting or silencing nodes or connections
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Definitions

  • the present application relates to the field of deep learning, and in particular, to a data set caching method, apparatus, device and storage medium for network training.
  • Deep learning has been widely used at present. Deep learning refers to the feature training of neural networks through a large amount of data to generate network models with the ability to identify corresponding data.
  • multiple training nodes are often used in a cluster mode that includes multiple training nodes and data set storage nodes to jointly utilize data.
  • the neural network is trained on the dataset in the set storage node.
  • the data sets cached in different training nodes may be different, and there are often situations where multiple training nodes need to be used for neural network training based on the same data set, and the overall reliability of neural network training is also in the current field. focus of attention.
  • the purpose of this application is to provide a data set caching method, device, device and storage medium for network training, so as to ensure the reliability of training nodes to cache the data set to be trained, thereby ensuring the overall reliability of network training.
  • the present application provides a data set caching method for network training, including:
  • the destination node is used to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
  • the method before using the destination node to cache the to-be-trained data set passed in by the source node, the method further includes:
  • use the destination node to cache the data set to be trained from the source node including:
  • the destination node is used to cache the data set to be trained passed in by the source node, including:
  • the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set includes:
  • the disk performance overhead is obtained by statistics.
  • the hardware performance parameters include the disk rotation speed, the disk average patrol time and the disk maximum transfer rate
  • the data attribute parameters of the data set to be trained include an average file size; the average file size is calculated based on the total data volume of the data set to be trained and the total number of files.
  • the method before using the destination node to cache the to-be-trained data set passed in by the source node, the method further includes:
  • the current performance parameter includes I/O queue length
  • the disk performance overhead includes IOPS overhead
  • select the destination node that meets the disk performance overhead in the network training cluster including:
  • a data set cache device for network training including:
  • the cost statistics module is used to count the disk performance overhead required by the training nodes in the network training cluster that do not cache the data set to be trained in the process of caching the data set to be trained;
  • the parameter monitoring module is used to monitor the current performance parameters of the training nodes that do not cache the data set to be trained
  • the node selection module is used to select the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters;
  • the node cache module is used for using the destination node to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
  • this application also provides a data set caching device for network training, including:
  • the processor is configured to implement the steps of the above-mentioned data set caching method for network training when executing the computer program.
  • the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above-mentioned network training data set caching method are implemented.
  • the data set caching method for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data sets to be trained.
  • the current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node.
  • the node performs network training on the training dataset.
  • the method obtains the disk performance overhead required for caching the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
  • the present application also provides a data set cache device, equipment and storage medium for network training, the beneficial effects are the same as above.
  • FIG. 1 is a flowchart of a method for caching data sets for network training disclosed in an embodiment of the present application
  • FIG. 2 is a schematic structural diagram of a data set cache device for network training disclosed in an embodiment of the present application.
  • multiple training nodes are often used in a cluster mode that includes multiple training nodes and data set storage nodes to jointly utilize data.
  • the neural network is trained on the dataset in the set storage node.
  • the data sets cached in different training nodes may be different, and there are often situations where multiple training nodes need to be used for neural network training based on the same data set, and the overall reliability of neural network training is also in the current field. focus of attention.
  • the core of this application is to provide a data set caching method for network training, so as to ensure the reliability of the training node's caching of the data set to be trained, thereby ensuring the overall reliability of network training.
  • an embodiment of the present application discloses a data set caching method for network training, including:
  • Step S10 Count the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set.
  • the execution body of this embodiment may be any node with computing capability in the network training cluster.
  • This step firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, and the disk performance overhead required in the process of caching the corresponding data set to be trained, where the data set to be trained refers to the data set used in the network training process.
  • a collection of sample data which counts the disk performance overhead corresponding to the training nodes that do not cache the data set to be trained. In essence, it estimates the overhead of communication and operating resources occupied by the training nodes in the process of caching the corresponding data set to be trained.
  • different The disk performance overhead when a training node caches the same dataset to be trained can vary depending on the hardware parameters of the corresponding training nodes.
  • Step S11 Monitor the current performance parameters of the training nodes that do not cache the data set to be trained.
  • this step further monitors the current performance parameters of the training node that does not cache the data set to be trained, and the purpose is to perform the following steps according to the data set of the training node.
  • the current performance parameter selects the destination node that can carry the disk performance overhead corresponding to the data set to be trained.
  • Step S12 Based on the current performance parameters, a destination node that satisfies the disk performance overhead is selected in the network training cluster.
  • this step further selects a destination node that satisfies the disk performance overhead in the network cluster based on the current performance parameters. training dataset.
  • Step S13 Use the destination node to cache the data set to be trained transmitted by the source node, so as to perform network training based on the data set to be trained by the destination node.
  • this step further uses the destination node to cache the data set to be trained from the source node, so as to perform network training on the training data set based on the destination node , the purpose is to ensure that when the data set to be trained is cached to the destination node, the destination node can reliably store the data set to be trained.
  • the data set caching method for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data sets to be trained.
  • the current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node.
  • the node performs network training on the training dataset.
  • the method obtains the disk performance overhead required for caching the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
  • the method before using the destination node to cache the data set to be trained transmitted by the source node, the method further includes:
  • use the destination node to cache the data set to be trained from the source node including:
  • the destination node when it is determined that there is a source training node that caches the data set to be trained in the network training cluster, that is, the training node that has stored the data set to be trained in the network training cluster, in this case, use Before the destination node caches the data set to be trained from the source node, it first selects the target source training node with the largest idle network bandwidth between the source training node and the destination node, and then uses the destination node to cache the data set passed in by the source node. When the data set is to be trained, the target node caches the data set to be trained from the source training node.
  • the bandwidth between the target source training node and the destination node is relatively high, the distance between the target source training node and the destination node is relatively high.
  • the network transmission efficiency is relatively high, and this embodiment further improves the overall efficiency of sharing the data set to be trained among the training nodes.
  • the destination node when there is no source training node that caches the data set to be trained in the network training cluster, the destination node is used to cache the data set to be trained passed in by the source node, including:
  • the network training cluster includes a data set storage node, and the data set storage node is used to store the to-be-trained data set.
  • the destination node When it is determined that there is no source training node that caches the to-be-trained data set in the network training cluster, the destination node is specifically used to cache the data set to be trained passed in by the data set storage node in the network training cluster, so as to further ensure that the destination node has The reliability of the acquisition of the dataset to be trained.
  • statistics of the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set include:
  • the disk performance overhead is obtained by statistics.
  • the hardware performance parameters of the training node and the data attribute parameters of the data set to be trained are counted to obtain the performance disk overhead.
  • Data feature parameters Since the hardware performance parameters can accurately represent the efficiency of the training node for data caching, and the data attribute parameters can accurately represent the data magnitude and data type of the data set to be trained, this embodiment is based on the hardware performance parameters of the training node and the data to be trained.
  • the data attribute parameters of the training data set are calculated to obtain the disk performance overhead, which can further improve the accuracy of the calculated disk performance overhead.
  • the hardware performance parameters include the disk rotation speed, the disk average patrol time, and the disk maximum transfer rate
  • the data attribute parameters of the data set to be trained include an average file size; the average file size is calculated based on the total data volume of the data set to be trained and the total number of files.
  • the hardware performance parameters include the rotational speed of the disk, the average track time of the disk, and the maximum transfer rate of the disk.
  • the disk rotation speed refers to the maximum number of revolutions that the disk platter can complete in one minute. It is one of the key factors determining the internal transfer rate of the disk and directly affects the speed of the disk to a large extent;
  • Track time refers to the average time taken by the magnetic head to move from the beginning to the track where the data is located after the disk receives the system command. It reflects the ability of the disk to read data to a certain extent and affects the internal data transmission of the disk.
  • the maximum transfer rate of the disk refers to the speed at which data is transferred from the head of the disk to the cache, which affects the overall efficiency of data caching in the disk.
  • the data attribute parameter of the data set to be trained includes the average file size, and the average file size is calculated based on the total amount of data in the data set to be trained and the total amount of files, that is, the data of the data set to be trained is calculated.
  • the total that is, the total size of the data and the total number of files, that is, the number of files, is obtained by quotient operation, and the average file size refers to the average disk storage space occupied by each file in the data set to be trained.
  • the hardware performance parameters of the training nodes and the data attribute parameters of the data set to be trained are refined, thereby further improving the accuracy of the statistical obtained disk performance overhead.
  • the method before using the destination node to cache the data set to be trained transmitted by the source node, the method further includes:
  • the destination node before using the destination node to cache the data set to be trained from the source node, it is further determined whether there is free space in the cache queue of the destination node, that is, it is determined whether the cache queue can work normally.
  • the data set to be trained is stored, and when there is free space in the cache queue of the destination node, the step of using the destination node to cache the data set to be trained from the source node is further performed.
  • the current performance parameter includes the I/O queue length
  • the disk performance overhead includes the IOPS overhead
  • select the destination node that meets the disk performance overhead in the network training cluster including:
  • the current performance parameters include I/O (abbreviation for Input/Output, that is, input and output ports) queue length
  • the disk performance overhead includes IOPS overhead
  • IOPS Input/Output, Operations Per Second
  • HDDs magnetic disks
  • SSDs solid-state disks
  • SANs storage area networks
  • the destination node When selecting the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters, specifically select the destination node whose I/O queue length is less than the IOPS overhead in the network cluster, in order to ensure that the data set to be trained can be reliably cached to The selected destination node further ensures the overall reliability of the data set caching process.
  • the present application also provides a scenario embodiment under a specific application scenario for further description.
  • disk rotation speed rpm
  • disk average seek time avgSeekTime
  • maxTransRate disk maximum transfer rate
  • the wrqm/s value, rrqm/s and avgqu-sz value of the data set cache disk can be obtained by monitoring.
  • wrqm/s represents the current number of writes per second of the disk after merging data (merge)
  • rrqm/s represents the current number of reads per second after merging data (merge) of the disk
  • avgqu-sz represents the I/ O queue length.
  • an embodiment of the present application provides a data set cache device for network training, including:
  • the overhead statistics module 10 is used to count the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set;
  • the parameter monitoring module 11 is used to monitor the current performance parameters of the training nodes that do not cache the data set to be trained;
  • the node selection module 12 is used to select the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameter;
  • the node caching module 13 is configured to use the destination node to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
  • the data set caching device for network training provided by the present application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data set to be trained.
  • the current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node.
  • the node performs network training on the training dataset.
  • the device Since the device obtains the disk performance overhead required for caching the data set to be trained according to the estimation, selects the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained imported from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
  • this application also provides a data set caching device for network training, including:
  • the processor is configured to implement the steps of the above-mentioned data set caching method for network training when executing the computer program.
  • the data set caching device for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data set to be trained.
  • the current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node.
  • the node performs network training on the training dataset.
  • the device Since the device obtains the disk performance overhead required to cache the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
  • the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above-mentioned network training data set caching method are implemented.
  • the computer-readable storage medium provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data to be trained.
  • the current performance parameters of the training nodes in the set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node.
  • the training dataset performs network training.
  • the destination node whose current performance parameters meet the disk performance overhead is selected from the training nodes of the network training cluster to the source node to be trained.
  • the data set is cached, which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of network training.

Abstract

A method and apparatus for data set caching in network training, a device, and a storage medium. The method comprises the steps of: collecting statistics about disk performance overhead required by training nodes, which have not cached a data set to be trained, in a network training cluster during a process of caching a data set to be trained; monitoring current performance parameters of the training nodes that have not cached a data set to be trained; selecting, on the basis of the current performance parameters, from the network training cluster a target node meeting the disk performance overhead requirement; and using the target node to cache a data set to be trained transmitted from a source node, so as to perform network training on said data set on the basis of the target node. The present method can ensure the reliability of the training node caching the data set to be trained, thereby ensuring the overall reliability of the network training. The apparatus for data set caching, the device, and the storage medium also have the beneficial effects as described above.

Description

一种网络训练的数据集缓存方法、装置、设备及存储介质A data set caching method, device, device and storage medium for network training
本申请要求于2020年11月27日提交中国专利局、申请号为202011357904.1、发明名称为“一种网络训练的数据集缓存方法、装置、设备及存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of the Chinese patent application filed on November 27, 2020, with the application number of 202011357904.1 and the invention titled "A method, device, equipment and storage medium for data set caching for network training", which The entire contents of this application are incorporated by reference.
技术领域technical field
本申请涉及深度学习领域,特别是涉及一种网络训练的数据集缓存方法、装置、设备及存储介质。The present application relates to the field of deep learning, and in particular, to a data set caching method, apparatus, device and storage medium for network training.
背景技术Background technique
深度学习在当前得到了广泛的运用,深度学习指的是通过大量数据对神经网络进行特征训练,产生具有识别相应数据能力的网络模型。Deep learning has been widely used at present. Deep learning refers to the feature training of neural networks through a large amount of data to generate network models with the ability to identify corresponding data.
由于神经网络训练的过程中所使用的样本数据集的多少,直接影响深度学习的效果,因此当前往往以包含有多个训练节点以及数据集存储节点的集群方式,采用多个训练节点共同利用数据集存储节点中的数据集对神经网络进行训练。在训练过程中,不同训练节点中缓存的数据集可能存在差异,并且当前往往存在需要使用多个训练节点基于相同数据集进行神经网络训练的情况,而神经网络训练的整体可靠性也是当前本领域所关注的重点。Since the number of sample data sets used in the process of neural network training directly affects the effect of deep learning, currently, multiple training nodes are often used in a cluster mode that includes multiple training nodes and data set storage nodes to jointly utilize data. The neural network is trained on the dataset in the set storage node. During the training process, the data sets cached in different training nodes may be different, and there are often situations where multiple training nodes need to be used for neural network training based on the same data set, and the overall reliability of neural network training is also in the current field. focus of attention.
由此可见,提供一种网络训练的数据集缓存方法,以确保训练节点缓存待训练数据集的可靠性,进而确保网络训练的整体可靠性,是本领域技术人员需要解决的问题。It can be seen that it is a problem to be solved by those skilled in the art to provide a data set caching method for network training to ensure the reliability of training nodes to cache the data set to be trained, thereby ensuring the overall reliability of network training.
发明内容SUMMARY OF THE INVENTION
本申请的目的是提供一种网络训练的数据集缓存方法、装置、设备及存储介质,以确保训练节点缓存待训练数据集的可靠性,进而确保网络训练的整体可靠性。The purpose of this application is to provide a data set caching method, device, device and storage medium for network training, so as to ensure the reliability of training nodes to cache the data set to be trained, thereby ensuring the overall reliability of network training.
为解决上述技术问题,本申请提供一种网络训练的数据集缓存方法,包括:In order to solve the above technical problems, the present application provides a data set caching method for network training, including:
统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销;Calculate the disk performance overhead required by the training nodes in the network training cluster that do not cache the data set to be trained in the process of caching the data set to be trained;
监控未缓存待训练数据集的训练节点的当前性能参数;Monitor the current performance parameters of training nodes that do not cache the data set to be trained;
基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点;Select the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters;
利用目的节点缓存由源节点传入的待训练数据集,以基于目的节点对待训练数据集执行网络训练。The destination node is used to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
优选地,在利用目的节点缓存由源节点传入的待训练数据集之前,方法还包括:Preferably, before using the destination node to cache the to-be-trained data set passed in by the source node, the method further includes:
判断网络训练集群中是否存在缓存有待训练数据集的源训练节点;Determine whether there is a source training node that caches the data set to be trained in the network training cluster;
若网络训练集群中存在缓存有待训练数据集的源训练节点,则在源训练节点中选取与目的节点之间的空闲网络带宽最大的目标源训练节点;If there is a source training node that caches the data set to be trained in the network training cluster, select the target source training node with the largest idle network bandwidth between the source training node and the destination node;
相应的,利用目的节点缓存由源节点传入的待训练数据集,包括:Correspondingly, use the destination node to cache the data set to be trained from the source node, including:
利用目的节点缓存由目标源训练节点传入的待训练数据集。Use the destination node to cache the data set to be trained passed in by the target source training node.
优选地,当网络训练集群中不存在缓存有待训练数据集的源训练节点时,利用目的节点缓存由源节点传入的待训练数据集,包括:Preferably, when there is no source training node that caches the data set to be trained in the network training cluster, the destination node is used to cache the data set to be trained passed in by the source node, including:
利用目的节点缓存由网络训练集群中的数据集存储节点传入的待训练数据集。Use the destination node to cache the data set to be trained that is passed in by the data set storage node in the network training cluster.
优选地,统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销,包括:Preferably, the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set includes:
基于未缓存待训练数据集的训练节点的硬件性能参数以及待训练数据集的数据属性参数统计得到磁盘性能开销。Based on the hardware performance parameters of the training nodes that do not cache the data set to be trained and the data attribute parameters of the data set to be trained, the disk performance overhead is obtained by statistics.
优选地,硬件性能参数包括磁盘转速、磁盘平均巡道时间以及磁盘最大传输速率;Preferably, the hardware performance parameters include the disk rotation speed, the disk average patrol time and the disk maximum transfer rate;
待训练数据集的数据属性参数包括平均文件大小;平均文件大小基于待训练数据集的数据总量以及文件总量运算得到。The data attribute parameters of the data set to be trained include an average file size; the average file size is calculated based on the total data volume of the data set to be trained and the total number of files.
优选地,在利用目的节点缓存由源节点传入的待训练数据集之前,方法还包括:Preferably, before using the destination node to cache the to-be-trained data set passed in by the source node, the method further includes:
判断目的节点的缓存队列是否存在空闲空间;Determine whether there is free space in the cache queue of the destination node;
若是,则执行利用目的节点缓存由源节点传入的待训练数据集的步骤;If so, execute the step of using the destination node to cache the data set to be trained passed in by the source node;
否则,将缓存队列中执行次数最少的目标待训练数据集删除,并执行利用目的节点缓存由源节点传入的待训练数据集的步骤。Otherwise, delete the target data set to be trained with the least number of executions in the cache queue, and perform the step of using the destination node to cache the data set to be trained from the source node.
优选地,当前性能参数包括I/O队列长度,磁盘性能开销包括IOPS开销;Preferably, the current performance parameter includes I/O queue length, and the disk performance overhead includes IOPS overhead;
基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点,包括:Based on the current performance parameters, select the destination node that meets the disk performance overhead in the network training cluster, including:
在网络集群中选取I/O队列长度小于IOPS开销的目的节点。Select the destination node whose I/O queue length is less than the IOPS cost in the network cluster.
此外,本申请还提供一种网络训练的数据集缓存装置,包括:In addition, the present application also provides a data set cache device for network training, including:
开销统计模块,用于统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销;The cost statistics module is used to count the disk performance overhead required by the training nodes in the network training cluster that do not cache the data set to be trained in the process of caching the data set to be trained;
参数监控模块,用于监控未缓存待训练数据集的训练节点的当前性能参数;The parameter monitoring module is used to monitor the current performance parameters of the training nodes that do not cache the data set to be trained;
节点选取模块,用于基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点;The node selection module is used to select the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters;
节点缓存模块,用于利用目的节点缓存由源节点传入的待训练数据集,以基于目的节点对待训练数据集执行网络训练。The node cache module is used for using the destination node to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
此外,本申请还提供一种网络训练的数据集缓存设备,包括:In addition, this application also provides a data set caching device for network training, including:
存储器,用于存储计算机程序;memory for storing computer programs;
处理器,用于执行计算机程序时实现如上述的网络训练的数据集缓存方法的步骤。The processor is configured to implement the steps of the above-mentioned data set caching method for network training when executing the computer program.
此外,本申请还提供一种计算机可读存储介质,计算机可读存储介质上存储有计算机程序,计算机程序被处理器执行时实现如上述的网络训练的数据集缓存方法的步骤。In addition, the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above-mentioned network training data set caching method are implemented.
本申请所提供的网络训练的数据集缓存方法,首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存待训练数据集的过程中所需的磁盘性能开销,进而监控未缓存有待训练数据集的训练节点的当前性能参数,并基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点,进而利用目的节点缓存由源节点传入的待训练数据集,以此基于目 的节点对待训练数据集执行网络训练。由于本方法根据预估得到缓存待训练数据集所需的磁盘性能开销,在网络训练集群的训练节点中选取当前性能参数满足磁盘性能开销的目的节点对源节点传入的待训练数据集进行缓存,相对确保了训练节点缓存待训练数据集的可靠性,进而确保了网络训练的整体可靠性。此外,本申请还提供一种网络训练的数据集缓存装置、设备及存储介质,有益效果同上所述。The data set caching method for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data sets to be trained. The current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node. The node performs network training on the training dataset. Since the method obtains the disk performance overhead required for caching the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training. In addition, the present application also provides a data set cache device, equipment and storage medium for network training, the beneficial effects are the same as above.
附图说明Description of drawings
为了更清楚地说明本申请实施例,下面将对实施例中所需要使用的附图做简单的介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to describe the embodiments of the present application more clearly, the following will briefly introduce the drawings that are used in the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application, which are not relevant to ordinary skills in the art. As far as personnel are concerned, other drawings can also be obtained from these drawings on the premise of no creative work.
图1为本申请实施例公开的一种网络训练的数据集缓存方法的流程图;FIG. 1 is a flowchart of a method for caching data sets for network training disclosed in an embodiment of the present application;
图2为本申请实施例公开的一种网络训练的数据集缓存装置的结构示意图。FIG. 2 is a schematic structural diagram of a data set cache device for network training disclosed in an embodiment of the present application.
具体实施方式Detailed ways
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下,所获得的所有其他实施例,都属于本申请保护范围。The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application. Obviously, the described embodiments are only a part of the embodiments of the present application, rather than all the embodiments. All other embodiments obtained by those of ordinary skill in the art based on the embodiments in the present application without creative work fall within the protection scope of the present application.
由于神经网络训练的过程中所使用的样本数据集的多少,直接影响深度学习的效果,因此当前往往以包含有多个训练节点以及数据集存储节点的集群方式,采用多个训练节点共同利用数据集存储节点中的数据集对神经网络进行训练。在训练过程中,不同训练节点中缓存的数据集可能存在差异,并且当前往往存在需要使用多个训练节点基于相同数据集进行神经网络训练的情况,而神经网络训练的整体可靠性也是当前本领域所关注的 重点。Since the number of sample data sets used in the process of neural network training directly affects the effect of deep learning, currently, multiple training nodes are often used in a cluster mode that includes multiple training nodes and data set storage nodes to jointly utilize data. The neural network is trained on the dataset in the set storage node. During the training process, the data sets cached in different training nodes may be different, and there are often situations where multiple training nodes need to be used for neural network training based on the same data set, and the overall reliability of neural network training is also in the current field. focus of attention.
为此,本申请的核心是提供一种网络训练的数据集缓存方法,以确保训练节点缓存待训练数据集的可靠性,进而确保网络训练的整体可靠性。To this end, the core of this application is to provide a data set caching method for network training, so as to ensure the reliability of the training node's caching of the data set to be trained, thereby ensuring the overall reliability of network training.
为了使本技术领域的人员更好地理解本申请方案,下面结合附图和具体实施方式对本申请作进一步的详细说明。In order to make those skilled in the art better understand the solution of the present application, the present application will be further described in detail below with reference to the accompanying drawings and specific embodiments.
请参见图1所示,本申请实施例公开了一种网络训练的数据集缓存方法,包括:Referring to FIG. 1, an embodiment of the present application discloses a data set caching method for network training, including:
步骤S10:统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销。Step S10: Count the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set.
需要说明的是,本实施例的执行主体可以为网络训练集群中的具有运算能力的任意节点。本步骤首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存相应待训练数据集过程中所需要的磁盘性能开销,其中,待训练数据集指的是网络训练过程中所使用的样本数据的集合,统计未缓存待训练数据集的训练节点对应的磁盘性能开销,本质上是预估训练节点在缓存相应待训练数据集过程中被占用的通信及运行资源的开销,另外,不同训练节点缓存同一待训练数据集时的磁盘性能开销可以根据相应训练节点的硬件参数差异而有所不同。It should be noted that the execution body of this embodiment may be any node with computing capability in the network training cluster. This step firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, and the disk performance overhead required in the process of caching the corresponding data set to be trained, where the data set to be trained refers to the data set used in the network training process. A collection of sample data, which counts the disk performance overhead corresponding to the training nodes that do not cache the data set to be trained. In essence, it estimates the overhead of communication and operating resources occupied by the training nodes in the process of caching the corresponding data set to be trained. In addition, different The disk performance overhead when a training node caches the same dataset to be trained can vary depending on the hardware parameters of the corresponding training nodes.
步骤S11:监控未缓存待训练数据集的训练节点的当前性能参数。Step S11: Monitor the current performance parameters of the training nodes that do not cache the data set to be trained.
在统计得到训练节点在缓存待训练数据集的过程中所需的磁盘性能开销之后,本步骤进一步监控未缓存有待训练数据集的训练节点的当前性能参数,目的是在后续步骤中根据训练节点的当前性能参数选取能够承载与待训练数据集对应的磁盘性能开销的目的节点。After the disk performance overhead required by the training node in the process of caching the data set to be trained is obtained by statistics, this step further monitors the current performance parameters of the training node that does not cache the data set to be trained, and the purpose is to perform the following steps according to the data set of the training node. The current performance parameter selects the destination node that can carry the disk performance overhead corresponding to the data set to be trained.
步骤S12:基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点。Step S12: Based on the current performance parameters, a destination node that satisfies the disk performance overhead is selected in the network training cluster.
在监控未缓存待训练数据集的训练节点的当前性能参数之后,本步骤进一步基于当前性能参数在网络集群中选取满足磁盘性能开销的目的节点,目的是在后续步骤中,通过目的节点缓存该待训练数据集。After monitoring the current performance parameters of the training nodes that do not cache the data set to be trained, this step further selects a destination node that satisfies the disk performance overhead in the network cluster based on the current performance parameters. training dataset.
步骤S13:利用目的节点缓存由源节点传入的待训练数据集,以基于 目的节点对待训练数据集执行网络训练。Step S13: Use the destination node to cache the data set to be trained transmitted by the source node, so as to perform network training based on the data set to be trained by the destination node.
在基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点之后,本步骤进一步利用目的节点缓存由源节点传入的待训练数据集,以此基于目的节点对待训练数据集执行网络训练,目的是确保将待训练数据集缓存至目的节点时,目的节点能够对待训练数据集进行可靠存储。After selecting the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters, this step further uses the destination node to cache the data set to be trained from the source node, so as to perform network training on the training data set based on the destination node , the purpose is to ensure that when the data set to be trained is cached to the destination node, the destination node can reliably store the data set to be trained.
本申请所提供的网络训练的数据集缓存方法,首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存待训练数据集的过程中所需的磁盘性能开销,进而监控未缓存有待训练数据集的训练节点的当前性能参数,并基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点,进而利用目的节点缓存由源节点传入的待训练数据集,以此基于目的节点对待训练数据集执行网络训练。由于本方法根据预估得到缓存待训练数据集所需的磁盘性能开销,在网络训练集群的训练节点中选取当前性能参数满足磁盘性能开销的目的节点对源节点传入的待训练数据集进行缓存,相对确保了训练节点缓存待训练数据集的可靠性,进而确保了网络训练的整体可靠性。The data set caching method for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data sets to be trained. The current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node. The node performs network training on the training dataset. Since the method obtains the disk performance overhead required for caching the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
在上述实施例的基础上,作为一种优选的实施方式,在利用目的节点缓存由源节点传入的待训练数据集之前,方法还包括:On the basis of the above embodiment, as a preferred implementation, before using the destination node to cache the data set to be trained transmitted by the source node, the method further includes:
判断网络训练集群中是否存在缓存有待训练数据集的源训练节点;Determine whether there is a source training node that caches the data set to be trained in the network training cluster;
若网络训练集群中存在缓存有待训练数据集的源训练节点,则在源训练节点中选取与目的节点之间的空闲网络带宽最大的目标源训练节点;If there is a source training node that caches the data set to be trained in the network training cluster, select the target source training node with the largest idle network bandwidth between the source training node and the destination node;
相应的,利用目的节点缓存由源节点传入的待训练数据集,包括:Correspondingly, use the destination node to cache the data set to be trained from the source node, including:
利用目的节点缓存由目标源训练节点传入的待训练数据集。Use the destination node to cache the data set to be trained passed in by the target source training node.
需要说明的是,在本实施方式中,当判定网络训练集群中存在缓存有待训练数据集的源训练节点时,即网络训练集群中已存储有待训练数据集的训练节点,在此情况下,利用目的节点缓存由源节点传入的待训练数据集之前,首先在源训练节点中选取与目的节点之间的空闲网络带宽最大的目标源训练节点,进而在利用目的节点缓存由源节点传入的待训练数据集时,具体是通过目的节点缓存由源训练节点传入的待训练数据集,由于目 标源训练节点与目的节点之间的带宽相对较高,因此目标源训练节点与目的节点之间的网络传输效率较高,进而本实施方式进一步提高了训练节点之间共享待训练数据集的整体效率。It should be noted that, in this embodiment, when it is determined that there is a source training node that caches the data set to be trained in the network training cluster, that is, the training node that has stored the data set to be trained in the network training cluster, in this case, use Before the destination node caches the data set to be trained from the source node, it first selects the target source training node with the largest idle network bandwidth between the source training node and the destination node, and then uses the destination node to cache the data set passed in by the source node. When the data set is to be trained, the target node caches the data set to be trained from the source training node. Since the bandwidth between the target source training node and the destination node is relatively high, the distance between the target source training node and the destination node is relatively high. The network transmission efficiency is relatively high, and this embodiment further improves the overall efficiency of sharing the data set to be trained among the training nodes.
更进一步的,作为一种优选的实施方式,当网络训练集群中不存在缓存有待训练数据集的源训练节点时,利用目的节点缓存由源节点传入的待训练数据集,包括:Further, as a preferred embodiment, when there is no source training node that caches the data set to be trained in the network training cluster, the destination node is used to cache the data set to be trained passed in by the source node, including:
利用目的节点缓存由网络训练集群中的数据集存储节点传入的待训练数据集。Use the destination node to cache the data set to be trained that is passed in by the data set storage node in the network training cluster.
需要说明的是,在本实施方式中,网络训练集群包含有数据集存储节点,数据集存储节点用于存储待训练数据集,当判定网络训练集群中不存在缓存有待训练数据集的源训练节点时,则在利用目的节点缓存由源节点传入的待训练数据集时,具体利用目的节点缓存由网络训练集群中的数据集存储节点传入的待训练数据集,以此进一步确保目的节点对于待训练数据集进行获取的可靠性。It should be noted that, in this embodiment, the network training cluster includes a data set storage node, and the data set storage node is used to store the to-be-trained data set. When it is determined that there is no source training node that caches the to-be-trained data set in the network training cluster When using the destination node to cache the data set to be trained passed in from the source node, the destination node is specifically used to cache the data set to be trained passed in by the data set storage node in the network training cluster, so as to further ensure that the destination node has The reliability of the acquisition of the dataset to be trained.
在上述实施例的基础上,作为一种优选的实施方式,统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销,包括:On the basis of the above embodiment, as a preferred implementation, statistics of the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set include:
基于未缓存待训练数据集的训练节点的硬件性能参数以及待训练数据集的数据属性参数统计得到磁盘性能开销。Based on the hardware performance parameters of the training nodes that do not cache the data set to be trained and the data attribute parameters of the data set to be trained, the disk performance overhead is obtained by statistics.
需要说明的是,本实施方式在统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销时,具体是基于未缓存待训练数据集的训练节点的硬件性能参数以及待训练数据集的数据属性参数统计得到性能磁盘开销,其中,硬件性能参数指的是训练节点的硬件运行指标参数,数据属性参数指的是待训练数据集所具有的数据特征参数。由于硬件性能参数能够准确表征训练节点进行数据缓存时的效率,而数据属性参数能够准确表征待训练数据集的数据量级以及数据类型等特征,因此本实施方式基于训练节点的硬件性能参数以及待训练数据 集的数据属性参数统计得到磁盘性能开销,能够进一步提高统计得到的磁盘性能开销的准确性。It should be noted that, in this embodiment, when calculating the disk performance overhead required by the training nodes that do not cache the data set to be trained in the network training cluster in the process of caching the data set to be trained, it is specifically based on the uncached data set to be trained. The hardware performance parameters of the training node and the data attribute parameters of the data set to be trained are counted to obtain the performance disk overhead. Data feature parameters. Since the hardware performance parameters can accurately represent the efficiency of the training node for data caching, and the data attribute parameters can accurately represent the data magnitude and data type of the data set to be trained, this embodiment is based on the hardware performance parameters of the training node and the data to be trained. The data attribute parameters of the training data set are calculated to obtain the disk performance overhead, which can further improve the accuracy of the calculated disk performance overhead.
更进一步的,作为一种优选的实施方式,硬件性能参数包括磁盘转速、磁盘平均巡道时间以及磁盘最大传输速率;Further, as a preferred embodiment, the hardware performance parameters include the disk rotation speed, the disk average patrol time, and the disk maximum transfer rate;
待训练数据集的数据属性参数包括平均文件大小;平均文件大小基于待训练数据集的数据总量以及文件总量运算得到。The data attribute parameters of the data set to be trained include an average file size; the average file size is calculated based on the total data volume of the data set to be trained and the total number of files.
本实施方式中,硬件性能参数包括磁盘转速、磁盘平均巡道时间以及磁盘最大传输速率。其中,磁盘转速指的是,磁盘盘片在一分钟内所能完成的最大转数,它是决定磁盘内部传输率的关键因素之一,在很大程度上直接影响到磁盘的速度;磁盘平均巡道时间,是指磁盘在接收到系统指令后,磁头从开始移动到数据所在的磁道所花费时间的平均值,它在一定程度上体现了磁盘读取数据的能力,是影响磁盘内部数据传输率的重要参数;磁盘最大传输速率,指的是数据由磁盘的磁头到高速缓存之间传输的速度,影响数据在磁盘中缓存的整体效率。In this implementation manner, the hardware performance parameters include the rotational speed of the disk, the average track time of the disk, and the maximum transfer rate of the disk. Among them, the disk rotation speed refers to the maximum number of revolutions that the disk platter can complete in one minute. It is one of the key factors determining the internal transfer rate of the disk and directly affects the speed of the disk to a large extent; Track time refers to the average time taken by the magnetic head to move from the beginning to the track where the data is located after the disk receives the system command. It reflects the ability of the disk to read data to a certain extent and affects the internal data transmission of the disk. The maximum transfer rate of the disk refers to the speed at which data is transferred from the head of the disk to the cache, which affects the overall efficiency of data caching in the disk.
另外,本实施方式中,待训练数据集的数据属性参数包括平均文件大小,平均文件大小基于待训练数据集的数据总量以及文件总量运算得到,也就是说,通过待训练数据集的数据总量,即数据总大小与文件总量,即文件个数进行商运算得到,平均文件大小指的是待训练数据集中的每一个文件需要占用的磁盘存储空间平均值。In addition, in this embodiment, the data attribute parameter of the data set to be trained includes the average file size, and the average file size is calculated based on the total amount of data in the data set to be trained and the total amount of files, that is, the data of the data set to be trained is calculated. The total, that is, the total size of the data and the total number of files, that is, the number of files, is obtained by quotient operation, and the average file size refers to the average disk storage space occupied by each file in the data set to be trained.
本实施方式通过对训练节点的硬件性能参数以及待训练数据集的数据属性参数进行了细化,以此进一步提高了统计得到的磁盘性能开销的准确性。In this implementation manner, the hardware performance parameters of the training nodes and the data attribute parameters of the data set to be trained are refined, thereby further improving the accuracy of the statistical obtained disk performance overhead.
在上述实施例的基础上,作为一种优选的实施方式,在利用目的节点缓存由源节点传入的待训练数据集之前,方法还包括:On the basis of the above embodiment, as a preferred implementation, before using the destination node to cache the data set to be trained transmitted by the source node, the method further includes:
判断目的节点的缓存队列是否存在空闲空间;Determine whether there is free space in the cache queue of the destination node;
若是,则执行利用目的节点缓存由源节点传入的待训练数据集的步骤;If so, execute the step of using the destination node to cache the data set to be trained passed in by the source node;
否则,将缓存队列中执行次数最少的目标待训练数据集删除,并执行利用目的节点缓存由源节点传入的待训练数据集的步骤。Otherwise, delete the target data set to be trained with the least number of executions in the cache queue, and perform the step of using the destination node to cache the data set to be trained from the source node.
需要说明的是,在本实施方式中,在利用目的节点缓存由源节点传入的待训练数据集之前,进一步判断目的节点的缓存队列中是否存在空闲空间,也就是判断缓存队列中是否能够正常存储待训练数据集,进而当目的节点的缓存队列存在空闲空间时,则进一步执行利用目的节点缓存由源节点传入的待训练数据集的步骤,相反的,当目的节点的缓存队列不存在空闲空间时,则将缓存队列中执行次数最少的目标待训练数据集删除,并执行利用目的节点缓存由源节点传入的待训练数据集的步骤,以此确保待训练数据集能够正常存储至目的节点的缓存队列,进一步确保了数据集缓存过程的整体可靠性。It should be noted that, in this embodiment, before using the destination node to cache the data set to be trained from the source node, it is further determined whether there is free space in the cache queue of the destination node, that is, it is determined whether the cache queue can work normally. The data set to be trained is stored, and when there is free space in the cache queue of the destination node, the step of using the destination node to cache the data set to be trained from the source node is further performed. On the contrary, when there is no free space in the cache queue of the destination node When there is no space, delete the target data set to be trained with the least number of executions in the cache queue, and perform the step of using the destination node to cache the data set to be trained from the source node, so as to ensure that the data set to be trained can be stored normally to the destination The node's cache queue further ensures the overall reliability of the data set caching process.
在上述一系列实施方式的基础上,作为一种优选的实施方式,当前性能参数包括I/O队列长度,磁盘性能开销包括IOPS开销;Based on the above series of implementations, as a preferred implementation, the current performance parameter includes the I/O queue length, and the disk performance overhead includes the IOPS overhead;
基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点,包括:Based on the current performance parameters, select the destination node that meets the disk performance overhead in the network training cluster, including:
在网络集群中选取I/O队列长度小于IOPS开销的目的节点。Select the destination node whose I/O queue length is less than the IOPS cost in the network cluster.
需要说明的是,在本实施方式中,当前性能参数包括I/O(Input/Output的缩写,即输入输出端口)队列长度,磁盘性能开销包括IOPS开销,IOPS(Input/Output,Operations Per Second)是一个用于计算机存储设备(如磁盘(HDD)、固态磁盘(SSD)或存储区域网络(SAN))性能测试的测量方式,可以视为是每秒的读写次数。在基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点时,具体是在网络集群中选取I/O队列长度小于IOPS开销的目的节点,目的是确保待训练数据集能够可靠缓存至所选取的目的节点,进而进一步确保了数据集缓存过程的整体可靠性。It should be noted that, in this implementation manner, the current performance parameters include I/O (abbreviation for Input/Output, that is, input and output ports) queue length, the disk performance overhead includes IOPS overhead, IOPS (Input/Output, Operations Per Second) It is a measurement used for performance testing of computer storage devices such as magnetic disks (HDDs), solid-state disks (SSDs), or storage area networks (SANs), and can be considered as the number of reads and writes per second. When selecting the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameters, specifically select the destination node whose I/O queue length is less than the IOPS overhead in the network cluster, in order to ensure that the data set to be trained can be reliably cached to The selected destination node further ensures the overall reliability of the data set caching process.
为了进一步加深对于上述实施例的理解,本申请还提供一种具体应用场景下的场景实施例做进一步说明。In order to further deepen the understanding of the above embodiments, the present application also provides a scenario embodiment under a specific application scenario for further description.
对于集群中的每个训练节点,可获得用户进行数据集缓存的磁盘的下列参数:磁盘转速(rpm),磁盘平均寻道时间(avgSeekTime),磁盘最大传输速率(maxTransRate)。通过以上4个参数、数据集的平均文件大小(avgSize) 以及下式,计算该磁盘在该数据集的缓存时的单次IO时间,即IOTime:For each training node in the cluster, the following parameters of the disk used for data set caching can be obtained: disk rotation speed (rpm), disk average seek time (avgSeekTime), and disk maximum transfer rate (maxTransRate). Through the above 4 parameters, the average file size of the dataset (avgSize) and the following formula, calculate the single IO time of the disk in the cache of the dataset, that is, IOTime:
Figure PCTCN2021109237-appb-000001
Figure PCTCN2021109237-appb-000001
根据上式可得在该数据集下,每次IO操作耗时的近似值。According to the above formula, the approximate value of the time consuming of each IO operation under this data set can be obtained.
对于该数据集在当前的节点进行缓存时的最大IOPS能力,可以通过下式近似计算:For the maximum IOPS capability of the data set when the current node is cached, it can be approximated by the following formula:
Figure PCTCN2021109237-appb-000002
Figure PCTCN2021109237-appb-000002
同时,对于每个训练节点,通过监控可得到数据集缓存磁盘的wrqm/s值、rrqm/s和avgqu-sz值。其中wrqm/s代表该磁盘当前的每秒合并数据(merge)后的写次数,rrqm/s代表该磁盘当前的每秒合并数据(merge)后的读次数,avgqu-sz代表该磁盘的I/O队列长度。考虑到IOTime的增加与IOPS的增加并不是一个完全线性的关系,事实上在达到某个IOPS值之后,IOTime会随着IOPS的增加而显著提升。故本方法对于avgqu-sz<IOPS*70%的节点,做为本次数据集缓存的目的节点,用于缓存待训练数据集。At the same time, for each training node, the wrqm/s value, rrqm/s and avgqu-sz value of the data set cache disk can be obtained by monitoring. Where wrqm/s represents the current number of writes per second of the disk after merging data (merge), rrqm/s represents the current number of reads per second after merging data (merge) of the disk, and avgqu-sz represents the I/ O queue length. Considering that the increase of IOTime is not a completely linear relationship with the increase of IOPS, in fact, after reaching a certain IOPS value, IOTime will increase significantly with the increase of IOPS. Therefore, in this method, the node with avgqu-sz<IOPS*70% is used as the destination node of this data set cache, which is used to cache the data set to be trained.
请参见图2所示,本申请实施例提供了一种网络训练的数据集缓存装置,包括:Referring to FIG. 2, an embodiment of the present application provides a data set cache device for network training, including:
开销统计模块10,用于统计网络训练集群中未缓存待训练数据集的训练节点在缓存待训练数据集的过程中所需的磁盘性能开销;The overhead statistics module 10 is used to count the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set;
参数监控模块11,用于监控未缓存待训练数据集的训练节点的当前性能参数;The parameter monitoring module 11 is used to monitor the current performance parameters of the training nodes that do not cache the data set to be trained;
节点选取模块12,用于基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点;The node selection module 12 is used to select the destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameter;
节点缓存模块13,用于利用目的节点缓存由源节点传入的待训练数据集,以基于目的节点对待训练数据集执行网络训练。The node caching module 13 is configured to use the destination node to cache the to-be-trained data set passed in by the source node, so as to perform network training based on the destination node's to-be-trained data set.
本申请所提供的网络训练的数据集缓存装置,首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存待训练数据集的过程中所需的磁盘性能开销,进而监控未缓存有待训练数据集的训练节点的当前性能参数,并基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的 节点,进而利用目的节点缓存由源节点传入的待训练数据集,以此基于目的节点对待训练数据集执行网络训练。由于本装置根据预估得到缓存待训练数据集所需的磁盘性能开销,在网络训练集群的训练节点中选取当前性能参数满足磁盘性能开销的目的节点对源节点传入的待训练数据集进行缓存,相对确保了训练节点缓存待训练数据集的可靠性,进而确保了网络训练的整体可靠性。The data set caching device for network training provided by the present application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data set to be trained. The current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node. The node performs network training on the training dataset. Since the device obtains the disk performance overhead required for caching the data set to be trained according to the estimation, selects the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained imported from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
此外,本申请还提供一种网络训练的数据集缓存设备,包括:In addition, this application also provides a data set caching device for network training, including:
存储器,用于存储计算机程序;memory for storing computer programs;
处理器,用于执行计算机程序时实现如上述的网络训练的数据集缓存方法的步骤。The processor is configured to implement the steps of the above-mentioned data set caching method for network training when executing the computer program.
本申请所提供的网络训练的数据集缓存设备,首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存待训练数据集的过程中所需的磁盘性能开销,进而监控未缓存有待训练数据集的训练节点的当前性能参数,并基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点,进而利用目的节点缓存由源节点传入的待训练数据集,以此基于目的节点对待训练数据集执行网络训练。由于本设备根据预估得到缓存待训练数据集所需的磁盘性能开销,在网络训练集群的训练节点中选取当前性能参数满足磁盘性能开销的目的节点对源节点传入的待训练数据集进行缓存,相对确保了训练节点缓存待训练数据集的可靠性,进而确保了网络训练的整体可靠性。The data set caching device for network training provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data set to be trained. The current performance parameters of the training node of the training data set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node. The node performs network training on the training dataset. Since the device obtains the disk performance overhead required to cache the data set to be trained according to the estimation, select the destination node whose current performance parameters meet the disk performance overhead in the training nodes of the network training cluster to cache the data set to be trained input from the source node , which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of the network training.
此外,本申请还提供一种计算机可读存储介质,计算机可读存储介质上存储有计算机程序,计算机程序被处理器执行时实现如上述的网络训练的数据集缓存方法的步骤。In addition, the present application also provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the steps of the above-mentioned network training data set caching method are implemented.
本申请所提供的计算机可读存储介质,首先统计网络训练集群中未缓存待训练数据集的训练节点,在缓存待训练数据集的过程中所需的磁盘性能开销,进而监控未缓存有待训练数据集的训练节点的当前性能参数,并基于当前性能参数在网络训练集群中选取满足磁盘性能开销的目的节点, 进而利用目的节点缓存由源节点传入的待训练数据集,以此基于目的节点对待训练数据集执行网络训练。由于本计算机可读存储介质根据预估得到缓存待训练数据集所需的磁盘性能开销,在网络训练集群的训练节点中选取当前性能参数满足磁盘性能开销的目的节点对源节点传入的待训练数据集进行缓存,相对确保了训练节点缓存待训练数据集的可靠性,进而确保了网络训练的整体可靠性。The computer-readable storage medium provided by this application firstly counts the training nodes in the network training cluster that do not cache the data set to be trained, the disk performance overhead required in the process of caching the data set to be trained, and then monitors the uncached data to be trained. The current performance parameters of the training nodes in the set, and based on the current performance parameters, the destination node that satisfies the disk performance overhead is selected in the network training cluster, and then the destination node is used to cache the data set to be trained from the source node. The training dataset performs network training. Since the computer-readable storage medium obtains the disk performance overhead required for caching the data set to be trained according to the estimation, the destination node whose current performance parameters meet the disk performance overhead is selected from the training nodes of the network training cluster to the source node to be trained. The data set is cached, which relatively ensures the reliability of the training node to cache the data set to be trained, thereby ensuring the overall reliability of network training.
以上对本申请所提供的一种网络训练的数据集缓存方法、装置、设备及存储介质进行了详细介绍。说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例公开的装置而言,由于其与实施例公开的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。应当指出,对于本技术领域的普通技术人员来说,在不脱离本申请原理的前提下,还可以对本申请进行若干改进和修饰,这些改进和修饰也落入本申请权利要求的保护范围内。The data set caching method, device, device and storage medium for network training provided by the present application have been described in detail above. The various embodiments in the specification are described in a progressive manner, and each embodiment focuses on the differences from other embodiments, and the same and similar parts between the various embodiments can be referred to each other. For the device disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant part can be referred to the description of the method. It should be pointed out that for those of ordinary skill in the art, without departing from the principles of the present application, several improvements and modifications can also be made to the present application, and these improvements and modifications also fall within the protection scope of the claims of the present application.
还需要说明的是,在本说明书中,诸如第一和第二等之类的关系术语仅仅用来将一个实体或者操作与另一个实体或操作区分开来,而不一定要求或者暗示这些实体或操作之间存在任何这种实际的关系或者顺序。而且,术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、方法、物品或者设备不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、方法、物品或者设备所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括所述要素的过程、方法、物品或者设备中还存在另外的相同要素。It should also be noted that, in this specification, relational terms such as first and second, etc. are only used to distinguish one entity or operation from another entity or operation, and do not necessarily require or imply these entities or operations. There is no such actual relationship or sequence between operations. Moreover, the terms "comprising", "comprising" or any other variation thereof are intended to encompass a non-exclusive inclusion such that a process, method, article or device comprising a list of elements includes not only those elements, but also includes not explicitly listed or other elements inherent to such a process, method, article or apparatus. Without further limitation, an element qualified by the phrase "comprising a..." does not preclude the presence of additional identical elements in a process, method, article or apparatus that includes the element.

Claims (10)

  1. 一种网络训练的数据集缓存方法,其特征在于,包括:A data set caching method for network training, characterized in that it includes:
    统计网络训练集群中未缓存待训练数据集的训练节点在缓存所述待训练数据集的过程中所需的磁盘性能开销;Counting the disk performance overhead required by the training nodes that do not cache the data set to be trained in the process of caching the data set to be trained in the network training cluster;
    监控未缓存待训练数据集的所述训练节点的当前性能参数;monitoring the current performance parameters of the training nodes that do not cache the data set to be trained;
    基于所述当前性能参数在所述网络训练集群中选取满足所述磁盘性能开销的目的节点;Selecting a destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameter;
    利用所述目的节点缓存由源节点传入的所述待训练数据集,以基于所述目的节点对所述待训练数据集执行网络训练。The destination node is used to cache the to-be-trained data set passed in by the source node, so as to perform network training on the to-be-trained data set based on the destination node.
  2. 根据权利要求1所述的网络训练的数据集缓存方法,其特征在于,在所述利用所述目的节点缓存由源节点传入的所述待训练数据集之前,所述方法还包括:The data set caching method for network training according to claim 1, characterized in that, before using the destination node to cache the to-be-trained data set passed in by the source node, the method further comprises:
    判断所述网络训练集群中是否存在缓存有所述待训练数据集的源训练节点;Judging whether there is a source training node that caches the data set to be trained in the network training cluster;
    若所述网络训练集群中存在缓存有所述待训练数据集的源训练节点,则在所述源训练节点中选取与所述目的节点之间的空闲网络带宽最大的目标源训练节点;If there is a source training node that caches the data set to be trained in the network training cluster, select a target source training node with the largest idle network bandwidth between the source training nodes and the destination node;
    相应的,所述利用所述目的节点缓存由源节点传入的所述待训练数据集,包括:Correspondingly, using the destination node to cache the to-be-trained data set passed in by the source node includes:
    利用所述目的节点缓存由所述目标源训练节点传入的所述待训练数据集。The to-be-trained data set passed in by the target source training node is cached by the destination node.
  3. 根据权利要求2所述的网络训练的数据集缓存方法,其特征在于,当所述网络训练集群中不存在缓存有所述待训练数据集的源训练节点时,所述利用所述目的节点缓存由源节点传入的所述待训练数据集,包括:The data set caching method for network training according to claim 2, wherein when there is no source training node that caches the to-be-trained data set in the network training cluster, the cache using the destination node The to-be-trained data set passed in by the source node includes:
    利用所述目的节点缓存由所述网络训练集群中的数据集存储节点传入的所述待训练数据集。The destination node is used to cache the to-be-trained data set passed in by the data set storage node in the network training cluster.
  4. 根据权利要求1所述的网络训练的数据集缓存方法,其特征在于,所述统计网络训练集群中未缓存待训练数据集的训练节点在缓存所述待训练数据集的过程中所需的磁盘性能开销,包括:The data set caching method for network training according to claim 1, characterized in that, a training node in the network training cluster that does not cache the data set to be trained needs a disk in the process of caching the data set to be trained. Performance overhead, including:
    基于未缓存所述待训练数据集的所述训练节点的硬件性能参数以及所述待训练数据集的数据属性参数统计得到所述磁盘性能开销。The disk performance overhead is statistically obtained based on hardware performance parameters of the training nodes that do not cache the to-be-trained data set and data attribute parameters of the to-be-trained data set.
  5. 根据权利要求4所述的网络训练的数据集缓存方法,其特征在于,所述硬件性能参数包括磁盘转速、磁盘平均巡道时间以及磁盘最大传输速率;The data set caching method for network training according to claim 4, wherein the hardware performance parameters include disk rotation speed, disk average patrol time and disk maximum transfer rate;
    所述待训练数据集的数据属性参数包括平均文件大小;所述平均文件大小基于所述待训练数据集的数据总量以及文件总量运算得到。The data attribute parameter of the data set to be trained includes an average file size; the average file size is calculated based on the total amount of data and the total amount of files in the data set to be trained.
  6. 根据权利要求1所述的网络训练的数据集缓存方法,其特征在于,在所述利用所述目的节点缓存由源节点传入的所述待训练数据集之前,所述方法还包括:The data set caching method for network training according to claim 1, characterized in that, before using the destination node to cache the to-be-trained data set passed in by the source node, the method further comprises:
    判断所述目的节点的缓存队列是否存在空闲空间;Determine whether there is free space in the cache queue of the destination node;
    若是,则执行所述利用所述目的节点缓存由源节点传入的所述待训练数据集的步骤;If so, execute the step of using the destination node to cache the data set to be trained passed in by the source node;
    否则,将所述缓存队列中执行次数最少的目标待训练数据集删除,并执行所述利用所述目的节点缓存由源节点传入的所述待训练数据集的步骤。Otherwise, delete the target data set to be trained with the least number of executions in the cache queue, and execute the step of using the destination node to cache the data set to be trained transmitted by the source node.
  7. 根据权利要求1至6任意一项所述的网络训练的数据集缓存方法,其特征在于,所述当前性能参数包括I/O队列长度,所述磁盘性能开销包括IOPS开销;The data set caching method for network training according to any one of claims 1 to 6, wherein the current performance parameter includes an I/O queue length, and the disk performance overhead includes an IOPS overhead;
    所述基于所述当前性能参数在所述网络训练集群中选取满足所述磁盘性能开销的目的节点,包括:The selecting a destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameter includes:
    在所述网络集群中选取所述I/O队列长度小于IOPS开销的所述目的节点。The destination node whose I/O queue length is less than the IOPS overhead is selected in the network cluster.
  8. 一种网络训练的数据集缓存装置,其特征在于,包括:A data set cache device for network training, comprising:
    开销统计模块,用于统计网络训练集群中未缓存待训练数据集的训练节点在缓存所述待训练数据集的过程中所需的磁盘性能开销;An overhead statistics module, used to count the disk performance overhead required by the training nodes in the network training cluster that do not cache the to-be-trained data set in the process of caching the to-be-trained data set;
    参数监控模块,用于监控未缓存待训练数据集的所述训练节点的当前性能参数;a parameter monitoring module, used to monitor the current performance parameters of the training nodes that do not cache the data set to be trained;
    节点选取模块,用于基于所述当前性能参数在所述网络训练集群中选 取满足所述磁盘性能开销的目的节点;A node selection module is used to select a destination node that satisfies the disk performance overhead in the network training cluster based on the current performance parameter;
    节点缓存模块,用于利用所述目的节点缓存由源节点传入的所述待训练数据集,以基于所述目的节点对所述待训练数据集执行网络训练。A node caching module, configured to use the destination node to cache the to-be-trained data set passed in by the source node, so as to perform network training on the to-be-trained data set based on the destination node.
  9. 一种网络训练的数据集缓存设备,其特征在于,包括:A data set cache device for network training, characterized in that it includes:
    存储器,用于存储计算机程序;memory for storing computer programs;
    处理器,用于执行所述计算机程序时实现如权利要求1至7任一项所述的网络训练的数据集缓存方法的步骤。The processor is configured to implement the steps of the data set caching method for network training according to any one of claims 1 to 7 when executing the computer program.
  10. 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至7任一项所述的网络训练的数据集缓存方法的步骤。A computer-readable storage medium, characterized in that, a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the network training according to any one of claims 1 to 7 is implemented. The steps of the dataset caching method.
PCT/CN2021/109237 2020-11-27 2021-07-29 Method and apparatus for data set caching in network training, device, and storage medium WO2022110861A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011357904.1 2020-11-27
CN202011357904.1A CN112446490A (en) 2020-11-27 2020-11-27 Network training data set caching method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
WO2022110861A1 true WO2022110861A1 (en) 2022-06-02

Family

ID=74737918

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/109237 WO2022110861A1 (en) 2020-11-27 2021-07-29 Method and apparatus for data set caching in network training, device, and storage medium

Country Status (2)

Country Link
CN (1) CN112446490A (en)
WO (1) WO2022110861A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112446490A (en) * 2020-11-27 2021-03-05 苏州浪潮智能科技有限公司 Network training data set caching method, device, equipment and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103905517A (en) * 2012-12-28 2014-07-02 中国移动通信集团公司 Data storage method and equipment
CN109710406A (en) * 2018-12-21 2019-05-03 腾讯科技(深圳)有限公司 Data distribution and its model training method, device and computing cluster
CN111459506A (en) * 2020-03-02 2020-07-28 平安科技(深圳)有限公司 Deployment method, device, medium and electronic equipment of deep learning platform cluster
CN111625696A (en) * 2020-07-28 2020-09-04 北京升鑫网络科技有限公司 Distributed scheduling method, computing node and system of multi-source data analysis engine
CN112446490A (en) * 2020-11-27 2021-03-05 苏州浪潮智能科技有限公司 Network training data set caching method, device, equipment and storage medium

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805259A (en) * 2018-05-23 2018-11-13 北京达佳互联信息技术有限公司 neural network model training method, device, storage medium and terminal device
CN110502487B (en) * 2019-08-09 2022-11-22 苏州浪潮智能科技有限公司 Cache management method and device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103905517A (en) * 2012-12-28 2014-07-02 中国移动通信集团公司 Data storage method and equipment
CN109710406A (en) * 2018-12-21 2019-05-03 腾讯科技(深圳)有限公司 Data distribution and its model training method, device and computing cluster
CN111459506A (en) * 2020-03-02 2020-07-28 平安科技(深圳)有限公司 Deployment method, device, medium and electronic equipment of deep learning platform cluster
CN111625696A (en) * 2020-07-28 2020-09-04 北京升鑫网络科技有限公司 Distributed scheduling method, computing node and system of multi-source data analysis engine
CN112446490A (en) * 2020-11-27 2021-03-05 苏州浪潮智能科技有限公司 Network training data set caching method, device, equipment and storage medium

Also Published As

Publication number Publication date
CN112446490A (en) 2021-03-05

Similar Documents

Publication Publication Date Title
US9811474B2 (en) Determining cache performance using a ghost cache list indicating tracks demoted from a cache list of tracks in a cache
US10592432B2 (en) Adjusting active cache size based on cache usage
EP3229142B1 (en) Read cache management method and device based on solid state drive
US8521986B2 (en) Allocating storage memory based on future file size or use estimates
US10097464B1 (en) Sampling based on large flow detection for network visibility monitoring
CN111064808B (en) Load balancing method and device based on distributed storage system
US9979624B1 (en) Large flow detection for network visibility monitoring
US10536360B1 (en) Counters for large flow detection
WO2017117734A1 (en) Cache management method, cache controller and computer system
US9792231B1 (en) Computer system for managing I/O metric information by identifying one or more outliers and comparing set of aggregated I/O metrics
AU2012204481A1 (en) Selecting storage locations for storing data based on storage location attributes and data usage statistics
US20150127905A1 (en) Cache Modeling Using Random Sampling and a Timestamp Histogram
EP3282643A1 (en) Method and apparatus of estimating conversation in a distributed netflow environment
CN109086009B (en) Monitoring management method and device and computer readable storage medium
WO2021088404A1 (en) Data processing method, apparatus and device, and readable storage medium
Shankar et al. High-performance hybrid key-value store on modern clusters with RDMA interconnects and SSDs: Non-blocking extensions, designs, and benefits
WO2022110861A1 (en) Method and apparatus for data set caching in network training, device, and storage medium
EP2811410B1 (en) Monitoring record management method and device
US20220342848A1 (en) Snapshot space reporting using a probabilistic data structure
CN111427920B (en) Data acquisition method, device, system, computer equipment and storage medium
WO2023165543A1 (en) Shared cache management method and apparatus, and storage medium
WO2020224125A1 (en) Method and device for monitoring io delay of distributed file system, and storage medium
US20210141738A1 (en) Method and Apparatus for Characterizing Workload Sequentiality for Cache Policy Optimization
US20220414154A1 (en) Community generation based on a common set of attributes
US9354820B2 (en) VSAM data set tier management

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21896368

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21896368

Country of ref document: EP

Kind code of ref document: A1