WO2015070539A1 - Dma-based data compression chip structure and implementation method therefor - Google Patents

Dma-based data compression chip structure and implementation method therefor Download PDF

Info

Publication number
WO2015070539A1
WO2015070539A1 PCT/CN2014/071244 CN2014071244W WO2015070539A1 WO 2015070539 A1 WO2015070539 A1 WO 2015070539A1 CN 2014071244 W CN2014071244 W CN 2014071244W WO 2015070539 A1 WO2015070539 A1 WO 2015070539A1
Authority
WO
WIPO (PCT)
Prior art keywords
data
dma
module
compression
descriptor
Prior art date
Application number
PCT/CN2014/071244
Other languages
French (fr)
Chinese (zh)
Inventor
王恩东
胡雷钧
李仁刚
Original Assignee
浪潮电子信息产业股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 浪潮电子信息产业股份有限公司 filed Critical 浪潮电子信息产业股份有限公司
Publication of WO2015070539A1 publication Critical patent/WO2015070539A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F13/00Interconnection of, or transfer of information or other signals between, memories, input/output devices or central processing units
    • G06F13/14Handling requests for interconnection or transfer
    • G06F13/20Handling requests for interconnection or transfer for access to input/output bus
    • G06F13/28Handling requests for interconnection or transfer for access to input/output bus using burst mode transfer, e.g. direct memory access DMA, cycle steal

Definitions

  • the present invention relates to the field of computer system design and integrated circuit design, and in particular to a chip structure and an implementation method thereof, and more particularly to a data compression chip structure based on DMA and an implementation method thereof. Background technique
  • the compression chip uses DMA technology to improve the processing efficiency of system data, and the DMA technology uses bandwidth matching design. The method works together with a plurality of compression algorithm cores to complete data compression processing and storage operations. Summary of the invention
  • the present invention provides a data compression chip structure based on DMA and an implementation method thereof, which adopts a data compression chip based on the DMA principle, compresses large data, and adopts a bandwidth matching DMA technology. With multiple compression algorithm cores, it can realize the compression processing and storage of big data together, and improve the utilization of system storage resources.
  • the technical solutions adopted by the present invention are as follows:
  • a data compression chip structure based on DMA including system to device DMA module, device to system DMA module, transmitting end arbitration module, data storage module interface module and compression algorithm core module, Among them, the system-to-device DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; device-to-system DMA module adopts descriptor DMA and data DMA. Improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; the arbitration module at the sending end controls the arbitration mechanism for sending packets by giving priority analysis, judgment, and generation, which is used to solve multi-message transmission and transmission.
  • a DMA-based data compression chip structure implementation method is to communicate with a system through a PCI-E interface, realize system and device data transmission through DMA technology, respectively design a system to device and device to system DMA engine, and send The end design arbitration logic realizes the efficient transmission of the transmitted data.
  • the descriptor DMA and the data DMA are respectively implemented in the DMA module in two directions to improve the data transmission efficiency, and the received data is connected to the data compression algorithm module through the interface storage module, and the data is connected.
  • the compression module performs compression processing, and the compressed data is then transferred to the system via DMA.
  • the implementation method of the present invention specifically includes the following steps:
  • the system initiates a read request
  • the data compression chip sends a read request and a read command through its sender, and sends it to the system through the PCI-E port, and at the receiving end, receives descriptor information sent by the system, and the information passes through the system to the device.
  • the descriptor DMA mode controls the transmission, that is, the start and end address information of the descriptor, and is transmitted by DMA, and finally written to the descriptor register, located in the system-to-device descriptor, and the data transmission of the corresponding descriptor information is controlled by the system-to-device DMA engine. ;
  • the received DMA data is transmitted to the data storage interface module through the system-to-device interface, so that the compression algorithm core module performs data compression application, and the data storage module is a highly parallel packet memory array, which can support multiple data compression cores to read in parallel.
  • the core module of the compression algorithm reads data from the data storage interface module, performs data compression operations, and collects data information for compression control to improve compression efficiency.
  • the data matching information in the data compression process is saved by the matching information module for compression.
  • Module use D. Determine the core number of the core module of the compression algorithm according to the PCI-E interface, the DMA transmission capability, the data storage interface, and the bandwidth of each interface and module of the compressed core module capability system; the core modules of each compression algorithm are completely independent and parallel work;
  • the compressed data after compression is controlled by the device to the system DMA module to complete the data transmission from the device to the system; the device to system descriptor DMA mode controls the transmission of multiple descriptor information to improve system performance.
  • the transmitted descriptor information is registered in the device to system descriptor register, and is transmitted by the device to the system DMA mode;
  • the arbitrating module of the transmitting end controls the priority control when multiple message data are transmitted, thereby avoiding the conflict problem when multiple messages are simultaneously transmitted, and finally the data is completed by the PCI-E interface to the system. Send transmission.
  • the system-to-device DMA feature mainly refers to the compression chip receiving data from the system by means of DMA, including descriptor DMA and data DMA, wherein the data DMA includes a DMA descriptor and a DMA engine, and the description DMA is to transfer the descriptor information of the system by DMA to improve performance.
  • the DMA descriptor refers to the descriptor register
  • the DMA engine refers to the DMA method for data transfer.
  • the device-to-system DMA feature mainly refers to the compression chip transmitting compressed data to the system by means of DMA, including descriptor DMA and data DMA, wherein the data DMA includes a DMA descriptor and a DMA engine, and the descriptor DMA is a system.
  • the descriptor information is transmitted by DMA to improve performance.
  • the DMA descriptor refers to the descriptor register
  • the DMA engine refers to the data transfer by DMA
  • the characteristics of the sender arbitration mainly refers to the priority.
  • the generated arbitration method controls and resolves conflicts when multiple messages are sent simultaneously.
  • the characteristics of the data storage module interface mainly refer to the use of a configurable packet storage array to form a data buffer with high parallel storage for multiple compression.
  • the algorithm core stores system-to-device data and device-to-system compressed data;
  • the core characteristics of the compression algorithm mainly refer to the core of multiple sets of compression algorithms, so that each module and interface The bandwidth matching ensures the utilization of resources, thereby achieving the purpose of improving system performance.
  • the core module of the compression algorithm includes a data compression module, a compression control module, and a matching information module, and the data compression module mainly implements data compression function, and the compression control module mainly The control information is received, and the compression control information is generated, and the compression module is controlled to work.
  • the matching information module mainly stores the matching information in the compression process for use by the compression module.
  • the DMA-based data compression chip structure and the implementation method thereof have the above advantages, which make up for the deficiency of the traditional software data compression design structure seriously affecting the system CPU computing resources, and does not significantly affect the system. Under the condition of performance, it has many features such as configurability, hardware resource saving, and flexible expansion of the core of the compression algorithm, making it more suitable for big data storage and analysis systems, so it has a very broad development prospect. High technical value. DRAWINGS
  • FIG. 1 is a logic structural diagram of a DMA-based data compression chip structure and an implementation method thereof according to the present invention
  • FIG. 2 is a schematic diagram of a logic structure of a transmitting end arbitration of a DMA-based data compression chip structure and an implementation method thereof according to the present invention.
  • a data compression chip structure based on DMA comprising a system to device DMA module, a device to system DMA module, a transmitting end arbitration module, a data storage module interface module, and a compression algorithm core module, wherein
  • the system-to-device DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; device-to-system DMA module uses descriptor DMA and data DMA to improve the system.
  • the efficiency that is, the descriptor information and the transmission data are all DMA transmission;
  • the arbitrating module at the transmitting end controls the arbitration mechanism for sending packets by giving priority analysis, determination, and generation, and is used to solve the conflict problem of multi-message transmission and transmission.
  • the data storage module interface module uses a high parallel multi-group storage array to implement data caching; the compression algorithm core module adopts a data path bandwidth matching method to configure multiple compression algorithm cores.
  • a DMA-based data compression chip structure implementation method is to communicate with a system through a PCI-E interface, realize system and device data transmission through DMA technology, respectively design a system to device and device to system DMA engine, and send The end design arbitration logic realizes the efficient transmission of the transmitted data.
  • the descriptor DMA and the data DMA are respectively implemented in the DMA module in two directions to improve the data transmission efficiency, and the received data is connected to the data compression algorithm module through the interface storage module, and the data is connected.
  • the compression module performs compression processing, and the compressed data is then transferred to the system via DMA.
  • the implementation method of the present invention specifically includes the following steps:
  • the system initiates a read request
  • the data compression chip sends a read request and a read command through its sender, and sends it to the system through the PCI-E port, and at the receiving end, receives descriptor information sent by the system, and the information passes through the system to the device.
  • the descriptor DMA mode controls the transmission, that is, the start and end address information of the descriptor, and is transmitted by DMA, and finally written to the descriptor register, located in the system-to-device descriptor, and the data transmission of the corresponding descriptor information is controlled by the system-to-device DMA engine. ;
  • the received DMA data is transmitted to the data storage interface module through the system-to-device interface, so that the compression algorithm core module performs data compression application, and the data storage module is a highly parallel packet memory array, which can support multiple data compression cores to read in parallel.
  • the core module of the compression algorithm reads data from the data storage interface module, performs data compression operations, and collects data information for compression control to improve compression efficiency.
  • the data matching information in the data compression process is saved by the matching information module for compression.
  • Module use D. Determine the core number of the core module of the compression algorithm according to the PCI-E interface, the DMA transmission capability, the data storage interface, and the bandwidth of each interface and module of the compressed core module capability system; the core modules of each compression algorithm are completely independent and parallel work;
  • the compressed data after compression is controlled by the device to the system DMA module to complete the data transmission from the device to the system; the device to system descriptor DMA mode controls the transmission of multiple descriptor information to improve system performance.
  • the transmitted descriptor information is registered in the device to system descriptor register, and is transmitted by the device to the system DMA mode;
  • the arbitrating module of the transmitting end controls the priority control when multiple message data are transmitted, thereby avoiding the conflict problem when multiple messages are simultaneously transmitted, and finally the data is completed by the PCI-E interface to the system. Send transmission.

Abstract

A DMA-based data compression chip structure and an implementation method therefor. The data compression chip structure comprises a system to device DMA module, a device to system DMA module, a sending-end mediation module, a data storage module interface module and a compression algorithm core module. The DMA-based data compression chip and the DMA technology with matched bandwidth in combination with a plurality of compression algorithm cores are used together to realize the compression processing and storage of large data, thereby increasing the utilization rate of system storage resources.

Description

一种基于 DMA的数据压縮芯片结构及其实现方法  Data compression chip structure based on DMA and implementation method thereof
技术领域 Technical field
本发明涉及计算机系统设计领域和集成电路设计领域, 具体是涉及一种芯 片结构及其实现方法,特别涉及一种基于 DMA的数据压缩芯片结构及其实现方 法。 背景技术  The present invention relates to the field of computer system design and integrated circuit design, and in particular to a chip structure and an implementation method thereof, and more particularly to a data compression chip structure based on DMA and an implementation method thereof. Background technique
随着计算机技术的飞速发展, 为了满足各领域的应用需求, 高性能、 高存 储能力越来越成为计算机系统的基本特征, 不断的增加存储设备会使设备成本 极高, 并且占用大量的机房面积, 代价极高, 这就需要采用新型的处理设备, 高效的实施大数据的压缩处理。 但是另一方面也陷入了因额外的数据处理所带 来的性能损耗和 CPU 计算资源损耗, 因此采用满足带宽匹配的 DMA (Direct Memory Access, 直接内存存取) 技术, 可以有效提高系统性能, 并且尽可能的 节约 CPU的计算资源, 保障数据处理导致的系统性能影响至最低。 对存储数据进行压缩处理, 可以节约 1 倍以上的存储设备资源, 根据不同 的应用环境设备节约的效果也不同, 在数据冗余度较高的应用环境中, 存储空 间节约可达数十倍。 额外的数据压缩处理势必会造成系统性能的降低, 在不采 用性能分析优化的条件下, 其性能损耗代价难以接受, 因此压缩芯片采用 DMA 技术提高系统数据的处理效率, DMA技术采用带宽匹配的设计方法, 配合多个 压缩算法核心共同完成数据的压缩处理和存储操作。 发明内容  With the rapid development of computer technology, in order to meet the application needs of various fields, high performance and high storage capacity have become the basic characteristics of computer systems. Increasing the storage equipment will make the equipment cost extremely high and occupy a large amount of equipment room area. , the cost is very high, which requires the use of new processing equipment, efficient implementation of big data compression processing. However, on the other hand, it suffers from performance loss and CPU computing resource loss caused by additional data processing. Therefore, DMA (Direct Memory Access) technology that satisfies bandwidth matching can effectively improve system performance, and Save CPU computing resources as much as possible to ensure minimal system performance impact caused by data processing. Compressing the stored data can save more than 1 time of storage device resources. The effect of device saving is different according to different application environments. In an application environment with high data redundancy, storage space savings can reach tens of times. Additional data compression processing will inevitably lead to system performance degradation. Without performance analysis optimization, the performance loss cost is unacceptable. Therefore, the compression chip uses DMA technology to improve the processing efficiency of system data, and the DMA technology uses bandwidth matching design. The method works together with a plurality of compression algorithm cores to complete data compression processing and storage operations. Summary of the invention
为了解决现有技术的问题,本发明提供了一种基于 DMA的数据压缩芯片结 构及其实现方法, 其采用基于 DMA原理的数据压缩芯片,对大数据进行压缩处 理, 并且采用带宽匹配的 DMA技术, 配合多个压缩算法核心, 共同实现大数据 的压缩处理和存储, 提高系统存储资源的利用率。 本发明所采用的技术方案如下:  In order to solve the problems of the prior art, the present invention provides a data compression chip structure based on DMA and an implementation method thereof, which adopts a data compression chip based on the DMA principle, compresses large data, and adopts a bandwidth matching DMA technology. With multiple compression algorithm cores, it can realize the compression processing and storage of big data together, and improve the utilization of system storage resources. The technical solutions adopted by the present invention are as follows:
一种基于 DMA的数据压缩芯片结构, 包括系统到设备 DMA模块、 设备到 系统 DMA模块、发送端仲裁模块、数据存储模块接口模块和压缩算法核心模块, 其中, 系统到设备 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; 设备到系统 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; 发送端仲裁模块采用给予优先级分析、 判定、 生成的方式控制发送报文的 仲裁机制, 用于解决多报文发送传输的冲突问题; 数据存储模块接口模块采用高并行多组存储阵列的方式实现数据缓存; 压缩算法核心模块采用数据通路带宽匹配的方法, 配置多个压缩算法核心。 一种基于 DMA的数据压缩芯片结构的实现方法,是通过 PCI-E接口与系统 进行通信, 通过 DMA技术实现系统和设备的数据传输,分别设计系统到设备和 设备到系统的 DMA引擎,在发送端设计仲裁逻辑实现发送数据的高效传输,在 两个方向的 DMA模块中分别实现描述符 DMA和数据 DMA以提高数据传输效 率, 接收到的数据通过接口存储模块与数据压缩算法模块进行连接, 数据在压 缩模块进行压缩处理, 压缩后的数据再经由 DMA完成到系统的传输。 本发明的实现方法具体包括以下步骤: A data compression chip structure based on DMA, including system to device DMA module, device to system DMA module, transmitting end arbitration module, data storage module interface module and compression algorithm core module, Among them, the system-to-device DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; device-to-system DMA module adopts descriptor DMA and data DMA. Improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; the arbitration module at the sending end controls the arbitration mechanism for sending packets by giving priority analysis, judgment, and generation, which is used to solve multi-message transmission and transmission. Conflicts; Data storage module interface module uses high parallel multi-group storage array to achieve data caching; compression algorithm core module uses data path bandwidth matching method to configure multiple compression algorithm cores. A DMA-based data compression chip structure implementation method is to communicate with a system through a PCI-E interface, realize system and device data transmission through DMA technology, respectively design a system to device and device to system DMA engine, and send The end design arbitration logic realizes the efficient transmission of the transmitted data. The descriptor DMA and the data DMA are respectively implemented in the DMA module in two directions to improve the data transmission efficiency, and the received data is connected to the data compression algorithm module through the interface storage module, and the data is connected. The compression module performs compression processing, and the compressed data is then transferred to the system via DMA. The implementation method of the present invention specifically includes the following steps:
A、 系统发起读请求, 数据压缩芯片通过其发送端发送读请求和读命令, 并 通过 PCI-E端口发送至系统, 在接收端, 接收系统发送来的描述符信息, 该信 息通过系统到设备描述符 DMA方式控制传输, 即描述符的起止地址信息, 并通 过 DMA的方式传输, 最终写入描述符寄存器, 位于系统到设备描述符, 通过系 统到设备 DMA引擎控制对应描述符信息的数据传输; A, the system initiates a read request, the data compression chip sends a read request and a read command through its sender, and sends it to the system through the PCI-E port, and at the receiving end, receives descriptor information sent by the system, and the information passes through the system to the device. The descriptor DMA mode controls the transmission, that is, the start and end address information of the descriptor, and is transmitted by DMA, and finally written to the descriptor register, located in the system-to-device descriptor, and the data transmission of the corresponding descriptor information is controlled by the system-to-device DMA engine. ;
B、 接收到的 DMA数据通过系统到设备接口传输到数据存储接口模块, 以 便压缩算法核心模块进行数据压缩应用, 数据存储模块是一个高并行的分组存 储器阵列, 可支持多个数据压缩核心并行读写操作; B. The received DMA data is transmitted to the data storage interface module through the system-to-device interface, so that the compression algorithm core module performs data compression application, and the data storage module is a highly parallel packet memory array, which can support multiple data compression cores to read in parallel. Write operation
C、 压缩算法核心模块从数据存储接口模块读取数据, 进行数据压缩操作, 并收集数据信息进行压缩控制, 以提高压缩效率, 数据压缩过程中的数据匹配 信息由匹配信息模块保存, 以供压缩模块使用; D、 根据 PCI-E接口、 DMA传输能力、 数据存储接口、 压缩核心模块能力 系统的各个接口和模块的带宽确定压缩算法核心模块的核心数; 各个压缩算法 核心模块是完全独立的并行工作; C. The core module of the compression algorithm reads data from the data storage interface module, performs data compression operations, and collects data information for compression control to improve compression efficiency. The data matching information in the data compression process is saved by the matching information module for compression. Module use; D. Determine the core number of the core module of the compression algorithm according to the PCI-E interface, the DMA transmission capability, the data storage interface, and the bandwidth of each interface and module of the compressed core module capability system; the core modules of each compression algorithm are completely independent and parallel work;
E、 压缩后的压缩数据经由数据存储接口模块, 由设备到系统 DMA模块控 制完成设备到系统的数据传输;同样由设备到系统描述符 DMA方式控制多个描 述符信息的传输, 以提高系统性能, 传输的描述符信息寄存在设备到系统描述 符寄存器中, 由设备到系统 DMA方式传输;  E. The compressed data after compression is controlled by the device to the system DMA module to complete the data transmission from the device to the system; the device to system descriptor DMA mode controls the transmission of multiple descriptor information to improve system performance. The transmitted descriptor information is registered in the device to system descriptor register, and is transmitted by the device to the system DMA mode;
F、 在发送端数据发送时, 由发送端仲裁模块控制多个报文数据发送时的优 先级控制, 避免了多个报文同时发送时的冲突问题, 最后数据由 PCI-E接口完 成向系统的发送传输。 F. When the data is transmitted at the transmitting end, the arbitrating module of the transmitting end controls the priority control when multiple message data are transmitted, thereby avoiding the conflict problem when multiple messages are simultaneously transmitted, and finally the data is completed by the PCI-E interface to the system. Send transmission.
在系统工作过程中, 内部中断和外部中断由终端模块控制并实现向系统的 传输。  During system operation, internal and external interrupts are controlled by the terminal module and transmitted to the system.
本发明的技术方案中, 系统到设备 DMA 的特性, 主要是指压缩芯片通过 DMA的方式接收来自系统的数据, 包括描述符 DMA和数据 DMA, 其中数据 DMA又包括 DMA描述符和 DMA引擎, 描述符 DMA是将系统的描述符信息 采用 DMA的方式进行传递, 达到提高性能的目的, DMA描述符是指描述符寄 存器, DMA引擎是指采用 DMA的方式进行数据搬运。 设备到系统 DMA的特性,主要是指压缩芯片通过 DMA的方式向系统发送 压缩后的数据, 包括描述符 DMA和数据 DMA, 其中数据 DMA又包括 DMA 描述符和 DMA引擎, 描述符 DMA是将系统的描述符信息采用 DMA的方式进 行传递, 达到提高性能的目的, DMA描述符是指描述符寄存器, DMA引擎是 指采用 DMA的方式进行数据搬运; 发送端仲裁的特性, 主要是指采用优先级生成的仲裁方式, 控制和解决多 个报文同时发送时的冲突问题; 数据存储模块接口的特性, 主要是指采用可配的分组存储阵列, 形成高并 行存储的数据缓冲区, 为多个压缩算法核心存储系统到设备的数据和设备到系 统的压缩数据;  In the technical solution of the present invention, the system-to-device DMA feature mainly refers to the compression chip receiving data from the system by means of DMA, including descriptor DMA and data DMA, wherein the data DMA includes a DMA descriptor and a DMA engine, and the description DMA is to transfer the descriptor information of the system by DMA to improve performance. The DMA descriptor refers to the descriptor register, and the DMA engine refers to the DMA method for data transfer. The device-to-system DMA feature mainly refers to the compression chip transmitting compressed data to the system by means of DMA, including descriptor DMA and data DMA, wherein the data DMA includes a DMA descriptor and a DMA engine, and the descriptor DMA is a system. The descriptor information is transmitted by DMA to improve performance. The DMA descriptor refers to the descriptor register, the DMA engine refers to the data transfer by DMA, and the characteristics of the sender arbitration mainly refers to the priority. The generated arbitration method controls and resolves conflicts when multiple messages are sent simultaneously. The characteristics of the data storage module interface mainly refer to the use of a configurable packet storage array to form a data buffer with high parallel storage for multiple compression. The algorithm core stores system-to-device data and device-to-system compressed data;
压缩算法核心的特性, 主要是指多组压缩算法核心, 使各个模块以及接口 的带宽匹配, 保障资源的利用率, 从而达到提高系统性能的目的, 其中压缩算 法核心模块包括数据压缩模块、 压缩控制模块、 匹配信息模块, 数据压缩模块 主要实现数据的压缩功能, 压缩控制模块主要接收控制信息, 并生成压缩控制 信息, 控制压缩模块工作, 匹配信息模块主要存储压缩过程中的匹配信息, 供 压缩模块使用。 The core characteristics of the compression algorithm mainly refer to the core of multiple sets of compression algorithms, so that each module and interface The bandwidth matching ensures the utilization of resources, thereby achieving the purpose of improving system performance. The core module of the compression algorithm includes a data compression module, a compression control module, and a matching information module, and the data compression module mainly implements data compression function, and the compression control module mainly The control information is received, and the compression control information is generated, and the compression module is controlled to work. The matching information module mainly stores the matching information in the compression process for use by the compression module.
本发明的一种基于 DMA的数据压缩芯片结构及其实现方法,其结构设计所 具有的上述优点, 使得其弥补了传统软件数据压缩设计结构严重影响系统 CPU 计算资源的不足, 在不显著影响系统性能的条件下, 使其所带来的可配置性、 硬件资源节约、 压缩算法核心灵活扩展等诸多特性, 使其更适用于大数据存储、 分析系统, 因而具有非常广阔的发展前景, 具有很高的技术价值。 附图说明  The DMA-based data compression chip structure and the implementation method thereof have the above advantages, which make up for the deficiency of the traditional software data compression design structure seriously affecting the system CPU computing resources, and does not significantly affect the system. Under the condition of performance, it has many features such as configurability, hardware resource saving, and flexible expansion of the core of the compression algorithm, making it more suitable for big data storage and analysis systems, so it has a very broad development prospect. High technical value. DRAWINGS
为了更清楚地说明本发明实施例中的技术方案, 下面将对实施例描述中所 需要使用的附图作简单地介绍, 显而易见地, 下面描述中的附图仅仅是本发明 的一些实施例, 对于本领域普通技术人员来讲, 在不付出创造性劳动的前提下, 还可以根据这些附图获得其他的附图。  In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings used in the description of the embodiments will be briefly described. It is obvious that the drawings in the following description are only some embodiments of the present invention. Other drawings may also be obtained from those of ordinary skill in the art in view of the drawings.
图 1为本发明的一种基于 DMA的数据压缩芯片结构及其实现方法的逻辑结 构图;  1 is a logic structural diagram of a DMA-based data compression chip structure and an implementation method thereof according to the present invention;
图 2为本发明的一种基于 DMA的数据压缩芯片结构及其实现方法的发送端 仲裁逻辑结构图。  2 is a schematic diagram of a logic structure of a transmitting end arbitration of a DMA-based data compression chip structure and an implementation method thereof according to the present invention.
具体实施方式 detailed description
为使本发明的目的、 技术方案和优点更加清楚, 下面将结合附图对本发明 实施方式作进一步地详细描述。 实施例一  The embodiments of the present invention will be further described in detail below with reference to the accompanying drawings. Embodiment 1
一种基于 DMA的数据压缩芯片结构, 包括系统到设备 DMA模块、 设备到 系统 DMA模块、发送端仲裁模块、数据存储模块接口模块和压缩算法核心模块, 其中, 系统到设备 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; 设备到系统 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; 发送端仲裁模块采用给予优先级分析、 判定、 生成的方式控制发送报文的 仲裁机制, 用于解决多报文发送传输的冲突问题; 数据存储模块接口模块采用高并行多组存储阵列的方式实现数据缓存; 压缩算法核心模块采用数据通路带宽匹配的方法, 配置多个压缩算法核心。 实施例二 A data compression chip structure based on DMA, comprising a system to device DMA module, a device to system DMA module, a transmitting end arbitration module, a data storage module interface module, and a compression algorithm core module, wherein The system-to-device DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, descriptor information and transmission data are all DMA transmission; device-to-system DMA module uses descriptor DMA and data DMA to improve the system. The efficiency, that is, the descriptor information and the transmission data are all DMA transmission; the arbitrating module at the transmitting end controls the arbitration mechanism for sending packets by giving priority analysis, determination, and generation, and is used to solve the conflict problem of multi-message transmission and transmission. The data storage module interface module uses a high parallel multi-group storage array to implement data caching; the compression algorithm core module adopts a data path bandwidth matching method to configure multiple compression algorithm cores. Embodiment 2
一种基于 DMA的数据压缩芯片结构的实现方法,是通过 PCI-E接口与系统 进行通信, 通过 DMA技术实现系统和设备的数据传输,分别设计系统到设备和 设备到系统的 DMA引擎,在发送端设计仲裁逻辑实现发送数据的高效传输,在 两个方向的 DMA模块中分别实现描述符 DMA和数据 DMA以提高数据传输效 率, 接收到的数据通过接口存储模块与数据压缩算法模块进行连接, 数据在压 缩模块进行压缩处理, 压缩后的数据再经由 DMA完成到系统的传输。 本发明的实现方法具体包括以下步骤:  A DMA-based data compression chip structure implementation method is to communicate with a system through a PCI-E interface, realize system and device data transmission through DMA technology, respectively design a system to device and device to system DMA engine, and send The end design arbitration logic realizes the efficient transmission of the transmitted data. The descriptor DMA and the data DMA are respectively implemented in the DMA module in two directions to improve the data transmission efficiency, and the received data is connected to the data compression algorithm module through the interface storage module, and the data is connected. The compression module performs compression processing, and the compressed data is then transferred to the system via DMA. The implementation method of the present invention specifically includes the following steps:
A、 系统发起读请求, 数据压缩芯片通过其发送端发送读请求和读命令, 并 通过 PCI-E端口发送至系统, 在接收端, 接收系统发送来的描述符信息, 该信 息通过系统到设备描述符 DMA方式控制传输, 即描述符的起止地址信息, 并通 过 DMA的方式传输, 最终写入描述符寄存器, 位于系统到设备描述符, 通过系 统到设备 DMA引擎控制对应描述符信息的数据传输; A, the system initiates a read request, the data compression chip sends a read request and a read command through its sender, and sends it to the system through the PCI-E port, and at the receiving end, receives descriptor information sent by the system, and the information passes through the system to the device. The descriptor DMA mode controls the transmission, that is, the start and end address information of the descriptor, and is transmitted by DMA, and finally written to the descriptor register, located in the system-to-device descriptor, and the data transmission of the corresponding descriptor information is controlled by the system-to-device DMA engine. ;
B、 接收到的 DMA数据通过系统到设备接口传输到数据存储接口模块, 以 便压缩算法核心模块进行数据压缩应用, 数据存储模块是一个高并行的分组存 储器阵列, 可支持多个数据压缩核心并行读写操作; B. The received DMA data is transmitted to the data storage interface module through the system-to-device interface, so that the compression algorithm core module performs data compression application, and the data storage module is a highly parallel packet memory array, which can support multiple data compression cores to read in parallel. Write operation
C、 压缩算法核心模块从数据存储接口模块读取数据, 进行数据压缩操作, 并收集数据信息进行压缩控制, 以提高压缩效率, 数据压缩过程中的数据匹配 信息由匹配信息模块保存, 以供压缩模块使用; D、 根据 PCI-E接口、 DMA传输能力、 数据存储接口、 压缩核心模块能力 系统的各个接口和模块的带宽确定压缩算法核心模块的核心数; 各个压缩算法 核心模块是完全独立的并行工作; C. The core module of the compression algorithm reads data from the data storage interface module, performs data compression operations, and collects data information for compression control to improve compression efficiency. The data matching information in the data compression process is saved by the matching information module for compression. Module use; D. Determine the core number of the core module of the compression algorithm according to the PCI-E interface, the DMA transmission capability, the data storage interface, and the bandwidth of each interface and module of the compressed core module capability system; the core modules of each compression algorithm are completely independent and parallel work;
E、 压缩后的压缩数据经由数据存储接口模块, 由设备到系统 DMA模块控 制完成设备到系统的数据传输;同样由设备到系统描述符 DMA方式控制多个描 述符信息的传输, 以提高系统性能, 传输的描述符信息寄存在设备到系统描述 符寄存器中, 由设备到系统 DMA方式传输;  E. The compressed data after compression is controlled by the device to the system DMA module to complete the data transmission from the device to the system; the device to system descriptor DMA mode controls the transmission of multiple descriptor information to improve system performance. The transmitted descriptor information is registered in the device to system descriptor register, and is transmitted by the device to the system DMA mode;
F、 在发送端数据发送时, 由发送端仲裁模块控制多个报文数据发送时的优 先级控制, 避免了多个报文同时发送时的冲突问题, 最后数据由 PCI-E接口完 成向系统的发送传输。 F. When the data is transmitted at the transmitting end, the arbitrating module of the transmitting end controls the priority control when multiple message data are transmitted, thereby avoiding the conflict problem when multiple messages are simultaneously transmitted, and finally the data is completed by the PCI-E interface to the system. Send transmission.
在系统工作过程中, 内部中断和外部中断由终端模块控制并实现向系统的 传输。  During system operation, internal and external interrupts are controlled by the terminal module and transmitted to the system.
以上所述仅为本发明的较佳实施例, 并不用以限制本发明, 凡在本发明的 精神和原则之内, 所作的任何修改、 等同替换、 改进等, 均应包含在本发明的 保护范围之内。  The above is only the preferred embodiment of the present invention, and is not intended to limit the present invention. Any modifications, equivalent substitutions, improvements, etc., which are within the spirit and scope of the present invention, should be included in the protection of the present invention. Within the scope.

Claims

权利要求书 claims
1、 一种基于 DMA的数据压缩芯片结构, 包括系统到设备 DMA模块、 设 备到系统 DMA模块、发送端仲裁模块、数据存储模块接口模块和压缩算法核心 模块, 其中, 1. A data compression chip structure based on DMA, including system-to-device DMA module, device-to-system DMA module, sender arbitration module, data storage module interface module and compression algorithm core module, where,
系统到设备 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; The system-to-device DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, both descriptor information and transmission data use DMA transmission;
设备到系统 DMA模块采用描述符 DMA和数据 DMA两种方式提高系统效 率, 即描述符信息和传输数据均采用 DMA传输的方式; 发送端仲裁模块采用给予优先级分析、 判定、 生成的方式控制发送报文的 仲裁机制, 用于解决多报文发送传输的冲突问题; 数据存储模块接口模块采用高并行多组存储阵列的方式实现数据缓存; 压缩算法核心模块采用数据通路带宽匹配的方法, 配置多个压缩算法核心。 The device-to-system DMA module uses descriptor DMA and data DMA to improve system efficiency, that is, descriptor information and transmission data are both transmitted using DMA; the sending-end arbitration module uses priority analysis, judgment, and generation to control sending. The message arbitration mechanism is used to solve the conflict problem of multiple message sending and transmission; the data storage module interface module uses highly parallel multi-group storage arrays to implement data caching; the compression algorithm core module uses the data path bandwidth matching method, with multiple configurations. A compression algorithm core.
2、 一种基于 DMA的数据压缩芯片结构的实现方法, 是通过 PCI-E接口与 系统进行通信,通过 DMA技术实现系统和设备的数据传输,分别设计系统到设 备和设备到系统的 DMA 引擎, 在发送端设计仲裁逻辑实现发送数据的高效传 输, 在两个方向的 DMA模块中分别实现描述符 DMA和数据 DMA以提高数据 传输效率, 接收到的数据通过接口存储模块与数据压缩算法模块进行连接, 数 据在压缩模块进行压缩处理, 压缩后的数据再经由 DMA完成到系统的传输。 2. An implementation method of a DMA-based data compression chip structure is to communicate with the system through the PCI-E interface, realize data transmission between the system and the device through DMA technology, and design system-to-device and device-to-system DMA engines respectively. Arbitration logic is designed at the sending end to achieve efficient transmission of sent data. Descriptor DMA and data DMA are implemented in the DMA modules in both directions to improve data transmission efficiency. The received data is connected to the data compression algorithm module through the interface storage module. , the data is compressed in the compression module, and the compressed data is then transmitted to the system via DMA.
3、根据权利要求 2所述的一种基于 DMA的数据压缩芯片结构的实现方法, 其具体包括以下步骤: 3. A method for implementing a DMA-based data compression chip structure according to claim 2, which specifically includes the following steps:
A、 系统发起读请求, 数据压缩芯片通过其发送端发送读请求和读命令, 并 通过 PCI-E端口发送至系统, 在接收端, 接收系统发送来的描述符信息, 该信 息通过系统到设备描述符 DMA方式控制传输, 即描述符的起止地址信息, 并通 过 DMA的方式传输, 最终写入描述符寄存器, 位于系统到设备描述符, 通过系 统到设备 DMA引擎控制对应描述符信息的数据传输; A. The system initiates a read request. The data compression chip sends a read request and a read command through its sending end and sends them to the system through the PCI-E port. At the receiving end, the descriptor information sent by the system is received. This information is sent to the device through the system. The descriptor DMA method controls the transmission, that is, the start and end address information of the descriptor, and is transmitted through DMA, and finally written to the descriptor register, which is located in the system to device descriptor, and controls the data transmission of the corresponding descriptor information through the system to device DMA engine. ;
B、 接收到的 DMA数据通过系统到设备接口传输到数据存储接口模块, 以 便压缩算法核心模块进行数据压缩应用, 数据存储模块是一个高并行的分组存 c、 压缩算法核心模块从数据存储接口模块读取数据, 进行数据压缩操作, 并收集数据信息进行压缩控制, 以提高压缩效率, 数据压缩过程中的数据匹配 信息由匹配信息模块保存, 以供压缩模块使用; B. The received DMA data is transmitted to the data storage interface module through the system to device interface, so that the compression algorithm core module can perform data compression applications. The data storage module is a highly parallel group memory. c. The compression algorithm core module reads data from the data storage interface module, performs data compression operations, and collects data information for compression control to improve compression efficiency. The data matching information during the data compression process is saved by the matching information module for compression. module usage;
D、 根据 PCI-E接口、 DMA传输能力、 数据存储接口、 压缩核心模块能力 系统的各个接口和模块的带宽确定压缩算法核心模块的核心数; D. Determine the number of cores of the compression algorithm core module based on the PCI-E interface, DMA transmission capability, data storage interface, compression core module capability, each interface of the system and the bandwidth of the module;
E、 压缩后的压缩数据经由数据存储接口模块, 由设备到系统 DMA模块控 制完成设备到系统的数据传输;同样由设备到系统描述符 DMA方式控制多个描 述符信息的传输, 以提高系统性能, 传输的描述符信息寄存在设备到系统描述 符寄存器中, 由设备到系统 DMA方式传输; E. The compressed data passes through the data storage interface module and is controlled by the device-to-system DMA module to complete device-to-system data transmission; the device-to-system descriptor DMA method also controls the transmission of multiple descriptor information to improve system performance. , the transferred descriptor information is stored in the device to system descriptor register, and is transmitted from the device to the system in DMA mode;
F、 在发送端数据发送时, 由发送端仲裁模块控制多个报文数据发送时的优 先级控制, 避免了多个报文同时发送时的冲突问题, 最后数据由 PCI-E接口完 成向系统的发送传输。 F. When the sending end data is sent, the sending end arbitration module controls the priority control when sending multiple message data, avoiding the conflict problem when multiple messages are sent at the same time. Finally, the data is transferred to the system through the PCI-E interface. Send transmission.
4、根据权利要求 3所述的一种基于 DMA的数据压缩芯片结构的实现方法, 其在系统工作过程中, 内部中断和外部中断由终端模块控制并实现向系统的传 输。 4. A method for implementing a DMA-based data compression chip structure according to claim 3, in which during system operation, internal interrupts and external interrupts are controlled by the terminal module and transmitted to the system.
5、根据权利要求 3所述的一种基于 DMA的数据压缩芯片结构的实现方法, 所述步骤 D中, 各个压缩算法核心模块是完全独立的并行工作。 5. A method for implementing a DMA-based data compression chip structure according to claim 3, in step D, each compression algorithm core module works completely independently in parallel.
PCT/CN2014/071244 2013-11-14 2014-01-23 Dma-based data compression chip structure and implementation method therefor WO2015070539A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201310564780.8 2013-11-14
CN201310564780.8A CN103593437A (en) 2013-11-14 2013-11-14 Data compression chip structure based on DMA and implementation method thereof

Publications (1)

Publication Number Publication Date
WO2015070539A1 true WO2015070539A1 (en) 2015-05-21

Family

ID=50083578

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2014/071244 WO2015070539A1 (en) 2013-11-14 2014-01-23 Dma-based data compression chip structure and implementation method therefor

Country Status (2)

Country Link
CN (1) CN103593437A (en)
WO (1) WO2015070539A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9977619B2 (en) * 2015-11-06 2018-05-22 Vivante Corporation Transfer descriptor for memory access commands
JP6900691B2 (en) * 2017-02-07 2021-07-07 オムロン株式会社 Control and communication equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5845151A (en) * 1996-04-08 1998-12-01 Vlsi Technology, Inc. System using descriptor and having hardware state machine coupled to DMA for implementing peripheral device bus mastering via USB controller or IrDA controller
CN1552028A (en) * 2001-09-07 2004-12-01 ض� Method and apparatus for distributed direct memory access for systems on chip
CN101334761A (en) * 2008-06-11 2008-12-31 南京磐能电力科技股份有限公司 Point to multiple point data allocation DMA controller realization method
CN101937406A (en) * 2009-06-29 2011-01-05 研祥智能科技股份有限公司 Method and system for driving 1394 devices in VxWorks operating system
CN102650976A (en) * 2012-04-01 2012-08-29 中国科学院计算技术研究所 Control device and method supporting single IO (Input/Output) virtual user level interface
US20130254449A1 (en) * 2012-03-20 2013-09-26 Sung-Jung Wang Collaborative bus arbitration multiplex architecture and method of arbitration of data access based on the architecture

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101197773B (en) * 2007-10-25 2010-12-15 华为技术有限公司 Packet compressed transmitting and received de-compression method and system based on multilink
US8527689B2 (en) * 2010-10-28 2013-09-03 Lsi Corporation Multi-destination direct memory access transfer
CN102946570B (en) * 2012-09-21 2015-03-04 上海交通大学 Transmission system and method for multi-stream streaming media of self-adaption network bandwidth

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5845151A (en) * 1996-04-08 1998-12-01 Vlsi Technology, Inc. System using descriptor and having hardware state machine coupled to DMA for implementing peripheral device bus mastering via USB controller or IrDA controller
CN1552028A (en) * 2001-09-07 2004-12-01 ض� Method and apparatus for distributed direct memory access for systems on chip
CN101334761A (en) * 2008-06-11 2008-12-31 南京磐能电力科技股份有限公司 Point to multiple point data allocation DMA controller realization method
CN101937406A (en) * 2009-06-29 2011-01-05 研祥智能科技股份有限公司 Method and system for driving 1394 devices in VxWorks operating system
US20130254449A1 (en) * 2012-03-20 2013-09-26 Sung-Jung Wang Collaborative bus arbitration multiplex architecture and method of arbitration of data access based on the architecture
CN102650976A (en) * 2012-04-01 2012-08-29 中国科学院计算技术研究所 Control device and method supporting single IO (Input/Output) virtual user level interface

Also Published As

Publication number Publication date
CN103593437A (en) 2014-02-19

Similar Documents

Publication Publication Date Title
CN108062285B (en) NVMe storage device and method for accessing same
WO2018076793A1 (en) Nvme device, and methods for reading and writing nvme data
US9430432B2 (en) Optimized multi-root input output virtualization aware switch
US11169743B2 (en) Energy management method and apparatus for processing a request at a solid state drive cluster
CN108132897B (en) SRIO controller based on ZYNQ platform soft core
CN108107827B (en) SRIO control method based on ZYNQ platform soft core
CN102185833B (en) Fiber channel (FC) input/output (I/O) parallel processing method based on field programmable gate array (FPGA)
CN110297797B (en) Heterogeneous protocol conversion device and method
WO2023155526A1 (en) Data stream processing method, storage control node, and nonvolatile readable storage medium
CN103136163B (en) Protocol processor chip capable of allocating and achieving FC-AE-ASM and FC-AV protocol
CN109992543A (en) A kind of PCI-E data efficient transmission method based on ZYZQ-7000
CN114168520B (en) Optical fiber communication bus device, equipment and system
CN111177065A (en) Multi-chip interconnection method and device
CN114546913A (en) Method and device for high-speed data interaction among multiple hosts based on PCIE interface
CN104767697A (en) Aviation full-duplex switched Ethernet controller and control method thereof
CN104699641A (en) EDMA (enhanced direct memory access) controller concurrent control method in multinuclear DSP (digital signal processor) system
Qiu et al. Full-kv: Flexible and ultra-low-latency in-memory key-value store system design on cpu-fpga
WO2015070539A1 (en) Dma-based data compression chip structure and implementation method therefor
CN109743350B (en) Unloading implementation method for switching communication mode of scientific computing application image area
US20230132724A1 (en) Broadcast adapters in a network-on-chip
WO2023035722A1 (en) Rdma data transmission system, rdma data transmission method, and network device
EP3926482A1 (en) System and method for performing transaction aggregation in a network-on-chip (noc)
TW202008172A (en) Memory system
CN112148453A (en) Computing chip for privacy computation and network computing system
CN116136790A (en) Task processing method and device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14862904

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14862904

Country of ref document: EP

Kind code of ref document: A1