WO2021031137A1 - 人工智能应用开发系统、计算机设备及存储介质 - Google Patents
人工智能应用开发系统、计算机设备及存储介质 Download PDFInfo
- Publication number
- WO2021031137A1 WO2021031137A1 PCT/CN2019/101684 CN2019101684W WO2021031137A1 WO 2021031137 A1 WO2021031137 A1 WO 2021031137A1 CN 2019101684 W CN2019101684 W CN 2019101684W WO 2021031137 A1 WO2021031137 A1 WO 2021031137A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- neural network
- subsystem
- model
- artificial intelligence
- module
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/06—Physical realisation, i.e. hardware implementation of neural networks, neurons or parts of neurons
Definitions
- This application relates to the field of artificial intelligence technology, in particular to artificial intelligence application development systems, computer equipment and storage media.
- the purpose of the embodiments of the present application is to propose an artificial intelligence application development system, computer equipment, and storage medium, so as to lower the threshold of artificial intelligence application development and improve development efficiency.
- an embodiment of the present application provides an artificial intelligence application development system, which adopts the following technical solutions:
- the artificial intelligence application development system includes:
- Neural network generation subsystem used to construct, train and verify neural network models
- the neural network hardware execution subsystem is used to accept data input to the neural network model, and output the result after calculation by the neural network model;
- the deployment subsystem is used to compile the neural network model generated by the neural network generation subsystem and deploy it to the neural network hardware execution subsystem.
- the neural network generation subsystem is also used to provide training data for the neural network model and label the training data.
- neural network hardware execution subsystem is implemented based on FPGA.
- deployment subsystem includes:
- Compilation module used to analyze the neural network model and generate model structure files and data files
- the running module is used to allocate hardware computing resources according to the structure files and data files of the model
- the driving module is configured to call the corresponding hardware computing resource according to the allocation result of the running module, and the hardware computing resource includes the neural network hardware execution subsystem implemented based on FPGA.
- the allocation of hardware computing resources by the running module according to the structure file and data file of the model includes:
- the FPGA-based neural network hardware execution subsystem includes an FPGA core module and an extension module.
- the FPGA core module includes a core chip, a memory chip, a SAMTEC interface, and a JTAG interface.
- the expansion module includes a network interface, a UART port, a GPIO port, and a SAMTEC interface, and the FPGA core module and the expansion module are connected and communicated through the SAMTEC interface.
- the embodiments of the present application also provide a computer device, which adopts the following technical solutions:
- the computer device includes a memory and a processor, and a computer program is stored in the memory.
- the processor executes the computer program, the processor implements the artificial intelligence application development system described in any one of the embodiments of the present application.
- the embodiments of the present application also provide a computer-readable storage medium, which adopts the following technical solutions:
- a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, the function of the artificial intelligence application development system mentioned in any one of the embodiments of the present application is realized.
- an artificial intelligence application development system includes a neural network generation subsystem for constructing, training and verifying neural network models, and for receiving input.
- the neural network hardware execution subsystem that outputs the results after calculating the neural network model data and the neural network model is used to compile and deploy the neural network model generated by the neural network generation subsystem to the neural network hardware execution sub-system
- the deployment subsystem of the system Construct a training neural network model through the visual neural network generation subsystem, and automatically deploy the trained neural network model to the neural network hardware execution subsystem for execution through the deployment subsystem, which can lower the threshold for artificial intelligence application development and improve development effectiveness.
- FIG. 1 shows a schematic structural diagram of an embodiment of an artificial intelligence application development system 100 according to the present application
- FIG. 2 shows a schematic structural diagram of an embodiment of the deployment subsystem 103 of the artificial intelligence application development system according to the present application
- FIG. 3 shows a schematic structural diagram of an embodiment of the neural network hardware execution subsystem 102 of the artificial intelligence application development system according to the present application
- Fig. 4 is a schematic structural diagram of an embodiment of a computer device according to the present application.
- Fig. 1 shows a schematic structural diagram of an embodiment of an artificial intelligence application development system according to the present application.
- the artificial intelligence application development system 100 includes:
- the neural network generation subsystem 101 is used to construct, train and verify the neural network model.
- neural network model training is used The annotated data is iteratively trained on the neural network algorithm model built to minimize the loss of the model; neural network model verification refers to the effect verification of the trained neural network model through the verification data, and the user can upload image data and voice Data etc. are used as the input of the model, and the output result after model detection and recognition can verify the effect of the model, as well as the accuracy and speed of recognition.
- the neural network generation subsystem 101 can provide a visual interface through WEB (web page) technology to help developers quickly develop a neural network model, and train and verify the model, that is, the developer accesses the aforementioned neural network generation through a web page Subsystem 101 provides interfaces to obtain corresponding services, such as the construction of neural network models.
- WEB web page
- the construction, training and verification of visual neural network models provided through web pages can improve the development efficiency of developers.
- the neural network hardware execution subsystem 102 is configured to accept data input to the neural network model, and output the result after calculation by the neural network model.
- the neural network hardware execution subsystem 102 may be a general-purpose processor (such as a CPU, etc.) that has stored and can execute the above-mentioned neural network model 101, or may be a dedicated processor (such as an FPGA, etc.) that solidifies the above-mentioned neural network model 101.
- the neural network hardware execution subsystem 102 can also provide a network interface or other interfaces to receive and store data input from the outside world, and then input it into the neural network model for calculation, that is, extract features, classification or clustering , Regression or prediction, etc., to obtain the prediction or recognition result.
- the deployment subsystem 103 is configured to compile the neural network model generated by the neural network generation subsystem 101 and deploy it to the neural network hardware execution subsystem 102.
- the neural network model includes the neural network diagram (neural network structure) and the parameters corresponding to the structure.
- the structure of the neural network is based on the layer as the computing unit, including but not limited to convolutional layer, pooling layer, ReLU (activation function) ), fully connected layer, etc.
- each layer in the neural network structure also has a large number of parameters, including but not limited to: weight (weight), bias (bias), etc.
- the above neural network model is compiled into a model file (including the structure file and data file of the model) by a compiler (such as TVM, etc.), and the hardware resources required by the corresponding model are automatically allocated according to the above model file, for example
- the calculation unit, the cache unit, and the pipeline unit that can perform timing optimization, etc., that is, the aforementioned hardware resource is called from the aforementioned neural network hardware execution subsystem 102 and then executed.
- an artificial intelligence application development system including a neural network generation subsystem for building, training and verifying neural network models, and for receiving data input to the neural network model and passing the neural network
- the neural network hardware execution subsystem that outputs the result after model calculation is used to compile and deploy the neural network model generated by the neural network generation subsystem to the deployment subsystem of the neural network hardware execution subsystem. Construct a training neural network model through the visual neural network generation subsystem, and automatically deploy the trained neural network model to the neural network hardware execution subsystem for execution through the deployment subsystem, which can lower the threshold for artificial intelligence application development and improve development effectiveness.
- the neural network generation subsystem 101 is also used to provide training data for the neural network model and label the training data.
- the neural network generation subsystem 101 can also provide developers with functional modules such as creating a new database, uploading data, data annotation, etc., to prepare data for subsequent neural network model training, and the marked data can make the model update Train well quickly.
- neural network hardware execution subsystem 102 is implemented based on FPGA.
- FPGA has programmability. Developers can connect the logic blocks inside FPGA through programming according to their own needs, and realize the corresponding functions more freely and flexibly.
- GPU acceleration design is that the algorithm model adapts to the hardware structure
- the FPGA acceleration design is the hardware structure adapts to the algorithm model, that is, the corresponding hardware structure is designed (or called) according to the algorithm model. This accelerated design method can accelerate the depth more quickly Learn the neural network algorithm model.
- FPGAs compared to GPUs, FPGAs have a better energy efficiency ratio.
- ASIC is superior to FPGA in performance and power consumption, it needs to go through a lot of verification and physical design during design and manufacturing, which leads to a long development cycle.
- ASIC is a dedicated hardware and hardware structure designed for a certain type of application. It cannot be changed after generation.
- the deep learning neural network algorithm is currently in a stage of rapid development. For some widely used but immature application scenarios, it is very difficult to design a high-performance general-purpose ASIC to adapt to all application scenarios.
- FPGA is more suitable for accelerating the deep learning neural network algorithm model that is currently in the rapid development stage. Therefore, the neural network hardware execution subsystem 102 in this embodiment utilizes FPGA to accelerate the execution efficiency of the deep learning neural network.
- FIG. 2 shows a schematic structural diagram of an embodiment of the deployment subsystem 103 of the artificial intelligence application development system according to the present application.
- the deployment subsystem 103 includes:
- the compiling module 1031 is used to analyze the neural network model and generate the structure file and data file of the model;
- the running module 1032 is used to allocate hardware computing resources according to the structure files and data files of the model;
- the driving module 1033 is configured to call the corresponding hardware computing resource according to the allocation result of the running module, and the hardware computing resource includes the neural network hardware execution subsystem implemented based on FPGA.
- the compilation module 1031 can call a neural network compiler (such as TVM) to analyze the neural network model according to the structure of the neural network model generated by the neural network generation subsystem 101, and extract the network structure and The weight data is saved to a file to obtain the structure file and data file of the model.
- a neural network compiler such as TVM
- the file format can be json or xml, etc.; the running module 1032 can automatically allocate hardware computing resources, including calculations, according to the structure file and data file of the above neural network model Units and cache units, and pipeline units that can be optimized for timing; then through the drive module 1033, the corresponding hardware computing resources provided by the neural network hardware execution subsystem 102 implemented by FPGA are called to perform calculations and output the calculation results; among them, the neural network outputs
- the result of is the feature value, which can be understood as an abstract representation of the input picture or data, and then through some calculation methods, the abstract representation, that is, the feature value, is converted into a meaningful output, such as the picture category and the corresponding probability in the classification problem , In the detection problem, the target category, probability and coordinates contained in the picture.
- the allocation of hardware computing resources by the running module according to the structure file and data file of the model includes:
- the structure of the neural network model is based on the layer as the computing unit, including but not limited to the input layer, convolutional layer, pooling layer, ReLU (activation function), fully connected layer, etc.
- different neural networks pass Different types and different numbers of layers are combined to form a neural network structure with different functions; each layer in the neural network structure has a large number of parameters in addition to receiving the data stream output by the previous layer. These parameters include but are not limited to: weight (Weight), bias, etc.
- the network structure and parameter data of the model can be stored in files, read as node information when calculating each node of each layer, and according to the node information, the hardware resources required by the corresponding node can be dynamically allocated, for example, according to The calculation function and data type of the node, the corresponding calculation unit and storage unit are allocated for calculation operations, and the calculation results are stored in the register cache unit, which is convenient for the next layer to read quickly, saves data copy time, and accelerates the calculation of the neural network Speed, timing optimization of neural network calculations can also be done through pipeline units, which can improve the efficiency of neural network calculations.
- FIG. 3 shows a schematic structural diagram of an embodiment of the neural network hardware execution subsystem 102 of the artificial intelligence application development system according to the present application.
- the FPGA-based neural network hardware execution subsystem 102 includes an FPGA core module 1021 and an expansion module 1022.
- the FPGA core module 1021 includes a core chip 10211, a memory chip 10212, a SAMTEC interface 10214, and a 6-pin JTAG interface 10213;
- the expansion module 1022 includes a network interface 10222, a 3-pin UART port 10223, and a 40-pin GPIO port 10224 and a SAMTEC interface 10221.
- the FPGA core module 1021 and the expansion module 1022 are connected and communicated through the SAMTEC interface 10214 of the core module 1021 and the SAMTEC interface 10221 of the expansion module 1022.
- the above-mentioned core chip is used to provide computing resources and realize the calculation of the neural network
- Intel Arria 10Soc FPGA can be used as the core chip
- the memory chip is used to store the weight and other parameter data of the neural network and intermediate calculation data
- the JTAG interface can be used for data transmission between the core module 1021 and other devices, for example, can be used to download the initial program of the FPGA.
- the network interface of the expansion module 1022 is used to communicate with the host computer, program downloading and data transmission, etc., for example, it can be used to obtain the data input to the above neural network model through the network, etc.
- the network interface can be an RJ45 Ethernet interface (USB- C.
- the USB port replaces RJ45, the universality of the extended interface); the UART port is used to debug the expansion module 1022 and print related debugging information; the GPIO port can provide additional I/O interfaces for remote serial communication or control, for example, The camera or microphone is controlled through the GPIO port; the core module 1021 and the expansion module 1022 are connected and communicated through the SAMTEC interface, so that the core module 1021 can call the resources of the expansion module 1022 to implement corresponding functions.
- the implementation of all or part of the subsystems in the above-mentioned embodiment system can be accomplished by instructing relevant hardware through a computer program.
- the computer program can be stored in a computer readable storage medium.
- the aforementioned storage medium may be a non-volatile storage medium such as a magnetic disk, an optical disc, a read-only memory (Read-Only Memory, ROM), or a random access memory (Random Access Memory, RAM), etc.
- FIG. 4 is a block diagram of the basic structure of the computer device in this embodiment.
- the computer device 2 includes a memory 21, a processor 22, and a network interface 23 that communicate with each other through a system bus. It should be pointed out that the figure only shows the computer device 2 with components 21-23, but it should be understood that it is not required to implement all the components shown, and more or fewer components may be implemented instead. Among them, those skilled in the art can understand that the computer device here is a device that can automatically perform numerical calculation and/or information processing in accordance with pre-set or stored instructions. Its hardware includes but is not limited to microprocessors, dedicated Integrated Circuit (Application Specific Integrated Circuit, ASIC), Programmable Gate Array (Field-Programmable Gate Array, FPGA), Digital Processor (Digital Signal Processor, DSP), embedded devices, etc.
- ASIC Application Specific Integrated Circuit
- FPGA Field-Programmable Gate Array
- DSP Digital Processor
- the computer device may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
- the computer device can interact with the user through a keyboard, a mouse, a remote control, a touch panel, or a voice control device.
- the memory 21 includes at least one type of readable storage medium, the readable storage medium includes flash memory, hard disk, multimedia card, card type memory (for example, SD or DX memory, etc.), random access memory (RAM), static memory Random access memory (SRAM), read only memory (ROM), electrically erasable programmable read only memory (EEPROM), programmable read only memory (PROM), magnetic memory, magnetic disks, optical disks, etc.
- the memory 21 may be an internal storage unit of the computer device 2, such as a hard disk or memory of the computer device 2.
- the memory 21 may also be an external storage device of the computer device 2, for example, a plug-in hard disk, a smart media card (SMC), and a secure digital device equipped on the computer device 2.
- the memory 21 may also include both the internal storage unit of the computer device 2 and its external storage device.
- the memory 21 is generally used to store an operating system and various application software installed in the computer device 2, such as program codes of an artificial intelligence application development system.
- the memory 21 can also be used to temporarily store various types of data that have been output or will be output.
- the processor 22 may be a central processing unit (Central Processing Unit, CPU), a controller, a microcontroller, a microprocessor, or other data processing chips in some embodiments.
- the processor 22 is generally used to control the overall operation of the computer device 2.
- the processor 22 is used to run the program code or process data stored in the memory 21, for example, to run the program code of the artificial intelligence application development system.
- the network interface 23 may include a wireless network interface or a wired network interface, and the network interface 23 is generally used to establish a communication connection between the computer device 2 and other electronic devices.
- This application also provides another implementation manner, that is, a computer-readable storage medium storing a program of an artificial intelligence application development system, and the program of the artificial intelligence application development system can be at least One processor executes, so that the at least one processor executes the steps of the program of the artificial intelligence application development system described above to realize corresponding functions.
- the method of the above embodiments can be implemented by means of software plus the necessary general hardware platform. Of course, it can also be implemented by hardware, but in many cases the former is better. ⁇
- the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM, magnetic disk, The optical disc) includes several instructions to enable a terminal device (which can be a mobile phone, a computer, a server, an air conditioner, or a network device, etc.) to execute the method described in each embodiment of the present application.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Neurology (AREA)
- Stored Programmes (AREA)
Abstract
Description
Claims (10)
- 一种人工智能应用开发系统,其特征在于,包括:神经网络生成子系统,用于构建、训练并验证神经网络模型;神经网络硬件执行子系统,用于接受输入所述神经网络模型的数据,经所述神经网络模型计算后输出结果;部署子系统,用于将所述神经网络生成子系统生成的神经网络模型编译后部署到所述神经网络硬件执行子系统。
- 如权利要求1所述的系统,其特征在于,所述神经网络生成子系统还用于为所述神经网络模型提供训练数据并对训练数据进行标注。
- 如权利要求1所述的系统,其特征在于,所述神经网络硬件执行子系统基于FPGA实现。
- 如权利要求3所述的系统,其特征在于,所述部署子系统包括:编译模块,用于对所述神经网络模型进行解析并生成模型的结构文件和数据文件;运行模块,用于根据所述模型的结构文件和数据文件分配硬件计算资源;驱动模块,用于根据所述运行模块的分配结果调用对应的硬件计算资源,所述硬件计算资源包括所述基于FPGA实现的神经网络硬件执行子系统。
- 如权利要求4所述的系统,其特征在于,所述运行模块的根据所述模型的结构文件和数据文件分配硬件计算资源,包括:根据所述模型的结构文件和数据文件获取每个计算节点的信息;基于所述每个计算节点的信息给每个计算节点分配硬件计算资源。
- 如权利要求5所述的系统,其特征在于,所述基于FPGA实现的神经网络硬件执行子系统包括FPGA核心模块和扩展模块。
- 如权利要求6所述的系统,其特征在于,所述FPGA核心模块包括核心芯片、内存芯片、SAMTEC接口以及JTAG接口。
- 如权利要求7所述的系统,其特征在于,所述扩展模块包括网络接口、UART口、GPIO口以及SAMTEC接口,所述FPGA核心模块和扩展模块通过所 述SAMTEC接口进行连接和通信。
- 一种计算机设备,其特征在于,包括存储器和处理器,所述存储器中存储有计算机程序,所述处理器执行所述计算机程序时实现如权利要求1至8中任一项所述的人工智能应用开发系统的功能。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1至8中任一项所述的人工智能应用开发系统的功能。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201980066985.9A CN113168552A (zh) | 2019-08-21 | 2019-08-21 | 人工智能应用开发系统、计算机设备及存储介质 |
PCT/CN2019/101684 WO2021031137A1 (zh) | 2019-08-21 | 2019-08-21 | 人工智能应用开发系统、计算机设备及存储介质 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2019/101684 WO2021031137A1 (zh) | 2019-08-21 | 2019-08-21 | 人工智能应用开发系统、计算机设备及存储介质 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021031137A1 true WO2021031137A1 (zh) | 2021-02-25 |
Family
ID=74659981
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/101684 WO2021031137A1 (zh) | 2019-08-21 | 2019-08-21 | 人工智能应用开发系统、计算机设备及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN113168552A (zh) |
WO (1) | WO2021031137A1 (zh) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113900734B (zh) * | 2021-10-11 | 2023-09-22 | 北京百度网讯科技有限公司 | 一种应用程序文件配置方法、装置、设备及存储介质 |
CN114449295A (zh) * | 2022-01-30 | 2022-05-06 | 京东方科技集团股份有限公司 | 视频处理方法、装置、电子设备及存储介质 |
CN114282641B (zh) * | 2022-03-07 | 2022-07-05 | 麒麟软件有限公司 | 一种通用异构加速框架的构建方法 |
CN117709400A (zh) * | 2022-09-02 | 2024-03-15 | 深圳忆海原识科技有限公司 | 层次化系统、运算方法、运算装置、电子设备及存储介质 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106022472A (zh) * | 2016-05-23 | 2016-10-12 | 复旦大学 | 一种嵌入式深度学习处理器 |
CN107992299A (zh) * | 2017-11-27 | 2018-05-04 | 郑州云海信息技术有限公司 | 神经网络超参数提取转换方法、系统、装置及存储介质 |
CN108881446A (zh) * | 2018-06-22 | 2018-11-23 | 深源恒际科技有限公司 | 一种基于深度学习的人工智能平台系统 |
CN108920177A (zh) * | 2018-06-28 | 2018-11-30 | 郑州云海信息技术有限公司 | 深度学习模型配置文件到fpga配置文件的映射方法 |
CN108921289A (zh) * | 2018-06-20 | 2018-11-30 | 郑州云海信息技术有限公司 | 一种fpga异构加速方法、装置及系统 |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10664766B2 (en) * | 2016-01-27 | 2020-05-26 | Bonsai AI, Inc. | Graphical user interface to an artificial intelligence engine utilized to generate one or more trained artificial intelligence models |
CN108762768B (zh) * | 2018-05-17 | 2021-05-18 | 烽火通信科技股份有限公司 | 网络服务智能化部署方法及系统 |
-
2019
- 2019-08-21 WO PCT/CN2019/101684 patent/WO2021031137A1/zh active Application Filing
- 2019-08-21 CN CN201980066985.9A patent/CN113168552A/zh active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106022472A (zh) * | 2016-05-23 | 2016-10-12 | 复旦大学 | 一种嵌入式深度学习处理器 |
CN107992299A (zh) * | 2017-11-27 | 2018-05-04 | 郑州云海信息技术有限公司 | 神经网络超参数提取转换方法、系统、装置及存储介质 |
CN108921289A (zh) * | 2018-06-20 | 2018-11-30 | 郑州云海信息技术有限公司 | 一种fpga异构加速方法、装置及系统 |
CN108881446A (zh) * | 2018-06-22 | 2018-11-23 | 深源恒际科技有限公司 | 一种基于深度学习的人工智能平台系统 |
CN108920177A (zh) * | 2018-06-28 | 2018-11-30 | 郑州云海信息技术有限公司 | 深度学习模型配置文件到fpga配置文件的映射方法 |
Also Published As
Publication number | Publication date |
---|---|
CN113168552A (zh) | 2021-07-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021031137A1 (zh) | 人工智能应用开发系统、计算机设备及存储介质 | |
Kuchaiev et al. | Nemo: a toolkit for building ai applications using neural modules | |
US11900113B2 (en) | Data flow processing method and related device | |
US10025566B1 (en) | Scheduling technique to transform dataflow graph into efficient schedule | |
CN101231589B (zh) | 用于原位开发嵌入式软件的系统和方法 | |
CN112748914B (zh) | 一种应用程序开发方法、装置、电子设备和存储介质 | |
US11568232B2 (en) | Deep learning FPGA converter | |
WO2020062086A1 (zh) | 选择处理器的方法和装置 | |
JPH11513512A (ja) | ディジタル信号プロセッサの製造方法 | |
JP2014504767A (ja) | 正規表現をコンパイルするための方法および装置 | |
US10761822B1 (en) | Synchronization of computation engines with non-blocking instructions | |
US8725486B2 (en) | Apparatus and method for simulating a reconfigurable processor | |
CN114548384A (zh) | 具有抽象资源约束的脉冲神经网络模型构建方法和装置 | |
US10970449B2 (en) | Learning framework for software-hardware model generation and verification | |
Hao et al. | The implementation of a deep recurrent neural network language model on a Xilinx FPGA | |
US20230120227A1 (en) | Method and apparatus having a scalable architecture for neural networks | |
US20210247997A1 (en) | Method for data center storage evaluation framework simulation | |
CN113435582B (zh) | 基于句向量预训练模型的文本处理方法及相关设备 | |
CN115269204B (zh) | 一种用于神经网络编译的内存优化方法及装置 | |
CN112148276A (zh) | 用于深度学习的可视化编程 | |
TW202111610A (zh) | 定義和執行用於指定神經網路架構的程式碼之技術 | |
US7984416B2 (en) | System and method for providing class definitions in a dynamically typed array-based language | |
CN118014022A (zh) | 面向深度学习的fpga通用异构加速方法及设备 | |
CN116796678A (zh) | 一种基于解析式技术的fpga布局方法 | |
US11061654B1 (en) | Synchronization of concurrent computation engines |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19941813 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19941813 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 090822) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19941813 Country of ref document: EP Kind code of ref document: A1 |