TWI823488B - Method for implementing edge-optimized incremental learning for deep neural network and computer system - Google Patents

Method for implementing edge-optimized incremental learning for deep neural network and computer system Download PDF

Info

Publication number
TWI823488B
TWI823488B TW111127482A TW111127482A TWI823488B TW I823488 B TWI823488 B TW I823488B TW 111127482 A TW111127482 A TW 111127482A TW 111127482 A TW111127482 A TW 111127482A TW I823488 B TWI823488 B TW I823488B
Authority
TW
Taiwan
Prior art keywords
deep neural
neural network
network model
fully connected
incremental learning
Prior art date
Application number
TW111127482A
Other languages
Chinese (zh)
Other versions
TW202405700A (en
Inventor
蔡宗漢
歐海珊
Original Assignee
國立中央大學
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 國立中央大學 filed Critical 國立中央大學
Priority to TW111127482A priority Critical patent/TWI823488B/en
Application granted granted Critical
Publication of TWI823488B publication Critical patent/TWI823488B/en
Publication of TW202405700A publication Critical patent/TW202405700A/en

Links

Images

Abstract

A method for implementing edge-optimized incremental learning for deep neural network and a computer system are provided. The method is performed in a computer system that operates a deep neural network. In the method, a deep neural network model that connects with fully-connected layers and an output layer is prepared. In the beginning, the connections between the fully-connected layers and the deep neural network model are disconnected for a purpose of making the fully-connected layers not to participate in an incremental learning process and updating training parameters. Next, the fully-connected layers are cloned as new fully-connected layers, in which one of the new fully-connected layers is connected with a new output layer. The fully-connected layers and the output layer are connected with the deep neural network model. In the incremental learning process, through the deep neural network model, the newly-added incremental dataset are trained for updating the model.

Description

用於深度神經網路之邊緣優化增量學習方法與電腦系統Edge optimization incremental learning method and computer system for deep neural networks

說明書公開一種應用於深度神經網路之增量學習技術,特別是指採用分享學習機制以降低模型訓練複雜性與資源需求的用於深度神經網路之邊緣優化增量學習方法與電腦系統。The description discloses an incremental learning technology applied to deep neural networks, specifically an edge-optimized incremental learning method and computer system for deep neural networks that uses a shared learning mechanism to reduce model training complexity and resource requirements.

深度卷積神經網路已在許多應用領域得到廣泛而成功的應用,尤其是在計算機視覺領域,深度神經網路(deep neural network,DNN)已經自動化實現高準確度人臉識別與物件檢測與分類。Deep convolutional neural networks have been widely and successfully used in many application fields, especially in the field of computer vision. Deep neural networks (DNN) have automatically achieved high-accuracy face recognition and object detection and classification. .

部署深度神經網路模型主要有三個步驟,先開發出深度神經網路模型,接著給定數據集訓練已開發模型,以及為特定應用程序部署經過訓練的深度神經網路模型。深度神經網路模型在訓練階段面臨高計算量的挑戰,因為大量的深度神經網路模型參數需要較長的訓練時間,因此在大多數情況下,深度神經網路模型的訓練是在圖形處理單元(GPU)與張量處理單元(TPU)上進行的。There are three main steps to deploying a deep neural network model, first developing the deep neural network model, then training the developed model given a data set, and deploying the trained deep neural network model for a specific application. Deep neural network models face the challenge of high computational load in the training phase because a large number of deep neural network model parameters require a long training time. Therefore, in most cases, the training of deep neural network models is performed on the graphics processing unit. (GPU) and Tensor Processing Unit (TPU).

對於部署和訓練的深度神經網路模型,當數據發生變化並且需要重新訓練模型時會出現問題,這是大型神經網路的主要課題,因為大型神經網路需要很長的訓練時間。而這個問題都是通過使用遷移學習技術來解決。在所述遷移學習中,訓練後的神經網路被用作固定特徵提取器模型,並為新數據集重新訓練全連接層或是全連接層與深度神經網路模型末端的少量幾層卷積層,而這些訓練過程將使得深度神經網路模型的訓練時間減少了數倍。但是,當此類模型部署在應用程序時,它們的輸出種類是不能被改變,導致這些模型無法了解新數據,又因為種類是固定的,因此在許多應用中,數據是隨時間增量到達的。For deployed and trained deep neural network models, problems arise when the data changes and the model needs to be retrained, which is a major issue for large neural networks because they require a long training time. This problem is solved by using transfer learning technology. In the described transfer learning, the trained neural network is used as a fixed feature extractor model and the fully connected layer or the fully connected layer and a few convolutional layers at the end of the deep neural network model are retrained for the new data set. , and these training processes will reduce the training time of the deep neural network model several times. However, when such models are deployed in an application, their output categories cannot be changed, rendering these models unable to learn about new data. And because the categories are fixed, in many applications, data arrives in increments over time. .

在特定應用中,習知提出一種分揀機器人,利用機器人使用計算機視覺技術對不同物體進行分揀,分揀機器人在起初被訓練來對固定數量種類的物體進行分揀,如果使用者希望機器人的分揀種類能增添更多種類時,需要使用舊數據集(舊種類)和新數據集(欲添加種類)來重新訓練整個模型,就產生深度神經網路模需要具備增量學習能力的問題。In a specific application, Xi Zhi proposed a sorting robot that uses computer vision technology to sort different objects. The sorting robot is initially trained to sort a fixed number of types of objects. If the user wants the robot to When more categories can be added to the sorting category, the entire model needs to be retrained using the old data set (old category) and the new data set (to be added), which leads to the problem that the deep neural network model needs to have incremental learning capabilities.

種類增量學習(incremental learning)算法應具有三個屬性以實現高性能,第一,當新數據隨時到達時,增量學習算法應該能夠隨時訓練模型,不用從頭開始訓練,而是在已經訓練過的模型上進行訓練;第二,對於一個已經完成訓練的模型可以對種類進行高準確度分類;第三,當向模型添加新種類時,記憶體占用空間和計算需求應該很低或增長緩慢。Incremental learning algorithms should have three properties to achieve high performance. First, when new data arrives at any time, the incremental learning algorithm should be able to train the model at any time without starting from scratch, but after it has been trained. second, a trained model can classify categories with high accuracy; third, when new categories are added to the model, the memory footprint and computational requirements should be low or grow slowly.

增量學習技術旨在提高深度神經網路(DNN)模型在預訓練模型中添加新種類的能力,模型可以通過可用性來增量學習新種類,增量學習與常規的深度神經網路模型訓練不同,對於常規深度神經網路模型的訓練,整套訓練數據在深度神經網路模型訓練時必須全部參與訓練。如圖1顯示現行增量學習模型的運作示意圖,其中顯示有一經過訓練得出的增量學習模型100,在增量學習模型100的增量學習中,用於模型訓練的數據是在不同的時間參與訓練,如圖式中在不同時間引入不同新種類(class)的數據,示意顯示有種類一11、種類二12與種類三13。Incremental learning technology aims to improve the ability of deep neural network (DNN) models to add new categories in pre-trained models. The model can incrementally learn new categories through availability. Incremental learning is different from conventional deep neural network model training. , For the training of conventional deep neural network models, the entire set of training data must be fully involved in the training of the deep neural network model. Figure 1 shows a schematic diagram of the operation of the current incremental learning model, which shows a trained incremental learning model 100. In the incremental learning of the incremental learning model 100, the data used for model training are at different times. Participate in training. As shown in the figure, data of different new categories (classes) are introduced at different times, schematically showing that there are category one 11, category two 12 and category three 13.

然而,儘管近年來計算機影像領域有不錯的進展,但所述深度神經網路的演算法在增量學習過程中有侷限性,大多數深度神經網路演算法能夠對固定數量的種類執行訓練和推理,並且在執行深度神經網路模型的訓練階段時要求所有訓練數據都可用,若僅使用新數據對深度神經網路模型進行訓練,這會導致分類準確度下降,稱災難性的遺忘。However, despite good progress in the field of computer imaging in recent years, the deep neural network algorithms have limitations in the incremental learning process. Most deep neural network algorithms can perform training and inference on a fixed number of categories. , and all training data are required to be available when executing the training phase of the deep neural network model. If only new data is used to train the deep neural network model, this will lead to a decrease in classification accuracy, which is called catastrophic forgetting.

現有的增量學習技術試圖通過使用預練習的數據同時在模型中添加新種類數據或設計複雜的模型架構來減少災難性遺忘的影響,如此將導致高設計複雜性和記憶體儲存空間要求,使得增量學習無法在記憶體儲存空間和計算資源有限的邊緣設備(edge device)上實現。Existing incremental learning techniques attempt to reduce the impact of catastrophic forgetting by using pre-trained data while adding new types of data to the model or designing complex model architectures, which results in high design complexity and memory storage space requirements, making Incremental learning cannot be implemented on edge devices with limited memory storage space and computing resources.

因此,有鑒於習知深度神經網路在增量學習過程中需要耗費大量的記憶體空間與計算資源,有需要提出新的學習技術,以能在儲存空間與計算資源有限的裝置上可以實施增量學習法。Therefore, in view of the fact that deep neural networks consume a large amount of memory space and computing resources during the incremental learning process, it is necessary to propose new learning technologies that can implement incremental learning on devices with limited storage space and computing resources. Quantitative learning method.

為了在儲存空間與計算資源有限的裝置上實現增量學習法,揭露書提出一種用於深度神經網路之邊緣優化增量學習方法,而此方法可運行於一電腦系統中。In order to implement incremental learning methods on devices with limited storage space and computing resources, the disclosure proposes an edge-optimized incremental learning method for deep neural networks, and this method can be run on a computer system.

在所提出的運行深度神經網路的電腦系統中,電腦系統包括具有特定算力與儲存資源的處理器與記憶體,其中執行的邊緣優化增量學習方法流程包括,預備一深度神經網路模型,包括一或多個全連接層以及一輸出層,先斷開一或多個全連接層與深度神經網路模型原本的連結,使一或多個全連接層不參與增量學習訓練參數更新,之後複製一或多個全連接層為一或多個新全連接層,並使得其中之一新全連接層連接到一新輸出層。接著使一或多個新全連接層與新輸出層連接到深度神經網路模型,在一增量學習訓練過程中,通過深度神經網路模型,以新添加的增量數據集參與訓練,並更新深度神經網路模型。In the proposed computer system for running a deep neural network, the computer system includes a processor and a memory with specific computing power and storage resources. The edge optimization incremental learning method process executed includes preparing a deep neural network model. , including one or more fully connected layers and an output layer, first disconnect the original connection between the one or more fully connected layers and the deep neural network model, so that the one or more fully connected layers do not participate in incremental learning training parameter updates , and then copy one or more fully connected layers into one or more new fully connected layers, and connect one of the new fully connected layers to a new output layer. Then connect one or more new fully connected layers and new output layers to the deep neural network model. In an incremental learning training process, the deep neural network model participates in training with the newly added incremental data set, and Update deep neural network models.

進一步地,所述新輸出層的大小等於在一給定時間內所要增量學習的種類數量。Further, the size of the new output layer is equal to the number of categories to be incrementally learned in a given time.

並且,在所述增量學習訓練過程中,僅有新添加的增量數據集參與訓練,而原先深度神經網路模型已完成訓練的數據集不參與新的訓練,針對不同的應用與不同數量的種類可以在每次增量學習訓練過程中逐步添加到該深度神經網路模型中。Moreover, during the incremental learning and training process, only the newly added incremental data sets participate in the training, while the data sets that have been trained by the original deep neural network model do not participate in the new training. For different applications and different quantities, categories can be gradually added to the deep neural network model during each incremental learning training process.

進一步地,當完成增量學習並更新深度神經網路模型後,在一模型推理階段,先前與深度神經網路模型斷開連結的一或多個全連接層再次連接更新後的深度神經網路模型,使得原先學習種類的一或多個全連接層以及增量學習種類的一或多個新全連接層共享更新後的深度神經網路模型。Further, after incremental learning is completed and the deep neural network model is updated, in a model inference stage, one or more fully connected layers that were previously disconnected from the deep neural network model are reconnected to the updated deep neural network. model, so that one or more fully connected layers of the original learning type and one or more new fully connected layers of the incremental learning type share the updated deep neural network model.

為使能更進一步瞭解本發明的特徵及技術內容,請參閱以下有關本發明的詳細說明與圖式,然而所提供的圖式僅用於提供參考與說明,並非用來對本發明加以限制。In order to further understand the features and technical content of the present invention, please refer to the following detailed description and drawings of the present invention. However, the drawings provided are only for reference and illustration and are not used to limit the present invention.

以下是通過特定的具體實施例來說明本發明的實施方式,本領域技術人員可由本說明書所公開的內容瞭解本發明的優點與效果。本發明可通過其他不同的具體實施例加以施行或應用,本說明書中的各項細節也可基於不同觀點與應用,在不悖離本發明的構思下進行各種修改與變更。另外,本發明的附圖僅為簡單示意說明,並非依實際尺寸的描繪,事先聲明。以下的實施方式將進一步詳細說明本發明的相關技術內容,但所公開的內容並非用以限制本發明的保護範圍。The following is a specific example to illustrate the implementation of the present invention. Those skilled in the art can understand the advantages and effects of the present invention from the content disclosed in this specification. The present invention can be implemented or applied through other different specific embodiments, and various details in this specification can also be modified and changed based on different viewpoints and applications without departing from the concept of the present invention. In addition, the drawings of the present invention are only simple schematic illustrations and are not depictions based on actual dimensions, as is stated in advance. The following embodiments will further describe the relevant technical content of the present invention in detail, but the disclosed content is not intended to limit the scope of the present invention.

應當可以理解的是,雖然本文中可能會使用到“第一”、“第二”、“第三”等術語來描述各種元件或者訊號,但這些元件或者訊號不應受這些術語的限制。這些術語主要是用以區分一元件與另一元件,或者一訊號與另一訊號。另外,本文中所使用的術語“或”,應視實際情況可能包括相關聯的列出項目中的任一個或者多個的組合。It should be understood that although terms such as “first”, “second” and “third” may be used herein to describe various components or signals, these components or signals should not be limited by these terms. These terms are mainly used to distinguish one component from another component, or one signal from another signal. In addition, the term "or" used in this article shall include any one or combination of more of the associated listed items depending on the actual situation.

因為增量學習方法在訓練階段需要大量的計算能力,因此鑒於在計算能力、記憶體儲存空間與能耗的限制下難以在邊緣設備上進行此類算法的訓練,揭露書提出一種用於深度神經網路之邊緣優化增量學習方法與電腦系統,通過簡化的訓練算法降低整體增量學習算法的複雜性,使得增量學習算法可在任何給定的深度神經網路模型的部分參數中執行學習。為了對深度神經網路模型的部分參數進行訓練,利用一種遷移學習的概念設計增量學習算法。Because incremental learning methods require a large amount of computing power during the training phase, it is difficult to train such algorithms on edge devices due to limitations in computing power, memory storage space, and energy consumption. The disclosure proposes a method for deep neural The edge of the network optimizes the incremental learning method and computer system, reducing the complexity of the overall incremental learning algorithm through a simplified training algorithm, so that the incremental learning algorithm can perform learning in some parameters of any given deep neural network model . In order to train some parameters of the deep neural network model, an incremental learning algorithm is designed using a transfer learning concept.

遷移學習是一種僅重新訓練模型的全連接層或少數二維卷積層與全連接層的方法,與重新訓練整個深度神經網路模型相比,遷移學習使得模型在訓練時需要更少的記憶體儲存空間。遷移學習的概念可參考圖2顯示的示意圖。Transfer learning is a method that only retrains the fully connected layer or a few two-dimensional convolutional layers and fully connected layers of the model. Compared with retraining the entire deep neural network model, transfer learning makes the model require less memory during training. body storage space. The concept of transfer learning can be referred to the schematic diagram shown in Figure 2.

圖2顯示的來源數據集(source dataset)21為已存在的知識或已學習的數據集,通過神經網路學習演算法210後形成一個神經網路模型。之後提出要學習與訓練的目標數據集(target dataset)22,在進行新神經網路學習演算法220時,可以採用遷移學習,可以把已經訓練好的神經網路模型參數遷移到新神經網路學習演算法220上,使得到新神經網路模型不需要從零開始訓練一個新的模型。The source dataset 21 shown in Figure 2 is existing knowledge or a learned dataset, which forms a neural network model through the neural network learning algorithm 210. Then a target data set (target dataset) to be learned and trained 22 is proposed. When performing the new neural network learning algorithm 220, transfer learning can be used to transfer the trained neural network model parameters to the new neural network. The learning algorithm 220 makes it possible to obtain a new neural network model without training a new model from scratch.

經參考所述遷移學習的概念,揭露書提出的用於深度神經網路之邊緣優化增量學習方法是基於克隆(cloning)分享的概念實作共享學習(learning with sharing,LwS)的增量學習方法,所述克隆是指複製已經在某些數據上訓練的深度神經網路模型中的全連接層,如此,當應用在用於深度神經網路之邊緣優化增量學習方法中執行全連接(fully-connected)層的克隆和共享,能在保留原有種類的知識下,同時在神經網路模型中逐步添加新種類,其中無需儲存來自先前預訓練種類的數據,與習知僅專注於實現高準確度的增量學習方法相比,所提出的邊緣優化增量學習方法可實現高速模型訓練和推理以及低能耗的目的,並能維持傳統算法或是更高的分類準確度。With reference to the concept of transfer learning, the edge optimization incremental learning method for deep neural networks proposed in the disclosure is based on the concept of cloning and sharing to implement incremental learning of learning with sharing (LwS). Method, the cloning refers to copying the fully connected layer in the deep neural network model that has been trained on some data, so that when the application performs the fully connected layer in the edge optimization incremental learning method for deep neural network ( The cloning and sharing of fully-connected layers can gradually add new categories to the neural network model while retaining the knowledge of the original categories. There is no need to store data from previously pre-trained categories, and the knowledge only focuses on implementation. Compared with high-accuracy incremental learning methods, the proposed edge-optimized incremental learning method can achieve high-speed model training and inference and low energy consumption, and can maintain traditional algorithms or higher classification accuracy.

根據邊緣優化增量學習方法的實施例,在深度神經網路中,分類器與特徵提取架構可用以實現多種不同類型的深度神經網路模型,並參考上述遷移學習中共享卷積層以及只訓練全連結層的方法,以應付新種類數據隨時間增加而有重複訓練的需求。進一步地,所述用於深度神經網路之邊緣優化增量學習方法的主要特徵之一是,在不需要先前訓練數據的前提下保持對原先種類的分類能力,當所述增量學習算法使深度神經網路模型學習新種類時,並不使用深度神經網路模型已經訓練過的舊數據。According to embodiments of the edge-optimized incremental learning method, in deep neural networks, classifiers and feature extraction architectures can be used to implement multiple different types of deep neural network models, and refer to the above-mentioned transfer learning of shared convolutional layers and only training all The connection layer method is used to cope with the need for repeated training as new types of data increase over time. Furthermore, one of the main features of the edge optimization incremental learning method for deep neural networks is to maintain the classification ability of the original categories without the need for previous training data. When the incremental learning algorithm uses When the deep neural network model learns new categories, it does not use the old data that the deep neural network model has been trained on.

邊緣優化增量學習方法所應用的遷移學習的優點是,當訓練新種類數據集(如圖2顯示的目標數據集22)時,只需要重新訓練全連接層或部分深度神經網路模型,邊緣優化增量學習方法基於遷移學習中的共享屬性的增量學習訓練方法,實現共享學習方法,其中架構可參考圖3所示在訓練階段的邊緣優化增量學習方法的神經網路架構的實施例示意圖。The advantage of transfer learning applied by the edge optimization incremental learning method is that when training a new kind of data set (such as the target data set 22 shown in Figure 2), only the fully connected layer or part of the deep neural network model needs to be retrained, and the edge The optimized incremental learning method is based on the incremental learning training method of shared attributes in transfer learning to implement the shared learning method. The architecture can be referred to the embodiment of the neural network architecture of the edge-optimized incremental learning method in the training phase shown in Figure 3. Schematic diagram.

根據圖3顯示的神經網路架構的實施例示意圖,並請參考圖4顯示的邊緣優化增量學習方法中訓練階段實施例流程圖。預備基礎的深度神經網路模型30,即特徵提取層,其中包括一或多個全連接層,圖中示意表示為全連接層FC1與FC2,以及輸出層301,一開始,將深度神經網路模型30的所有層都被凍結(步驟S401),特別為模型所連結的一或多個全連接層,使之不參與增量學習訓練參數更新,圖例顯示斷開全連接層FC1、FC2與深度神經網路模型30原本的連結(步驟S403)。According to the schematic diagram of an embodiment of the neural network architecture shown in Figure 3, please refer to the flow chart of the training phase in the edge optimization incremental learning method shown in Figure 4. Prepare the basic deep neural network model 30, that is, the feature extraction layer, which includes one or more fully connected layers, schematically represented in the figure as fully connected layers FC1 and FC2, and the output layer 301. Initially, the deep neural network All layers of model 30 are frozen (step S401), especially one or more fully connected layers connected to the model, so that they do not participate in incremental learning training parameter updates. The legend shows that the fully connected layers FC1, FC2 and depth are disconnected. The original connection of the neural network model 30 (step S403).

接著,複製原始的深度神經網路模型30的一或多個全連接層,形成一或多個新全連接層,如圖例中將全連接層FC1、FC2複製為新全連接層FC1’與FC2’(步驟S405),其中之一連接到新輸出層,如此例顯示複製後的新全連接層FC2’連接到新輸出層303。其中新輸出層303的大小等於在給定時間內所要增量學習的種類數量。舉例來說,添加兩個種類將導致新全連接層FC2’連接到兩個節點的新輸出層303。Next, one or more fully connected layers of the original deep neural network model 30 are copied to form one or more new fully connected layers. In the figure, the fully connected layers FC1 and FC2 are copied into new fully connected layers FC1' and FC2. ' (step S405), one of them is connected to the new output layer. This example shows that the copied new fully connected layer FC2' is connected to the new output layer 303. The size of the new output layer 303 is equal to the number of categories to be incrementally learned within a given time. For example, adding two categories will result in a new fully connected layer FC2' connected to a new output layer 303 of both nodes.

之後,複製後的新全連接層FC1’與FC2’連同新插入的新輸出層303連接到深度神經網路模型30的凍結層(即特徵提取層)(步驟S407)。值得一提的是,上述步驟中複製原來的全連接層FC1與FC2,而不是用隨機值實例化全連接層,這是因為增量添加的種類與基本模型(即深度神經網路模型30)的種類有相似的特徵。與使用隨機啟動的全連接層進行訓練相比,複製全連接層的方式有助於深度神經網路模型30快速實現高分類準確度。Afterwards, the copied new fully connected layers FC1' and FC2', together with the newly inserted new output layer 303, are connected to the frozen layer (ie, feature extraction layer) of the deep neural network model 30 (step S407). It is worth mentioning that in the above steps, the original fully connected layers FC1 and FC2 are copied instead of instantiating the fully connected layers with random values. This is because of the type of incremental addition and the basic model (i.e., deep neural network model 30) types have similar characteristics. Compared with using randomly started fully connected layers for training, replicating the fully connected layers helps the deep neural network model 30 quickly achieve high classification accuracy.

接著,複製後的新全連接層FC1’與FC2’與新輸出層303執行增量學習訓練(步驟S409),持續更新深度神經網路模型30,以實現適應新種類的深度神經網路模型30(步驟S411)。在增量學習訓練過程中,僅有新的欲添加的增量數據集參與訓練,而原先深度神經網路模型30已完成訓練的數據集不參與新的訓練。值得一提的是,對於不同的應用、不同數量的種類可以在每次增量學習訓練過程中逐步添加到深度神經網路模型30中,並在向深度神經網路模型30添加新種類時,從初始的深度神經網路模型30中複製其中全連接層FC1與FC2,並可以有可變數量的種類增量添加到深度神經網路模型30中。Then, the copied new fully connected layers FC1' and FC2' and the new output layer 303 perform incremental learning training (step S409), and continuously update the deep neural network model 30 to achieve a deep neural network model 30 adapted to new types. (Step S411). During the incremental learning training process, only new incremental data sets to be added participate in the training, and the data sets that have been trained by the original deep neural network model 30 do not participate in new training. It is worth mentioning that for different applications, different numbers of categories can be gradually added to the deep neural network model 30 during each incremental learning training process, and when adding new categories to the deep neural network model 30, The fully connected layers FC1 and FC2 are copied from the initial deep neural network model 30, and a variable number of types can be added to the deep neural network model 30 in increments.

圖5接著顯示在推理階段的邊緣優化增量學習方法的神經網路架構實施例示意圖,同時參考圖6所示邊緣優化增量學習方法中推理階段的實施例流程圖。Figure 5 then shows a schematic diagram of an embodiment of the neural network architecture of the edge optimization incremental learning method in the inference stage, while referring to the flow chart of an embodiment of the inference stage of the edge optimization incremental learning method shown in Figure 6 .

在上述訓練階段完成後,通過增量學習實現更新深度神經網路模型50,接著在模型推理階段,之前與深度神經網路模型斷開連結的一或多個全連接層,即上述實施例顯示的全連接層FC1與FC2,將再次連結到更新後深度神經網路模型50(步驟S601),使得原先學習種類的全連接層FC1與FC2以及增量學習種類的新全連接層FC1’與FC2’可以共享更新後深度神經網路模型50。在增量學習過程得出的更新後深度神經網路模型50將可執行分類原先學習的種類以及新添加的種類(步驟S603)。After the above training phase is completed, the deep neural network model 50 is updated through incremental learning, and then in the model inference phase, one or more fully connected layers that were previously disconnected from the deep neural network model, as shown in the above embodiment The fully connected layers FC1 and FC2 will be connected to the updated deep neural network model 50 again (step S601), so that the original learning type fully connected layers FC1 and FC2 and the incremental learning type new fully connected layers FC1' and FC2 'The updated deep neural network model 50 can be shared. The updated deep neural network model 50 obtained in the incremental learning process will be able to perform classification of the originally learned categories and the newly added categories (step S603).

由於神經網路中的深度神經網路模型即一種特徵提取層,在增量學習過程中將凍結而不參與增量學習訓練參數的更新,因此在增量學習訓練中斷開連接後,即便經歷增量學習,最後產生的更新後深度神經網路模型50依然能夠保持分類原先學習的種類的能力。如果在之後欲增加更多種類,仍可採用上述相同增量學習方法。在增量學習訓練階段完成後,還採用相同模型推理方法。進一步地,隨著更多種類被添加到深度神經網路模型中,通過揭露書提出的用於深度神經網路之邊緣優化增量學習方法,重複上述流程,持續更新深度神經網路模型,使之適應新舊種類的應用。Since the deep neural network model in the neural network is a feature extraction layer, it will be frozen during the incremental learning process and will not participate in the update of the incremental learning training parameters. Therefore, after being disconnected during the incremental learning training, even after experiencing Through incremental learning, the finally updated deep neural network model 50 can still maintain the ability to classify the originally learned categories. If you want to add more categories later, you can still use the same incremental learning method mentioned above. After the incremental learning training phase is completed, the same model inference method is also used. Further, as more types are added to the deep neural network model, through the edge optimization incremental learning method for deep neural networks proposed in the disclosure, the above process is repeated to continuously update the deep neural network model, so that It adapts to new and old types of applications.

所述用於深度神經網路之邊緣優化增量學習方法執行於一電腦系統中,可參考圖7,其中顯示運行一深度神經網路的電腦系統70,較佳地可以為算力較低與記憶體儲存空間不大的邊緣裝置,其中包括可運用特定算力與儲存資源的處理器71與記憶體72,例如在Nvidia Jetson TX-2主機板設有ARM A57處理器、4GB RAM以及32GB的儲存空間,通過以上實施例流程可知,在所述遷移學習的概念下可簡化電腦系統中深度神經網路模型700的訓練過程,並降低增量學習方法在訓練階段算力與記憶體需求,更新後的深度神經網路模型700可以同時適應舊種類數據701與新種類數據702。The edge-optimized incremental learning method for deep neural networks is executed in a computer system. Refer to Figure 7 , which shows a computer system 70 running a deep neural network. Preferably, it can be a computer system with low computing power and Edge devices with small memory storage space include processors 71 and memory 72 that can use specific computing power and storage resources. For example, the Nvidia Jetson TX-2 motherboard has an ARM A57 processor, 4GB RAM, and 32GB Storage space, as can be seen from the above embodiment process, the concept of transfer learning can simplify the training process of the deep neural network model 700 in the computer system, and reduce the computing power and memory requirements of the incremental learning method in the training phase, update The resulting deep neural network model 700 can adapt to both old type data 701 and new type data 702 at the same time.

在完成增量學習並更新深度神經網路模型後,最後的結果可於一些著名的深度神經網路模型上驗證成功,如MobileNetv3、FBNet、MNasNet等,還與知名的數據集(如Cifar-100、Caltech-101)比較其他最先進的增量學習方法,如LwS、LwM、PNS、LwF與ICaRL等,都獲得驗證。After completing incremental learning and updating the deep neural network model, the final results can be successfully verified on some famous deep neural network models, such as MobileNetv3, FBNet, MNasNet, etc., and also with well-known data sets (such as Cifar-100 , Caltech-101), compared with other state-of-the-art incremental learning methods, such as LwS, LwM, PNS, LwF and ICaRL, etc., all have been verified.

表一顯示通過用於深度神經網路之邊緣優化增量學習方法實作的深度神經網路模型在每次增量學習的步驟準確度下降比率(%),在Cifar-100數據集上,分別是每批增量種類大小設置(B)為5、10、20和50的種類,4個不同場景下進行增量學習的準確度分析比較。其中顯示了揭露書提出的方法採用的共享學習(LwS)應用在不同深度神經網路模型上與其他增量學習方法在Caltech-101數據集每次增量10個類別的準確度比較後的可用性。Table 1 shows the accuracy degradation rate (%) at each incremental learning step of the deep neural network model implemented by the edge-optimized incremental learning method for deep neural networks, on the Cifar-100 dataset, respectively. The incremental category size of each batch (B) is set to 5, 10, 20 and 50 categories, and the accuracy analysis and comparison of incremental learning is performed in 4 different scenarios. It shows the usability of shared learning (LwS) applied by the method proposed in the disclosure on different deep neural network models compared with other incremental learning methods on the Caltech-101 data set with accuracy of 10 categories per increment. .

表一: 方法 B=5 B=10 B=20 B=50 LwM[14] - 6.84 8.5 9.8 LwF[12] 3.9 7 12.25 20.8 iCaRL[10] 3.33 5.43 10.15 14.5 DMC[25] 3.6 5.25 11.25 15 EWC[30] 4.1 7.26 14.12 23 MAS[31] 4.0 7.77 14 22 PODNET[26] 2.65 4.67 8.12 16.9 LwS(MNV3L) 3.4 5.8 8.72 19.72 LwS(FBNet) 3.1 5.3 9.7 16.3 LwS(MNasNet) 3.2 5.3 9.21 17.03 Table I: method B=5 B=10 B=20 B=50 LwM[14] - 6.84 8.5 9.8 LwF[12] 3.9 7 12.25 20.8 iCaRL[10] 3.33 5.43 10.15 14.5 DMC[25] 3.6 5.25 11.25 15 EWC[30] 4.1 7.26 14.12 twenty three MAS[31] 4.0 7.77 14 twenty two PODNET[26] 2.65 4.67 8.12 16.9 LwS(MNV3L) 3.4 5.8 8.72 19.72 LwS(FBNet) 3.1 5.3 9.7 16.3 LwS(MNasNet) 3.2 5.3 9.21 17.03

就模型儲存空間而言,揭露書提出的用於深度神經網路之邊緣優化增量學習方法所採用的共享學習(LwS)技術與最先進的技術DMC和LwM相比有所減少,因為增量訓練的方法不同,LwM在訓練過程中需要2倍儲存空間進行模型訓練;DMC在訓練時,需要整個模型重新訓練,導致也需要2倍儲存空間。表二顯示不同深度神經網路模型隨著增量訓練模型參數的比較。In terms of model storage space, the shared learning (LwS) technology used in the edge optimization incremental learning method for deep neural networks proposed in the disclosure is reduced compared with the state-of-the-art technologies DMC and LwM, because the incremental The training methods are different. LwM requires 2 times the storage space for model training during the training process; DMC requires the entire model to be retrained during training, which also requires 2 times the storage space. Table 2 shows the comparison of different deep neural network models with incremental training of model parameters.

表二: 種類增加的數量 ResNet34 [13] MNV3-L/ MNasNet FBNet ResNet34[13]與MNV3-L/MNasNet(%)的差異 ResNet34[13]與FBNet(%)的差異 10 8518998 789770 1150218 90.7 86.49 20 8524128 1579540 2300436 81.5 73.0 30 8529258 2369310 3450654 72.2 59.4 Table II: Increased number of types ResNet34 [13] MNV3-L/ MNasNet FBNet The difference between ResNet34[13] and MNV3-L/MNasNet(%) The difference between ResNet34[13] and FBNet(%) 10 8518998 789770 1150218 90.7 86.49 20 8524128 1579540 2300436 81.5 73.0 30 8529258 2369310 3450654 72.2 59.4

表三則顯示了其他先進技術在訓練過程中所需要最低記憶體空間需求的比較,還包括模型中那些層需要參與增量學習訓練。其中假設每個參數的大小為32位浮點數。Table 3 shows a comparison of the minimum memory space requirements required by other advanced technologies during the training process, including which layers in the model need to participate in incremental learning training. Which assumes the size of each parameter is a 32-bit floating point number.

表三:   與基礎深度神經網路模型比較在訓練中最低記憶體要求 那些層需要參與增量訓練? A:全部模型 B:全連接層與卷積層 C:僅全連接層 LwM[14], LwF[12], DMC[25], LSIL[22], GEM[27], PODNET[26], DCM[28], PNN[24], iCaRL[10] 2x A ST[29] 2x B PNS[13] 1.3-1.6x B PackNet[43] 1.05-1.6x A, B(依照任務而定) MAS[31], EWC[30], DAN[21], FTDNN[32] 1.05-1.6x B, C(依照任務而定) LwS 1.05-1.1x C Table 3: Minimum memory requirements for training compared to base deep neural network models Which layers need to participate in incremental training? A: All models B: Fully connected layers and convolutional layers C: Only fully connected layers LwM[14], LwF[12], DMC[25], LSIL[22], GEM[27], PODNET[26], DCM[28], PNN[24], iCaRL[10] 2x A ST[29] 2x B PNS[13] 1.3-1.6x B PackNet[43] 1.05-1.6x A, B (depending on the task) MAS[31], EWC[30], DAN[21], FTDNN[32] 1.05-1.6x B, C (depending on the task) wxya 1.05-1.1x C

表四顯示隨著每次增量所添加2、5、10類進行增量學習,用於深度神經網路之邊緣優化增量學習方法所採用的共享學習(LwS)技術與其他先進技術針對訓練時間(training time)、能耗(power consumption)的比較分析。其中所有方法的訓練時間、能耗均以共享學習(LwS)作標準化。Table 4 shows that as 2, 5, and 10 categories are added in each increment for incremental learning, the shared learning (LwS) technology used in the edge optimization incremental learning method of deep neural networks and other advanced technologies are targeted at training. Comparative analysis of time (training time) and energy consumption (power consumption). The training time and energy consumption of all methods are standardized by shared learning (LwS).

表四:   B=2 B=5 B=10 方法 訓練時間 能耗 訓練時間 能耗 訓練時間 能耗 PNS[13] 1.4x 1.6x 1.2x 1.4x 1.2x 1.5x iCaRL[10] 3.1x 2.4x 5.6x 4.4x 10.8x 9.0x EWC[30] 24.4x 29.0x 24.1x 29.0x 22.8x 28.9x PNDNET[26] 6.4x 7.3x 4.2x 6.6x 4.3x 7.1x LwF[12] 1.3x 1.2x 1.2x 1.2x 1.3x 1.3x LwS 1x 1x 1x 1x 1x 1x Table 4: B=2 B=5 B=10 method training time Energy consumption training time Energy consumption training time Energy consumption PNS[13] 1.4x 1.6x 1.2x 1.4x 1.2x 1.5x iCaRL[10] 3.1x 2.4x 5.6x 4.4x 10.8x 9.0x EWC[30] 24.4x 29.0x 24.1x 29.0x 22.8x 28.9x PNDNET[26] 6.4x 7.3x 4.2x 6.6x 4.3x 7.1x LwF[12] 1.3x 1.2x 1.2x 1.2x 1.3x 1.3x wxya 1x 1x 1x 1x 1x 1x

綜上所述,根據揭露書提出的用於深度神經網路之邊緣優化增量學習方法與電腦系統的實施方式,在利用深度神經網路模型實現新的增量學習技術時,採用一種分享學習(learning with sharing)機制,可以達成的技術功效包括可降低模型訓練複雜性(complexity)、減少訓練時間、模型運算能耗和記憶體儲存空間需求,同時在增量學習過程(incremental learning process)中實現高準確度。In summary, according to the implementation of the edge-optimized incremental learning method and computer system for deep neural networks proposed in the disclosure, a shared learning method is adopted when using deep neural network models to implement new incremental learning technologies. (learning with sharing) mechanism, the technical benefits that can be achieved include reducing model training complexity, reducing training time, model computing energy consumption and memory storage space requirements, while at the same time in the incremental learning process (incremental learning process) Achieve high accuracy.

以上所公開的內容僅為本發明的優選可行實施例,並非因此侷限本發明的申請專利範圍,所以凡是運用本發明說明書及圖式內容所做的等效技術變化,均包含於本發明的申請專利範圍內。The contents disclosed above are only preferred and feasible embodiments of the present invention, and do not limit the scope of the patent application of the present invention. Therefore, all equivalent technical changes made by using the description and drawings of the present invention are included in the application of the present invention. within the scope of the patent.

100:增量學習模型100: Incremental learning model

11:種類一11: Type one

12:種類二12:Type two

13:種類三13: Type three

21:來源數據集21: Source data set

210:神經網路學習演算法210: Neural network learning algorithm

22:目標數據集22: Target data set

220:新神經網路學習演算法220: New Neural Network Learning Algorithm

30:深度神經網路模型30:Deep neural network model

FC1、FC2:全連接層FC1, FC2: fully connected layer

301:輸出層301:Output layer

303:新輸出層303: New output layer

FC1’、FC2’:新全連接層FC1’, FC2’: new fully connected layer

50:更新後深度神經網路模型50:Updated deep neural network model

701:舊種類數據701:Old type data

702:新種類數據702: New type of data

70:電腦系統70:Computer system

71:處理器71: Processor

72:記憶體72:Memory

步驟S401~S411:邊緣優化增量學習方法中訓練階段的流程Steps S401 to S411: The process of the training phase in the edge optimization incremental learning method

步驟S601~S603:邊緣優化增量學習方法中推理階段的流程Steps S601 to S603: Process of the inference phase in the edge optimization incremental learning method

圖1顯示現行增量學習模型的運作示意圖; 圖2顯示遷移學習的概念示意圖; 圖3顯示在訓練階段的邊緣優化增量學習方法的神經網路架構實施例示意圖; 圖4顯示邊緣優化增量學習方法中訓練階段的實施例流程圖; 圖5顯示在推理階段的邊緣優化增量學習方法的神經網路架構實施例示意圖; 圖6顯示邊緣優化增量學習方法中推理階段的實施例流程圖;以及 圖7顯示運行邊緣優化增量學習方法的電腦系統實施例示意圖。 Figure 1 shows the operation diagram of the current incremental learning model; Figure 2 shows a conceptual diagram of transfer learning; Figure 3 shows a schematic diagram of an embodiment of the neural network architecture of the edge optimization incremental learning method in the training stage; Figure 4 shows an embodiment flow chart of the training phase in the edge optimization incremental learning method; Figure 5 shows a schematic diagram of an embodiment of the neural network architecture of the edge optimization incremental learning method in the inference stage; Figure 6 shows an embodiment flow chart of the inference phase in the edge optimization incremental learning method; and Figure 7 shows a schematic diagram of an embodiment of a computer system running the edge optimization incremental learning method.

21:來源數據集 21: Source data set

210:神經網路學習演算法 210: Neural network learning algorithm

22:目標數據集 22: Target data set

220:新神經網路學習演算法 220: New Neural Network Learning Algorithm

Claims (8)

一種用於深度神經網路之邊緣優化增量學習方法,包括:預備一深度神經網路模型,原本連結一或多個全連接層以及一輸出層;斷開該一或多個全連接層與該深度神經網路模型原本的連結,使該一或多個全連接層不參與增量學習訓練參數更新;複製該一或多個全連接層為一或多個新全連接層,其中之一新全連接層連接到一新輸出層,其中該新輸出層的大小等於在一給定時間內所要增量學習的種類數量;該一或多個新全連接層與該新輸出層連接到該深度神經網路模型;以及在一增量學習訓練過程中,通過該深度神經網路模型,以新添加的增量數據集參與訓練,更新該深度神經網路模型。 An edge optimization incremental learning method for deep neural networks, including: preparing a deep neural network model, originally connecting one or more fully connected layers and an output layer; disconnecting the one or more fully connected layers and The original connection of the deep neural network model is such that the one or more fully connected layers do not participate in the incremental learning training parameter update; the one or more fully connected layers are copied into one or more new fully connected layers, one of which The new fully connected layer is connected to a new output layer, where the size of the new output layer is equal to the number of categories to be incrementally learned in a given time; the one or more new fully connected layers and the new output layer are connected to the a deep neural network model; and during an incremental learning training process, through the deep neural network model, participate in training with a newly added incremental data set, and update the deep neural network model. 如請求項1所述的用於深度神經網路之邊緣優化增量學習方法,其中複製原本的該一或多個全連接層的方式有助於該深度神經網路模型快速實現高分類準確度。 The edge-optimized incremental learning method for deep neural networks as described in claim 1, wherein copying the original one or more fully connected layers helps the deep neural network model quickly achieve high classification accuracy. . 如請求項1所述的用於深度神經網路之邊緣優化增量學習方法,其中,於該增量學習訓練過程中,僅有新添加的增量數據集參與訓練,而原先該深度神經網路模型已完成訓練的數據集不參與新的訓練。 The edge-optimized incremental learning method for deep neural networks as described in claim 1, wherein during the incremental learning training process, only newly added incremental data sets participate in the training, and the original deep neural network The data set for which the road model has completed training will not participate in new training. 如請求項3所述的用於深度神經網路之邊緣優化增量學習方法,其中,針對不同的應用與不同數量的種類可以在每次增量學習訓練過程中逐步添加到該深度神經網路模型中。 The edge-optimized incremental learning method for deep neural networks as described in claim 3, wherein different applications and different numbers of categories can be gradually added to the deep neural network during each incremental learning training process. in the model. 如請求項1至4中任一項所述的用於深度神經網路之邊緣優化增量學習方法,其中,於完成增量學習並更新該深度神經網路模型後,在一模型推理階段,先前與該深度神經網路模型斷開連結的該一或多個全連接層再次連接更新後的該深度神經網路模型,使得原先學習種類的該一或多個全連接層以及增量學習種類的該一或多個新全連接層共享更新後 的該深度神經網路模型。 The edge optimization incremental learning method for deep neural networks as described in any one of claims 1 to 4, wherein after completing incremental learning and updating the deep neural network model, in a model inference stage, The one or more fully connected layers that were previously disconnected from the deep neural network model are reconnected to the updated deep neural network model, so that the one or more fully connected layers of the original learning type and the incremental learning type are The one or more new fully connected layers share the updated The deep neural network model. 一種運行深度神經網路的電腦系統,其中該電腦系統包括具有特定算力與儲存資源的一處理器與一記憶體,其中執行的一邊緣優化增量學習方法包括:預備一深度神經網路模型,原本連結一或多個全連接層以及一輸出層;斷開該一或多個全連接層與該深度神經網路模型原本的連結,使該一或多個全連接層不參與增量學習訓練參數更新;複製該一或多個全連接層為一或多個新全連接層,其中之一新全連接層連接到一新輸出層,其中該新輸出層的大小等於在一給定時間內所要增量學習的種類數量;該一或多個新全連接層與該新輸出層連接到該深度神經網路模型;以及在一增量學習訓練過程中,通過該深度神經網路模型,以新添加的增量數據集參與訓練,更新該深度神經網路模型。 A computer system running a deep neural network, wherein the computer system includes a processor and a memory with specific computing power and storage resources, and an edge-optimized incremental learning method executed therein includes: preparing a deep neural network model , originally connecting one or more fully connected layers and an output layer; disconnecting the original connection between the one or more fully connected layers and the deep neural network model, so that the one or more fully connected layers do not participate in incremental learning Training parameter update; copy the one or more fully connected layers into one or more new fully connected layers, one of the new fully connected layers is connected to a new output layer, wherein the size of the new output layer is equal to a given time The number of categories to be incrementally learned within; the one or more new fully connected layers and the new output layer are connected to the deep neural network model; and during an incremental learning training process, through the deep neural network model, Participate in training with the newly added incremental data set and update the deep neural network model. 如請求項6所述的運行深度神經網路的電腦系統,其中,於該增量學習訓練過程中,僅有新添加的增量數據集參與訓練,而原先該深度神經網路模型已完成訓練的數據集不參與新的訓練,針對不同的應用與不同數量的種類可以在每次增量學習訓練過程中逐步添加到該深度神經網路模型中。 The computer system for running a deep neural network as described in claim 6, wherein during the incremental learning and training process, only the newly added incremental data set participates in the training, and the original deep neural network model has completed training The data set does not participate in new training. Different applications and different numbers of categories can be gradually added to the deep neural network model during each incremental learning training process. 如請求項6或7所述的運行深度神經網路的電腦系統,其中,於完成增量學習並更新該深度神經網路模型後,在一模型推理階段,先前與該深度神經網路模型斷開連結的該一或多個全連接層再次連接更新後的該深度神經網路模型,使得原先學習種類的該一或多個全連接層以及增量學習種類的該一或多個新全連接層共享更新後的該深度神經網路模型。The computer system for running a deep neural network as described in claim 6 or 7, wherein after completing incremental learning and updating the deep neural network model, in a model inference stage, the previous connection with the deep neural network model is disconnected. The one or more fully connected layers of the open connection are connected again to the updated deep neural network model, so that the one or more fully connected layers of the original learning type and the one or more new fully connected layers of the incremental learning type The layers share the updated deep neural network model.
TW111127482A 2022-07-22 2022-07-22 Method for implementing edge-optimized incremental learning for deep neural network and computer system TWI823488B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW111127482A TWI823488B (en) 2022-07-22 2022-07-22 Method for implementing edge-optimized incremental learning for deep neural network and computer system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW111127482A TWI823488B (en) 2022-07-22 2022-07-22 Method for implementing edge-optimized incremental learning for deep neural network and computer system

Publications (2)

Publication Number Publication Date
TWI823488B true TWI823488B (en) 2023-11-21
TW202405700A TW202405700A (en) 2024-02-01

Family

ID=89722727

Family Applications (1)

Application Number Title Priority Date Filing Date
TW111127482A TWI823488B (en) 2022-07-22 2022-07-22 Method for implementing edge-optimized incremental learning for deep neural network and computer system

Country Status (1)

Country Link
TW (1) TWI823488B (en)

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107590497A (en) * 2017-09-20 2018-01-16 重庆邮电大学 Off-line Handwritten Chinese Recognition method based on depth convolutional neural networks
CN110119689A (en) * 2019-04-18 2019-08-13 五邑大学 A kind of face beauty prediction technique based on multitask transfer learning
CN111612035A (en) * 2020-04-18 2020-09-01 华为技术有限公司 Method for training migration model, method and device for detecting fault
CN111967480A (en) * 2020-09-07 2020-11-20 上海海事大学 Multi-scale self-attention target detection method based on weight sharing
CN112801290A (en) * 2021-02-26 2021-05-14 中国人民解放军陆军工程大学 Multi-agent deep reinforcement learning method, system and application
TW202226070A (en) * 2020-12-21 2022-07-01 美商英特爾股份有限公司 Methods and apparatus to facilitate efficient knowledge sharing among neural networks

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107590497A (en) * 2017-09-20 2018-01-16 重庆邮电大学 Off-line Handwritten Chinese Recognition method based on depth convolutional neural networks
CN110119689A (en) * 2019-04-18 2019-08-13 五邑大学 A kind of face beauty prediction technique based on multitask transfer learning
CN111612035A (en) * 2020-04-18 2020-09-01 华为技术有限公司 Method for training migration model, method and device for detecting fault
CN111967480A (en) * 2020-09-07 2020-11-20 上海海事大学 Multi-scale self-attention target detection method based on weight sharing
TW202226070A (en) * 2020-12-21 2022-07-01 美商英特爾股份有限公司 Methods and apparatus to facilitate efficient knowledge sharing among neural networks
CN112801290A (en) * 2021-02-26 2021-05-14 中国人民解放军陆军工程大学 Multi-agent deep reinforcement learning method, system and application

Also Published As

Publication number Publication date
TW202405700A (en) 2024-02-01

Similar Documents

Publication Publication Date Title
Patricia et al. Learning to learn, from transfer learning to domain adaptation: A unifying perspective
TWI751458B (en) Neural network search method and device, processor, electronic equipment and computer readable storage medium
WO2020107351A1 (en) Model training method and nodes thereof, network and storage device
EP3785176A1 (en) Learning a truncation rank of singular value decomposed matrices representing weight tensors in neural networks
CN111882040A (en) Convolutional neural network compression method based on channel number search
US11625614B2 (en) Small-world nets for fast neural network training and execution
CN112418397B (en) Image classification method based on lightweight convolutional neural network
CN115578248A (en) Generalized enhanced image classification algorithm based on style guidance
Ying et al. EAGAN: Efficient two-stage evolutionary architecture search for GANs
TWI823488B (en) Method for implementing edge-optimized incremental learning for deep neural network and computer system
TW201839676A (en) Tracking axes during model conversion
US11915141B2 (en) Apparatus and method for training deep neural network using error propagation, weight gradient updating, and feed-forward processing
WO2024060839A1 (en) Object operation method and apparatus, computer device, and computer storage medium
Basirat et al. Learning Task-specific Activation Functions using Genetic Programming.
WO2019080844A1 (en) Data reasoning method and apparatus, and computer device
Verma et al. A novel framework for neural architecture search in the hill climbing domain
Ho et al. Adaptive communication for distributed deep learning on commodity GPU cluster
TWI705340B (en) Training method for phase image generator and training method of phase image classifier
Kawakami et al. A low-cost neural ODE with depthwise separable convolution for edge domain adaptation on fpgas
CN112633517A (en) Training method of machine learning model, computer equipment and storage medium
Korablyov et al. Dendritic Artificial Immune Network Model for Computing.
Li et al. Multi-task learning with attention: Constructing auxiliary tasks for learning to learn
Feng et al. Chebyshev polynomial broad learning system
WO2024078308A1 (en) Image optimization method and apparatus, electronic device, medium, and program product
Yu et al. Genetic-based K-means algorithm for selection of feature variables