CN111340182A - Input feature approximation low-complexity CNN training method and device - Google Patents

Input feature approximation low-complexity CNN training method and device Download PDF

Info

Publication number
CN111340182A
CN111340182A CN202010086794.3A CN202010086794A CN111340182A CN 111340182 A CN111340182 A CN 111340182A CN 202010086794 A CN202010086794 A CN 202010086794A CN 111340182 A CN111340182 A CN 111340182A
Authority
CN
China
Prior art keywords
matrix
training
model
data
characterization
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010086794.3A
Other languages
Chinese (zh)
Other versions
CN111340182B (en
Inventor
李斌
陈沛鋆
刘宏福
赵成林
许方敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute Of Sensing Technology And Business Beijing University Of Posts And Telecommunication
Beijing University of Posts and Telecommunications
Original Assignee
Institute Of Sensing Technology And Business Beijing University Of Posts And Telecommunication
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute Of Sensing Technology And Business Beijing University Of Posts And Telecommunication, Beijing University of Posts and Telecommunications filed Critical Institute Of Sensing Technology And Business Beijing University Of Posts And Telecommunication
Priority to CN202010086794.3A priority Critical patent/CN111340182B/en
Publication of CN111340182A publication Critical patent/CN111340182A/en
Application granted granted Critical
Publication of CN111340182B publication Critical patent/CN111340182B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Medical Informatics (AREA)
  • Complex Calculations (AREA)

Abstract

The invention provides a low-complexity CNN training method and a device for input feature approximation, which comprises the following steps: performing dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample; and training the CNN model by taking the low-dimensional representation as input data for training the model. According to the invention, through reducing the dimensionality of the data sample and using the low-dimensional representation of the data sample with the reduced data volume as the input data for training the model, the model is trained, the complexity of CNN model training can be reduced, the storage resources and the calculation resources required by the training model are reduced, the model correlation operation can be realized on the terminal equipment with lower configuration, and the application scene is expanded.

Description

Input feature approximation low-complexity CNN training method and device
Technical Field
The invention relates to the technical field of machine learning, in particular to a low-complexity CNN training method and device for input feature approximation.
Background
Machine learning is a multi-field subject combining technologies of mathematics, computers and the like, and is widely researched, popularized and applied in recent years. Various machine learning models have good application effects in the fields of images, videos, semantic analysis, machine translation, sequence processing and the like. However, with the development of deep learning technology, the model structure is more and more complex, the computation complexity of the training and reasoning process is more and more high, and the requirements on the storage and computation resources of hardware equipment are high, so that the model can only be deployed in a server and cannot realize the functions of the low-configuration terminal equipment.
Disclosure of Invention
In view of the above, the present invention provides a low complexity CNN training method and apparatus for input feature approximation, so as to solve the problem of high requirement on hardware due to complex model.
Based on the above purpose, the present invention provides a low complexity CNN training method for input feature approximation, which includes:
performing dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
and training the CNN model by taking the low-dimensional representation as input data for training the model.
Optionally, the performing dimension reduction processing on the data sample to obtain a low-dimensional characterization of the data sample includes:
converting the data sample tensor into a data matrix;
extracting a trip characterization matrix and a list characterization matrix from the data matrix;
and calculating to obtain a core characterization matrix according to the row characterization matrix and the list characterization matrix.
Optionally, training the CNN model by using the low-dimensional characterization as input data for training the model, including:
and taking the row characterization matrix, the list characterization matrix and the core characterization matrix as input data for training a model, and training a CNN model.
Optionally, the extracting a travel characterization matrix and a list characterization matrix from the data matrix includes:
constructing a row sampling matrix with only one value of 1 in each column, and calculating the row sampling matrix and the data matrix to obtain a row characterization matrix;
and constructing a column sampling matrix with only one value of 1 in each column, and calculating the column sampling matrix and the data matrix to obtain the column characterization matrix.
Optionally, the step of calculating a core characterization matrix according to the row characterization matrix and the list characterization matrix includes:
and selecting a superposition part of the row characterization matrix and the list characterization matrix to perform pseudo-inverse calculation to obtain the core characterization matrix.
Optionally, the method further includes:
and carrying out forward reasoning calculation by using the input data.
Optionally, the method further includes:
and performing back propagation calculation by using the input data.
Optionally, the method further includes:
updating model parameters using the input data.
The embodiment of the present invention further provides a low complexity CNN training device for input feature approximation, including:
the sample processing module is used for carrying out dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
and the model training module is used for training the CNN model by taking the low-dimensional representation as input data for training the model.
The embodiment of the invention also provides electronic equipment which comprises a memory, a processor and a computer program which is stored on the memory and can run on the processor, wherein the processor realizes the input feature approximating low-complexity CNN training method when executing the program.
From the above, it can be seen that the input feature approximated low-complexity CNN training method and apparatus provided by the present invention perform dimension reduction processing on the data sample to obtain a low-dimensional representation of the data sample, and train the CNN model by using the low-dimensional representation as input data for training the model. According to the invention, through reducing the dimensionality of the data sample and using the low-dimensional representation after the data volume is reduced as the input data for training the model, the model is trained, the complexity of model training can be reduced, the storage resources and the calculation resources required by the training model are reduced, the model correlation operation can be realized on the terminal equipment with lower configuration, and the application scene is expanded.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a schematic flow chart of a method according to an embodiment of the present invention;
FIG. 2 is a schematic flow chart of a method for computing a low-dimensional representation according to an embodiment of the present invention;
FIG. 3 is a diagram illustrating a data sample tensor conversion two-dimensional matrix according to an embodiment of the present invention;
FIG. 4 is a diagram illustrating a transformation of a weight matrix according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a sampling method according to an embodiment of the present invention;
FIG. 6 is a block diagram of an apparatus according to an embodiment of the present invention;
fig. 7 is a block diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is described in further detail below with reference to specific embodiments and the accompanying drawings.
It is to be noted that technical terms or scientific terms used in the embodiments of the present invention should have the ordinary meanings as understood by those having ordinary skill in the art to which the present disclosure belongs, unless otherwise defined. The use of "first," "second," and similar terms in this disclosure is not intended to indicate any order, quantity, or importance, but rather is used to distinguish one element from another. The word "comprising" or "comprises", and the like, means that the element or item listed before the word covers the element or item listed after the word and its equivalents, but does not exclude other elements or items. The terms "connected" or "coupled" and the like are not restricted to physical or mechanical connections, but may include electrical connections, whether direct or indirect. "upper", "lower", "left", "right", and the like are used merely to indicate relative positional relationships, and when the absolute position of the object being described is changed, the relative positional relationships may also be changed accordingly.
In some implementation manners, in the practical application process of some machine learning models with specific prediction functions, along with the development of machine learning technologies, the structures of the models become more complex, the models include a large number of redundant structures, the requirements on storage resources and computing resources required by the model training and reasoning process are high, in order to implement the prediction functions of the models, the models can be deployed in servers with higher hardware configurations, such as a central server or a cloud server, but the load of the servers can be increased. One solution is that a server executes a training model, and the trained model is deployed in edge equipment with a low configuration to execute model inference operation, however, in order to implement model training, data needs to be uploaded to the server, on one hand, there is a risk of data leakage and is not suitable for an application scenario with a high requirement on data security, and on the other hand, for an application scenario in which a mapping relationship and a task mode change, the edge equipment and the server need to frequently interact data, and the demand cannot be met for a scenario with limited network resources and a low delay requirement.
In order to solve the above problems, embodiments of the present invention provide a low-complexity CNN training method and apparatus with approximated input features, which can reduce the complexity of model training and inference operations and reduce the requirements on hardware configuration by reducing the dimensionality of input data samples, so that the training and inference operations of a model can be implemented by using hardware devices with lower configuration.
FIG. 1 is a schematic flow chart of a method according to an embodiment of the present invention. As shown in the figure, the low-complexity CNN training method for input feature approximation provided in the embodiment of the present invention includes:
s101: performing dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
s102: the CNN model is trained using the low-dimensional representations as input data for training the model.
In this embodiment, it is considered that some network compression methods can simplify the structure of the model, for example, compress the convolution kernel, however, it is not considered that the data samples used for training the model often contain a large amount of redundant data, which results in a complicated process for training the model. According to the CNN model training method provided by the embodiment of the invention, the dimensionality of the data sample is reduced, the low-dimensional representation with the reduced data volume is used as input data for training the model, the CNN model is trained, the complexity of model training can be reduced, the storage resources and the calculation resources required by the training model are reduced, and the training model operation can be realized on a terminal device with lower configuration.
FIG. 2 is a flow chart of a method for calculating a data characterization sample according to an embodiment of the present invention. As shown in the figure, in an embodiment, in the step S101, the dimension reduction processing is performed on the data sample to obtain a low-dimensional representation of the data sample, which includes
S201: converting the data sample tensor into a data matrix;
s202: extracting a trip characterization matrix and a list characterization matrix from the data matrix;
s203: and calculating to obtain a core characterization matrix according to the row characterization matrix and the list characterization matrix.
Then, the model is trained using the row characterization matrix, the list characterization matrix, and the core characterization matrix as input data for training the model.
In this embodiment, the original data samples are high-dimensional data sample tensors, in order to reduce the data volume of the data sample tensors, the data sample tensors are converted into data matrices, a row feature matrix and a list feature matrix are extracted from the data matrices, a core feature matrix is obtained through calculation based on the row feature matrix and the list feature matrix, the simplified row feature matrix, the simplified list feature matrix and the simplified core feature matrix are used as input data, a model is trained, and model training operations are simplified by optimizing and simplifying the data samples. According to the embodiment, the thought of approximate approximation of the high-dimensional matrix is used for reference, three low-dimensional characterization matrixes are extracted from the high-dimensional data sample through random sampling and approximate calculation, and the calculation complexity and the space complexity can be effectively reduced on the premise of not reducing the generalization performance of the model.
The low complexity CNN training method of the present invention is described below with reference to specific embodiments.
Let X ∈ R be the data sample tensor of the Convolutional layer of Convolutional Neural Network (CNN)m×h×w×cThe weight tensor is W ∈ Rk×l×c×nWherein m is the number of data samples, h and w are the height and width of the data sample tensor respectively, c is the number of channels of the data sample tensor, k and l are the height and width of the convolution kernel respectively, and n is the number of convolution kernels.
Unfolding a data sample tensor X into a two-dimensional matrix Xm,XmIs related to the height and width of the convolution kernel, the scan step size, and the edge filling approach. In one mode, taking the height and width of a convolution kernel as k and l, the scanning step length as 1, and the edge not filled as an example, the tensor X of the data sample is unfolded to obtain a two-dimensional matrix Xm∈Ra×bWhere a ═ m (h-k +1) (w-l +1)) and b ═ klc, as shown in fig. 3, the two-dimensional matrix XmExpressed as:
Xm=unfold(X,height=k,width=l,strides=1,padding=0) (1)
wherein, underfold is an expansion function, and parameters height, width, threads and padding respectively represent height, width, scanning step length and edge filling mode of a convolution kernel.
Obtaining a two-dimensional matrixXmAnd then, sampling is extracted from the two-dimensional matrix to obtain a row feature matrix and a list feature matrix. In some embodiments, a suitable sampling method is selected from the two-dimensional matrix XmAnd extracting a row characterization matrix and a list characterization matrix. Optionally, the sampling method includes: from a two-dimensional matrix XmIn the method, uniform row sampling is carried out, the sum of squares of elements in each column is calculated in a matrix obtained by sampling, the sum of squares of elements in each column accounts for the sum of squares of all elements in all columns of the matrix and is taken as the sampling probability of the column, and the performance of a model is hardly influenced by the sampling method after training. As shown in fig. 5, the sampling probability of the first column is:
Figure BDA0002382342870000051
for a row characterization matrix, let the number of sampling rows be t (t)<<a) Constructing a line sampling matrix according to the selected line sampling method
Figure BDA0002382342870000052
The constructed row sampling matrix SrHas only one element with a value of 1 and the other elements have a value of 0, and then the row sampling matrix S is usedrFor two-dimensional matrix XmSampling to obtain a row characterization matrix
Figure BDA0002382342870000053
R=Sr TXm(3)
For a list characterization matrix, let the number of sampling columns be s(s)<<b) Constructing a column sampling matrix according to the selected column sampling method
Figure BDA0002382342870000061
Constructed column sampling matrix ScHas only one element with a value of 1 and the other elements have a value of 0, and then the column sampling matrix S is usedcFor two-dimensional matrix XmSampling to obtain a list characterization matrix
Figure BDA0002382342870000062
C=XmSc(4)
Then, according to the row characterization matrix and the list characterization matrix, a core characterization matrix is constructed
Figure BDA0002382342870000063
In some embodiments, the coincidence of the row characterization matrix R and the column characterization matrix C may be selected
Figure BDA0002382342870000064
To compute the core characterization matrix U, i.e.:
V=Sr TXmSc(5)
Figure BDA0002382342870000065
it should be noted that the method for constructing the core characterization matrix U according to the row characterization matrix R and the column characterization matrix C is not limited to the above manner, and a part of the row characterization matrix and a part of the column characterization matrix may also be selected to construct the core characterization matrix, or an operation is performed according to an element of the row characterization matrix and an element of the column characterization matrix by using a specific algorithm to obtain the core characterization matrix.
In this embodiment, because the original data samples are very complex, the original data samples are simplified to reduce the data volume, the travel characterization matrix and the list characterization matrix are extracted, the core characterization matrix is obtained through calculation according to the row characterization matrix and the list characterization matrix, the row characterization matrix, the list characterization matrix and the core characterization matrix are used as input data of a training model to train the CNN model, the computation volume of the training model is greatly reduced, the storage resources and the computation resources required by the training model are further reduced, and the function of the training model can be realized even in a hardware device with low configuration.
In the embodiment of the invention, the forward reasoning calculation is carried out on the model obtained by training by using the input data. In some embodiments, the weight parameters of the network are processed to obtain a weight matrix, the input data and the weight matrix are multiplied to obtain a product result, the product result is added with a bias vector, and an output result of the model is obtained after the action of an activation function.
Referring to fig. 4, for the CNN network, the first three dimensions of the convolution kernel tensor W are first combined to obtain a two-dimensional weight matrix
Figure BDA0002382342870000066
Then, a row characterization matrix R, a column characterization matrix C, a core characterization matrix U and a weight matrix W are combinedmMultiplying to obtain product result, adding offset vector
Figure BDA0002382342870000067
Obtaining the output result of the model after the action of the activation function f
Figure BDA0002382342870000068
Expressed as:
Xm`=f(C(URWm)+b) (7)
as shown in formula (7), the core characterization matrix U is multiplied by the row characterization matrix R and then multiplied by the two-dimensional weight matrix WmThe multiplication sequence is calculated in a minimum amount by multiplying the obtained product by the list characterization matrix C, and the model calculation amount can be reduced to the maximum extent.
In the embodiment of the invention, the model obtained by training is subjected to backward propagation calculation by using the input data. In some embodiments, the reverse error tensor of the current layer is converted into an error matrix, a row error matrix, a column error matrix and a core error matrix are respectively obtained through calculation according to the row characterization matrix, the column characterization matrix, the core characterization matrix and the error matrix, and the reverse error tensor transmitted to the next layer is obtained according to the row error matrix, the column error matrix and the core error matrix.
Let the tensor of the error received by the current layer be
Figure BDA0002382342870000071
Firstly, the error tensor delta is converted into a two-dimensional error matrix
Figure BDA0002382342870000072
Then separately calculate the column error matrix deltac', row error matrix deltar"sum core error matrix deltauAccording to the sampling and pseudo-inverse operation process in the estimation, the method can obtain the following steps by a chain rule:
Figure BDA0002382342870000073
in the calculation process, the column sampling matrix S of right multiplicationcIs transposed Sc TWithout matrix multiplication, since the column sample matrix ScHas only one element with a value of 1, and the element values of the rest positions are all 0, and are multiplied by S rightc TCorresponding to the inverse of the column sampling, i.e. the columns of the matrix to which they are multiplied are placed back in an all-zero matrix according to the sampled index. Similarly, the left-hand row sampling matrix SrThe operation of (1) is the inverse of row sampling, and only the rows of the matrix multiplied by it need to be placed back into an all-zero matrix according to the sampled index. Compared with matrix multiplication, the complexity of the calculation process is low, and the calculation amount is greatly reduced.
And calculating a total error matrix of the back propagation according to the row error matrix, the column error matrix and the core error matrix as follows:
δm`=δc`+δu`+δr`,δm`∈-a×b(9)
then, the total error matrix delta is determinedmThe 'conversion to error tensor δ' ∈ Rm×h×w×cThe error tensor is passed to the next layer:
δ`=fold(δm`,height=k,width=l,strides=1,padding=0) (10)
wherein, fold is a folding function, and the meaning of the parameter in the function is the same as that of the parameter in the unfold function.
In the embodiment of the invention, the model parameters are updated by using the input data. In some embodiments, a weight gradient is calculated using the input data, and the model parameters are updated based on the weight gradient.
Will error matrix deltamMultiplying the obtained data by a row characterization matrix R, a list characterization matrix C and a core characterization matrix U to calculate the weight gradient
Figure BDA0002382342870000081
The calculation formula is as follows:
ΔWm=RT(UT(CTδm))/m (11)
the transpose C of the list eigen matrix C is shown in equation (11)TAnd error matrix deltamTranspose U of core characterization matrix U after multiplicationTMultiplying, the resulting product being then transposed R of the row characterization matrix RTMultiplication, such multiplication operations are performed in the order of the smallest calculation amount.
Calculating to obtain a weight gradient delta WmThen, it is converted into a gradient tensor
Figure BDA0002382342870000082
The model is then updated using a gradient descent based optimization algorithm.
According to the low-complexity CNN training method, the computational complexity of the training model calculation depends on the sampling calculation method, and is generally lower than the computational processes of forward propagation, backward propagation and model updating; and the computational complexity of the forward propagation calculation and the calculation of the updated model parameters are both O (asn + stn + btn), the computational complexity of the backward propagation calculation is about O (asn + stn + btn), and by synthesis, the computational complexity of the CNN model training and updating method can be reduced to O (asn + stn + btn) which is far lower than the computational complexity O (abn) of the existing model training, and the required computational resources are greatly reduced. In addition, because the input data samples are required to be used for updating the model parameter calculation, compared with the prior art in which all data samples are reserved, the method only needs to reserve the row feature matrix, the list feature matrix and the core feature matrix extracted from all the data samples, the spatial complexity is reduced to O (as + tb + st), which is far lower than the original spatial complexity O (ab), and the required storage resources are greatly reduced. The low-complexity CNN model training method provided by the embodiment of the invention can deploy the model to a high-configuration server and a low-configuration edge device, can realize model operations such as model training, model input prediction and model updating by using the server and/or the edge device, can meet the requirements of adaptivity, real-time property and data confidentiality in various application scenes, expands the application scenes of the model and improves the resource utilization rate.
It should be noted that the method of the embodiment of the present invention may be executed by a single device, such as a computer or a server. The method of the embodiment can also be applied to a distributed scene and completed by the mutual cooperation of a plurality of devices. In the case of such a distributed scenario, one of the multiple devices may only perform one or more steps of the method according to the embodiment of the present invention, and the multiple devices interact with each other to complete the method.
FIG. 6 is a block diagram of an apparatus according to an embodiment of the present invention. As shown in the figure, the low-complexity CNN training apparatus for input feature approximation provided in the embodiment of the present invention includes:
the sample processing module is used for carrying out dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
and the model training module is used for training the CNN model by taking the low-dimensional representation as input data for training the model.
The apparatus of the foregoing embodiment is used to implement the corresponding method in the foregoing embodiment, and has the beneficial effects of the corresponding method embodiment, which are not described herein again.
Fig. 7 is a schematic diagram illustrating a more specific hardware structure of an electronic device according to this embodiment, where the electronic device may include: a processor 1010, a memory 1020, an input/output interface 1030, a communication interface 1040, and a bus 1050. Wherein the processor 1010, memory 1020, input/output interface 1030, and communication interface 1040 are communicatively coupled to each other within the device via bus 1050.
The processor 1010 may be implemented by a general-purpose CPU (Central Processing Unit), a microprocessor, an Application Specific Integrated Circuit (ASIC), or one or more Integrated circuits, and is configured to execute related programs to implement the technical solutions provided in the embodiments of the present disclosure.
The Memory 1020 may be implemented in the form of a ROM (Read Only Memory), a RAM (Random access Memory), a static storage device, a dynamic storage device, or the like. The memory 1020 may store an operating system and other application programs, and when the technical solution provided by the embodiments of the present specification is implemented by software or firmware, the relevant program codes are stored in the memory 1020 and called to be executed by the processor 1010.
The input/output interface 1030 is used for connecting an input/output module to input and output information. The i/o module may be configured as a component in a device (not shown) or may be external to the device to provide a corresponding function. The input devices may include a keyboard, a mouse, a touch screen, a microphone, various sensors, etc., and the output devices may include a display, a speaker, a vibrator, an indicator light, etc.
The communication interface 1040 is used for connecting a communication module (not shown in the drawings) to implement communication interaction between the present apparatus and other apparatuses. The communication module can realize communication in a wired mode (such as USB, network cable and the like) and also can realize communication in a wireless mode (such as mobile network, WIFI, Bluetooth and the like).
Bus 1050 includes a path that transfers information between various components of the device, such as processor 1010, memory 1020, input/output interface 1030, and communication interface 1040.
It should be noted that although the above-mentioned device only shows the processor 1010, the memory 1020, the input/output interface 1030, the communication interface 1040 and the bus 1050, in a specific implementation, the device may also include other components necessary for normal operation. In addition, those skilled in the art will appreciate that the above-described apparatus may also include only those components necessary to implement the embodiments of the present description, and not necessarily all of the components shown in the figures.
Computer-readable media of the present embodiments, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device.
Those of ordinary skill in the art will understand that: the discussion of any embodiment above is meant to be exemplary only, and is not intended to intimate that the scope of the disclosure, including the claims, is limited to these examples; within the idea of the invention, also features in the above embodiments or in different embodiments may be combined, steps may be implemented in any order, and there are many other variations of the different aspects of the invention as described above, which are not provided in detail for the sake of brevity.
In addition, well known power/ground connections to Integrated Circuit (IC) chips and other components may or may not be shown within the provided figures for simplicity of illustration and discussion, and so as not to obscure the invention. Furthermore, devices may be shown in block diagram form in order to avoid obscuring the invention, and also in view of the fact that specifics with respect to implementation of such block diagram devices are highly dependent upon the platform within which the present invention is to be implemented (i.e., specifics should be well within purview of one skilled in the art). Where specific details (e.g., circuits) are set forth in order to describe example embodiments of the invention, it should be apparent to one skilled in the art that the invention can be practiced without, or with variation of, these specific details. Accordingly, the description is to be regarded as illustrative instead of restrictive.
While the present invention has been described in conjunction with specific embodiments thereof, many alternatives, modifications, and variations of these embodiments will be apparent to those of ordinary skill in the art in light of the foregoing description. For example, other memory architectures (e.g., dynamic ram (dram)) may use the discussed embodiments.
The embodiments of the invention are intended to embrace all such alternatives, modifications and variances that fall within the broad scope of the appended claims. Therefore, any omissions, modifications, substitutions, improvements and the like that may be made without departing from the spirit and principles of the invention are intended to be included within the scope of the invention.

Claims (10)

1. A low-complexity CNN training method for input feature approximation is characterized by comprising the following steps:
performing dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
and training the CNN model by taking the low-dimensional representation as input data for training the model.
2. The method of claim 1, wherein the performing the dimensionality reduction on the data sample to obtain the low-dimensional characterization of the data sample comprises:
converting the data sample tensor into a data matrix;
extracting a trip characterization matrix and a list characterization matrix from the data matrix;
and calculating to obtain a core characterization matrix according to the row characterization matrix and the list characterization matrix.
3. The method of claim 2, wherein training a CNN model using the low-dimensional tokens as input data for training the model comprises:
and taking the row characterization matrix, the list characterization matrix and the core characterization matrix as input data for training a model, and training a CNN model.
4. The method of claim 2, wherein extracting a row characterization matrix and a list characterization matrix from the data matrix comprises:
constructing a row sampling matrix with only one value of 1 in each column, and calculating the row sampling matrix and the data matrix to obtain a row characterization matrix;
and constructing a column sampling matrix with only one value of 1 in each column, and calculating the column sampling matrix and the data matrix to obtain the column characterization matrix.
5. The method of claim 4, wherein computing a core characterization matrix from the row characterization matrix and the column characterization matrix comprises:
and selecting a superposition part of the row characterization matrix and the list characterization matrix to perform pseudo-inverse calculation to obtain the core characterization matrix.
6. The method of claim 1, further comprising:
and carrying out forward reasoning calculation by using the input data.
7. The method of claim 1, further comprising:
and performing back propagation calculation by using the input data.
8. The method of claim 1, further comprising:
updating model parameters using the input data.
9. A low complexity CNN training device for input feature approximation, comprising:
the sample processing module is used for carrying out dimensionality reduction processing on the data sample to obtain a low-dimensional representation of the data sample;
and the model training module is used for training the CNN model by taking the low-dimensional representation as input data for training the model.
10. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the method according to any of claims 1 to 8 when executing the program.
CN202010086794.3A 2020-02-11 2020-02-11 Low-complexity CNN training method and device for input feature approximation Active CN111340182B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010086794.3A CN111340182B (en) 2020-02-11 2020-02-11 Low-complexity CNN training method and device for input feature approximation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010086794.3A CN111340182B (en) 2020-02-11 2020-02-11 Low-complexity CNN training method and device for input feature approximation

Publications (2)

Publication Number Publication Date
CN111340182A true CN111340182A (en) 2020-06-26
CN111340182B CN111340182B (en) 2024-04-02

Family

ID=71181459

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010086794.3A Active CN111340182B (en) 2020-02-11 2020-02-11 Low-complexity CNN training method and device for input feature approximation

Country Status (1)

Country Link
CN (1) CN111340182B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022042603A1 (en) * 2020-08-27 2022-03-03 International Business Machines Corporation Tensor comparison across distributed machine learning environment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008269573A (en) * 2007-02-09 2008-11-06 Denso Corp Arithmetic unit and program
CN109242028A (en) * 2018-09-19 2019-01-18 西安电子科技大学 SAR image classification method based on 2D-PCA and convolutional neural networks
CN110163261A (en) * 2019-04-28 2019-08-23 平安科技(深圳)有限公司 Unbalanced data disaggregated model training method, device, equipment and storage medium

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008269573A (en) * 2007-02-09 2008-11-06 Denso Corp Arithmetic unit and program
CN109242028A (en) * 2018-09-19 2019-01-18 西安电子科技大学 SAR image classification method based on 2D-PCA and convolutional neural networks
CN110163261A (en) * 2019-04-28 2019-08-23 平安科技(深圳)有限公司 Unbalanced data disaggregated model training method, device, equipment and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
雷恒鑫等: "利用CUR矩阵分解提高特征选择与矩阵恢复能", 计算机应用, vol. 37, no. 3, pages 640 - 646 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022042603A1 (en) * 2020-08-27 2022-03-03 International Business Machines Corporation Tensor comparison across distributed machine learning environment
GB2613316A (en) * 2020-08-27 2023-05-31 Ibm Tensor comparison across distributed machine learning environment
US11954611B2 (en) 2020-08-27 2024-04-09 International Business Machines Corporation Tensor comparison across a distributed machine learning environment

Also Published As

Publication number Publication date
CN111340182B (en) 2024-04-02

Similar Documents

Publication Publication Date Title
US10810483B2 (en) Superpixel methods for convolutional neural networks
CN111369440B (en) Model training and image super-resolution processing method, device, terminal and storage medium
KR20180101055A (en) Neural network device and operating method of neural network device
WO2020256704A1 (en) Real-time video ultra resolution
CN112883149B (en) Natural language processing method and device
CN111209933A (en) Network traffic classification method and device based on neural network and attention mechanism
CN109034206A (en) Image classification recognition methods, device, electronic equipment and computer-readable medium
CN112883227B (en) Video abstract generation method and device based on multi-scale time sequence characteristics
CN112561028A (en) Method for training neural network model, and method and device for data processing
CN112488923A (en) Image super-resolution reconstruction method and device, storage medium and electronic equipment
US20180005113A1 (en) Information processing apparatus, non-transitory computer-readable storage medium, and learning-network learning value computing method
CN112329808A (en) Optimization method and system of Deeplab semantic segmentation algorithm
EP4120141A1 (en) Method and apparatus for performing deconvolution processing on feature data by utilizing convolution hardware
CN113535912A (en) Text association method based on graph convolution network and attention mechanism and related equipment
Li et al. Fast principal component analysis for hyperspectral imaging based on cloud computing
CN111340182B (en) Low-complexity CNN training method and device for input feature approximation
CN112819155B (en) Deep neural network model hierarchical compression method and device applied to edge equipment
Li et al. Object detection network pruning with multi-task information fusion
CN111324860B (en) Lightweight CNN calculation method and device based on random matrix approximation
CN112561050A (en) Neural network model training method and device
CN113902107A (en) Data processing method, readable medium and electronic device for neural network model full connection layer
CN114943332A (en) Training method of lightweight YOLO model and related equipment
CN114611659A (en) STN-based efficient and safe federal learning method and related equipment
CN112580772B (en) Compression method and device for convolutional neural network
CN111445282B (en) Service processing method, device and equipment based on user behaviors

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant