CN109344958B - Object identification method and identification system based on feedback adjustment - Google Patents
Object identification method and identification system based on feedback adjustment Download PDFInfo
- Publication number
- CN109344958B CN109344958B CN201810936167.7A CN201810936167A CN109344958B CN 109344958 B CN109344958 B CN 109344958B CN 201810936167 A CN201810936167 A CN 201810936167A CN 109344958 B CN109344958 B CN 109344958B
- Authority
- CN
- China
- Prior art keywords
- feedback
- neural network
- deep neural
- layer
- tau
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 32
- 238000013528 artificial neural network Methods 0.000 claims abstract description 71
- 230000000007 visual effect Effects 0.000 claims abstract description 11
- 230000000644 propagated effect Effects 0.000 claims abstract description 6
- 230000015654 memory Effects 0.000 claims description 43
- 210000002569 neuron Anatomy 0.000 claims description 27
- 230000002441 reversible effect Effects 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 5
- 230000006870 function Effects 0.000 claims description 4
- 230000006386 memory function Effects 0.000 claims description 4
- 230000001629 suppression Effects 0.000 claims 2
- 239000010410 layer Substances 0.000 description 77
- 230000000694 effects Effects 0.000 description 16
- 241000282472 Canis lupus familiaris Species 0.000 description 10
- 238000002474 experimental method Methods 0.000 description 8
- 230000010387 memory retrieval Effects 0.000 description 8
- 230000008569 process Effects 0.000 description 8
- 241000282693 Cercopithecidae Species 0.000 description 6
- 241000282326 Felis catus Species 0.000 description 6
- 230000007246 mechanism Effects 0.000 description 6
- 230000001149 cognitive effect Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 5
- 230000009123 feedback regulation Effects 0.000 description 5
- 210000000653 nervous system Anatomy 0.000 description 5
- 230000000875 corresponding effect Effects 0.000 description 4
- 230000001537 neural effect Effects 0.000 description 4
- 210000000977 primary visual cortex Anatomy 0.000 description 4
- 230000010354 integration Effects 0.000 description 3
- 230000037361 pathway Effects 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 239000013598 vector Substances 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000005764 inhibitory process Effects 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 230000000638 stimulation Effects 0.000 description 2
- 210000000225 synapse Anatomy 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 210000000857 visual cortex Anatomy 0.000 description 2
- 241001465754 Metazoa Species 0.000 description 1
- 230000009471 action Effects 0.000 description 1
- 230000006978 adaptation Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000033228 biological regulation Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 230000002964 excitative effect Effects 0.000 description 1
- 238000010304 firing Methods 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 210000001926 inhibitory interneuron Anatomy 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000011229 interlayer Substances 0.000 description 1
- 210000001153 interneuron Anatomy 0.000 description 1
- 230000001404 mediated effect Effects 0.000 description 1
- 239000012528 membrane Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008904 neural response Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000000306 recurrent effect Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 241000894007 species Species 0.000 description 1
- 230000000946 synaptic effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses an object identification method based on feedback adjustment, which adopts a deep neural network to identify objects and comprises the following steps: visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; in the early stage of object identification, inter-class noise caused by mutual interference among different classes in the deep neural network is inhibited through forward feedback; in the later stage of object identification, objects of the same class in the deep neural network are inhibited through inverse feedback to mix up the identified intra-class noise due to the similarity between the objects. The invention also discloses an object identification system adopting the method.
Description
Technical Field
The invention relates to an object recognition method, in particular to an object recognition method for optimizing a memory retrieval effect in a neural network by combining forward feedback (push) and reverse feedback (pull), and also relates to a corresponding object recognition system, belonging to the technical field of deep neural networks.
Background
Research in cognitive neuroscience has demonstrated that there is a particular hierarchy of human cognitive recognition of objects. For example, the process we identify border shepherd dogs is such: first, it is judged to belong to the animal (highest grade); then, identify it as a certain dog (higher level); finally, it was identified as belonging to border shepherd dogs (lower grade). In this hierarchical structure, the commonalities of species belonging to the same class constitute its higher-level characteristics, for example, border shepherd dogs and central chinese garden dogs belong to the same class, and their commonalities constitute their higher-level characteristics. In the identification process, the greater the difference between the features of the higher categories, the easier it is to achieve accurate and rapid discrimination, and can help achieve identification of the lower categories.
Deep Neural Network (Deep Neural Network) is a multi-layer Neural Network with at least one hidden layer. The method simulates the hierarchical information processing process of a visual path, and achieves great success in the application fields of object identification and the like. Structurally, a deep neural network is mainly composed of feed-forward connections from lower layers to higher layers. However, experimental data of cognitive neuroscience show that there is also a very rich feedback connection from high to low layers in the real nervous system. At present, the degree of understanding of the role of feedback links in the nervous system is still insufficient.
In the chinese patent application with the application number of 201110185644.1, tianjin university proposes an object identification method based on a cascade micro neural network, which includes the following steps: calculating the number of cascaded convolutional layers in the deep neural network according to the size of the input sample image; constructing a deep cascade micro neural network; setting the training parameters of the deep cascade micro-neural network, and training by adopting a random gradient descent method; classifying by adopting a softmax classifier, and calculating a classification error by utilizing a forward propagation algorithm; updating the weight value of the parameter to be trained in the neural network by utilizing back propagation operation; the deep cascade micro-neural network is obtained, and the identification performance of the object identification system can be improved under the condition that parameters (operation complexity) are not increased.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides an object identification method which combines forward feedback (push) and reverse feedback (pull) and optimizes the memory retrieval effect in a neural network.
Another object of the present invention is to provide an object recognition system using the above method.
In order to achieve the purpose of the invention, the invention adopts the following technical scheme:
according to a first aspect of the embodiments of the present invention, there is provided an object recognition method based on feedback adjustment, which uses a deep neural network to perform object recognition, including the following steps:
visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; wherein,
in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback; in the later stage of object identification, the objects in the same class in the deep neural network are inhibited by reverse feedback to mix up the identified intra-class noise due to the similarity between the objects.
Preferably, in the deep neural network, the feedforward connection and the feedback connection are simultaneously arranged between each level of hierarchy, and the recursive connection between the neurons in the same layer is the connection with the associative memory function.
Preferably, the object classification information is stored by using the hierarchical structure of the deep neural network, so that the information of the memory module can be identified in a structured mode.
According to a second aspect of embodiments of the present invention, there is provided an object recognition system based on feedback adjustment, comprising a processor and a memory; the memory having stored thereon a computer program operable on the processor, the computer program when executed by the processor implementing the steps of:
visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; wherein,
in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback; in the later stage of object identification, the objects in the same class in the deep neural network are inhibited by reverse feedback to mix up the identified intra-class noise due to the similarity between the objects.
Compared with the prior art, the object identification method and the object identification system provided by the invention adopt an inter-level feedback regulation mechanism organically combining forward feedback (push) and reverse feedback (pull). When the object recognition task is completed, the image of the object to be recognized is input into the first layer of the deep neural network, information is integrated from bottom to top, and meanwhile recurrence is attempted in the layers. The invention provides a heuristic for developing a better artificial intelligence algorithm to carry out object recognition.
Drawings
FIG. 1 is a diagram of the activity of a neuron population at the primary visual cortex V1 level during contour integration tasks performed by monkeys;
FIG. 2 is a model diagram of a multi-layer neural network;
FIG. 3 is a diagram illustrating memory modules stored in a class hierarchy in the multi-layer neural network shown in FIG. 2;
FIG. 4 is an exemplary graph of layer 1 neuronal population activity with the "push-pull" phenomenon observed in the experiment;
FIG. 5 is a schematic diagram of object recognition based on feedback adjustment, in accordance with an embodiment of the present invention;
fig. 6 is a schematic structural diagram of an object recognition system provided in the present invention.
Detailed Description
The technical contents of the present invention will be further described in detail with reference to the accompanying drawings and specific embodiments.
In combination with the study of the cognitive neuroscience on the complex visual cortex pathway, we know that in a neural network, relatively higher levels process more comprehensive and abstract information, for example, the primary visual cortex V1 layer identifies local orientation, higher levels can identify contour, and then upward cortex identifies a specific certain face. Meanwhile, the phenomenon is also found in a deep neural network of object recognition tasks such as face recognition. This means that a bottom-up, gradual abstraction of the processed information is naturally formed in a hierarchical network.
On the other hand, experiments in cognitive neuroscience have shown that there are a large number of feedback connections in addition to feedforward connections in the nervous system. Existing experimental data indicate that feedback modulation in the visual system has a "push-pull" phenomenon. Figure 1 shows the neuronal population activity of V1 (primary visual cortex V1 layer) when the monkey was subjected to the contour integration task. Data were recorded at V1 of awake monkeys. The visual stimulus is a virtual contour hidden in noise, which occurs at time t-0. The blue curve shows that the neural response changes with time, rising early and falling late when the monkey is able to identify the contour. The red curve indicates the situation where the contour is not known by the monkey. The green curve is the neuron population response in the absence of visual stimulation. Under visual stimulation, the neural activity of V1 increased in the early stage, showing a "push (positive feedback)" phenomenon, and decreased in the later stage, showing a "pull (negative feedback)" phenomenon. It is noteworthy that the "push-pull" phenomenon only occurs if the monkey correctly recognizes the contour. In a similar multi-unit recording experiment, it was further confirmed that the "push-pull" phenomenon was due to feedback from V4 (primary visual cortex V4 layer) and was not a neural adaptation result.
In order to reproduce and fully utilize the above-mentioned "push-pull" phenomenon in the deep neural network, the inventor firstly constructs a multi-layer neural network with memory modules arranged in a hierarchy, and the memory modules in each hierarchy correspondingly store different types of information. Referring to fig. 2, in the established multi-layer neural network, both feedforward connection and feedback connection exist between each layer, and the recursive connection (recurrent connection) between neurons in the same layer is also a connection with associative memory function (this is to consider that there are abundant recursive connections between neurons in visual cortex). The recursive connection not only helps the neural network to realize the identification of the memory module, but also can reproduce a more complete and clear memory module from fuzzy or incomplete input information. For convenience of description, the inventor uses the Hopfield model to describe the function of each layer of neural network in the embodiment of the present invention, but the technical idea of the present invention is obviously applicable to other similar memory models. The advantage of the Hopfield model is that it can analyze information quantitatively at each layer, thereby clearly elucidating the role of feedback modulation.
Fig. 3 is a specific embodiment of the multi-layer neural network. In this example, the inventors have built a three-layer neural network. For convenience of description, three layers from top to bottom are respectively called an ancestor layer (layer 3), an ancestor layer (layer 2) and a child layer (layer 1) in the present invention, so as to reflect descending class relationships thereof. It stores memory modules with three-level hierarchical structure in one-to-one correspondence: each module belongs to a group of parent modules, which in turn belong to a set of grandparent modules. In order to simulate the role of the network layer in hierarchical memory retrieval, each layer has a corresponding level of associative memory function, and the layers contact their parent and child layers through feed-forward and feedback connections, respectively. It will be appreciated that it is natural to generalize the three-layer neural network described above to more than three layers in other embodiments.
For ease of analysis, the inventors first described using a discrete Hopfield model at each layer, which facilitates a clear elucidation of the role of feedback modulation. Later, the inventors will confirm the results using a more biologically reasonable continuous Hopfield model simulation.
It is assumed that each layer in a multi-layer neural network has the same number of neurons. Herein, useWherein i 1.. times.n, which represents the state of the neuron i of the l layers at the time t, and the value thereof is ± 1;recursive connecting weights representing symmetry between neurons i and j at level l;is the feed-forward connection weight from l layer to l +1 layer; corresponding toI.e. the feedback connection weight from l +1 layer to l layer. The dynamics of neurons follow the Hopfield model, namely:
the memory modules stored in the multilayer neural network are respectively as follows: grandfather's module is { ξ a }, where α ═ 1αThe parent module is { xiα,βWhere β 1.., PβAnd the child module is { xiα,β,γWhere γ 1.., Pγ。Pα,PβAnd PγRespectively representing the number of ancestor modules, the number of ancestor modules that are affiliated with the same ancestor, and the number of child modules that are affiliated with the same ancestor. Class relationships between memory modules are defined based on hierarchical properties, i.e., modules belonging to the same group have greater similarity than modules belonging to different groups. For example, child modules xi from the same ancestor, different ancestors, the same ancestor, or different ancestorsα,β,γAnd xiα′,β′,γ′Sigma of similarity betweeniξα,β,γξα′,β′,γ′N is eachOr is 0.
The connections between the neurons of the same layer generated by the Hebbian learning rule are:
the feedforward connection of the lower layer to the upper layer is as follows:
in the multilayer neural network, if the state of the layer 1 is a memory moduleThe information passed to layer 2 by the feed-forward connection is thenThat is, through the information integration process of the feed-forward connection, the parent module is retrieved at layer 2Is easy.
In order to quantify the performance of memory retrieval, the inventor defines a macroscopic variable m (t) which can well measure the degree of coincidence between the network state x (t) and the memory module. The degree of overlap m (t) on the different layers is calculated as follows:
next, the inventor will further explore the role of feedback adjustment in improving the network memory retrieval effect. Without loss of generality, the inventors focused on analyzing the feedback regulation effect of layer 2 on layer 1.
First, it is necessary to understand the memory retrieval of layer 1 itself without feedback adjustment. Like conventional stability analysis, the inventors consider the case of a specific memory module at layer 1 as the initial state of the network, that is, assume thatAnd then analyzing what factors affect the retrieval effect of the module. After one-step iteration, the neural network is rearranged to obtain:
in the above expression, the input received by the neuron is decomposed into two parts, signal and noise. The latter can be further decomposed into two terms: c. CiAndthey represent intra-class noise and inter-class noise, respectively. Wherein, the noise in class refers to the noise caused by the correlation between the modules of the same parent; accordingly, inter-class noise refers to noise introduced by those sibling modules of different parents of the same ancestor.
The morphology of the above distribution indicates that even if the initial input to the multi-layer neural network is noise free, it is trueThe noise generated by the correlation of the memory modules can also lead to errors in the eventual evolution of the network dynamics. The correlated noise includes both intra-class noise and inter-class noise, which reflects that the similarity between objects in the object identification process is likely to cause confusion.
To address this problem, the inventors first introduced a setting of forward feedback (push) from layer 2 to layer 1 to suppress inter-class noiseWherein,
specifically, if the parent module is retrieved at layer 2Then the "push" feedback gives the input to layer 1 asSuch feedback input can enhance the activity strength of all child modules belonging to this parent module, i.e., the activity strength of all child modules belonging to this parent moduleAnd will not be effective for other memory modules. Conversely, the "push" feedback can effectively suppress interlayer noise of layer 1.
Then, the feedback regulation mechanism of reverse feedback (pull) action is continuously introduced to suppress the noise C in classiThe following are specifically described:
assume that the "pull" feedback connection weight is set to:
if the state of layer 2 happens to be the correct parent module, i.e. theThen layer 1 is under "pull" feedback,
at this time, it is possible to obtain,
through analysis, it can be found that the sameSimilarity coefficient b between different children of parent1Within a certain range, "pull" feedback can reduce retrieval errors, i.e.,
essentially, the effect of the "pull" feedback is to actually subtract the values of the parent layer from the module information of the lower layer, consistent with the idea of predictive coding. In fact, the "pull" feedback can explain the endpoint phenomenon that is often used to demonstrate predictive coding theory as well.
The inventors have studied the different effects of the "push" feedback and the "pull" feedback, respectively, and analyzed the effects of the two, the former suppressing inter-class noise and the latter suppressing intra-class noise. That is, both forms of feedback interaction are necessary to maximize feedback gain.
In combination with theoretical analysis and simulation experimental results, the inventors believe that the optimal feedback modulation should be dynamic, that is, the feedback is positive at an early stage (typically 1 τ to 2.5 τ, about 50ms to 150ms, where τ is a time constant); in the later stage (typically 2 τ to 4 τ, approximately 150ms to 250ms, where τ is the time constant) the feedback becomes reversed. In this way, feedback modulation can suppress both inter-class noise and intra-class noise, thereby helping the multi-layer neural network to perform memory module retrieval from coarse to fine. The time constant tau is a parameter for regulating the network evolution rate and can be dynamically adjusted according to actual needs.
It should be noted that the early stage and the later stage may partially coincide, i.e. there is a coincidence time period (typically 2 τ to 2.5 τ). During this coincidence period, the "push" feedback and the "pull" feedback are simultaneously active. Relevant experiments prove that the retrieval effect is better under the condition that the overlapping time periods exist. Of course, by adjusting the time ranges of the early stage and the late stage, the coincidence period may be cancelled or extended.
On this basis, the inventors conducted a simulation experiment combining "push-pull (forward feedback-reverse feedback)". Here, the inventors only describe the results of using the continuous Hopfield model, because although the conclusions of the experiments on the discrete Hopfield model are also consistent in nature, the continuous model is more biologically reasonable than the discrete model, and can better illustrate the dynamics of the feedback modulation.
For brevity of the lines, the inventors herein only describe the neuron kinetic equations of layer 1, and the neuron kinetic equations of the other layers are similar in form, as shown below.
Wherein,andrepresenting the input and firing frequencies of the synapses, respectively, tau is a time constant,is the external input and actan (x) is the inverse function of tangent.
The feedback connection varies with time. At an early stage, the feedback is positive,a + is an adjustment coefficient, which is always a positive number. At a later stage, the feedback regulation is reversed, i.e.,a-is the adjustment coefficient, which is always a positive number. According to experimental data, the duration of the feedback adjustment set by the inventors is on the order of the membrane time constant, approximatelyAbout 1 tau to 2.5 tau.
FIG. 4 is a typical example of layer 1 neuron population activity, with the "push-pull" phenomenon observed in the experiment. Section B of fig. 4 shows the search accuracy of layer 1 and layer 2 in the same experiment as a function of time. It can be seen from figure 4 that the retrieval accuracy of the parent module at layer 2 quickly rises to a higher level (m is close to 1). And the retrieval precision of the layer 1 to the child modules is gradually increased under the feedback of 'push-pull', and finally, the accuracy is higher than the accuracy of the level without the feedback. This indicates that the "push-pull" feedback does improve the retrieval of the sub-module.
Based on the feedback modulation mechanism, the multilayer neural network provided by the invention can well reproduce the phenomenon of push-pull: in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback; in the later stage of object identification, the objects in the same class in the deep neural network are inhibited by reverse feedback to mix up the identified intra-class noise due to the similarity between the objects. The above process is a process of acquiring object information from a rough step to a fine step in terms of time. Again, this result is consistent with the "push-pull" phenomenon observed in the object recognition experiments.
In one embodiment of the invention, assuming the object recognition task is to identify cat or dog classes, in this example, a total of 18 cat or dog classes (9 classes each, sample pictures from the picture library ImageNet) are memorized, and the pictures are processed into column vectors by a pre-trained deep neural network (vgg), and it can be seen in conjunction with fig. 5 that the column vectors automatically form a hierarchical structure between them, more similar to the classes belonging to cats or dogs. The column vector for a particular picture is then entered into the network, and its category is identified and reproduced by the network. As shown in fig. 5, after inputting the picture of the object, the network evolution is started, a large category (cat category) is identified at layer 2, namely forward feedback, the similarity of the dog is suppressed, reverse feedback, and the similarity of other cats is suppressed in a feedback modulation mechanism, until the category identification of the cat or the dog is completed.
It should be noted that the real nervous system has sufficient capacity to implement a dynamic feedback modulation mechanism. For simplicity, the current model assumes that the "push" and "pull" feedback share the same neuron connection, but in practice they are most likely implemented through different signal paths. For example, "push" feedback may be accomplished by direct excitatory synaptic connections from high to low, and in such conditions, feedback cessation may be controlled by means of short-term inhibition of synapses. On the other hand, "pull" feedback may come from another pathway, such as inhibitory interneurons (inter-level inhibition is usually mediated by interneurons). This regulation is delayed compared to the direct connection pathway and is controlled by the response of the upper neurons. That is, the "pull" feedback is only initiated when a higher level object is identified at a higher level.
On the other hand, the conventional neural network with weighting constructed by the Hebbian learning rule cannot support too large storage capacity because the similarity between memory patterns can lead to rapid deterioration of retrieval. To address this deficiency, the present invention contemplates using the hierarchical features of the nervous system structure to classify information by objects stored in a hierarchical structure, thereby reducing the correlation between memory modules (since memory modules at higher levels are more independent). And the interference of intra-class and inter-class correlation between the lower-layer memory modes is restrained through 'push-pull' feedback during the memory retrieval, so that the memory retrieval effect is improved.
Further, the invention also provides an object recognition system based on feedback adjustment. As shown in fig. 6, the system includes a processor 12 and a memory 11 storing instructions executable by the processor 12;
The memory 11 is configured to store a program code, and transmit the program code to a CPU or a single chip Microcomputer (MCU). The memory 11 may include volatile memory, such as Random Access Memory (RAM); the memory 11 may also include a non-volatile memory, such as a read-only memory, a flash memory, a hard or solid state disk, an MCU, etc.; the memory 11 may also comprise a combination of memories of the kind described above.
Specifically, the object recognition system based on feedback adjustment provided by the embodiment of the invention comprises a processor 12 and a memory 11; the memory 11 has stored thereon a computer program operable on the processor 12, which when executed by the processor 12 performs the steps of:
visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; wherein,
in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback; in the later stage of object identification, the objects in the same class in the deep neural network are inhibited by reverse feedback to mix up the identified intra-class noise due to the similarity between the objects.
The embodiment of the invention also provides a computer readable storage medium. The computer-readable storage medium herein stores one or more programs. Among other things, computer-readable storage media may include volatile memory, such as random access memory; the memory may also include non-volatile memory, such as read-only memory, flash memory, hard or solid state disk, MCU, etc.; the memory may also comprise a combination of memories of the kind described above. When the one or more programs in the computer readable storage medium are executable by the one or more processors to perform some or all of the steps of the above-described feedback adjustment-based object recognition method.
Compared with the prior art, the object identification method and the object identification system provided by the invention adopt an inter-level feedback regulation mechanism organically combining forward feedback (push) and reverse feedback (pull). When the object recognition task is completed, the image of the object to be recognized is input into the first layer of the deep neural network, information is integrated from bottom to top, and meanwhile recurrence is attempted in the layers. The invention provides a heuristic for developing a better artificial intelligence algorithm to carry out object recognition.
The object recognition method and system based on feedback adjustment provided by the present invention are explained in detail above. Any obvious modifications to the invention, which would occur to those skilled in the art, without departing from the true spirit of the invention, would constitute a violation of the patent rights of the invention and would carry a corresponding legal responsibility.
Claims (10)
1. An object recognition method based on feedback adjustment adopts a deep neural network to perform object recognition, and is characterized by comprising the following steps:
visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; wherein,
in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback;
in the later stage of object identification, the identified intra-class noise is mixed up due to the similarity between the objects in the same class in the deep neural network through reverse feedback suppression,
the early stage is 1 tau to 2.5 tau, the later stage is 2 tau to 4 tau, where tau is a time constant,
the forward feedback connection weight from layer 2 to layer 1 between neuron i and neuron j is:
wherein α ═ 1.., PαParent moduleIs { xiα,βWhere β 1.., PβAnd the child module is { xiα,β,γWhere γ 1.., Pγ,Pα,PβAnd PγRespectively representing the number of ancestor modules;
2. The object recognition method according to claim 1, characterized by comprising the steps of:
the inter-class noise refers to noise caused by the table brother and sister modules of different parents of the same ancestor.
3. The object recognition method according to claim 1, characterized by comprising the steps of:
the intra-class noise refers to noise due to correlation between sub-modules of the same parent.
4. The object recognition method according to claim 1, characterized by comprising the steps of:
the forward feedback and the forward feedback act simultaneously during a coincidence time period of the early stage and the late stage.
5. The object recognition method according to claim 1, characterized by comprising the steps of:
in the deep neural network, feedforward connection and feedback connection are simultaneously arranged between each level, and recursive connection between neurons in the same layer is connection with associative memory function.
6. The object recognition method according to claim 1, characterized by comprising the steps of:
the forward feedback and the backward feedback share the same neuron connection.
7. The object recognition method according to claim 1, characterized by comprising the steps of:
if there is no direct connection between the forward feedback and the backward feedback, the backward feedback is only activated if a higher level object is identified at a higher level.
8. The object recognition method according to claim 1, characterized by comprising the steps of:
storing object classification information using a hierarchy of the deep neural network.
9. The object identification method according to any one of claims 1 to 8, characterized by comprising the steps of:
in the deep neural network, the function of each layer of neural network is described by a Hopfield model.
10. An object recognition system based on feedback adjustment, comprising a processor and a memory; the memory having stored thereon a computer program operable on the processor, the computer program when executed by the processor implementing the steps of:
visual information enters the deep neural network from a lower level and is propagated and integrated from bottom to top along the feedforward connection of the deep neural network; wherein,
in the early stage of object identification, inter-class noise caused by mutual interference among different classes in a deep neural network is inhibited through forward feedback; in the later stage of object identification, the identified intra-class noise is mixed up due to the similarity between the objects in the same class in the deep neural network through reverse feedback suppression,
the early stage is 1 tau to 2.5 tau, the later stage is 2 tau to 4 tau, where tau is a time constant,
the forward feedback connection weight from layer 2 to layer 1 between neuron i and neuron j is:
wherein α ═ 1.., PαThe parent module is { xiα,βWhere β 1.., PβAnd the child module is { xiα,β,γWhere γ 1.., Pγ,Pα,PβAnd PγRespectively representing the number of ancestor modules;
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810936167.7A CN109344958B (en) | 2018-08-16 | 2018-08-16 | Object identification method and identification system based on feedback adjustment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810936167.7A CN109344958B (en) | 2018-08-16 | 2018-08-16 | Object identification method and identification system based on feedback adjustment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109344958A CN109344958A (en) | 2019-02-15 |
CN109344958B true CN109344958B (en) | 2022-04-29 |
Family
ID=65291554
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810936167.7A Active CN109344958B (en) | 2018-08-16 | 2018-08-16 | Object identification method and identification system based on feedback adjustment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109344958B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105894012A (en) * | 2016-03-29 | 2016-08-24 | 天津大学 | Object identification method based on cascade micro neural network |
CN106203283A (en) * | 2016-06-30 | 2016-12-07 | 重庆理工大学 | Based on Three dimensional convolution deep neural network and the action identification method of deep video |
CN107977621A (en) * | 2017-11-29 | 2018-05-01 | 淮海工学院 | Shipwreck identification model construction method, device, electronic equipment and storage medium |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9633306B2 (en) * | 2015-05-07 | 2017-04-25 | Siemens Healthcare Gmbh | Method and system for approximating deep neural networks for anatomical object detection |
-
2018
- 2018-08-16 CN CN201810936167.7A patent/CN109344958B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105894012A (en) * | 2016-03-29 | 2016-08-24 | 天津大学 | Object identification method based on cascade micro neural network |
CN106203283A (en) * | 2016-06-30 | 2016-12-07 | 重庆理工大学 | Based on Three dimensional convolution deep neural network and the action identification method of deep video |
CN107977621A (en) * | 2017-11-29 | 2018-05-01 | 淮海工学院 | Shipwreck identification model construction method, device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN109344958A (en) | 2019-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Dawid et al. | Modern applications of machine learning in quantum sciences | |
US9053431B1 (en) | Intelligent control with hierarchical stacked neural networks | |
US11126913B2 (en) | Methods and systems for implementing deep spiking neural networks | |
Gallo | Artificial neural networks tutorial | |
Chakraborty et al. | Hellinger net: A hybrid imbalance learning model to improve software defect prediction | |
KR20160130398A (en) | Photo management | |
Han et al. | Efficient self-organizing multilayer neural network for nonlinear system modeling | |
Shi et al. | A curiosity-based learning method for spiking neural networks | |
Buscema et al. | Artificial Adaptive Systems Using Auto Contractive Maps | |
Iranmehr et al. | Bio-inspired evolutionary model of spiking neural networks in ionic liquid space | |
Chanamarn et al. | Enhancing efficient study plan for student with machine learning techniques | |
Subramanian et al. | A meta-cognitive interval type-2 fuzzy inference system classifier and its projection based learning algorithm | |
CN109344958B (en) | Object identification method and identification system based on feedback adjustment | |
Murtagh | Neural networks and related'massively parallel'methods for statistics: A short overview | |
CN114118357A (en) | Retraining method and system for replacing activation function in computer visual neural network | |
Mihai et al. | Optimizing a convolutional neural network using particle swarm optimization | |
Matheus et al. | Learning in artificial neural systems | |
Li et al. | LendingClub loan default and profitability prediction | |
Zhang et al. | Intelligent Information Processing with Matlab | |
Marchev et al. | Evolution of the Concept of Self-Organization by the Founding Fathers of AI | |
Motzev | Statistical learning networks in simulations for business training and education | |
Scholtes | Kohonen feature maps in natural language processing | |
Koromyslova et al. | About the effectiveness of evolutionary algorithms for multicriterial design of artificial neural networks | |
WO2024016945A1 (en) | Training method for image classification model, image classification method, and related device | |
Farci | The art of character recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |