EP0474222B1 - Learning machine with a hierarchical structure without mutual connections within levels thereof - Google Patents
Learning machine with a hierarchical structure without mutual connections within levels thereof Download PDFInfo
- Publication number
- EP0474222B1 EP0474222B1 EP91114955A EP91114955A EP0474222B1 EP 0474222 B1 EP0474222 B1 EP 0474222B1 EP 91114955 A EP91114955 A EP 91114955A EP 91114955 A EP91114955 A EP 91114955A EP 0474222 B1 EP0474222 B1 EP 0474222B1
- Authority
- EP
- European Patent Office
- Prior art keywords
- weight
- amount
- learning
- error
- output signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
Definitions
- This invention relates to a learning machine capable of learning by means of a hierarchical structure without mutual connections within levels thereof, and having plural multiple-input single-output signal processors connected in a network structure in such a manner that signals propagate only to a higher level.
- Fig. 8 shows a structure of a conventional learning machine.
- this conventional learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 2 which renews the value of the weight coefficient of the output signal calculation means 1 based on the output signal obtained therefrom.
- the output signal calculation means 1 is a multiple stage circuit network as shown in Fig. 9 comprising plural multiple-input single-output signal processors 3 and plural input means 4.
- a specific example of the multiple-input single-output signal processor 3 used in this output signal calculation means 1 is shown in Fig. 10. As shown in Fig.
- each multiple-input single-output signal processor 3 comprises plural input means 5, a memory 6 to store the weight coefficients which weight the plural inputs from the input means 5, plural multipliers 7 which multiply the weight coefficients from the memory 6 by the inputs from the input means 5, an adder 8 which adds the outputs from each of the multipliers 7, and a threshold value processor 9 which limits the output from the adder 8 to a value within a predetermined range.
- the input/output characteristics of the threshold value processor 9 are shown in Fig. 11.
- the weight coefficient renewing means 2 shown in Fig. 8 comprises a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- a conventional learning machine constructed as described above operates as described below.
- each of the multiple-input single-output signal processors 3 multiplies the output of each of the lower level multiple-input single-output signal processors 3 connected thereto by a weight coefficient, which represents the significance of the connection and is stored in the memory 6, by means of the multipliers 7 and then obtains the sum of the outputs from the multipliers 7 by means of the adder 8; this sum is converted by the threshold value processor 9 and the resulting value is output to the multiple-input single-output signal processors 3 one level higher.
- Fig. 11 is a graph of the input/output characteristics of a function f which expresses a threshold value process of the multiple-input single-output signal processor 3, wherein I in Fig. 11 is the input value to the threshold value processor 9.
- the teacher signal generator 10 in the weight coefficient renewing means 2 generates a desirable output signal for the number p input signal input from the input section 4 of the output signal calculation means 1; this output signal is used as the teacher signal t pk (where t pk expresses the teacher signal for the output of the number k multiple-input single-output signal processor in the highest level of the output signal calculation means 1).
- ⁇ p is the sum for all input signals
- ⁇ k is the sum of the outputs of all multiple-input single-output signal processors in the highest level of the output signal calculation means 1
- W is a vector which has the weight coefficient W ij as each component, and which hereinafter is called a weight vector.
- the error E becomes a function of the weight vector W .
- the teacher signal t pk is a value of either 0 or 1.
- the direction of steepest descent is a vector of which a component is a differential of the error E to the weight coefficient W ij .
- the weight modification means 14 changes the weight vector of the output signal calculation means according to the amount of change in the weight vector. The amount of error is reduced by thus repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- the weights of multiple-input single output signal processors are simply changed while an error for each of the multiple-input single-output signal processors becomes sufficiently small, if a total error, which is expressed as the total sum square error, decreases, even if there remains the other multiple-input single-output signal processor with large error, thus learning efficiency deteriorates.
- an object of the present invention is to provide a learning machine which requires a short period of time for learning.
- a learning machine of a second embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- a learning machine of a third embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- a learning machine of a fourth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- a learning machine of a fifth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- a learning machine of a sixth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- a learning machine of a seventh embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
- Fig. 1 is a block diagram of a first embodiment of a learning machine according to the present invention.
- this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 102.
- the weight coefficient renewing means 102 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 102 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, and a first amount of weight change controller 20.
- the output signal calculation means 1 calculates an output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- 2 d' is a conjugate gradient direction obtained from the previous learning cycle, and
- the threshold value controller 16 switches a threshold value T1 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- the weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T1 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13.
- the weight change direction 1 selection means 19 may select the direction of steepest descent for the first part of the learning cycle and the direction of the conjugate gradient for the later part of the learning cycle.
- the selected direction of weight change is the direction of the conjugate gradient
- the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
- the weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- Fig. 2 is a block diagram of a second embodiment of a learning machine according to the present invention.
- this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 202.
- the weight coefficient renewing means 202 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 202 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
- the output signal calculation means 1 calculates an output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value T1 in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of an error curve from the values of errors to learning rates with low error values, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
- the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T1, and outputs the result to the weight modification means 14.
- the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
- the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- Fig. 3 is a block diagram of a third embodiment of a learning machine according to the present invention.
- this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 302.
- the weight coefficient renewing means 302 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 302 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
- an output signal calculation means 1 calculates an output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- 2 d' is a conjugate gradient direction in the previous learning cycle, and
- the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
- the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T1, and outputs the result to the weight modification means 14.
- the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
- the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- Fig. 4 is a block diagram of a fourth embodiment of a learning machine according to the present invention.
- this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 402.
- the weight coefficient renewing means 402 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 402 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, a skip evaluation means 24, and a weight change controller 25.
- the output signal calculation means 1 calculates the output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- 2 d' is a conjugate gradient direction in the previous learning cycle, and
- the threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- the weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T2 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13.
- the weight change direction selection means 19 may select the direction of steepest descent for learning during early learning process and the direction of the conjugate gradient for the later learning process.
- the selected direction of weight change is the direction of the conjugate gradient
- the weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
- the first error signal evaluation means 17 determines whether the error
- the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
- the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the error of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but the threshold value T2 which is the evaluation standard for whether the weight is changed is also reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
- the direction of weight change is set to the direction of steepest descent, which is locally the most efficient, during the early learning process and is set to the direction of the conjugate gradient, which is generally the most efficient, during the later learning process to change the weight, learning efficiency improves and the time required for learning is shortened.
- Fig. 5 is a block diagram of a fifth embodiment of a learning machine according to the present invention.
- this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 502.
- the weight coefficient renewing means 502 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
- the output signal calculation means 1 calculates the output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
- the threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T2, and outputs the result to the weight modification means 14.
- the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
- the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
- the first error signal evaluation means 17 determines whether the error
- the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
- the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
- the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
- Fig. 6 is a block diagram of a sixth embodiment of a learning machine according to the present invention.
- this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 602.
- Te weight coefficient renewing means 602 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 602 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
- the output signal calculation means 1 calculates the output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
- the threshold value controller 16 switches the threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
- the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
- the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T2, and outputs the result to the weight modification means 14.
- the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
- the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
- the first error signal evaluation means 17 determines whether the error
- the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
- the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
- the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
- Fig. 7 is a block diagram of a seventh embodiment of a learning machine according to the present invention.
- this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 702.
- the weight coefficient renewing means 702 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
- Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
- the weight coefficient renewing means 702 includes a second error signal evaluation means 26, a teacher signal evaluation means 27, and a second amount of weight change controller 28.
- the output signal calculation means 1 calculates the output signal
- the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
- the second amount of weight change controller 28 multiplies by m1 (m1 > 1) times the amount of change for the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
- the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
- the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
- the multiple-input single-output signal processors 3 with a large error value can be more quickly converged and the time required for learning can be shortened because, of the multiple-input single-output signal processors 3 with a large error value, and the weight coefficient of these which should output a particularly slow convergence value "1" (multiple-input single-output signal processors 3 with a teacher signal value of "1") is actively changed.
- the second amount of weight change controller 28 in this embodiment may increase the amount of weight change m2 (m2 > 1) times for multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Life Sciences & Earth Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Feedback Control In General (AREA)
- Image Analysis (AREA)
- Air Conditioning Control Device (AREA)
Description
- the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value of the threshold value control means, a first amount of weight change control means which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the first error signal evaluation means, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, and a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means.
- the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a conjugate gradient direction determination means to obtain a direction of the conjugate gradient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a first amount of weight change control means which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the first error signal evaluation means, an amount of weight change selection means which selects either an output of an amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, and a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means.
- the weight coefficient renewing means further comprising a conjugate gradient direction determination means to obtain a direction of a conjugate gradient from the direction of steepest descent, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts the outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, a weight change direction selection means which selects either the direction of the conjugate gradient or the direction of the steepest descent as the direction of weight change according to the level of the threshold value, an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the selected direction of weight change, a weight modification means which changes the value of the weight coefficient stored in the memory according to the output of the amount of weight change calculation means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
- the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces the threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts the outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
- the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a conjugate gradient direction determination means to obtain a direction of the conjugate gradient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
- the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient stored in the memory from the direction of steepest descent, a second error signal evaluation means which determines whether the absolute value of an output of the error signal calculation means is greater than a threshold value, a teacher signal evaluation means which determines whether an output of the teacher signal generator is equal to "1", a second amount of weight change control means which controls an amount of change in the weight coefficient according to the outputs of the teacher signal evaluation means and the second error signal evaluation means, and a weight modification means which changes the value of the weight coefficient stored in the memory according to the output of the amount of weight change calculation means.
Claims (7)
- A learning machine having an output signal calculation means (1) comprising plural multiple-input single-output signal processors (3) network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means (2) which renews values of weight coefficients of the output signal calculation means (1) based on output signals obtained by the output signal calculation means (1),each multiple-input single-output signal processor (3) comprising a memory (6) to store plural weight coefficients, an input section (5) for inputting plural data, multiplying means (7) to weight the input data from the input section (5) by the weight coefficients stored in the memory (6), an adding means (8) to add the plural data weighted by the multiplying means (7), and a threshold value processor (9) to limit an output of the adding means (8) to a value within a predetermined range, andthe weight coefficient renewing means (2) comprising a teacher signal generator (10) to produce as a teacher signal a desirable value for the output signal of the output signal calculation means (1), an error signal calculation means (11) to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means (12) to obtain a direction of steepest descent from an output of the error signal calculation means (11), an amount of weight change calculation means (13) for calculating an amount of change in the weight coefficient from the direction of weight change, and a weight modification means (14) which changes the value of the weight coefficient stored in the memory (6) according to an output of the amount of weight change calculation means (13),the weight coefficient renewing means (102; 402) further comprises a conjugate gradient direction determination means (18) to obtain a direction of a conjugate gradient from the direction of steepest descent, a learning progress evaluation means (15) to evaluate progress of learning from the error output from the error signal calculation means (11), a threshold value control means (16) which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means (15), an error signal evaluation means (17) which evaluates whether the output of the error signal calculation means (11) is less than the threshold value of the threshold value control means (16), a weight change direction selection means (19) which selects either the direction of the conjugate gradient or the direction of the steepest descent as a direction of weight change according to the level of the threshold value, wherein the amount of weight change calculation means (13) calculates an amount of change in the weight coefficient from the selected direction of weight change, and an amount of weight change control means (20) which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the error signal evaluation means (17).
- A learning machine according to the preamble of claim 1,
characterized in thatthe weight coefficient renewing means (202, 502) further comprises a linear search means (21 ) which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means (22) which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means (21), a learning progress evaluation means (15) evaluating progress of learning from the error output by the error signal calculation means (11), a threshold value control means (16) which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means (15), an error signal evaluation means (17) which evaluates whether the output of the error signal calculation means (11) is less than the threshold value of the threshold value control means (16), an amount of weight change control means (20) which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the error signal evaluation means (17), and an amount of weight change selection means (23) which selects either an amount of change in the weight coefficient from the direction of steepest descent calculated by the amount of weight change calculation means (13) or an output of the parabolic line approximation means (22) according to the level of the threshold value, wherein the weight modification means (14) changes the value of the weight coefficient stored in the memory (6) according to an amount of weight change selected by the amount of weight change selection means (23). - The learning machine according to claim 2,
characterized in thatthe weight coefficient renewing means (302; 602) further comprises a conjugate gradient direction determination means (18) to obtain a direction of the conjugate gradient from the direction of steepest descent, and that the linear search means (21) outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient. - The learning machine according to anyone of the preceding claims,
further characterized in thatthe weight coefficient renewing means further comprises a skip evaluation means (24) which counts the outputs of the error signal evaluation means (17) and outputs a skip signal when all output signals are less than the threshold value, and a weight change control means (25) which causes the weight modification means (14) to skip a weight change operation according to the skip signal. - A learning machine according to the preamble of claim 1,
characterized in thatthe weight coefficient renewing means (702) further comprises an error signal evaluation means (26) which determines whether the absolute value of an output of the error signal calculation means (11) is greater than a threshold value, a teacher signal evaluation means (27) which determines whether an output of the teacher signal generator (10) is equal to "1", and an amount of weight change control means (28) which controls an amount of change in the weight coefficient according to the outputs of the teacher signal evaluation means (27) and the error signal evaluation means (26). - The learning machine according to claim 5 wherein the amount of weight change control means (28) multiplies by m1 (m1 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) in the highest level of the output signal calculation means (1) for which the value of the error is evaluated by the error signal evaluation means to be greater than the threshold value and the value of the teacher signal is evaluated to be "1" by the teacher signal evaluation means (27).
- The learning machine according to claim 5 wherein the amount of weight change control means (28) multiplies by m2 (m2 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) in the highest level of the output signal calculation means (1) for which the value of the error is evaluated by the error signal evaluation means (26) to be greater than the threshold value, and multiplies by m3 (m3 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) for which the value of the teacher signal is further evaluated to be "1" by the teacher signal evaluation means (27).
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP234287/90 | 1990-09-04 | ||
JP2234287A JPH04114263A (en) | 1990-09-04 | 1990-09-04 | Learning machine |
Publications (3)
Publication Number | Publication Date |
---|---|
EP0474222A2 EP0474222A2 (en) | 1992-03-11 |
EP0474222A3 EP0474222A3 (en) | 1994-05-18 |
EP0474222B1 true EP0474222B1 (en) | 1998-10-21 |
Family
ID=16968621
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP91114955A Expired - Lifetime EP0474222B1 (en) | 1990-09-04 | 1991-09-04 | Learning machine with a hierarchical structure without mutual connections within levels thereof |
Country Status (4)
Country | Link |
---|---|
US (1) | US5295228A (en) |
EP (1) | EP0474222B1 (en) |
JP (1) | JPH04114263A (en) |
DE (1) | DE69130387T2 (en) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0635888A (en) * | 1992-07-15 | 1994-02-10 | Mitsubishi Electric Corp | Learning method for neural network and electric system voltage/reactive power controller using the learning method |
US5726688A (en) * | 1995-09-29 | 1998-03-10 | Ncr Corporation | Predictive, adaptive computer interface |
US6652283B1 (en) | 1999-12-30 | 2003-11-25 | Cerego, Llc | System apparatus and method for maximizing effectiveness and efficiency of learning retaining and retrieving knowledge and skills |
US7319943B2 (en) * | 2002-12-16 | 2008-01-15 | Canon Kabushiki Kaisha | Optimum design method, and apparatus, and program for the same |
US8873813B2 (en) | 2012-09-17 | 2014-10-28 | Z Advanced Computing, Inc. | Application of Z-webs and Z-factors to analytics, search engine, learning, recognition, natural language, and other utilities |
US11195057B2 (en) | 2014-03-18 | 2021-12-07 | Z Advanced Computing, Inc. | System and method for extremely efficient image and pattern recognition and artificial intelligence platform |
US9916538B2 (en) | 2012-09-15 | 2018-03-13 | Z Advanced Computing, Inc. | Method and system for feature detection |
US11074495B2 (en) | 2013-02-28 | 2021-07-27 | Z Advanced Computing, Inc. (Zac) | System and method for extremely efficient image and pattern recognition and artificial intelligence platform |
US11914674B2 (en) | 2011-09-24 | 2024-02-27 | Z Advanced Computing, Inc. | System and method for extremely efficient image and pattern recognition and artificial intelligence platform |
CN102866179B (en) * | 2012-09-13 | 2014-06-18 | 重庆大学 | Online recognition and inhibition method based on non-target interference smell in electronic nose of artificial intelligent learning machine |
CN104200268B (en) * | 2014-09-03 | 2017-02-15 | 辽宁大学 | PSO (Particle Swarm Optimization) extremity learning machine based strip steel exit thickness predicting method |
CN115171453B (en) * | 2022-06-13 | 2023-05-30 | 广东电网有限责任公司 | Artificial customer service training method and system based on smart grid |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS62104324A (en) * | 1985-10-31 | 1987-05-14 | Toshiba Corp | Adaptive automatic equalizer |
US4912652A (en) * | 1988-12-14 | 1990-03-27 | Gte Laboratories Incorporated | Fast neural network training |
US5129038A (en) * | 1989-02-23 | 1992-07-07 | Matsushita Electric Industrial Co., Ltd. | Neural network with selective error reduction to increase learning speed |
EP0579341A2 (en) * | 1989-02-23 | 1994-01-19 | Matsushita Electric Industrial Co., Ltd. | Learning machine with multi-input single output circuits connected in hierarchical structure |
-
1990
- 1990-09-04 JP JP2234287A patent/JPH04114263A/en active Pending
-
1991
- 1991-09-04 US US07/754,517 patent/US5295228A/en not_active Expired - Fee Related
- 1991-09-04 EP EP91114955A patent/EP0474222B1/en not_active Expired - Lifetime
- 1991-09-04 DE DE69130387T patent/DE69130387T2/en not_active Expired - Fee Related
Also Published As
Publication number | Publication date |
---|---|
DE69130387D1 (en) | 1998-11-26 |
EP0474222A3 (en) | 1994-05-18 |
US5295228A (en) | 1994-03-15 |
JPH04114263A (en) | 1992-04-15 |
DE69130387T2 (en) | 1999-05-06 |
EP0474222A2 (en) | 1992-03-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5479571A (en) | Neural node network and model, and method of teaching same | |
EP0474222B1 (en) | Learning machine with a hierarchical structure without mutual connections within levels thereof | |
Gilbert et al. | Fast reference governors for systems with state and control constraints and disturbance inputs | |
Haykin et al. | Optimum nonlinear filtering | |
US5129038A (en) | Neural network with selective error reduction to increase learning speed | |
Lee et al. | Channel equalization using radial basis function network | |
US4674028A (en) | Identification method of a process parameter | |
US5535149A (en) | Duplex adaptive digital filter and method of operation | |
US6223194B1 (en) | Adaptive filter, step size control method thereof, and record medium therefor | |
US5168550A (en) | Neutral network with plural weight calculation methods and variation of plural learning parameters | |
US5384896A (en) | Learning machine | |
US5319587A (en) | Computing element for neural networks | |
US5701395A (en) | Method of programming a polynomial processor | |
US5898731A (en) | Auto-coefficient renewal digital channel equalizer | |
KR102574719B1 (en) | Electronice device and learning method for low complexity artificial intelligentce model learning based on selecting the dynamic prediction confidence thresholed | |
Su et al. | Processing element architecture design for deep reinforcement learning with flexible block floating point exploiting signal statistics | |
US5542005A (en) | Recognition unit and recognition apparatus | |
Ludermir | Logical networks capable of computing weighted regular languages | |
JP3328953B2 (en) | Nonlinear synaptic neuron, device thereof and prediction method using the same | |
Benureau et al. | Learning Timescales in MTRNNs | |
KR20220169085A (en) | Device for performing reinforcement learning and operation method thereof | |
Ammoumou et al. | A New FTF9N Stabilized Recursive Algorithm, Implementation on Finite-Precision | |
JPH1091604A (en) | Function learning device | |
CN114492748A (en) | Echo state network forward propagation acceleration algorithm based on activation function approximation | |
CN118211628A (en) | Model training method, device, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 19910905 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): DE FR GB |
|
PUAL | Search report despatched |
Free format text: ORIGINAL CODE: 0009013 |
|
AK | Designated contracting states |
Kind code of ref document: A3 Designated state(s): DE FR GB |
|
17Q | First examination report despatched |
Effective date: 19961031 |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE FR GB |
|
REF | Corresponds to: |
Ref document number: 69130387 Country of ref document: DE Date of ref document: 19981126 |
|
ET | Fr: translation filed | ||
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 19990901 Year of fee payment: 9 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 19990906 Year of fee payment: 9 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 19990909 Year of fee payment: 9 |
|
26N | No opposition filed | ||
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20000904 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20000904 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20010531 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20010601 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST |