EP0474222B1 - Learning machine with a hierarchical structure without mutual connections within levels thereof - Google Patents

Learning machine with a hierarchical structure without mutual connections within levels thereof Download PDF

Info

Publication number
EP0474222B1
EP0474222B1 EP91114955A EP91114955A EP0474222B1 EP 0474222 B1 EP0474222 B1 EP 0474222B1 EP 91114955 A EP91114955 A EP 91114955A EP 91114955 A EP91114955 A EP 91114955A EP 0474222 B1 EP0474222 B1 EP 0474222B1
Authority
EP
European Patent Office
Prior art keywords
weight
amount
learning
error
output signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
EP91114955A
Other languages
German (de)
French (fr)
Other versions
EP0474222A3 (en
EP0474222A2 (en
Inventor
Toshiyuki Koda
Yasuharu Shimeki
Shigeo Sakaue
Hiroshi Yamamoto
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Panasonic Holdings Corp
Original Assignee
Matsushita Electric Industrial Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Matsushita Electric Industrial Co Ltd filed Critical Matsushita Electric Industrial Co Ltd
Publication of EP0474222A2 publication Critical patent/EP0474222A2/en
Publication of EP0474222A3 publication Critical patent/EP0474222A3/en
Application granted granted Critical
Publication of EP0474222B1 publication Critical patent/EP0474222B1/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Definitions

  • This invention relates to a learning machine capable of learning by means of a hierarchical structure without mutual connections within levels thereof, and having plural multiple-input single-output signal processors connected in a network structure in such a manner that signals propagate only to a higher level.
  • Fig. 8 shows a structure of a conventional learning machine.
  • this conventional learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 2 which renews the value of the weight coefficient of the output signal calculation means 1 based on the output signal obtained therefrom.
  • the output signal calculation means 1 is a multiple stage circuit network as shown in Fig. 9 comprising plural multiple-input single-output signal processors 3 and plural input means 4.
  • a specific example of the multiple-input single-output signal processor 3 used in this output signal calculation means 1 is shown in Fig. 10. As shown in Fig.
  • each multiple-input single-output signal processor 3 comprises plural input means 5, a memory 6 to store the weight coefficients which weight the plural inputs from the input means 5, plural multipliers 7 which multiply the weight coefficients from the memory 6 by the inputs from the input means 5, an adder 8 which adds the outputs from each of the multipliers 7, and a threshold value processor 9 which limits the output from the adder 8 to a value within a predetermined range.
  • the input/output characteristics of the threshold value processor 9 are shown in Fig. 11.
  • the weight coefficient renewing means 2 shown in Fig. 8 comprises a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • a conventional learning machine constructed as described above operates as described below.
  • each of the multiple-input single-output signal processors 3 multiplies the output of each of the lower level multiple-input single-output signal processors 3 connected thereto by a weight coefficient, which represents the significance of the connection and is stored in the memory 6, by means of the multipliers 7 and then obtains the sum of the outputs from the multipliers 7 by means of the adder 8; this sum is converted by the threshold value processor 9 and the resulting value is output to the multiple-input single-output signal processors 3 one level higher.
  • Fig. 11 is a graph of the input/output characteristics of a function f which expresses a threshold value process of the multiple-input single-output signal processor 3, wherein I in Fig. 11 is the input value to the threshold value processor 9.
  • the teacher signal generator 10 in the weight coefficient renewing means 2 generates a desirable output signal for the number p input signal input from the input section 4 of the output signal calculation means 1; this output signal is used as the teacher signal t pk (where t pk expresses the teacher signal for the output of the number k multiple-input single-output signal processor in the highest level of the output signal calculation means 1).
  • ⁇ p is the sum for all input signals
  • ⁇ k is the sum of the outputs of all multiple-input single-output signal processors in the highest level of the output signal calculation means 1
  • W is a vector which has the weight coefficient W ij as each component, and which hereinafter is called a weight vector.
  • the error E becomes a function of the weight vector W .
  • the teacher signal t pk is a value of either 0 or 1.
  • the direction of steepest descent is a vector of which a component is a differential of the error E to the weight coefficient W ij .
  • the weight modification means 14 changes the weight vector of the output signal calculation means according to the amount of change in the weight vector. The amount of error is reduced by thus repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • the weights of multiple-input single output signal processors are simply changed while an error for each of the multiple-input single-output signal processors becomes sufficiently small, if a total error, which is expressed as the total sum square error, decreases, even if there remains the other multiple-input single-output signal processor with large error, thus learning efficiency deteriorates.
  • an object of the present invention is to provide a learning machine which requires a short period of time for learning.
  • a learning machine of a second embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • a learning machine of a third embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • a learning machine of a fourth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • a learning machine of a fifth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • a learning machine of a sixth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • a learning machine of a seventh embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • Fig. 1 is a block diagram of a first embodiment of a learning machine according to the present invention.
  • this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 102.
  • the weight coefficient renewing means 102 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 102 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, and a first amount of weight change controller 20.
  • the output signal calculation means 1 calculates an output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • 2 d' is a conjugate gradient direction obtained from the previous learning cycle, and
  • the threshold value controller 16 switches a threshold value T1 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • the weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T1 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13.
  • the weight change direction 1 selection means 19 may select the direction of steepest descent for the first part of the learning cycle and the direction of the conjugate gradient for the later part of the learning cycle.
  • the selected direction of weight change is the direction of the conjugate gradient
  • the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
  • the weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • Fig. 2 is a block diagram of a second embodiment of a learning machine according to the present invention.
  • this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 202.
  • the weight coefficient renewing means 202 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 202 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
  • the output signal calculation means 1 calculates an output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value T1 in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of an error curve from the values of errors to learning rates with low error values, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
  • the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T1, and outputs the result to the weight modification means 14.
  • the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
  • the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • Fig. 3 is a block diagram of a third embodiment of a learning machine according to the present invention.
  • this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 302.
  • the weight coefficient renewing means 302 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 302 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
  • an output signal calculation means 1 calculates an output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • 2 d' is a conjugate gradient direction in the previous learning cycle, and
  • the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
  • the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T1, and outputs the result to the weight modification means 14.
  • the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
  • the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • Fig. 4 is a block diagram of a fourth embodiment of a learning machine according to the present invention.
  • this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 402.
  • the weight coefficient renewing means 402 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 402 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, a skip evaluation means 24, and a weight change controller 25.
  • the output signal calculation means 1 calculates the output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • 2 d' is a conjugate gradient direction in the previous learning cycle, and
  • the threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • the weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T2 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13.
  • the weight change direction selection means 19 may select the direction of steepest descent for learning during early learning process and the direction of the conjugate gradient for the later learning process.
  • the selected direction of weight change is the direction of the conjugate gradient
  • the weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
  • the first error signal evaluation means 17 determines whether the error
  • the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
  • the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the error of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but the threshold value T2 which is the evaluation standard for whether the weight is changed is also reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
  • the direction of weight change is set to the direction of steepest descent, which is locally the most efficient, during the early learning process and is set to the direction of the conjugate gradient, which is generally the most efficient, during the later learning process to change the weight, learning efficiency improves and the time required for learning is shortened.
  • Fig. 5 is a block diagram of a fifth embodiment of a learning machine according to the present invention.
  • this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 502.
  • the weight coefficient renewing means 502 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
  • the output signal calculation means 1 calculates the output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
  • the threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T2, and outputs the result to the weight modification means 14.
  • the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
  • the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
  • the first error signal evaluation means 17 determines whether the error
  • the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
  • the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
  • the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
  • Fig. 6 is a block diagram of a sixth embodiment of a learning machine according to the present invention.
  • this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 602.
  • Te weight coefficient renewing means 602 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 602 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
  • the output signal calculation means 1 calculates the output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • the linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
  • the threshold value controller 16 switches the threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15.
  • the learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
  • the amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T2, and outputs the result to the weight modification means 14.
  • the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
  • the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
  • the first error signal evaluation means 17 determines whether the error
  • the skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error
  • the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses.
  • the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
  • Fig. 7 is a block diagram of a seventh embodiment of a learning machine according to the present invention.
  • this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 702.
  • the weight coefficient renewing means 702 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
  • Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove.
  • the weight coefficient renewing means 702 includes a second error signal evaluation means 26, a teacher signal evaluation means 27, and a second amount of weight change controller 28.
  • the output signal calculation means 1 calculates the output signal
  • the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10.
  • the second amount of weight change controller 28 multiplies by m1 (m1 > 1) times the amount of change for the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error
  • the weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
  • the amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
  • the multiple-input single-output signal processors 3 with a large error value can be more quickly converged and the time required for learning can be shortened because, of the multiple-input single-output signal processors 3 with a large error value, and the weight coefficient of these which should output a particularly slow convergence value "1" (multiple-input single-output signal processors 3 with a teacher signal value of "1") is actively changed.
  • the second amount of weight change controller 28 in this embodiment may increase the amount of weight change m2 (m2 > 1) times for multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Feedback Control In General (AREA)
  • Image Analysis (AREA)
  • Air Conditioning Control Device (AREA)

Description

BACKGROUND OF THE INVENTION
This invention relates to a learning machine capable of learning by means of a hierarchical structure without mutual connections within levels thereof, and having plural multiple-input single-output signal processors connected in a network structure in such a manner that signals propagate only to a higher level.
A conventional learning machine has been described by D.E. Rumelhart, G.E. Hinton, and R.J. Williams in their article "Learning Representations by Back-Propagating Errors" in the Oct. 9, 1986 issue of Nature (vol. 323, pp. 533 - 536).
Fig. 8 shows a structure of a conventional learning machine. As shown in the block diagram of Fig. 8, this conventional learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 2 which renews the value of the weight coefficient of the output signal calculation means 1 based on the output signal obtained therefrom. The output signal calculation means 1 is a multiple stage circuit network as shown in Fig. 9 comprising plural multiple-input single-output signal processors 3 and plural input means 4. A specific example of the multiple-input single-output signal processor 3 used in this output signal calculation means 1 is shown in Fig. 10. As shown in Fig. 10, each multiple-input single-output signal processor 3 comprises plural input means 5, a memory 6 to store the weight coefficients which weight the plural inputs from the input means 5, plural multipliers 7 which multiply the weight coefficients from the memory 6 by the inputs from the input means 5, an adder 8 which adds the outputs from each of the multipliers 7, and a threshold value processor 9 which limits the output from the adder 8 to a value within a predetermined range. The input/output characteristics of the threshold value processor 9 are shown in Fig. 11.
For example, the input/output characteristics of the threshold value processor 9 which limits the output to a value within the range of (0, 1) can be numerically expressed as f(I) = 1/(1 + exp(-I+)).
It is to be noted that the weight coefficient renewing means 2 shown in Fig. 8 comprises a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14.
A conventional learning machine constructed as described above operates as described below.
When an input signal is input to the input section 4 of the output signal calculation means 1, each of the multiple-input single-output signal processors 3 multiplies the output of each of the lower level multiple-input single-output signal processors 3 connected thereto by a weight coefficient, which represents the significance of the connection and is stored in the memory 6, by means of the multipliers 7 and then obtains the sum of the outputs from the multipliers 7 by means of the adder 8; this sum is converted by the threshold value processor 9 and the resulting value is output to the multiple-input single-output signal processors 3 one level higher. In other words, the multiple-input single-output signal processor 3 shown in Fig. 10 processes the equation oi = f (Σjwijoj) where oj is the input value to the input means 5 (the output of the lower level multiple-input single-output signal processor at position j) and wij is the weight coefficient stored in the memory 6 (the connection weight of the i position multiple-input single-output signal processor to the lower level j position multiple-input single-output signal processor). Fig. 11 is a graph of the input/output characteristics of a function f which expresses a threshold value process of the multiple-input single-output signal processor 3, wherein I in Fig. 11 is the input value to the threshold value processor 9.
The teacher signal generator 10 in the weight coefficient renewing means 2 generates a desirable output signal for the number p input signal input from the input section 4 of the output signal calculation means 1; this output signal is used as the teacher signal tpk (where tpk expresses the teacher signal for the output of the number k multiple-input single-output signal processor in the highest level of the output signal calculation means 1). The error signal calculator 11 then obtains the error from the difference between the teacher signal tpk and the actual output signal opk to the number p input signal output from the output signal calculation means 1 (where opk expresses the output of the number k multiple-input single-output signal processor in the highest level of the output signal calculation means 1) using the equation E = 0.5ΣpΣk (tpk - opk)2 = E ( W) and applies this value to evaluate the performance of the network at the present connection state (weight coefficient value). In equation (1), Σp is the sum for all input signals, Σk is the sum of the outputs of all multiple-input single-output signal processors in the highest level of the output signal calculation means 1, and W is a vector which has the weight coefficient Wij as each component, and which hereinafter is called a weight vector. The error E becomes a function of the weight vector W. Furthermore, the teacher signal tpk is a value of either 0 or 1.
Based on the obtained error E, the steepest descent direction determination means 12 computes g = ∂E/∂ W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1. The direction of steepest descent is a vector of which a component is a differential of the error E to the weight coefficient Wij.
The amount of weight change calculator 13 obtains the amount of change in the weight vector of the output signal calculation means from this direction of steepest descent by the equation ΔW = -ε * ∂E/∂ W + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle. The weight modification means 14 changes the weight vector of the output signal calculation means according to the amount of change in the weight vector. The amount of error is reduced by thus repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
Because learning is accomplished by minimizing the sum square error E in the conventional learning machine as described above, the weights of multiple-input single output signal processors are simply changed while an error for each of the multiple-input single-output signal processors becomes sufficiently small, if a total error, which is expressed as the total sum square error, decreases, even if there remains the other multiple-input single-output signal processor with large error, thus learning efficiency deteriorates.
Furthermore, minimizing of the sum square error is not limited to changing the weight so that the error of the multiple-input single-output signal processor with the greatest error is reduced, and if the error of only some of the multiple-input single-output signal processors in the highest level remains extremely high without converging and yet the block causing this high error is biased towards the multiple-input single-output signal processors which should output "1" (the teacher signal = 1), convergence speed thereof will deteriorate compared with the multiple-input single-output signal processors which should output "0" (teacher signal = 0).
Therefore, the problem exists with the conventional learning machine as described above that the time required for learning becomes greater.
SUMMARY OF THE INVENTION
Therefore, an object of the present invention is to provide a learning machine which requires a short period of time for learning.
To achieve the aforementioned object, a learning machine according to the present invention is described in claim 1.
A learning machine of a second embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value of the threshold value control means, a first amount of weight change control means which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the first error signal evaluation means, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, and a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means.
    A learning machine of a third embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a conjugate gradient direction determination means to obtain a direction of the conjugate gradient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a first amount of weight change control means which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the first error signal evaluation means, an amount of weight change selection means which selects either an output of an amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, and a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means.
    A learning machine of a fourth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising a conjugate gradient direction determination means to obtain a direction of a conjugate gradient from the direction of steepest descent, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts the outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, a weight change direction selection means which selects either the direction of the conjugate gradient or the direction of the steepest descent as the direction of weight change according to the level of the threshold value, an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the selected direction of weight change, a weight modification means which changes the value of the weight coefficient stored in the memory according to the output of the amount of weight change calculation means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
    A learning machine of a fifth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces the threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts the outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
    A learning machine of a sixth embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient from the direction of steepest descent, a conjugate gradient direction determination means to obtain a direction of the conjugate gradient from the direction of steepest descent, a linear search means which outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient, a parabolic line approximation means which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means, a learning progress evaluation means which evaluates progress of learning from the error output by the error signal calculation means, a threshold value control means which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means, a first error signal evaluation means which evaluates whether the output of the error signal calculation means is less than the threshold value, a skip evaluation means which counts outputs of the first error signal evaluation means and outputs a skip signal when all output signals are less than the threshold value, an amount of weight change selection means which selects either an output of the amount of weight change calculation means or an output of the parabolic line approximation means according to the level of the threshold value, a weight modification means which changes the value of the weight coefficient stored in the memory according to an amount of weight change selected by the amount of weight change selection means, and a weight change control means which causes the weight modification means to skip a weight change operation according to the skip signal.
    A learning machine of a seventh embodiment of the present invention comprises an output signal calculation means comprising plural multiple-input single-output signal processors network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means which renews values of weight coefficients of the output signal calculation means based on output signals obtained by the output signal calculation means,
  • each multiple-input single-output signal processor comprising a memory to store plural weight coefficients, an input section for inputting plural data, multiplying means to weight the input data from the input section by the weight coefficients stored in the memory, an adding means to add the plural data weighted by the multiplying means, and a threshold value processor to limit an output of the adding means to a value within a predetermined range, and
  • the weight coefficient renewing means comprising a teacher signal generator to produce as a teacher signal a desirable value (0 or 1) for the output signal of the output signal calculation means, an error signal calculation means to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means to obtain a direction of steepest descent from an output of the error signal calculation means,
  •    said learning machine characterized by
    • the weight coefficient renewing means further comprising an amount of weight change calculation means which calculates an amount of change in the weight coefficient stored in the memory from the direction of steepest descent, a second error signal evaluation means which determines whether the absolute value of an output of the error signal calculation means is greater than a threshold value, a teacher signal evaluation means which determines whether an output of the teacher signal generator is equal to "1", a second amount of weight change control means which controls an amount of change in the weight coefficient according to the outputs of the teacher signal evaluation means and the second error signal evaluation means, and a weight modification means which changes the value of the weight coefficient stored in the memory according to the output of the amount of weight change calculation means.
    BRIEF DESCRIPTION OF THE DRAWINGS
    The present invention will become more fully understood from the detailed description given hereinbelow and the accompanying drawings which are given by the way of illustration only, and thus are not limitative to the present invention, and wherein:
  • Fig. 1 is a block diagram of a learning machine according to a first embodiment of the present invention;
  • Fig. 2 is a block diagram of a learning machine according to a second embodiment of the present invention;
  • Fig. 3 is a block diagram of a learning machine according to a third embodiment of the present invention;
  • Fig. 4 is a block diagram of a learning machine according to a fourth embodiment of the present invention;
  • Fig. 5 is a block diagram of a learning machine according to a fifth embodiment of the present invention;
  • Fig. 6 is a block diagram of a learning machine according to a sixth embodiment of the present invention;
  • Fig. 7 is a block diagram of a learning machine according to a seventh embodiment of the present invention;
  • Fig. 8 is a block diagram of a conventional learning machine;
  • Fig. 9 is a block diagram of an output signal calculation means in a conventional learning machine;
  • Fig. 10 is a block diagram of a multiple-input single-output signal processor in the conventional learning machine; and
  • Fig. 11 is a graph of the input/output characteristics of a threshold value processor in the conventional learning machine.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
    Before the description of the present invention proceeds, it is to be noted that like parts are designated by like reference numerals throughout the accompanying drawings.
    (First Embodiment)
    Fig. 1 is a block diagram of a first embodiment of a learning machine according to the present invention. As shown in Fig. 1, this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 102. The weight coefficient renewing means 102 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 102 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, and a first amount of weight change controller 20.
    The operation of the learning machine according to the above first embodiment of the present invention is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates an output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The conjugate gradient direction determination means 18 calculates a conjugate gradient direction from the direction of steepest descent using the equation d = g + β * d' where β is a constant defined as β = |g|2 / |g'|2 d' is a conjugate gradient direction obtained from the previous learning cycle, and |g'| is the norm of the direction of steepest descent obtained from the previous learning cycle. It is to be noted that the direction of weight change in the first learning cycle is determined to be the direction of steepest descent.
    The threshold value controller 16 switches a threshold value T1 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T1 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13. For example, the weight change direction 1 selection means 19 may select the direction of steepest descent for the first part of the learning cycle and the direction of the conjugate gradient for the later part of the learning cycle. If the selected direction of weight change is the direction of steepest descent, the amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the following equation ΔW = -ε * g + α * Δ W' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle. However, if the selected direction of weight change is the direction of the conjugate gradient, the amount of change in the weight vector is calculated from the following equation ΔW = - ε * d
    At this time the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error |tpk - opk| is evaluated by the first error signal evaluation means 17 to be smaller than the threshold value T1 set by the threshold value controller 16. The weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, learning efficiency is improved and the time required for learning is shortened because not only is the weight not changed for multiple-input single-output signal processors of which the error is less than a threshold value T1, but also the threshold value T1 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the direction of weight change is set to the direction of steepest descent, which is locally the most efficient, during the early learning process and is set to the direction of the conjugate gradient, which is generally the most efficient, during the later learning process to change the weight, learning efficiency improves and the time required for learning is shortened.
    (Second Embodiment)
    Fig. 2 is a block diagram of a second embodiment of a learning machine according to the present invention.
    As shown in Fig. 2, this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 202. The weight coefficient renewing means 202 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 202 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
    The operation of the learning machine as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates an output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the direction of steepest descent using the following equation ΔW = -ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle.
    At this time the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error |tpk - opk| is evaluated by the first error signal evaluation means 17 to be smaller than a threshold value T1 set by the threshold value controller 16. It is to be noted that the threshold value controller 16 switches the threshold value T1 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors outputting an error greater than the threshold value T1 in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of an error curve from the values of errors to learning rates with low error values, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
    The amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T1, and outputs the result to the weight modification means 14. For example, the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
    The weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, learning efficiency is improved and the time required for learning is shortened because not only is the weight not changed for multiple-input single-output signal processors of which the error is less than a threshold value T1, but also the threshold value T1 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
    (Third Embodiment)
    Fig. 3 is a block diagram of a third embodiment of a learning machine according to the present invention.
    As shown in Fig. 3, this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 302. The weight coefficient renewing means 302 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 302 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a first amount of weight change controller 20, a linear search means 21, a parabolic line approximation means 22, and an amount of weight change selection means 23.
    The operation of the learning machine as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, an output signal calculation means 1 calculates an output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the direction of steepest descent using the following equation ΔW = -ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle.
    At this time the first amount of weight change controller 20 sets to "0" the amount of change in the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error |tpk - opk| 1 is evaluated by the first error signal evaluation means 17 to be smaller than a threshold value T1 set by the threshold value controller 16. It is to be noted that the threshold value controller 16 switches the threshold value T1 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The conjugate gradient direction determination means 18 calculates a conjugate gradient direction from the direction of steepest descent using the following equation d = g + β * d' where β is a constant defined as β = |g|2 / |g'|2 d' is a conjugate gradient direction in the previous learning cycle, and |g'| is the norm of a direction g' of steepest descent in the previous learning cycle. It is to be noted that the direction of weight change in the first learning cycle is determined to be the direction of steepest descent.
    The linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
    The amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T1, and outputs the result to the weight modification means 14. For example, the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
    The weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, learning efficiency is improved and the time required for learning is shortened because not only is the weight not changed for multiple-input single-output signal processors of which the error is less than a threshold value T1, but also the threshold value T1 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
    (Fourth Embodiment)
    Fig. 4 is a block diagram of a fourth embodiment of a learning machine according to the present invention.
    As shown in Fig. 4, this learning machine comprises an output signal calculation means 1 and a weight coefficient renewing means 402. The weight coefficient renewing means 402 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 402 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a weight change direction selection means 19, a skip evaluation means 24, and a weight change controller 25.
    The operation of the learning machine according to the fourth embodiment of the present invention as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates the output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The conjugate gradient direction determination means 18 calculates the conjugate gradient direction from the direction of steepest descent using the following equation d = g + β * d' where β is a constant defined as β = |g|2 / |g'|2 d' is a conjugate gradient direction in the previous learning cycle, and | g'| is the norm of the direction g' of steepest descent in the previous learning cycle. It is to be noted that the direction of weight change in the first learning cycle is determined to be the direction of steepest descent.
    The threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The weight change direction selection means 19 selects either the direction of steepest descent or the direction of the conjugate gradient according to the threshold value T2 from the threshold value controller 16, and outputs the result to the amount of weight change calculator 13. For example, the weight change direction selection means 19 may select the direction of steepest descent for learning during early learning process and the direction of the conjugate gradient for the later learning process. If the selected direction of weight change is the direction of steepest descent, the amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the following equation ΔW = -ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle. On the other hand, if the selected direction of weight change is the direction of the conjugate gradient, the amount of change in the weight vector is calculated from the following equation ΔW = -ε * d
    The weight modification means 14 changes the weight vector of the output signal calculation means 1 based on this calculated change in the weight vector.
    At this time the first error signal evaluation means 17 determines whether the error |tpk - opk| is less than the threshold value T2 set by the threshold value controller 16, and outputs a "0" or "1" to the skip evaluation means 24 if the error is greater or less than, respectively, the threshold value T2. The skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error |tpk - opk| of all multiple-input single-output signal processors 3 in the highest level is less than the threshold value T2. When the skip evaluation means 24 outputs the skip signal, the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the error of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but the threshold value T2 which is the evaluation standard for whether the weight is changed is also reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the direction of weight change is set to the direction of steepest descent, which is locally the most efficient, during the early learning process and is set to the direction of the conjugate gradient, which is generally the most efficient, during the later learning process to change the weight, learning efficiency improves and the time required for learning is shortened.
    (Fifth Embodiment)
    Fig. 5 is a block diagram of a fifth embodiment of a learning machine according to the present invention.
    As shown in Fig. 5, this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 502. The weight coefficient renewing means 502 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
    The operation of the learning machine as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates the output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The amount of weight change calculator 13 calculates and outputs to the amount of weight change selection means 23 the amount of change in the weight vector of the output signal calculation means 1 from the direction of steepest descent using the following equation ΔW = -ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle.
    The linear search means 21 calculates the amount of weight change to plural learning rates for the direction of steepest descent, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
    The threshold value controller 16 switches a threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent according to the threshold value T2, and outputs the result to the weight modification means 14. For example, the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent during the later learning process.
    The weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
    At this time the first error signal evaluation means 17 determines whether the error |tpk - opk| is less than the threshold value T2 set by the threshold value controller 16, and outputs a "0" or "1" to the skip evaluation means 24 if the error is greater or less than, respectively, the threshold value T2. The skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error |tpk - opk| of all multiple-input single-output signal processors 3 in the highest level is less than the threshold value T2. When the skip evaluation means 24 outputs the skip signal, the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of steepest descent is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
    (Sixth Embodiment)
    Fig. 6 is a block diagram of a sixth embodiment of a learning machine according to the present invention.
    As shown in Fig. 6, this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 602. Te weight coefficient renewing means 602 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 602 includes a learning progress evaluation means 15, a threshold value controller 16, a first error signal evaluation means 17, a conjugate gradient direction determination means 18, a linear search means 21, a parabolic line approximation means 22, an amount of weight change selection means 23, a skip evaluation means 24, and a weight change controller 25.
    The operation of the learning machine as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates the output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the direction of steepest descent using the following equation ΔW = - ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle.
    The conjugate gradient direction determination means 18 calculates the conjugate gradient direction from the direction of steepest descent using the following equation d = g + β * d' where β is a constant defined as β = |g|2 / |g'|2 d' is the conjugate gradient direction in the previous learning cycle, and |g'| is the norm of the direction g' of steepest descent in the previous learning cycle. It is to be noted that the direction of weight change in the first learning cycle is determined to be the direction of steepest descent.
    The linear search means 21 calculates the amount of weight change to plural learning rates for the direction of the conjugate gradient, and the parabolic line approximation means 22 approximates the parabolic line of the error curve from the value of the error to the learning rate with a low error value, and after selecting the amount of weight change at which the error is lowest, outputs this to the amount of weight change selection means 23.
    The threshold value controller 16 switches the threshold value T2 to a smaller value as learning progresses based on the evaluated result of learning progress output by the learning progress evaluation means 15. The learning progress evaluation means 15 evaluates the progress of the learning operation using a variety of means, including evaluating the sum of errors of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1, by referencing the number of learning cycles, by counting the number of multiple-input single-output signal processors 3 outputting an error greater than a threshold value in the highest level of the output signal calculation means 1, and by applying the maximum value in one learning cycle of the error signal calculator 11.
    The amount of weight change selection means 23 selects either the amount of weight change calculated from the direction of steepest descent or the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient according to the threshold value T2, and outputs the result to the weight modification means 14. For example, the amount of weight change selection means 23 may select the amount of weight change calculated from the direction of steepest descent during the early learning process, and select the amount of weight change determined by the parabolic line approximation means 22 from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient during the later learning process.
    The weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
    At this time the first error signal evaluation means 17 determines whether the error |tpk - opk| is less than the threshold value T2 set by the threshold value controller 16, and outputs a "0" or "1" to the skip evaluation means 24 if the error is greater or less than, respectively, the threshold value T2. The skip evaluation means 24 counts the evaluation result to the current input signal, and outputs a skip signal if the error |tpk - opk| of all multiple-input single-output signal processors 3 in the highest level is less than the threshold value T2. When the skip evaluation means 24 outputs the skip signal, the weight change controller 25 controls the weight modification means 14 so that the weight change operation is not applied to the current input signals.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, the data computation volume is greatly reduced and learning efficiency is therefore improved and the time required for learning is shortened because not only is the weight change operation of the weight modification means skipped when the sum of errors of all multiple-input single-output signal processors to the input data is less than a threshold value T2, but also the threshold value T2 which is the evaluation standard for whether the weight is changed is reduced as learning progresses, thereby enabling learning to progress in approximate degrees during the early learning process and gradually increasing precision as learning progresses. Moreover, because the amount of weight change calculated from the direction of steepest descent is selected during the early learning process, and the amount of weight change determined by the parabolic line approximation means from among the weight change amounts calculated from the plural learning rates for the direction of the conjugate gradient is selected during the later learning process, and the weight is changed, learning efficiency improves and the time required for learning is shortened.
    (Seventh Embodiment)
    Fig. 7 is a block diagram of a seventh embodiment of a learning machine according to the present invention.
    As shown in Fig. 7, this learning machine comprises a output signal calculation means 1 and a weight coefficient renewing means 702. The weight coefficient renewing means 702 includes a teacher signal generator 10, an error signal calculator 11, a steepest descent direction determination means 12, an amount of weight change calculator 13, and a weight modification means 14. Each of the output signal calculation means 1, the teacher signal generator 10, the error signal calculator 11, the steepest descent direction determination means 12, the amount of weight change calculator 13, and the weight modification means 14 is the same as the corresponding components of the conventional learning machine described hereinabove. In addition, the weight coefficient renewing means 702 includes a second error signal evaluation means 26, a teacher signal evaluation means 27, and a second amount of weight change controller 28.
    The operation of the learning machine as described hereinabove is described hereinbelow.
    As in the conventional learning machine described above, the output signal calculation means 1 calculates the output signal, and the error signal calculator 11 calculates an error E by applying the equation (1) to this output signal and the output from the teacher signal generator 10. Based on this calculated error E, the steepest descent direction determination means 12 computes g = ∂E/∂W to obtain the direction of steepest descent, which is the direction of change of the weight vector of the output signal calculation means 1.
    The amount of weight change calculator 13 calculates the amount of change in the weight vector of the output signal calculation means 1 from the direction of steepest descent using the following equation Δ W = - ε * g + α * ΔW' where ε is a positive constant called the learning rate, α is a positive constant called the acceleration parameter, and ΔW' is the amount of change in the weight vector in the previous learning cycle.
    At this time, the second amount of weight change controller 28 multiplies by m1 (m1 > 1) times the amount of change for the weight coefficient of the multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error |tpk - opk| has been determined to be greater than a threshold value T3 by the second error signal evaluation means 26 and for which the teacher signal has been determined to be "1" by the teacher signal evaluation means 27.
    The weight modification means 14 thus changes the weight vector of the output signal calculation means 1 based on the calculated amount of change for the weight vector.
    The amount of error is thus reduced by repeatedly renewing the weight, and when the error is sufficiently small, it is concluded that the output signal is sufficiently close to the desired value and learning stops.
    Thus, the multiple-input single-output signal processors 3 with a large error value can be more quickly converged and the time required for learning can be shortened because, of the multiple-input single-output signal processors 3 with a large error value, and the weight coefficient of these which should output a particularly slow convergence value "1" (multiple-input single-output signal processors 3 with a teacher signal value of "1") is actively changed.
    It is to be noted that the second amount of weight change controller 28 in this embodiment may increase the amount of weight change m2 (m2 > 1) times for multiple-input single-output signal processors 3 in the highest level of the output signal calculation means 1 for which the error |tpk - opk| has been determined to be greater than the threshold value T3 by the second error signal evaluation means 26, and may further increase the amount of weight change m3 (m3 > 1) times for those multiple-input single-output signal processors 3 for which the teacher signal has been determined to be "1" by the teacher signal evaluation means 27. In this case, the multiple-input single-output signal processors 3 with a large error value can be more quickly converged and the time required for learning can be reduced because not only is the weight coefficient of the multiple-input single-output signal processors 3 outputting a high error value actively changed, but also the weight coefficient is changed even more actively for those multiple-input single-output signal processors 3 which should output a slow convergence "1" (teacher signal = 1).
    The invention being thus described, it will be obvious that the same may be varied in many ways. All such modifications as would be obvious to one skilled in the art are intended to be included within the scope of the following claims.

    Claims (7)

    1. A learning machine having an output signal calculation means (1) comprising plural multiple-input single-output signal processors (3) network connected in a hierarchical structure without mutual connections within levels in such a manner that signals propagate only to a higher level, and a weight coefficient renewing means (2) which renews values of weight coefficients of the output signal calculation means (1) based on output signals obtained by the output signal calculation means (1),
      each multiple-input single-output signal processor (3) comprising a memory (6) to store plural weight coefficients, an input section (5) for inputting plural data, multiplying means (7) to weight the input data from the input section (5) by the weight coefficients stored in the memory (6), an adding means (8) to add the plural data weighted by the multiplying means (7), and a threshold value processor (9) to limit an output of the adding means (8) to a value within a predetermined range, and
      the weight coefficient renewing means (2) comprising a teacher signal generator (10) to produce as a teacher signal a desirable value for the output signal of the output signal calculation means (1), an error signal calculation means (11) to obtain an error between the output signal and the teacher signal, a steepest descent direction determination means (12) to obtain a direction of steepest descent from an output of the error signal calculation means (11), an amount of weight change calculation means (13) for calculating an amount of change in the weight coefficient from the direction of weight change, and a weight modification means (14) which changes the value of the weight coefficient stored in the memory (6) according to an output of the amount of weight change calculation means (13),
      characterized in that
      the weight coefficient renewing means (102; 402) further comprises a conjugate gradient direction determination means (18) to obtain a direction of a conjugate gradient from the direction of steepest descent, a learning progress evaluation means (15) to evaluate progress of learning from the error output from the error signal calculation means (11), a threshold value control means (16) which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means (15), an error signal evaluation means (17) which evaluates whether the output of the error signal calculation means (11) is less than the threshold value of the threshold value control means (16), a weight change direction selection means (19) which selects either the direction of the conjugate gradient or the direction of the steepest descent as a direction of weight change according to the level of the threshold value, wherein the amount of weight change calculation means (13) calculates an amount of change in the weight coefficient from the selected direction of weight change, and an amount of weight change control means (20) which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the error signal evaluation means (17).
    2. A learning machine according to the preamble of claim 1,
      characterized in that
      the weight coefficient renewing means (202, 502) further comprises a linear search means (21 ) which outputs amounts of weight change to plural learning rates for the direction of steepest descent, a parabolic line approximation means (22) which determines an amount of weight change by parabolic line approximation method from the plural weight change amounts output from the linear search means (21), a learning progress evaluation means (15) evaluating progress of learning from the error output by the error signal calculation means (11), a threshold value control means (16) which gradually reduces a threshold value as learning progresses according to an evaluation result of the learning progress evaluation means (15), an error signal evaluation means (17) which evaluates whether the output of the error signal calculation means (11) is less than the threshold value of the threshold value control means (16), an amount of weight change control means (20) which sets the amount of weight coefficient change to 0 when the error is determined to be less than the threshold value by the error signal evaluation means (17), and an amount of weight change selection means (23) which selects either an amount of change in the weight coefficient from the direction of steepest descent calculated by the amount of weight change calculation means (13) or an output of the parabolic line approximation means (22) according to the level of the threshold value, wherein the weight modification means (14) changes the value of the weight coefficient stored in the memory (6) according to an amount of weight change selected by the amount of weight change selection means (23).
    3. The learning machine according to claim 2,
      characterized in that
      the weight coefficient renewing means (302; 602) further comprises a conjugate gradient direction determination means (18) to obtain a direction of the conjugate gradient from the direction of steepest descent, and that the linear search means (21) outputs amounts of weight change to plural learning rates for the direction of the conjugate gradient.
    4. The learning machine according to anyone of the preceding claims,
      further characterized in that
      the weight coefficient renewing means further comprises a skip evaluation means (24) which counts the outputs of the error signal evaluation means (17) and outputs a skip signal when all output signals are less than the threshold value, and a weight change control means (25) which causes the weight modification means (14) to skip a weight change operation according to the skip signal.
    5. A learning machine according to the preamble of claim 1,
      characterized in that
      the weight coefficient renewing means (702) further comprises an error signal evaluation means (26) which determines whether the absolute value of an output of the error signal calculation means (11) is greater than a threshold value, a teacher signal evaluation means (27) which determines whether an output of the teacher signal generator (10) is equal to "1", and an amount of weight change control means (28) which controls an amount of change in the weight coefficient according to the outputs of the teacher signal evaluation means (27) and the error signal evaluation means (26).
    6. The learning machine according to claim 5 wherein the amount of weight change control means (28) multiplies by m1 (m1 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) in the highest level of the output signal calculation means (1) for which the value of the error is evaluated by the error signal evaluation means to be greater than the threshold value and the value of the teacher signal is evaluated to be "1" by the teacher signal evaluation means (27).
    7. The learning machine according to claim 5 wherein the amount of weight change control means (28) multiplies by m2 (m2 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) in the highest level of the output signal calculation means (1) for which the value of the error is evaluated by the error signal evaluation means (26) to be greater than the threshold value, and multiplies by m3 (m3 > 1) times the amount of weight change for the multiple-input single-output signal processors (3) for which the value of the teacher signal is further evaluated to be "1" by the teacher signal evaluation means (27).
    EP91114955A 1990-09-04 1991-09-04 Learning machine with a hierarchical structure without mutual connections within levels thereof Expired - Lifetime EP0474222B1 (en)

    Applications Claiming Priority (2)

    Application Number Priority Date Filing Date Title
    JP234287/90 1990-09-04
    JP2234287A JPH04114263A (en) 1990-09-04 1990-09-04 Learning machine

    Publications (3)

    Publication Number Publication Date
    EP0474222A2 EP0474222A2 (en) 1992-03-11
    EP0474222A3 EP0474222A3 (en) 1994-05-18
    EP0474222B1 true EP0474222B1 (en) 1998-10-21

    Family

    ID=16968621

    Family Applications (1)

    Application Number Title Priority Date Filing Date
    EP91114955A Expired - Lifetime EP0474222B1 (en) 1990-09-04 1991-09-04 Learning machine with a hierarchical structure without mutual connections within levels thereof

    Country Status (4)

    Country Link
    US (1) US5295228A (en)
    EP (1) EP0474222B1 (en)
    JP (1) JPH04114263A (en)
    DE (1) DE69130387T2 (en)

    Families Citing this family (12)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    JPH0635888A (en) * 1992-07-15 1994-02-10 Mitsubishi Electric Corp Learning method for neural network and electric system voltage/reactive power controller using the learning method
    US5726688A (en) * 1995-09-29 1998-03-10 Ncr Corporation Predictive, adaptive computer interface
    US6652283B1 (en) 1999-12-30 2003-11-25 Cerego, Llc System apparatus and method for maximizing effectiveness and efficiency of learning retaining and retrieving knowledge and skills
    US7319943B2 (en) * 2002-12-16 2008-01-15 Canon Kabushiki Kaisha Optimum design method, and apparatus, and program for the same
    US8873813B2 (en) 2012-09-17 2014-10-28 Z Advanced Computing, Inc. Application of Z-webs and Z-factors to analytics, search engine, learning, recognition, natural language, and other utilities
    US11195057B2 (en) 2014-03-18 2021-12-07 Z Advanced Computing, Inc. System and method for extremely efficient image and pattern recognition and artificial intelligence platform
    US9916538B2 (en) 2012-09-15 2018-03-13 Z Advanced Computing, Inc. Method and system for feature detection
    US11074495B2 (en) 2013-02-28 2021-07-27 Z Advanced Computing, Inc. (Zac) System and method for extremely efficient image and pattern recognition and artificial intelligence platform
    US11914674B2 (en) 2011-09-24 2024-02-27 Z Advanced Computing, Inc. System and method for extremely efficient image and pattern recognition and artificial intelligence platform
    CN102866179B (en) * 2012-09-13 2014-06-18 重庆大学 Online recognition and inhibition method based on non-target interference smell in electronic nose of artificial intelligent learning machine
    CN104200268B (en) * 2014-09-03 2017-02-15 辽宁大学 PSO (Particle Swarm Optimization) extremity learning machine based strip steel exit thickness predicting method
    CN115171453B (en) * 2022-06-13 2023-05-30 广东电网有限责任公司 Artificial customer service training method and system based on smart grid

    Family Cites Families (4)

    * Cited by examiner, † Cited by third party
    Publication number Priority date Publication date Assignee Title
    JPS62104324A (en) * 1985-10-31 1987-05-14 Toshiba Corp Adaptive automatic equalizer
    US4912652A (en) * 1988-12-14 1990-03-27 Gte Laboratories Incorporated Fast neural network training
    US5129038A (en) * 1989-02-23 1992-07-07 Matsushita Electric Industrial Co., Ltd. Neural network with selective error reduction to increase learning speed
    EP0579341A2 (en) * 1989-02-23 1994-01-19 Matsushita Electric Industrial Co., Ltd. Learning machine with multi-input single output circuits connected in hierarchical structure

    Also Published As

    Publication number Publication date
    DE69130387D1 (en) 1998-11-26
    EP0474222A3 (en) 1994-05-18
    US5295228A (en) 1994-03-15
    JPH04114263A (en) 1992-04-15
    DE69130387T2 (en) 1999-05-06
    EP0474222A2 (en) 1992-03-11

    Similar Documents

    Publication Publication Date Title
    US5479571A (en) Neural node network and model, and method of teaching same
    EP0474222B1 (en) Learning machine with a hierarchical structure without mutual connections within levels thereof
    Gilbert et al. Fast reference governors for systems with state and control constraints and disturbance inputs
    Haykin et al. Optimum nonlinear filtering
    US5129038A (en) Neural network with selective error reduction to increase learning speed
    Lee et al. Channel equalization using radial basis function network
    US4674028A (en) Identification method of a process parameter
    US5535149A (en) Duplex adaptive digital filter and method of operation
    US6223194B1 (en) Adaptive filter, step size control method thereof, and record medium therefor
    US5168550A (en) Neutral network with plural weight calculation methods and variation of plural learning parameters
    US5384896A (en) Learning machine
    US5319587A (en) Computing element for neural networks
    US5701395A (en) Method of programming a polynomial processor
    US5898731A (en) Auto-coefficient renewal digital channel equalizer
    KR102574719B1 (en) Electronice device and learning method for low complexity artificial intelligentce model learning based on selecting the dynamic prediction confidence thresholed
    Su et al. Processing element architecture design for deep reinforcement learning with flexible block floating point exploiting signal statistics
    US5542005A (en) Recognition unit and recognition apparatus
    Ludermir Logical networks capable of computing weighted regular languages
    JP3328953B2 (en) Nonlinear synaptic neuron, device thereof and prediction method using the same
    Benureau et al. Learning Timescales in MTRNNs
    KR20220169085A (en) Device for performing reinforcement learning and operation method thereof
    Ammoumou et al. A New FTF9N Stabilized Recursive Algorithm, Implementation on Finite-Precision
    JPH1091604A (en) Function learning device
    CN114492748A (en) Echo state network forward propagation acceleration algorithm based on activation function approximation
    CN118211628A (en) Model training method, device, electronic equipment and storage medium

    Legal Events

    Date Code Title Description
    PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

    Free format text: ORIGINAL CODE: 0009012

    17P Request for examination filed

    Effective date: 19910905

    AK Designated contracting states

    Kind code of ref document: A2

    Designated state(s): DE FR GB

    PUAL Search report despatched

    Free format text: ORIGINAL CODE: 0009013

    AK Designated contracting states

    Kind code of ref document: A3

    Designated state(s): DE FR GB

    17Q First examination report despatched

    Effective date: 19961031

    GRAG Despatch of communication of intention to grant

    Free format text: ORIGINAL CODE: EPIDOS AGRA

    GRAG Despatch of communication of intention to grant

    Free format text: ORIGINAL CODE: EPIDOS AGRA

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAH Despatch of communication of intention to grant a patent

    Free format text: ORIGINAL CODE: EPIDOS IGRA

    GRAA (expected) grant

    Free format text: ORIGINAL CODE: 0009210

    AK Designated contracting states

    Kind code of ref document: B1

    Designated state(s): DE FR GB

    REF Corresponds to:

    Ref document number: 69130387

    Country of ref document: DE

    Date of ref document: 19981126

    ET Fr: translation filed
    PLBE No opposition filed within time limit

    Free format text: ORIGINAL CODE: 0009261

    STAA Information on the status of an ep patent application or granted ep patent

    Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: GB

    Payment date: 19990901

    Year of fee payment: 9

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: DE

    Payment date: 19990906

    Year of fee payment: 9

    PGFP Annual fee paid to national office [announced via postgrant information from national office to epo]

    Ref country code: FR

    Payment date: 19990909

    Year of fee payment: 9

    26N No opposition filed
    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: GB

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20000904

    GBPC Gb: european patent ceased through non-payment of renewal fee

    Effective date: 20000904

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: FR

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20010531

    PG25 Lapsed in a contracting state [announced via postgrant information from national office to epo]

    Ref country code: DE

    Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES

    Effective date: 20010601

    REG Reference to a national code

    Ref country code: FR

    Ref legal event code: ST