US20180197080A1 - Learning apparatus and method for bidirectional learning of predictive model based on data sequence - Google Patents

Learning apparatus and method for bidirectional learning of predictive model based on data sequence Download PDF

Info

Publication number
US20180197080A1
US20180197080A1 US15/403,986 US201715403986A US2018197080A1 US 20180197080 A1 US20180197080 A1 US 20180197080A1 US 201715403986 A US201715403986 A US 201715403986A US 2018197080 A1 US2018197080 A1 US 2018197080A1
Authority
US
United States
Prior art keywords
model
learning
input
time point
input data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/403,986
Inventor
Hiroshi Kajino
Takayuki Osogami
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
International Business Machines Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US15/403,986 priority Critical patent/US20180197080A1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KAJINO, HIROSHI, OSOGAMI, TAKAYUKI
Priority to US15/804,702 priority patent/US20180197082A1/en
Priority to PCT/IB2017/057999 priority patent/WO2018130890A1/en
Priority to CN201780083071.4A priority patent/CN110168577A/en
Priority to JP2019535257A priority patent/JP7123938B2/en
Publication of US20180197080A1 publication Critical patent/US20180197080A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/047Probabilistic or stochastic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/02Knowledge representation; Symbolic representation
    • G06N5/022Knowledge engineering; Knowledge acquisition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Definitions

  • the present invention generally relates to predictive models, and more particularly to a learning apparatus and method for bidirectional learning of a predictive model based on a data sequence.
  • a neural network Boltzmann machine, and the like are known that can be applied to time-series data.
  • a dynamic Boltzmann machine is known that can learn a model corresponding to data input in time series through machine learning.
  • Such a Boltzmann machine or the like is desired to have improved expressive power, learning power, and the like by accurately learning a predictive model based on a data sequence, in order to be applied to various, fields such as moving images, languages, signal waveforms, and music.
  • a computer-implemented method for learning a first model.
  • the method includes generating, by a processor, a second model based on the first model.
  • the first model is configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and that are from a first input data sequence.
  • the second model is configured to learn a first learning, target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence.
  • the method further includes performing, by the processor, a learning process using both the first model and the second model.
  • the method also includes storing, in a memory device, the first model that has been learned.
  • a computer program product for predictive model learning.
  • the computer program product includes a non-transitory computer readable storage medium having program instructions embodied therewith.
  • the program instructions are executable by a computer to cause the computer to perform a method.
  • the method includes generating, by a processor, a second model based on a first model.
  • the first model is configured to perform a learning process based on sequentially inputting, each of a plurality of pieces of input data that include a plurality of input values and are from a first input data sequence.
  • the second model is configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence.
  • the method farther includes performing, by the processor, a learning process using both the first model and the second model.
  • the method also includes storing, in a memory device, the first model that has been learned.
  • a learning apparatus for learning a model corresponding to input data.
  • the learning apparatus includes a processor.
  • the learning apparatus further includes one or more computer readable mediums collectively including instructions that, when executed by the processor, cause the processor to generate a second model based on a first model.
  • the first model is configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and are from a first input data sequence.
  • the second model is configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence.
  • the one or more computer readable mediums collectively including instructions that, when executed by the processor, further cause the processor to perform a learning process using both the first model and the second model.
  • the one or more computer readable mediums collectively including instructions that, when executed by the processor, further cause the processor to store the first model that has been learned.
  • FIG. 1 shows an exemplary configuration of a learning apparatus 100 , according to an embodiment of the present invention
  • FIG. 2 shows an exemplary configuration of a model 10, according to an embodiment of the present invention
  • FIG. 3 shows a flow of an operation of the learning apparatus 100 , according to an embodiment of the present invention
  • FIG. 4 shows an example of structures of time-series data and training data for use in learning, according to an embodiment of the present invention
  • FIG. 5 shows a first modification of the learning apparatus 100 according to an embodiment of the present invention
  • FIG. 6 shows a modification of the model 10, according to an embodiment of the present invention.
  • FIG. 7 shows an example of a temporal change in the update parameter ⁇ i,l , according to an embodiment of the present invention
  • FIG. 8 shows an example of a temporal change in the update parameter ⁇ i,j,k , according; to an embodiment of the present invention
  • FIG. 9 shows an exemplary schematic configuration of the first model, according to an embodiment of the present invention.
  • FIG. 10 shows an exemplary schematic configuration of the second model, according to an embodiment of the present invention.
  • FIG. 11 shows an operational flow of the learning process using the first model and the second model performed by the learning apparatus 100 , according to an embodiment of the present invention.
  • FIG. 12 shows an exemplary hardware configuration of a computer, according to an embodiment of the present invention.
  • FIG. 1 shows an exemplary configuration of a learning apparatus 100 according to an embodiment.
  • the learning apparatus 100 may be an apparatus for learning a model 10 corresponding to time-series input data.
  • the learning apparatus 100 may be operable to learn a model based on a Boltzmann machine by supplying time-series data to nodes of the model.
  • the learning apparatus 100 includes an acquiring section 110 , a supplying section 120 , a storage section 130 , a computing section 140 , and a learning processing section 150 .
  • the acquiring section 110 may be operable to acquire time-series input data.
  • Time-series input data may be, for example, a data sequence in which a plurality of pieces of data are arranged along a time axis, such as moving image data.
  • the acquiring section 110 may be connected to a device operated by a user or a device (sensor) that detects and outputs time-series data and may acquire the time-series input data from such a device.
  • the acquiring section 110 may read and acquire time-series input data stored in a storage device in a predetermined format.
  • the acquiring section 110 may he connected to a network and acquire time-series input data via the network.
  • the acquiring section 110 may also store the acquired time-series input data in a storage device included in the learning apparatus 100 .
  • the supplying section 120 may be operable to supply a plurality of input values corresponding to input data at one time point in the time-series input data to a plurality of nodes of a model.
  • the supplying section 120 is connected to the acquiring section 110 and may handle, as training data input data at one time point in the received time-series input data and supply input values at the one time point to corresponding nodes of the model, input data at one time point may be the temporally newest data in a training data set for use in learning. Alternatively, input data at one time point may be temporally intermediate data in a training data set for use in learning. That is, input data at one time point may be selected arbitrarily from the time-series data.
  • the storage section 130 may be operable to store values of hidden nodes of the model in correspondence with a plurality of time points in the time series input data.
  • the storage section 130 may sample the values of hidden nodes corresponding to one time: point and store these values in the hidden nodes corresponding to this time point.
  • the storage section 130 may store the sampled values respectively in the hidden nodes corresponding to the time points for each time point.
  • the computing section 140 may be operational to compute a conditional probability of each input value at one time point on a condition that an input data sequence has occurred.
  • the pieces of data input to the respective nodes of the model at one time point are referred to as input values at one time point
  • the pieces of data input to the respective nodes at the time points before the one time point are referred to as the input data sequence.
  • a model used by the learning apparatus 100 may have a weight parameter between (i) a plurality of hidden nodes and a plurality of input values corresponding to input data at each time point prior to the one time point in an input data sequence and (ii) a plurality of hidden nodes corresponding to the one time point and a plurality of input nodes.
  • the computing section 140 may be operable to compute a conditional probability of each input value at one time point, on the basis of an input data sequence before the one time point in the time-series input data, the stored values of hidden nodes, and the weight parameter of the model. Furthermore, the computing section 140 may be operable to compute a conditional probability of the value of each hidden node at one time point on a condition that an input data sequences has occurred, based on an, input data sequence before the one time point in the time-series input data and the weight parameter of the model.
  • the learning processing section 150 may be operable to increase a conditional probability of input data at one time point occurring on a condition that the input data sequence has occurred, by adjusting the weight parameter of the model.
  • the learning processing section 150 may further adjust bias parameters which are given respectively to the plurality of nodes and hidden nodes of the model.
  • the learning processing section 150 may supply the adjusted weight parameter and bias parameters of the model to a storage device, such as an external database 1000 , to store these parameters in the storage device.
  • the above-described learning apparatus 100 may be operable to learn the model by adjusting the weight parameter and bias parameters of the model, based on input data at one time point in the time-series input data.
  • the model according to the present embodiment is described with reference to FIG. 2 .
  • FIG. 2 shows an exemplary configuration of a model 10 according to the present embodiment.
  • the model 10 includes a plurality of common layers 12 .
  • FIG, 2 shows an example including a total of T common layers 12 .
  • the model 10 may include a finite number of common layers 12 .
  • Each common layer 12 includes an input layer 14 and a hidden layer 16 .
  • Each input layer 14 may be a layer corresponding to the time-series data. Each input layer 14 may correspond to a respective time point in the time-series data. Each input layer 14 may include a predetermined number of nodes. For example, the 0-th input layer may be a layer corresponding to input data at one time point in the time-series data. The 0-th input layer may include a plurality of nodes corresponding to the number of input values in this input data.
  • a total of T ⁇ 1 input layers 14 other than the 0-th input layer 14 among the plurality of input layers may he input layers 14 corresponding to the input data sequence before the one time point in the time-series, input data.
  • the ⁇ 1st input layer may correspond to input data at a time point that temporally precedes the one time point by one time point
  • the ( ⁇ )-th input: layer may correspond to input data at a time point that temporally precedes the one time point by ⁇ time points. That is, a total of T ⁇ 1 input layers other than the 0-th input layer each have the same number of nodes as the 0-th input layer and are respectively supplied with input values of corresponding input data values in the input data sequence, for example.
  • Each hidden layer 16 may correspond to a respective time point in the time-series data.
  • the 0-th hidden layer may be a layer corresponding to one time point in the time-series data.
  • FIG. 2 shows an example including, a total of T hidden layers 16 .
  • Each hidden layer 16 may include one or more hidden nodes, and the storage section 130 may store the values sampled at the one time point.
  • a total of T ⁇ 1 hidden layers other than the 0-th hidden layer among the plurality of hidden layers 16 may be hidden layers 16 corresponding to time points before the one time point in the time-series data.
  • the ⁇ 1st hidden layer corresponds to a time point that temporally precedes the input data of the one time point by one time point
  • the storage section 130 stores the values sampled at the time point that temporally precedes the one time point by one time point.
  • the ( ⁇ )-th hidden layer may correspond to a time point that temporally precedes the input data of the one time point by ⁇ time points
  • the storage section 130 may store the values sampled at the time point that temporally precedes the one time point by ⁇ time points. That is, a total of T ⁇ 1 hidden layers other than the 0-th hidden layer each have the same number of nodes as the 0-th hidden layer and are respectively supplied with values of corresponding hidden nodes, for example.
  • the time-series input data is moving image data
  • the last image data of the moving image data corresponds to the 0-th input layer
  • a plurality of nodes of the 0-th input layer each receive corresponding pixel data of the image data.
  • the 0-th hidden layer corresponds to the final time point of the moving image data
  • the storage section 130 may store values sampled at this final time point in the hidden nodes of the 0-th hidden layer.
  • the ⁇ 1st input layer is supplied with image data that immediately precedes the last image data, and ⁇ a plurality of nodes of the ⁇ 1st input layer each receive corresponding pixel data of the immediately preceding image data.
  • the ⁇ 1st hidden layer corresponds to the time point that immediately precedes the final time point, and for each of the plurality of nodes of the ⁇ 1st hidden layer, the storage section 130 may store the values sampled at this immediately preceding time point.
  • the plurality of nodes of the ( ⁇ )-th input layer each receive corresponding pixel data of image data that precedes the last image data by ⁇ images
  • the plurality of nodes of the ( ⁇ )-th hidden layer each store corresponding sampling values at the time point that precedes the last time point by ⁇ time points.
  • FIG. 2 shows an example in which each common layer 12 includes an input layer 14 and a hidden layer 16 , but instead, one or more common layers 12 need not include a hidden layer 16 .
  • the 0-th common layer to the ( ⁇ m)-th common layer include input layers 14 and hidden layers 16
  • the ( ⁇ m ⁇ 1)-th common layer to ( ⁇ T+1)-th common layer may include input layers 14 .
  • the plurality of nodes in the 0-th input layer and/or the plurality of hidden nodes in the 0-th hidden layer may each have a bias parameter.
  • the j-th node j in the common layer 12 has a bias parameter b j .
  • the plurality of nodes in the 0-th input layer and the nodes of the hidden layer corresponding to the input data sequence and layers corresponding to the input data sequence before the one time point may respectively have weight parameters there between. There need not be weight parameters between the plurality of nodes in each input layer 14 and hidden layer 16 .
  • the plurality of nodes in the 0-th hidden layer and the nodes of the hidden layer corresponding to the input data sequence and layers corresponding to the input data sequence before the one time point may respectively have weight parameters there between. That is, the plurality of nodes of the 0-th common layer and the nodes of the plurality of common layers before the one time point may respectively have weight parameters there between.
  • FIG. 2 shows a concept of a weight parameter W ij [ ⁇ ] between the node j of the 0-th input layer and a node i of the ( ⁇ )-th layer.
  • FIG. 2 shows an example in which the model 10 has the same number of input layers 14 and layers 16 , each input layer 14 includes I nodes, and each hidden layer 16 includes H hidden nodes.
  • the input layers 14 and hidden layers 16 are expressed by one common layer 12 that has a plurality of nodes x j [t] .
  • the first to I-th nodes (1 ⁇ j ⁇ 1) of the common layer 12 indicate the nodes of the input layer 14
  • the (I+1)-th to (I+H)-th nodes (I+1, j, I+H) indicate hidden nodes.
  • the weight parameter W ij [ ⁇ ] shown in FIG. 2 indicates a weight parameter between two nodes of an input layer 14 .
  • the weight parameter W ij [ ⁇ ] may include weight parameters from an input node to a hidden node, from a hidden node to an input node, and between to hidden nodes.
  • a specific example of a weight parameter W ij [ ⁇ ] is as shown in the Expression below.
  • a weight parameter from an input node or hidden node to an input node is referred to as a “weight parameter to an input node,” and a weight parameter from an input node or hidden node to a hidden node is referred to as a ‘weight parameter to a hidden node.”
  • u i, j, k and v i, j, l are learning parameters that are learning targets, for example.
  • the weight parameter W ij [ ⁇ ] may be a parameter based on a positive value, which is based on a product of the first learning parameter u i, j, k and the first predefined parameter ⁇ k t1 , and a negative value, which is based on a product of the second learning parameter v i, j, l and a second predefined parameter ⁇ l t2 .
  • the weight parameter W ij [ ⁇ ] may be a positive value based on a product of the first learning parameter u i, j, k and the first, predefined parameter ⁇ k t1 .
  • the weight parameter W ij [ ⁇ ] may be a negative value based cm a product of the second learning parameter v i, j, l and the second predefined parameter ⁇ l t2 .
  • the weight parameter W ij [ ⁇ ] may be equal to 0.
  • the weight parameter W ij [ ⁇ ] may be, based on a plurality of positive values that are based on the products u i, j, k ⁇ k t1 of a plurality of sets of the first learning parameter u i, j, k and the first predefined parameter ⁇ k t1 respectively from among the plurality of first learning parameters i, j, k and the plurality of first predefined parameters ⁇ k t1 .
  • the weight parameter W ij [ ⁇ ] may be based on a plurality of negative values that are based on products v i, j, l ⁇ 1 t2 of a plurality of sets of the second learning parameter v i, j, l and the second predefined parameter ⁇ l t2 respectively from among the plurality of second learning parameters v i, j, l and the plurality of second predefined parameters ⁇ l t2 .
  • a predefined parameter may be a parameter based on a value obtained by raising a predetermined constant to the power of a value based on the time point difference ⁇ .
  • the first predefined parameter ⁇ k t1 is a parameter whose value gradually decreases as the time point difference ⁇ increases, for example.
  • the second predefined parameter ⁇ l t2 may be a parameter whose value gradually decreases as the time point difference ⁇ increases, for example.
  • the above-described model 10 may be operable to form a Boltzmann machine. That is, the model 10 may be a Boltzmann machine to which time-series data is applied.
  • the model 10 may be a Boltzmann machine that includes hidden layers into which are, input values differing from the time-series data, in addition to the input layers into which the time-series data is input.
  • the learning apparatus 100 learns the model 10 by adjusting the learning parameters u i, j, k and v i, j, l and the bias parameter b j while sampling and storing the values of the hidden nodes, by using, as training data, input data at one time point that is supplied to the 0-th input layer of the model 10.
  • a learning operation of the learning apparatus 100 is described with reference to FIG. 3 .
  • FIG. 3 shows a flow of an operation of the learning apparatus 100 according to the present embodiment.
  • the learning apparatus 100 may be operable to learn the model 10 corresponding to time-series input data and determine the learning parameters u i, j, k and v i, j, l and the bias parameter b j , by executing the processing steps of S 310 to S 360 .
  • the determination of the weight parameters to the hidden nodes and the weight parameters to the input nodes is perforated by the learning apparatus 100 using substantially the same operation.
  • the acquiring section 110 may acquire time-series data (S 310 ).
  • the acquiring section 110 may acquire time-series data of a duration equivalent to a total of T layers from the 0-th layer to the ( ⁇ T+1)-th layer of the model 10.
  • the acquiring section 110 acquires, for example, T pieces of image data in time-series that form the moving image data.
  • the supplying section 120 may supply a plurality of input values corresponding to the input data of the time-series input data at one time point to the plurality of input nodes of the 0-th input layer of the model 10 (S 320 ).
  • the supplying section 120 supplies, for example, I input values x j [0] corresponding to input data x [1,I] [0] of the time-series input data at the most recent time point to the corresponding nodes j of the 0-th input layer (1 ⁇ j ⁇ 1).
  • the supplying section 120 supplies I pieces of pixel data included in the last piece of image data of T pieces of image data arranged in time series to form the moving image data to I nodes of the 0-th input layer.
  • the supplying section 120 may supply a value of 1 or 0 as the pixel data to each node of the 0-th input layer.
  • the supplying section 120 may supply the data to a number of layers from the 0-th input layer corresponding to the length of the time series, and may supply a value of 0, for example, to the nodes of the rest of the layers.
  • the supplying section 120 may supply a plurality of input values corresponding to the input data sequence before the one time point to the plurality of nodes included in respective layers from the ⁇ 1st input layer to the ( ⁇ T+1)-th input layer of the model 10.
  • x j ( ⁇ T, ⁇ 1] denote input data, supplied to layers from the ⁇ 1st input layer to the ( ⁇ T+1)-th input layer (1 ⁇ j ⁇ 1).
  • the term ( ⁇ T, ⁇ 1] indicates layers from the ( ⁇ T+1)-th layer to the ⁇ 1st layer. That is, the input data x j ( ⁇ T, ⁇ 1] in the tune-series data denotes a history up to the input data x j [0] , for example.
  • the storage section 130 samples the values of a plurality of hidden nodes corresponding to the one time point, and respectively stores these values in the corresponding plurality of hidden nodes of the 0-th hidden layer (S 330 ).
  • the storage section 130 may arbitrarily or randomly input values of 1 or 0.
  • the storage section 130 stores sampled values in the corresponding hidden nodes j of the 0-th hidden layer, for example (I+1 ⁇ j ⁇ I+H).
  • the storage section 130 may store the values of the hidden nodes before the one time point respectively in a plurality of nodes in each of the corresponding hidden layers from the ⁇ 1st hidden layer to the ( ⁇ T+1)-th hidden layer of the model 10.
  • x j ( ⁇ T, ⁇ 1] denote the values of the hidden nodes stored in the layers from the ⁇ 1st hidden layer to the ( ⁇ T+1)-th hidden layer (I+1 ⁇ j ⁇ I+H). That is, the values x j ( ⁇ T, ⁇ 1] input to the nodes of each common layer 12 before the one time point denote a history up to the input values x j [0] input to the nodes of the 0-th common layer, for example (1 ⁇ j ⁇ I+H).
  • the computing section 140 may compute conditional probabilities of each input value x j [0] (1 ⁇ j ⁇ I) of an input node at the one time point, based on the input values x j ( ⁇ T, ⁇ 1] (1 ⁇ j ⁇ I+H) of the plurality of nodes of the ( ⁇ T+1) common layer to the ⁇ 1st common layer and the weight parameter W ij [ ⁇ ] (S 340 ).
  • the computing section 140 computes a probability ⁇ x j [0] > ⁇ of the input value x j [0] (1 ⁇ j ⁇ I) of the j-th node of the 0-th input layer being equal to 1 by substituting 1 for x j [0] in the following expression, based on the history x ( ⁇ T, ⁇ 1] (1 ⁇ j ⁇ I+H) of the plurality of nodes of the common layer 12 .
  • the computing section 140 may compute the conditional probabilities of the values x j [0] (I+1 ⁇ j ⁇ I+H) of the 0-th hidden, layer.
  • Expression 2 is derived as a Boltzmann machine from a known probability formula.
  • may be a parameter that is dependent on a known “system temperature” of the Boltzmann machine, and may be preset by a user or the like.
  • x ⁇ ( ⁇ T, ⁇ 1] ) of Expression 2 is computed by using the following expression.
  • T denotes a transpose
  • :.j indicates extraction of the j-th column
  • ⁇ i, j, k , ⁇ i, j, l , and ⁇ i,1 are denoted by the following expressions.
  • x j ( ⁇ T, ⁇ 1] ) obtained by substituting 1 for x j [0] of Expression 2 can be computed from Expression 3 by substituting 1 for x j [0] in expression 5.
  • predetermined initial values for example, 0
  • the computing section 140 can compute a conditional probability ⁇ x j [0] > ⁇ of each input value x j [0] at the one time point which is denoted by Expression 2.
  • the learning processing section 150 may determine a direction of the change in the bias parameter b, by using the following expression.
  • x j [0] on the right side of Expression 8 denotes an input value supplied as training data by the supplying section 120
  • ⁇ x j [0] > ⁇ on the right side denotes a probability computed by using Expression 2 (1 ⁇ j ⁇ I).
  • the bias parameter b j for each input node (1 ⁇ j ⁇ I) may be adjusted and updated as denoted by the following expression by using Expression 8.
  • a coefficient c is a parameter predetermined by the user or the like.
  • the learning processing section 150 adjusts the bias parameter b j so as to increase the conditional probability of the input value x j [0] of the node of the 0-th input layer occurring, on a condition that the history x ( ⁇ T, ⁇ 1] of the common layer 12 has occurred.
  • the learning processing section 150 may iteratively perform updating of the bias parameter b j denoted by Expression 9 and computing of the probability ⁇ x j [0] > ⁇ denoted by Expression 2, to determine the bias parameter b j .
  • the learning processing section 150 stops updating the bias parameter b j and determines the bias parameter b if a difference in the bias parameter b j before and after updating is less than or equal to a predetermined threshold. If a bias parameter b j is also set for a hidden node, the learning processing section 150 may determine the bias parameter b j of the hidden node in the same manner.
  • the learning processing section 150 may decide upon the bias parameter b j by iteratively updating the bias parameter b j a predetermined number of times. If a difference in the bias parameter b j before and after updating is greater than or equal to the predetermined threshold even after the bias parameter b j has been updated the predetermined number of times, the learning processing section 150 may stop updating the bias parameter b j and inform the user that the parameter does not converge.
  • the learning processing section 150 may determine the direction of a change in the learning parameter u u, j, k by using the following expression.
  • the learning processing section 150 may determine the direction of a change in the learning parameter u i, j, k using the following expression.
  • the learning processing section 150 may iteratively perform updating of the learning parameters u i, j, k and v i, j, l corresponding to the input nodes (1 ⁇ j ⁇ I) and computing of the probability ⁇ x j [0] > ⁇ to determine the learning parameters u i, j, k and v i, j, l corresponding to the input nodes (1 ⁇ j ⁇ I).
  • the learning processing section 150 can decide upon the learning parameters u i, j, k and v i, j, l and the bias parameter b j through learning.
  • the learning apparatus 100 may then determine whether to continue learning (S 360 ).
  • the learning apparatus 100 may continue learning until it performs the learning process a predetermined number of times, or may continue learning until a stop command is input by the user. Alternatively, the learning apparatus 100 may continue learning until it can no longer acquire time-series data.
  • the process may return to step S 310 , in which the acquiring section 110 acquires the next time-series data, and the learning apparatus 100 may then perform leaning of the model 10 based on the next time-series data.
  • the supplying section 120 supplies the 0-th input layer with the next image data in the image data acquired by the acquiring section 110 .
  • the storage section 130 samples the values of the hidden layers and stores these values in the 0-th hidden layer. Then, values held in the t-th common layer ( ⁇ T ⁇ t ⁇ 0) may be supplied to the (t ⁇ 1)-th common layer. The values held in the ( ⁇ T+1)-th layer may be deleted.
  • the learning apparatus 100 may perform learning by using image data supplied to the layers from the 0-th input layer to the ( ⁇ T+1)-th input layer as training data and using the values stored in the layers from the 0-th hidden layer to the ( ⁇ T+1)-th hidden layer.
  • the supplying section 120 and the storage section 130 may sequentially acquire new input values x j [0] at the next time point corresponding to;the nodes of the 0-th common layer in the model 10. Then, the computing section 140 may compute a conditional probability ⁇ x j [0] > ⁇ of the new input value x j [0] on a condition that the history has occurred for each common layer before the next time point
  • the learning processing section 150 may adjust the weight parameter so as to increase the conditional probability of the new input value occurring on the condition that this history has occurred,
  • the learning processing section 150 may output the learning parameters u i, j, k and v i, j, l and the bias parameter b j that have been determined and store the parameters in the external database 1000 or the like.
  • the learning apparatus 100 may be operable to apply, to time-series input data that is input in time series, a model having a total of T layers by associating one time point with the 0-th common layer and an input data sequence before the one time, point with T ⁇ 1 layers.
  • the learning apparatus 100 may be operable to apply a model having hidden nodes to each common layer 12 . That is, the learning apparatus 100 may be operable to form a time-evolution Boltzmann machine that predicts input data at one time point on the basis of the input data sequence and hidden node values —
  • the learning apparatus 100 is able to learn the model by computing a conditional probability of the input value x j [0] at the one time point occurring, based on the input value x ( ⁇ T, ⁇ 1] , which is a history, for a model that takes time evolution into consideration. Furthermore, since the learning apparatus 100 learns the model using hidden nodes in addition to the time-series input data, the expressive ability and learning ability can be improved.
  • the learning apparatus 100 may acquire time-series input data of a predetermined duration and then adjust the weight parameters.
  • the learning processing section 150 adjusts the weight parameters collectively for a plurality of tune points in response to acquisition of new input data at a plurality of time points corresponding to D) layers.
  • FIG. 4 shows an example of structures of time-series data and training data for use in learning in the present embodiment.
  • the horizontal axis denotes time.
  • FIG. 4 shows an example in which the learning apparatus 100 uses time-series data y [1, L] having a duration L that is longer than a duration T of time-series data y [1, T] used as training data by the learning apparatus 100 during learning.
  • the learning processing section 150 may be operable to adjust weight parameters for a plurality of time points all together, in response to input data at a plurality of time points being newly acquired.
  • the learning apparatus 100 first performs learning using, as first training data, a time-series data segment of the time-series data from a time 1 to a time T.
  • the learning apparatus 100 may perform learning by setting the time-series data and corresponding, hidden layer values from the time 1 to the time T as each input value of the common layer 12 at the one time point in order, and incrementally shifting the time points one at a time toward the future.
  • the learning apparatus 100 may use data at a time T as each input value x j [0] at the one time point, and continue learning until the time-series data from the time 1 to a time T ⁇ 1 becomes the input data sequence x ( ⁇ T, ⁇ 1] (i.e. the history).
  • the learning apparatus 100 performs learning using, as second training data, a time-series data segment of the time-series data from a time 2 to a time T+1.
  • the learning apparatus 100 may sequentially use each of D pieces of data in the second training data as the input value x j [0] at the one time point.
  • the learning apparatus 100 may shift the time point in the interval from the time 2 to the time T+1 one time point at a time toward the future and use, as the history, the corresponding time-series data and hidden nodes of the inter al from the, time 2 to the time T.
  • the learning apparatus 100 may adjust the parameters D times for the D input values x j [0] and the corresponding D histories. That is, the learning apparatus 100 may use a stochastic gradient technique>in which the learning method described with Expressions 8 to 11 is performed.
  • the learning apparatus 100 may acquire D time-series data sets, generate a plurality of training data sets from time-sequence data segments of a duration of L, and collectively perform learning for D layers. Specifically, the learning apparatus 100 may perform the stochastic gradient technique described using Expressions 8 to 11 collectively for D layers, by using the following expression.
  • FIG. 5 shows a first modification of the learning apparatus 100 according to the present embodiment.
  • Components of the learning apparatus 100 shown in FIG. 5 that perform substantially the same operations as those of the learning apparatus 100 according to the embodiment illustrated in FIG. 1 are denoted by the same reference numerals, and a description thereof is omitted.
  • the learning apparatus 100 according to the present modification may be operable to efficiently update parameters by using FIFO memories and learn a model corresponding to the time-series input data.
  • the learning apparatus 100 according to the present modification further includes FIFO memories 160 and an updating section 170 .
  • Each of the FIFO memories 160 may sequentially store input data and output the stored data after'a predetermined number of storages have been performed.
  • Each of the FIFO memories 160 may be a memory that first outputs data that has been stored first (FIFO: First in, First Out).
  • Each of the FIFO memories 160 may sequentially store an input value of the common layer 12 and output the input value after a predetermined number of storages have been performed.
  • the learning apparatus 100 may include a plurality of FIFO memories 160 , the number of which is greater than or equal to the number of nodes n of the model.
  • the plurality of FIFO memories 160 is desirably provided to have a one-to-one correspondence with the plurality of nodes of the common layer 12 . That is, each of the plurality of FIFO memories 160 may be provided in a manner to store a history for a respective node of the common layer 12 or to update the history thereof.
  • the plurality of FIFO memories 160 are connected, to the acquiring section 110 and the storage section 130 , and sequentially store input values corresponding to new input data of the common layer 12 .
  • the plurality of FIFO memories 160 are also connected to the updating section 170 and sequentially supply the data stored therein to the updating section 170 .
  • the updating section 170 may he operable to update a plurality of update parameters that are based on the hidden nodes and the input data sequence of the time-series input data before the one time point, from values at a prior time point to values at the one time point, on the basis of values of the update parameters and values of the hidden nodes and input values corresponding to the input data to be reflected next.
  • the updating section 170 may update the update parameters by using values input to the FIFO memories 160 and values output from the FIFO memories 160 .
  • the updating section 170 may be connected to the acquiring section 110 and the storage section 130 , and may receive values input to the FIFO memories 160 .
  • the updating section 170 may receive values input to the FIFO memories 160 from the acquiring section 110 via the supplying section 120 .
  • the update parameters are ⁇ i, j, k and ⁇ i,l shown in Expressions 5 and 7.
  • the update parameters are based on input values i (1 ⁇ i ⁇ I) corresponding to input data of the input data sequence at each time point and the predefined parameters ⁇ k t1 and ⁇ l t2 of the weight parameter W ij [ ⁇ ] between this input value i and the target input node j (1 ⁇ j ⁇ I) or hidden node j (I+1 ⁇ j+I+H), for example.
  • the update parameters are based on the hidden node i (I+1 ⁇ i ⁇ I+H) at each time point and the predefined parameters and ⁇ k t1 and ⁇ l t2 of the weight parameter W ij [ ⁇ ] between this hidden node i and the target input node j (1 ⁇ j ⁇ I) or hidden node j (I+1 ⁇ j ⁇ I+H), for example.
  • the update parameters may be updated every time the acquisition of the time-series input data by the acquiring section 110 and the storage of the sampling values by the storage section 130 are performed sequentially.
  • the above-described learning apparatus 100 according to the present modification may be operable to learn a modification of the model 10.
  • the modification of the model 10 is described with reference to FIG. 6 .
  • FIG. 6 shows a modification of the model 10 according to the present embodiment.
  • the model 10 according to the modification needs not have the layered structure including T layers shown in FIG. 2 .
  • FIG. 6 shows an example of a model corresponding to one of the FIFO memories 160 .
  • the overall configuration of the model 10 according to the present modification includes a storage area that is equivalent to the 0-th common layer in FIG. 2 including the training data, and a number of the configurations illustrated in FIG. 6 equal to the number of nodes n I+H). Neurons i and j and a FIFO sequence 20 of the model 10 according to the present modification are described below.
  • the neuron i may be equivalent to the'input terminal of the FIFO memory 160 .
  • An input value y i [t] (1 ⁇ i ⁇ I) of each node in the input data of the input data sequence at each time point t and a corresponding value y i [t] the values y i [t] (I+1 ⁇ i ⁇ i+H) of the hidden nodes at each time point are sequentially input to the neuron i.
  • the neuron i may set the value y i [t] input thereto as the current input value.
  • the neuron i may supply the input value y i [t] input at the time point t to the updating section 170 and to the FIFO sequence 20 as the previous input value and may hold the input value y i [t+1] at the time point t+1 as the current input value.
  • the FIFO sequence 20 may store d ij ⁇ 1 of the latest input values received from the neuron i.
  • the FIFO sequence 20 may supply the d ij ⁇ 1 input values stored therein to the updating section 170 .
  • the updating section 170 may be operable to compute the values of the update parameters denoted by Expression 6 by using the input values supplied by the FIFO sequence. If the FIFO sequence 20 holds input values from the time point t ⁇ 1 to the time point t ⁇ d ij ⁇ 1, the FIFO sequence 20 is denoted by the following expression.
  • the FIFO sequence 20 may supply the input value y i [t1] to the neuron j.
  • the input value y i [t1] supplied to the neuron j at the time point t 4 is immediately supplied to the updating section 170 at the time point t 4 .
  • the input value y i [t1] that the neuron j has received from the FIFO sequence 20 at the time point t 4 does not serve as an input for the neuron j, and the input value y i [t4] may be input to the neuron j at the time point t 4 .
  • the neuron j may be equivalent to the output terminal of the FIFO memory 160 , and the neuron j may receive the input value y i [t1] input to the neuron i at the time point t 1 , via the FIFO sequence after the time period d ij , i.e. at the time point t 1 +d ij . That is, the model 10 from the neuron i to the neuron j via the FIFO sequence 20 may correspond to the FIFO memory 160 that, stores d ij pieces of input data.
  • the neuron i of the model 10 according to the modification may correspond to, for example, a node for an input data sequence such as a node i of the ( ⁇ )-th common layer of the model 10 shown in FIG. 2 , and in this case the neuron j may correspond to, for example, the node j of the 0-th common layer.
  • the neuron j may supply the received input value y [t1] to the updating section 170 .
  • the model 10 may supply the input values at the time point t ⁇ 1 and the time point t ⁇ d ij +1 to the updating section 170 at the time point tin: this way, the updating section 170 can update the update parameters by adding the corresponding, input value in the input data to be reflected next to the update parameters for the time point before the one time point, and then multiplying the resulting sum by a predetermined constant.
  • the update parameters denoted by Expression 8 may be computed in accordance with Expression 8 by using the input values stored in the FIFO sequence 20 that are supplied to the updating section 170 .
  • the update parameter ⁇ i,l denoted by Expression 7 can be updated by using the input values supplied to the updating section 170 and the second predefined parameter.
  • the updating section 170 can compute the update parameter ⁇ i,l to be used in the current learning by performing computing at the time point t according to the following expression by using, the prior update parameter ⁇ i,l and the input value y i [t ⁇ 1] received from the neuron i at the time point t.
  • FIG. 7 shows an example of a temporal change in the update parameter ⁇ i,l according to the present embodiment.
  • FIG. 7 shows an example in which values greater than 0 (for example, 1) are input to the neuron i as the input value at time points t ⁇ 5, t ⁇ 2, and t ⁇ 1, and these input values are supplied to the updating section 170 at time points t ⁇ 4, t ⁇ 1, and t.
  • the second predefined parameter ⁇ l is a parameter whose value gradually decreases as the time point difference increases. Accordingly, the update parameter ⁇ i,l computed by the updating section 170 tends to decrease as time passes from when the input value of 1 is input to when the next input is given.
  • the update parameter ⁇ i, j, k denoted by Expression 5 can be updated by using the input values supplied to the updating section 170 and the first predefined parameter ⁇ k .
  • the updating section 170 can compute the update parameter ⁇ i, j, k to be used in the current learning by performing computing at the time point t according to the following expression, by using the prior update parameter ⁇ i, j, k and the input value y i [t ⁇ dij] from the neuron j at the time point t.
  • FIG. 8 shows an example of a temporal change in the update parameter ⁇ i, j, k according to the present embodiment.
  • FIG. 8 shows an example in which values greater than 0 (for example, 1) are supplied to the neuron j as the input value at time points t ⁇ 3, t ⁇ 1, and t.
  • the first predefined parameter ⁇ k is a parameter whose value gradually decreases as the time point difference increases. Accordingly, the update parameter ⁇ i, j, k computed by the updating section 170 tends to decrease as time passes from when the input value of 1 is input to when the next input is given.
  • the learning apparatus 100 can update the update parameters ⁇ i, j, k and ⁇ i,l by applying the model 10 shown in FIG. 6 using the FIFO memories 160 and the updating section 170 .
  • the updating section 170 can apply the model 10 according to the present modification, for example, by acquiring the input values x i [t ⁇ 1] at the time point t ⁇ 1 from the input data input to the FIFO memories 160 and acquiring the input values x i [t ⁇ dij] at the time point t ⁇ d ij from the output of the FIFO memories 160 .
  • the learning apparatus 100 may update the parameter ⁇ i,j, l through substantially the same operation as the operation described in FIG. 3 .
  • the computing section 140 can compute the parameter ⁇ i,j, l by determining the sum of products of the second predefined parameter ⁇ l and the input value x i (y i in the present modification) for time points from t ⁇ 1 to t ⁇ d ij +1 as indicated by Expression 6.
  • the computing section 140 can compute, by using the plurality of update parameters, conditional probabilities of input data values at one time point on the condition that the hidden node values and input data sequence have occurred. Then, the learning processing section 150 can determine the learning parameters u i, j, k and v i, j, l and the bias parameter b j by performing substantially the same operation as the operation described in FIG. 3
  • the learning apparatus 100 can determine the weight parameter and bias parameters in a manner to increase the probability of predicting the input value to be input to the input layer 14 of the common layer 12 , based on the past values that have been input to the common layer 12 of the model 10 before the one time point. Furthermore, the learning apparatus 100 can improve the prediction accuracy, the expressive ability, the learning efficiency, and the like of the input values input to the input layer 14 by having the common layer 12 include the hidden layer 16 in addition to the input layer 14 .
  • the learning apparatus 100 is an example in which a value that is unrelated to the prediction made by the learning apparatus 100 is sampled and input as the hidden node value to be input to the hidden layer 16 .
  • the learning apparatus 100 may determine the hidden node value by using a history of the conditional probability of the values of the nodes of the common layer 12 .
  • the learning apparatus 100 may determine the weight parameter to a hidden node by using this conditional probability history.
  • the learning apparatus 100 can improve the prediction accuracy by using the conditional probability history of nodes of the common layer 12 to determine the weight parameter to the hidden node and the hidden node value.
  • the computing section 140 may compute the conditional probability p j,t of the value of a node j of the common layer 12 at one time point t based on the values input to the corresponding node j of the common layer 12 at each time point before the one time point t, and store this conditional probability in the storage section or the like.
  • the computing section 140 may compute the conditional probability of each hidden node in the layer 16 at the one time point in the same manner. That is, the computing section 140 may use the plurality of update parameters to compute the conditional probability of the value of each hidden node and each input data value at the one time point on the condition that an input data sequence has occurred.
  • the computing section 140 may store the conditional probability p j,t in a FIFO or the like.
  • the computing section 140 may be operable to compute a total likelihood, after the learning by the learning apparatus 100 has continued.
  • the computing section 140 computes the total likelihood p j as shown in the following expression, based on the conditional probabilities p j,t ⁇ K+1 , p j,t ⁇ K+2 , . . . , p j,t computed by K instances of learning from the time point t ⁇ K+1 to the time point t, for example.
  • the total likelihood p j in Expression 16 indicates a total sum of the conditional probabilities, as an example, but the total likelihood p j may be at least one of a sum, weighted sum, product, or weighted product of the conditional probabilities.
  • K may be an integer greater than or equal to 2
  • the computing section 140 stores the conditional probabilities p j,t in a FIFO or the like, the length of the FIFO sequence may be equal to the value of K.
  • the computing section 140 may supply the total likelihood p j to the storage section 130 .
  • the storage section 130 may sample the values x j [t] of the hidden nodes of the hidden layer 16 at the one time point, based on the most recent likelihood p j,t . That is, the storage section 130 according to the present embodiment may be operable to sample the value of each hidden node at the one time point, by using the conditional probability of the value of each hidden node at the one time point. For example, the storage section 130 samples the values of the hidden nodes based on the history of the conditional probabilities computed by the computing section 140 . That is, the storage seal m 130 may sample the values of the hidden nodes after the learning operation of the learning processing section 150 has been performed a plurality of times. The storage section 130 may store a value of 0 in the hidden nodes as the sampling value, until the leaching operation of the learning processing section 150 has been performed a plurality of times.
  • the storage section 130 may store a value of 1 or 0 in the hidden node j as the sampling result, according to the result of a comparison between the value of the total likelihood p j and a threshold value. In this way, when predicting the time series data to be input to the input layer 14 , the storage section 130 can store a more preferable value as the hidden node value by performing sampling based on the history of past conditional probabilities.
  • the learning processing section 150 may be operable to determine the weight parameter based on the total likelihood p j .
  • the learning processing section 150 may compute update amounts and ⁇ u i, j, k and ⁇ v i, j, k for the learning parameters u i, j, k and v i, j, k in the weight parameter for one hidden node j at the one time point.
  • the learning processing section 150 may compute these update amounts ⁇ u i, j, k [t] and ⁇ v i, j, k [t] as shown in the following expression, based on the value x j [t] of the one hidden node j at the one time point t and on the conditional probability ⁇ X j [t] > of the value of this hidden node j at the one time point t on the condition that the input data'sequence has occurred (I+1 ⁇ j ⁇ I+H).
  • the update amount ⁇ v i, j, k [t] is equal to ⁇ u i, j, k (1)[t] + ⁇ u i, j, k (2)[t] .
  • the conditional probability ⁇ X j [t] > of the value of the hidden node j may be computed by the computing section 140 using Expression 2.
  • the learning processing section 150 may store the computed update amounts ⁇ v i, j, k [t] , ⁇ u i, j, k (1)[t] , and ⁇ u i, j, k (2)[t] in the storage section or the like.
  • the learning processing section 150 may be operable to store the update amounts ⁇ v i, j, k [t] , ⁇ u i, j, k (1)[t] , and ⁇ u i, j, k (2)[t] computed for one time point in the FIFO sequence. That is, the learning processing section 150 may be operable to update the learning parameters based on update amounts computed in the past.
  • the learning, processing section 150 changes the ratio by which the update amounts are reflected in the learning parameters, according to the conditional probability of input data occurring at a following time point that is after the one time point t, for example.
  • the learning processing section 150 may change the ratio by which the update amounts are reflected in the learning parameters according to the conditional probability of a bidden node value occurring at a plurality of following time points that are after the one tune point.
  • the learning processing section 150 changes the ratio by which the update amounts are reflected in the learning parameters according to the total likelihood p j after the one time point
  • the learning processing section 150 may update the learning parameters as shown in the following expression, based on the total likelihood p j computed by K instances of teaming from the time point to the time point t and on the update amount at the time point t ⁇ K+1, for example.
  • K may be an integer greater than or equal to 2
  • the length of the FIFO sequence may be equal to the value of K.
  • ⁇ 1 may be a constant for adjusting the update amount.
  • ⁇ 1 may be a coefficient whose value becomes smaller according to an increase in the number of updates.
  • ⁇ l may have a value of substantially 1 at the stage when the learning processing section 150 begins learning, and may be a coefficient whose value becomes smaller according to the amount of learning occurring as time progresses from the time point t.
  • ⁇ 1 ⁇ 10 /t 2 .
  • ⁇ 1 may be a coefficient whose value becomes smaller according to the update amount.
  • ⁇ 1 ⁇ 10 /( ⁇ u i, j, k 1/2 ).
  • ⁇ 10 may be a predetermined constant.
  • the learning processing section 150 may update the learning parameters of a hidden node at one time point according to the conditional probabilities computed at time points before the one time point.
  • the learning apparatus 100 can more strongly reflect the update amounts at time points before the one time in the learning parameters, in response to the predicted probability of an input value of an input node being large due to the weight parameters at time points before the one time point. That is, the learning apparatus 100 can update the weight parameters of the hidden nodes in a mariner to increase the conditional probabilities.
  • the learning processing section 150 may extract from the FIFO sequence the update amounts ⁇ v i, j, k [t ⁇ K+1 , ⁇ u i, j, k (1)(t ⁇ K+1] , and ⁇ u i, j, k (2)[t ⁇ K+1] of a past time point (e.g. t ⁇ K+1) at the following time point (e.g. t) or a time point thereafter.
  • the learning processing section 150 may update the ratio by which the update extracted from the FIFO sequence are reflected in the learning parameters according to the conditional probabilities of hidden node values occurring at the following time point t. For example, the learning processing section 150 may multiply the total likelihood p j respectively by each update amount. In this way, the learning processing section 150 can efficiently perform the update of the weight parameters as described above.
  • the learning apparatus 100 can learn a model 10 in a manner to predict input data at one time point, based on an input data series before the one time point in the time-series input data. Furthermore, the learning apparatus 100 can improve the expressive power, learning power, and the like by including hidden nodes. With such a learning apparatus 100 , there are cases where it is relatively difficult to perform the learning.
  • the learning apparatus 100 may be operable to further improve the learning by changing the order in which the time-series input data is input to the model 10 and performing a bidirectional learning process.
  • the present embodiment describes an example in which the learning apparatus 100 is operable to learn a model 10 in a manner to predict input data at one time point, based on the input data series before the one time point in the time-series input data.
  • the learning apparatus 100 creates a first model that is a forward model that performs a learning process in which the time-series input data is input to the model 10 in order from past data to future data.
  • FIG. 9 shows an exemplary schematic configuration of the first model according to the present embodiment.
  • FIG. 9 shows a schematic configuration of the first model corresponding to the model 10 shown in FIG. 2 .
  • the value of the node j in the input layer 14 at the one time point t is indicated by x j [t] (1 ⁇ j ⁇ I)
  • the value of the node j in the hidden layer 16 at the one time point t is indicated by h j [t] (I+1 ⁇ j ⁇ H).
  • the value of the node i in a plurality of input layers 14 before the one time point t is indicated by x i [ ⁇ 1] (1 ⁇ i ⁇ I), and the value of the node i in the hidden layer 16 before the one time point t is indicated by h i [:t ⁇ 1] (I+1 ⁇ i ⁇ H).
  • the first model may include a plurality of input nodes j that sequentially input a plurality of input values at each time point in a time-series input data sequence, and a weight parameter W ij [ ⁇ ] between each input node j and each input value x i [1 ⁇ 1] at the time points before the time point: corresponding to the plurality of input nodes j.
  • W ij [ ⁇ ] the weight parameter between a plurality of input layers 14 before the one time point and the input layer 14 at the one time point.
  • the first model may include a weight parameter W ij [ ⁇ ] between each input node j and each of a plurality of hidden nodes i corresponding to time points before the time point corresponding to the plurality of input nodes j.
  • W ij [ ⁇ ] the weight parameter between the plurality of hidden layers 16 before the one time point t and the input layer 14 at the one time point t is W HV .
  • the first model may further include a weight parameter W ij [ ⁇ ] between each input node i corresponding, to a time point before the time point corresponding to the plurality of input nodes j and each hidden node j corresponding to the time point corresponding to the plurality of input nodes j.
  • the weight parameter between the plurality of input layers 14 before the one time point t and the hidden layer 16 at the one time point t is W VH .
  • the first model may further include a weight parameter W ij [ ⁇ ] between each of the plurality of hidden nodes i corresponding to a time point before the time point corresponding to the plurality of input nodes j and each hidden node j corresponding to the time point corresponding to the plurality of input nodes j.
  • the weight parameter between the plurality of hidden layers 16 before the one time point t and the hidden layer 16 at the one time point t is W HH .
  • conditional probability p of the, input node j and the hidden node j at the one time point may be calculated based on the following expression.
  • the update of each parameter in the first model is as described above.
  • the order in which this time-series input data is input to the model 10 is not limited to the forward order from past data to future data, as in the first model.
  • the learning apparatus 100 may be operable to learns the model 10 in a manner to predict input data at one time point based on the input data series after the one time point i the'time-series input data. In this way, the learning apparatus 100 creates a second model that is a backward model that performs learning process in which the time-series input data is input to the model 10 m order from future data to past data.
  • FIG. 10 shows an exemplary schematic configuration of the second model according to the present embodiment.
  • FIG. 10 shows a schematic configuration of the second model corresponding to the model 10 shown in FIG. 2
  • the value of the node j in the input layer 14 at the one time point t is indicated by x j [t] (1 ⁇ j ⁇ I)
  • the value of the node j in the hidden layer 16 at the one time point t is indicated by h j [t] (I+1 ⁇ j ⁇ H).
  • the value of the node i in a plurality of input layers 14 after the one time point t is indicated by x i [t ⁇ 1:] (1 ⁇ i ⁇ I), and the value of the node i in the hidden layer 16 after the one time point t is indicated by h i [t ⁇ 1:] (I+1 ⁇ i ⁇ H).
  • the learning apparatus 100 inputs the data the model 10 from the temporally newest data in the time-series input data.
  • the second model may include a plurality of input nodes i that sequentially input, in a backwards order, a plurality of input values at each time point in the time-series input data sequence, and a weight parameter W ji [ ⁇ ] between each input value x j [t] at time points after the tune point corresponding to the plurality of input nodes i and each input node i.
  • W VV the weight parameter between the input layer 14 at the one time point and a plurality of input layers 14 after the one time point.
  • the second model may include a weight parameter W ji [ ⁇ ] between each of a plurality of hidden nodes j corresponding to time points after the time point corresponding to the plurality of input nodes i and each input node i.
  • W ji [ ⁇ ] the weight parameter between the input layer 14 at the one time point t and the plurality of hidden layers 16 after the one time point t is W VH .
  • the second model may further include a weight parameter W ji [ ⁇ ] between each input node j corresponding to a time point after the time point corresponding to the plurality of input nodes i and each hidden node i corresponding to a time point corresponding to the plurality of input nodes i.
  • the weight parameter between the hidden layer 16 at the one time point t and the plurality of input layers 14 after the one time point t is w HV .
  • the second model may further include a weight parameter W ij [ ⁇ ] between each of the plurality of hidden nodes j corresponding to a time point after the time point corresponding to the plurality of input nodes i and each hidden node i corresponding to the time point corresponding to the plurality of input nodes i.
  • the weight parameter between the hidden layer 16 at the one time point t and the plurality of hidden layers 16 after the one time point t is W HH .
  • the learning apparatus 100 may be operable to learn each parameter in common for the first model and the second model. Furthermore, the learning operation of the learning apparatus 100 using the second model may be operable to be performed using the same operation as the learning operation using the first model.
  • the bias parameters for the first model and the second model may be substantially the same parameter.
  • the weight parameters W ij and W ji for the first model and the second model may be operable to be converted according to a transpositional relationship there between.
  • the learning parameters u i,j,k and v i,j,l of the first model ma be operable to be converted to the learning parameters u i,j,k and v j,i,l of the second model by switching i and j.
  • the predefined parameters ⁇ k t1 and ⁇ l t2 of the first model and the second model may be substantially the same parameter. Since the value of the predefined parameter is determined according to the time point difference, the predefined parameter value may change differently according to each time point in the first model and the second model.
  • the parameters of the first model and the second model are shared. Accordingly, the learning result of the first model and the learning result of the second model based on the same time-series input data sequence will theoretically match. However, there are cases where differences in the ease of learning occur, according to the directional order in which the data is input.
  • the value of the input layer 14 at the one time point t and the plurality of values of the input layer 14 before the one time point t are already known as the time-series input data.
  • the plurality of values of the hidden layer 16 before the one time point t are values obtained after a certain amount of time points have passed from when these values were sampled, for example, and therefore an evaluation result is already obtained for most of these values.
  • the weight parameters W VV and W HV relating to the input layer 14 at the one time point t in the first model are parameters that can be accurately learned.
  • the value of the hidden layer 16 at the one time point t is a sampled value, for example, and therefore an evaluation result of this sampled value becomes known after several instances of learning have been completed. Accordingly, there are cases where the weight parameters W VH and W HH relating to the hidden layer 16 at the one time point t in the first model are difficult to learn.
  • the weight parameter W HV that is difficult to learn in the first model is set as a first learning target parameter, and the weight parameter W HV that can be accurately learned is set as a second learning target parameter.
  • the weight parameters W VV and W VH relating to the input layer 14 at the one time point t are parameters that can be accurately learned.
  • the weight parameters W HV and W HH relating to the hidden layer 16 at the one time point t are difficult to learn. Therefore, the weight parameter W VH that can be accurately learned in the second model is set as a first learning target parameter, and the weight parameter W HV that is difficult to learn is set as a second learning target parameter.
  • the first learning target parameter W VH of the first model may he expressed relatively as the weight parameter between a past input layer 14 and a fixture hidden layer 16 .
  • the first learning target parameter W VH of the second model may also be expressed relatively as the weight parameter between a past input layer 14 and a future hidden layer 16 .
  • the first learning target parameter W VH of the first model and the first learning target parameter W VH of the second model may be a common parameter, or the first learning target parameter W VH of the first model may be operable to be converted into the first learning target parameter W VH of the second model. Accordingly, by learning the first learning target parameter W VH using the second model that can accurately learn this parameter, the learning apparatus 100 can easily calculate the first learning target parameter W VH that can be difficult to calculate with the first model.
  • the second learning target parameter W HV of the first model may be expressed relatively as the weight parameter between a past hidden layer 16 and a future input layer 14 .
  • the second learning target parameter W HV of the second model may also be expressed relatively as the weight parameter between a past hidden layer 16 and a future input layer 14 .
  • the second learning target parameter W HV of the first model and the second learning target parameter W HV of the second model may be a common parameter, or the second learning target parameter W HV of the first model may be operable to be converted into the second learning target parameter Wuv of the second model. Accordingly, by learning the second learning target parameter W HV using the first model that can accurately learn this parameter, the learning apparatus 100 can easily calculate the second learning target parameter W HV that can be difficult to calculate with the second model,
  • the learning apparatus 100 may be operable to perform efficient learning by using the first model and the second model to learn a parameter that is difficult for one of the models to learn using a parameter that can be accurately learned by the other model.
  • the following describes the learning operation of such a learning apparatus 100 .
  • FIG. 11 shows an operational flow of the learning process using the first model and the second model performed by the learning apparatus 100 according to the present embodiment.
  • the learning apparatus 100 may be operable to accurately learn a model 10 corresponding to time-series input data by performing the processes from S 410 to S 470 .
  • the acquiring section 110 may acquire a time-series input data sequence (S 410 ).
  • the acquiring section 110 acquires, as the time-series input data sequence, time series data in an interval L that is longer than an interval T of the time-series data used for one instance of learning.
  • the acquiring section 110 acquires L pieces of image data arranged in time series to form moving image data, for example.
  • the learning apparatus 100 may learn the weight parameter between each input value and each hidden node corresponding to a time point before a time point corresponding to a plurality of input nodes in the first model, using a first model learning process (S 420 ).
  • the supplying section 120 may sequentially input the time-series input data sequence into the model 10 in order from the older pieces of input data, and the storage section 130 may sample the values of the plurality of hidden nodes corresponding to the respective time points and store the values respectively in the corresponding plurality of hidden nodes.
  • the learning apparatus 100 may perform this learning using the first model. Specifically, the calculating section 140 may calculate the conditional probability of each input value of the input node at the one time point, and the learning processing section 150 may update the parameter ⁇ . Furthermore, the updating section 170 may update the update parameter.
  • the learning operation of the learning apparatus 100 has already been described above, and therefore is not provided here.
  • the learning apparatus 100 may be operable to perform the learning process a predetermined number of times using the first model.
  • the learning apparatus 100 may convert the parameter learned using the first model into a parameter corresponding to the second model (S 430 ).
  • the updating section 170 includes a converting section, and the converting section;may be operable to convert the parameter of the first model into the parameter of the second model.
  • the converting section may convert the parameters by performing a transposition, element replacement, and the like.
  • the learning apparatus 100 may learn the weight parameter between each of the plurality of hidden nodes and each input node corresponding to the time points after the time point corresponding to the plurality of input nodes in the second model, by using the learning process with the second model using the converted parameter (S 440 ).
  • the supplying section 120 may sequentially input the time-series input data sequence into the model 10 in order from the newer pieces of input data, and the storage section 130 may sample the values of the plurality of hidden nodes corresponding to each time point and store the sampled values respectively in the corresponding plurality of hidden nodes.
  • the supplying section 120 and the storage section 130 may store each piece of data in a FIFO memory 160 used for learning by the second model, which is different from a FIFO memory 160 used for learning by the first model.
  • the learning, apparatus 100 may be operable to perform the learning process a predetermined number of times using the second model.
  • the learning apparatus 100 may convert the parameter learned using the second model into the parameter corresponding to the first model (S 450 ).
  • the converting section may be operable to convert the parameter of the second model into the parameter of the first model.
  • the converting section may convert the parameters using a transposition, element replacement, and the like.
  • the learning apparatus 100 may judge whether the learning is to continue (S 460 ).
  • the learning apparatus 100 may continue the learning until the learning process has been performed a predetermined number of times, or may instead continue the learning until a stop command from the user is input. As another example, the learning apparatus 100 may continue the learning until there is no more data in the time-series input data sequence that can be input.
  • the learning apparatus 100 returns to step S 410 , and if there is no more time-series data to be supplied to the model 10, the acquiring section 110 may acquire the next piece of data in the time-series data sequence.
  • the learning apparatus 100 may sequentially perform learning with the first model and the second model based on the next piece of time-series data.
  • the teaming processing section 150 may perform a final learning process using the first model (S 470 ).
  • the learning processing section 150 may be operable to acquire the learned first model after performing the learning process with the first model one or more times.
  • the learning processing section 150 may be operable to output the learned first model to the database 1000 or the like.
  • the learning processing section 150 may output the first model and end the learning process.
  • the learning apparatus 100 may output, as the learned model, the model used to perform the last instance of learning corresponding to the model to be output from among the first model and the second model.
  • FIG. 11 describes an example in which the learned first model is output, but if the learned second model is output, the learned second model may be output without performing the process of S 450 for convening the parameters.
  • the learning apparatus 100 may continue with the learning while focusing more heavily on learning corresponding to the model to be output from among the first model and the second model. For example, the learning apparatus 100 may perform the learning process with the first model a greater number of times than the learning process with the second model. Furthermore, the learning apparatus 100 may be operable to perform the learning using a higher learning rate for the learning process with the first model than for the learning process with the second model, by adjusting a coefficient or the like. In this way, the learning apparatus 100 performs learning in a focused manner corresponding to the model to be output, and can therefore output a more suitable learned parameter model using this model.
  • the learning process of the learning apparatus 100 performs a bidirectional learning process with the first model and the second model, using common parameters. Accordingly, even if parameters having learning accuracies that fluctuate according to the learning direction are included, the learning apparatus 100 can learn these parameters with higher accuracy. Furthermore, by performing, bidirectional learning, the learning apparatus 100 can prevent the occurrence of overfitting caused when model learning is performed using only one of the first model and the second model.
  • the learning process of the learning apparatus 100 is an example of a bidirectional learning process performed using a forward model and a backward model.
  • the learning apparatus 100 may make a distinction in the parameters being learned according to the model.
  • the first learning target parameter W VH of the first model may be operable to be learned with higher accuracy for learning using the second model than for learning using the first model.
  • the second learning target parameter %Tv of the first model may be operable to be learned with higher accuracy for learning using the first model than for learning using the second model.
  • the learning using the first model performed by the learning apparatus 100 may be operable to learn the second learning target parameter W HV without changing the first learning target parameter W VH .
  • the learning apparatus 100 may update the second learning target parameter W HV without updating the first learning target parameter W VH .
  • the learning using the second model performed by he learning apparatus 100 may be operable to learn the first learning target parameter W VH without changing the second learning target parameter W HV .
  • the learning apparatus 100 may update the first learning target parameter W VH but need does not need to update the second learning target parameter W HV .
  • the learning apparatus 100 may increase or decrease the update amount, according to the model used for the learning. For example, when learning with the first model, the learning apparatus 100 may perform the update by multiplying the update amount of the first learning target parameter W VH by a coefficient ⁇ , and multiplying the update amount of the second learning target parameter W HV (1 ⁇ ).
  • the coefficient p may be a positive value that is less than 0.5.
  • the learning apparatus 100 may perform the update by multiplying the update amount of the first learning target parameter W VH by (1 ⁇ ), and multiplying, the update amount of the second learning target parameter W HV by the coefficient ⁇ .
  • the learning apparatus 100 can perform the learning process more;accurately by encouraging parameter updates that are suitable for the learning, according to the model used for the learning. Accordingly, by including the hidden nodes, the learning apparatus 100 can improve the expressive power, learning power, and the like, and can perform the learning more accurately by performing the bidirectional learning process.
  • the learning apparatus 100 is an example in which the learning apparatus 100 acquires and learns a time-series input data sequence, but the present invention is not limited to time-series data.
  • n-pixel column data arranged in a column direction may be'used as the data corresponding to the input data at the one time point.
  • the plurality of pieces of column data arranged in the row direction may be the data corresponding to the input data before or after the one time point
  • the learning apparatus 100 may perform the bidirectional learning for a case where a plurality of pieces of column data are input to the model 10 in one row direction and a case where the plurality of pieces of column data are input to the model 10 in the other row direction.
  • the learning apparatus 100 may perform the bidirectional learning for waveform data, image pattern data, language data and the like, in the same manner.
  • the learning apparatus 100 is an example in which the learning apparatus 100 acquires one input data sequence and generates training data for the bidirectional learning, based on this one input data sequence, but the data acquired in the present invention is not limited to one input data sequence.
  • the learning apparatus 100 may acquire a plurality of data sequences. In this case, the learning apparatus 100 may use learning with the first model for the first input data sequence and use learning with the second model for a second input data sequence that differs from the first input data sequence.
  • the first input data sequence and the second input data sequence are preferably at least partially related. Furthermore, if the data sequences are the same type or have substantially the same trend, the first input data sequence and the second input data sequence may be input data sequences for learning that are different from each other and are included in a plurality of input data sequences for learning.
  • the learning apparatus 100 may be operable to, based on the first model, sequentially input each piece of input data from the first input data sequence, in which are arranged pieces of input data including a plurality of input values, into the input nodes of the model 10, to perform the learning process using the first model.
  • the learning apparatus 100 may be operable to input each piece of input data from the second input data sequence, in which are arranged pieces of input data including a plurality of input values, into the input nodes of the model 10 in a different order than, in the first model, to generate the second model for learning the first learning target parameter included in the first model.
  • the learning apparatus 100 may be operable to input each piece of input data from the second input data sequence in a backwards order, to generate the second model for learning the learning target parameter.
  • the first and second input data sequences may be time-series input data sequences.
  • the first model may input the first input data sequence in order from the older pieces of input data
  • the second model may input the second input data sequence in order from the newer pieces of input data.
  • the learning apparatus 100 may be operable to perform the learning process using both the first model and the second model.
  • the learning apparatus 100 may be operable to output the learned first model after performing the bidirectional learning with the first model and the second model.
  • the learning apparatus 100 may be operable to delete the learned second model and output the learned first model as the predictive model based on the input data sequence. In this way, the learning apparatus 100 may be operable to acquire a plurality of input data sequences for learning and perform the learning process.
  • the learning apparatus 100 is an example in which the learning apparatus 100 learns by applying substantially the same learning; process to the first learning target parameter and the second learning target parameter. Instead of this, the learning apparatus 100 may learn by applying different learning processes respectively to the first learning target parameter and the second learning target parameter. In other words, the learning apparatus 100 may be operable to perform an update by decreasing the update amount for a parameter that can be difficult to learn.
  • the learning apparatus 100 may be operable to independently perform learning with ⁇ v as an update parameter than can be accurately updated and with On as an update parameter than can be difficult to learn.
  • bv i indicates the bias parameter corresponding to each node of the input layer 14 .
  • b Hj indicates the bias parameter corresponding to each node of the hidden layer 16 .
  • a lower bound for the total likelihood can be expressed as shown in the following expression from Jensen's inequality
  • the value ⁇ tilde over (h) ⁇ of the hidden node may be sampled in correspondence with p ⁇ ( ⁇ tilde over (h) ⁇
  • the stochastic gradient may be calculated as shown in the following Expression, according to the parameter to be updated.
  • the learning apparatus 100 may be operable to, when learning using the first model, update the parameter ⁇ v that can be learned accurately as described above.
  • the update of the parameter ⁇ v can be expressed as shown in the expression below.
  • the lemming apparatus 100 may be operable to, when learning using the first model, gradually update the parameter ⁇ H that can be difficult to learn, by using a gentle gradient.
  • the update of the parameter ⁇ H can be expressed as shown in the expression below.
  • the parameter g is a gradient parameter indicating the direction of the gradient.
  • the constant ⁇ may be a value less than 1, and may be a value closer to 0, for example.
  • the log p ⁇ operating On ⁇ g in the expression for updating ⁇ H need not be included.
  • may be a constant for adjusting the update amount or may be a coefficient whose value becomes smaller according to an increase in the number of updates.
  • the learning apparatus 100 may be operable to, when learning using the second model, update the parameter ⁇ v that can be accurately learned, as shown in the following expression.
  • the learning apparatus 100 may be operable to, when learning using the second model, update the parameter ⁇ H that can be difficult to learn, as shown in the following expression. °
  • the log p ⁇ operating on ⁇ g in the expression for updating ⁇ H need not be included.
  • the learning apparatus 100 can perform the learning process more accurately by changing the method used according to the ease of learning the parameter.
  • the bidirectional learning process of the learning apparatus 100 according to the present embodiment described above is an example suitable for a Boltzmann machine, but the present invention is not limited to this.
  • the present invention can be applied to any learning apparatus that is operable to bidirectionally learn a data series and capable of changing the ease of the learning according to the learning direction
  • FIG. 12 shows an example of a computer 800 in which aspects of the present invention may be wholly or partly embodied.
  • a program that is installed in the computer 800 can cause the computer 800 to function as or perform operations associated with apparatuses of the embodiments of the present invention or one or more sections (including modules, components, elements, etc.) thereof, and/or cause the computer 800 to perform processes of the embodiments of the present invention or steps thereof.
  • Such a program may be executed by the CPU 800 - 12 to cause the computer 800 to perform certain operations associated with some or all of the blocks of flowcharts and block diagrams described herein.
  • the computer 800 includes a CPU 800 - 12 , a RAM 800 - 14 , a graphics controller 80016 ., and a display device 800 - 18 , winch are mutually connected by a host controller 800 - 10 .
  • the computer 800 also includes input/output units such as a communication interface 800 - 22 , a hard disk drive 800 - 24 , a DVD-ROM drive 800 - 26 and an IC card drive, which are connected to the host controller 800 - 10 via an input/output controller 800 - 20 .
  • the computer also includes legacy input/output units such as a RUM 800 - 30 and a keyboard 800 - 42 , which are connected to the input/output controller 800 - 20 through an input/output chip 800 - 40 .
  • the CPU 800 - 12 operates according to programs stored in the ROM 800 - 30 and the RAM 800 - 14 , thereby controlling each unit.
  • the graphics controller 800 - 16 obtains image data generated by the CPU 800 - 12 on a frame buffer or the like provided in the RAM 800 - 14 or in itself, and causes the it tape data to be displayed on the display device 800 - 18 .
  • the communication interface 800 - 22 communicates with other electronic devices via a network 800 - 50 .
  • the hard disk drive 800 - 24 stores programs and data used by the CPU 800 - 12 within the computer 800 .
  • the DVD-ROM drive 800 - 26 reads the programs or the data from the DVD-ROM 800 - 01 , and provides the hard disk drive 800 - 24 with the programs or the data via the RAM 800 - 14 .
  • the IC card drive reads programs and data from an IC card, and/or writes programs and data into the IC card.
  • the ROM 800 - 30 stores therein a boot program or the like executed by the computer 800 at the time of activation, and/or a program depending on the hardware of the computer 800 .
  • the input/output chip 800 - 40 may also connect various input/output units via a parallel port, a serial port, a keyboard port, a mouse port, and the like to the input/output controller 800 - 20 .
  • a program is provided by computer readable media such as the DVD-ROM 800 - 01 or the IC card.
  • the program is read from the computer readable media, installed into the hard disk drive 800 - 24 , RAM 800 - 14 , or ROM 800 - 30 , which are also examples of computer readable media, and executed by the CPU 800 - 12 .
  • the information processing described in these programs is read into the computer 800 , resulting in cooperation between a program and the above-mentioned various types of hardware resources.
  • An apparatus or method may be constituted by realizing the operation or processing of information in accordance with the usage of the computer 800 -
  • the CPU 800 - 12 may execute a communication program loaded onto the RAM 800 - 14 to instruct communication processing to the communication interface 800 - 22 , based on the processing described in the communication program.
  • the communication interface 800 - 22 under control of the CPU 800 - 12 , reads transmission data stored on a transmission buffering region provided in a recording medium such as the RAM 800 - 14 , the hard disk drive 800 - 24 , the DVD-ROM 800 - 01 , or the IC card, and transmits the read transmission data to network 800 - 50 or writes reception data received from network 800 - 50 to a reception buffering region or the like provided on the recording medium.
  • the CPU 800 - 12 may cause all or a necessary portion of a file or a database to be read into the RAM 800 - 14 , the file or the database having been stored in an external recording medium such as the hard disk drive 800 - 24 , the DVD-ROM drive 800 - 26 (DVD-ROM 800 - 01 ), the IC card, etc., and perform various types of processing on the data on the RAM 800 - 14 .
  • the CPU 800 - 12 may then write back the processed data to the external recording medium.
  • the CPU 800 - 12 may perform various types of processing on the, data read from the RAM 800 - 14 , which includes various, types of operations, processing of information, condition judging, conditional branch, unconditional branch, search/replace of information, etc., as described throughout this disclosure and designated by an instruction sequence of programs, and writes the result back to the RAM 800 - 14 .
  • the CPU 800 - 12 may search for information in a file, a database, etc., in the recording medium.
  • the CPU 800 - 12 may search for an entry matching the condition whose attribute value of the first attribute is designated, from among the plurality of entries, and reads the attribute value of the second attribute stored in the entry, thereby obtaining the attribute value of the second attribute associated with the first attribute satisfying the predetermined condition.
  • the above-explained program or software modules may be stored in the computer readable media on or near the computer 800 .
  • a recording medium such as a hard disk or a RAM provided in a server system connected dedicated communication network or the Internet can be used as the computer readable media, thereby providing the program to the computer 800 via the network.
  • a model in accordance with the present invention can be used for a myriad of applications including, but not limited to, classification, recognition (e.g., speech recognition, speaker recognition, pattern recognition, etc.), robotics (e.g., robotic control, robotic decision making), machine control (e.g., controlling a movement of a machine (e.g., an assembly line machine), or powering down the machine, or changing the operational state of the machine (e.g., faster, slower, on, off low-power, etc.) based on an output of the model, and so forth.
  • classification systems e.g., speech recognition systems, speaker recognition systems, pattern recognition systems, etc.
  • machine control systems or machine controllers
  • the present invention may be a system, a method, and/or a computer program product.
  • the computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
  • the computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device.
  • the computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing.
  • a non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing.
  • RAM random access memory
  • ROM read-only memory
  • EPROM or Flash memory erasable programmable read-only memory
  • SRAM static random access memory
  • CD-ROM compact disc read-only memory
  • DVD digital versatile disk
  • memory stick a floppy disk
  • a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon
  • a computer readable storage medium is not to be construed as being transitory signals per se, such as radio eaves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire,
  • Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network.
  • the network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers.
  • a network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
  • Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages.
  • the computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the users computer and partly on a remote computer or entirely on the remote computer or server.
  • the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider).
  • electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute Me computer readable program instructions by utilizing state information of the computer readable program instructions to individualize the electronic circuitry, in order to perform aspects of the present invention.
  • These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
  • the computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s).
  • the functions noted in the block may occur out of the order noted in the figures.
  • two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes, be executed in the reverse order, depending upon the functionality involved.

Abstract

A computer-implemented method and an apparatus are provided for learning a first model. The method includes generating a second model based on the first model. The first model is configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and that are from a first input data sequence. The second model is configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence. The method further includes performing a learning process using both the first model and the second model. The method also includes storing the first model that has been learned.

Description

    BACKGROUND Technical Field
  • The present invention generally relates to predictive models, and more particularly to a learning apparatus and method for bidirectional learning of a predictive model based on a data sequence.
  • Description of Related Art
  • Conventionally, a neural network, Boltzmann machine, and the like are known that can be applied to time-series data. Furthermore, a dynamic Boltzmann machine is known that can learn a model corresponding to data input in time series through machine learning.
  • Such a Boltzmann machine or the like is desired to have improved expressive power, learning power, and the like by accurately learning a predictive model based on a data sequence, in order to be applied to various, fields such as moving images, languages, signal waveforms, and music.
  • SUMMARY
  • According to an aspect of the present invention, a computer-implemented method is provided for learning a first model. The method includes generating, by a processor, a second model based on the first model. The first model is configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and that are from a first input data sequence. The second model is configured to learn a first learning, target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence. The method further includes performing, by the processor, a learning process using both the first model and the second model. The method also includes storing, in a memory device, the first model that has been learned.
  • According to another aspect of the present invention, a computer program product is provided for predictive model learning. The computer program product includes a non-transitory computer readable storage medium having program instructions embodied therewith. The program instructions are executable by a computer to cause the computer to perform a method. The method includes generating, by a processor, a second model based on a first model. The first model is configured to perform a learning process based on sequentially inputting, each of a plurality of pieces of input data that include a plurality of input values and are from a first input data sequence. The second model is configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence. The method farther includes performing, by the processor, a learning process using both the first model and the second model. The method also includes storing, in a memory device, the first model that has been learned.
  • According to yet another aspect of the present invention, a learning apparatus is provided for learning a model corresponding to input data. The learning apparatus includes a processor. The learning apparatus further includes one or more computer readable mediums collectively including instructions that, when executed by the processor, cause the processor to generate a second model based on a first model. The first model is configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and are from a first input data sequence. The second model is configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence. The one or more computer readable mediums collectively including instructions that, when executed by the processor, further cause the processor to perform a learning process using both the first model and the second model. The one or more computer readable mediums collectively including instructions that, when executed by the processor, further cause the processor to store the first model that has been learned.
  • These and other features and advantages will become apparent from the following detailed description of illustrative embodiments thereof, which is to be read in connection with the accompanying drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The following description will provide details of preferred embodiments with reference to the following figures wherein:
  • FIG. 1 shows an exemplary configuration of a learning apparatus 100, according to an embodiment of the present invention;
  • FIG. 2 shows an exemplary configuration of a model 10, according to an embodiment of the present invention;
  • FIG. 3 shows a flow of an operation of the learning apparatus 100, according to an embodiment of the present invention;
  • FIG. 4 shows an example of structures of time-series data and training data for use in learning, according to an embodiment of the present invention;
  • FIG. 5 shows a first modification of the learning apparatus 100 according to an embodiment of the present invention;
  • FIG. 6 shows a modification of the model 10, according to an embodiment of the present invention;
  • FIG. 7 shows an example of a temporal change in the update parameter γi,l, according to an embodiment of the present invention;
  • FIG. 8 shows an example of a temporal change in the update parameter αi,j,k, according; to an embodiment of the present invention;
  • FIG. 9 shows an exemplary schematic configuration of the first model, according to an embodiment of the present invention;
  • FIG. 10 shows an exemplary schematic configuration of the second model, according to an embodiment of the present invention;
  • FIG. 11 shows an operational flow of the learning process using the first model and the second model performed by the learning apparatus 100, according to an embodiment of the present invention; and
  • FIG. 12 shows an exemplary hardware configuration of a computer, according to an embodiment of the present invention.
  • DETAILED DESCRIPTION
  • Hereinafter, some embodiments of the present invention will be described. The embodiments do not limit the invention according to the claims, and all the combinations of the features described in the embodiments are not necessarily essential to means provided by aspects of the invention.
  • FIG. 1 shows an exemplary configuration of a learning apparatus 100 according to an embodiment. The learning apparatus 100 may be an apparatus for learning a model 10 corresponding to time-series input data. The learning apparatus 100 may be operable to learn a model based on a Boltzmann machine by supplying time-series data to nodes of the model. The learning apparatus 100 includes an acquiring section 110, a supplying section 120, a storage section 130, a computing section 140, and a learning processing section 150.
  • The acquiring section 110 may be operable to acquire time-series input data. Time-series input data may be, for example, a data sequence in which a plurality of pieces of data are arranged along a time axis, such as moving image data. The acquiring section 110 may be connected to a device operated by a user or a device (sensor) that detects and outputs time-series data and may acquire the time-series input data from such a device. Alternatively, the acquiring section 110 may read and acquire time-series input data stored in a storage device in a predetermined format. Alternatively, the acquiring section 110 may he connected to a network and acquire time-series input data via the network. The acquiring section 110 may also store the acquired time-series input data in a storage device included in the learning apparatus 100.
  • The supplying section 120 may be operable to supply a plurality of input values corresponding to input data at one time point in the time-series input data to a plurality of nodes of a model. The supplying section 120 is connected to the acquiring section 110 and may handle, as training data input data at one time point in the received time-series input data and supply input values at the one time point to corresponding nodes of the model, input data at one time point may be the temporally newest data in a training data set for use in learning. Alternatively, input data at one time point may be temporally intermediate data in a training data set for use in learning. That is, input data at one time point may be selected arbitrarily from the time-series data.
  • The storage section 130 may be operable to store values of hidden nodes of the model in correspondence with a plurality of time points in the time series input data. The storage section 130 may sample the values of hidden nodes corresponding to one time: point and store these values in the hidden nodes corresponding to this time point. The storage section 130 may store the sampled values respectively in the hidden nodes corresponding to the time points for each time point.
  • The computing section 140 may be operational to compute a conditional probability of each input value at one time point on a condition that an input data sequence has occurred. Here, in the time-series input data, the pieces of data input to the respective nodes of the model at one time point are referred to as input values at one time point, and the pieces of data input to the respective nodes at the time points before the one time point are referred to as the input data sequence. A model used by the learning apparatus 100 may have a weight parameter between (i) a plurality of hidden nodes and a plurality of input values corresponding to input data at each time point prior to the one time point in an input data sequence and (ii) a plurality of hidden nodes corresponding to the one time point and a plurality of input nodes.
  • The computing section 140 may be operable to compute a conditional probability of each input value at one time point, on the basis of an input data sequence before the one time point in the time-series input data, the stored values of hidden nodes, and the weight parameter of the model. Furthermore, the computing section 140 may be operable to compute a conditional probability of the value of each hidden node at one time point on a condition that an input data sequences has occurred, based on an, input data sequence before the one time point in the time-series input data and the weight parameter of the model.
  • The learning processing section 150 may be operable to increase a conditional probability of input data at one time point occurring on a condition that the input data sequence has occurred, by adjusting the weight parameter of the model. The learning processing section 150 may further adjust bias parameters which are given respectively to the plurality of nodes and hidden nodes of the model. The learning processing section 150 may supply the adjusted weight parameter and bias parameters of the model to a storage device, such as an external database 1000, to store these parameters in the storage device.
  • The above-described learning apparatus 100 according to the present embodiment may be operable to learn the model by adjusting the weight parameter and bias parameters of the model, based on input data at one time point in the time-series input data. The model according to the present embodiment is described with reference to FIG. 2.
  • FIG. 2 shows an exemplary configuration of a model 10 according to the present embodiment. The model 10 includes a plurality of common layers 12. FIG, 2 shows an example including a total of T common layers 12. The model 10 may include a finite number of common layers 12. Each common layer 12 includes an input layer 14 and a hidden layer 16.
  • Each input layer 14 may be a layer corresponding to the time-series data. Each input layer 14 may correspond to a respective time point in the time-series data. Each input layer 14 may include a predetermined number of nodes. For example, the 0-th input layer may be a layer corresponding to input data at one time point in the time-series data. The 0-th input layer may include a plurality of nodes corresponding to the number of input values in this input data.
  • A total of T−1 input layers 14 other than the 0-th input layer 14 among the plurality of input layers may he input layers 14 corresponding to the input data sequence before the one time point in the time-series, input data. For example, the −1st input layer may correspond to input data at a time point that temporally precedes the one time point by one time point, and the (−δ)-th input: layer may correspond to input data at a time point that temporally precedes the one time point by δ time points. That is, a total of T−1 input layers other than the 0-th input layer each have the same number of nodes as the 0-th input layer and are respectively supplied with input values of corresponding input data values in the input data sequence, for example.
  • Each hidden layer 16 may correspond to a respective time point in the time-series data. For example, the 0-th hidden layer may be a layer corresponding to one time point in the time-series data. FIG. 2 shows an example including, a total of T hidden layers 16. Each hidden layer 16 may include one or more hidden nodes, and the storage section 130 may store the values sampled at the one time point.
  • A total of T−1 hidden layers other than the 0-th hidden layer among the plurality of hidden layers 16 may be hidden layers 16 corresponding to time points before the one time point in the time-series data. For example, the −1st hidden layer corresponds to a time point that temporally precedes the input data of the one time point by one time point, and the storage section 130 stores the values sampled at the time point that temporally precedes the one time point by one time point. Furthermore, the (−δ)-th hidden layer may correspond to a time point that temporally precedes the input data of the one time point by δ time points, and the storage section 130 may store the values sampled at the time point that temporally precedes the one time point by δ time points. That is, a total of T−1 hidden layers other than the 0-th hidden layer each have the same number of nodes as the 0-th hidden layer and are respectively supplied with values of corresponding hidden nodes, for example.
  • As an example, in the case where the time-series input data is moving image data, the last image data of the moving image data corresponds to the 0-th input layer, and a plurality of nodes of the 0-th input layer each receive corresponding pixel data of the image data. Furthermore, the 0-th hidden layer corresponds to the final time point of the moving image data, and the storage section 130 may store values sampled at this final time point in the hidden nodes of the 0-th hidden layer.
  • In addition, the −1st input layer is supplied with image data that immediately precedes the last image data, and <a plurality of nodes of the −1st input layer each receive corresponding pixel data of the immediately preceding image data. Furthermore, the −1st hidden layer corresponds to the time point that immediately precedes the final time point, and for each of the plurality of nodes of the −1st hidden layer, the storage section 130 may store the values sampled at this immediately preceding time point. Similarly, the plurality of nodes of the (−δ)-th input layer each receive corresponding pixel data of image data that precedes the last image data by δ images, and the plurality of nodes of the (−δ)-th hidden layer each store corresponding sampling values at the time point that precedes the last time point by δ time points.
  • FIG. 2 shows an example in which each common layer 12 includes an input layer 14 and a hidden layer 16, but instead, one or more common layers 12 need not include a hidden layer 16. In such a case, the 0-th common layer to the (−m)-th common layer include input layers 14 and hidden layers 16, and the (−m−1)-th common layer to (−T+1)-th common layer may include input layers 14.
  • The plurality of nodes in the 0-th input layer and/or the plurality of hidden nodes in the 0-th hidden layer may each have a bias parameter. For example, the j-th node j in the common layer 12 has a bias parameter bj.
  • The plurality of nodes in the 0-th input layer and the nodes of the hidden layer corresponding to the input data sequence and layers corresponding to the input data sequence before the one time point may respectively have weight parameters there between. There need not be weight parameters between the plurality of nodes in each input layer 14 and hidden layer 16.
  • Similarly, the plurality of nodes in the 0-th hidden layer and the nodes of the hidden layer corresponding to the input data sequence and layers corresponding to the input data sequence before the one time point may respectively have weight parameters there between. That is, the plurality of nodes of the 0-th common layer and the nodes of the plurality of common layers before the one time point may respectively have weight parameters there between.
  • FIG. 2 shows a concept of a weight parameter Wij [δ] between the node j of the 0-th input layer and a node i of the (−δ)-th layer. FIG. 2 shows an example in which the model 10 has the same number of input layers 14 and layers 16, each input layer 14 includes I nodes, and each hidden layer 16 includes H hidden nodes. In the present embodiment, the input layers 14 and hidden layers 16 are expressed by one common layer 12 that has a plurality of nodes xj [t]. The first to I-th nodes (1≤j≤1) of the common layer 12 indicate the nodes of the input layer 14, and the (I+1)-th to (I+H)-th nodes (I+1, j, I+H) indicate hidden nodes.
  • For example, the weight parameter Wij [δ] shown in FIG. 2 indicates a weight parameter between two nodes of an input layer 14. Similarly, the weight parameter Wij [δ] may include weight parameters from an input node to a hidden node, from a hidden node to an input node, and between to hidden nodes. A specific example of a weight parameter Wij [δ] is as shown in the Expression below. In the present embodiment, a weight parameter from an input node or hidden node to an input node is referred to as a “weight parameter to an input node,” and a weight parameter from an input node or hidden node to a hidden node is referred to as a ‘weight parameter to a hidden node.”
  • W i j [ δ ] = W ^ i j [ δ ] + W ^ i j [ - δ ] W ^ i j [ δ ] = { k K 0 u i , j , k λ k δ - d i j if δ = 0 if δ d i j l L - v i , j , l μ k - δ otherwise Expression 1
  • Here, ui, j, k and vi, j, l are learning parameters that are learning targets, for example. Furthermore, λk t1 and μl t2 are predefined parameters that change in a predetermined manner in accordance with a time point difference >between the hidden nodes and input data in the input data sequence before the one time point and the hidden nodes and input data at the one time point (t1=δ−dij, t2=−δ). That is, the weight parameter Wij [δ] may be a parameter based on the learning parameters ui, j, k and vi, j, l the predefined parameters λk t1 and μl t2.
  • The weight parameter Wij [δ] may be a parameter based on a positive value, which is based on a product of the first learning parameter ui, j, k and the first predefined parameter λk t1, and a negative value, which is based on a product of the second learning parameter vi, j, l and a second predefined parameter μl t2. Specifically, in the case where the time point difference δ is greater than or equal to a predetermined delay constant dij, the weight parameter Wij [δ] may be a positive value based on a product of the first learning parameter ui, j, k and the first, predefined parameter λk t1. In the case where the time point difference δ is less than the delay constant dij and is not equal to 0, the weight parameter Wij [δ] may be a negative value based cm a product of the second learning parameter vi, j, l and the second predefined parameter μl t2. In addition, in the case where the time point difference δ is equal to 0, the weight parameter Wij [δ] may be equal to 0.
  • In addition, in the case where the time point difference δ is greater than or equal to the predetermined delay constant dij, the weight parameter Wij [δ] may be, based on a plurality of positive values that are based on the products ui, j, k·λk t1 of a plurality of sets of the first learning parameter ui, j, k and the first predefined parameter λk t1 respectively from among the plurality of first learning parameters i, j, k and the plurality of first predefined parameters λk t1. In addition, in the case where the time point difference δ is less than the predetermined delay constant dij and is not equal to 0, the weight parameter Wij [δ] may be based on a plurality of negative values that are based on products vi, j, l·μ1 t2 of a plurality of sets of the second learning parameter vi, j, l and the second predefined parameter μl t2 respectively from among the plurality of second learning parameters vi, j, l and the plurality of second predefined parameters μl t2.
  • A predefined parameter may be a parameter based on a value obtained by raising a predetermined constant to the power of a value based on the time point difference δ. The first predefined parameter λk t1 is a parameter whose value gradually decreases as the time point difference δ increases, for example. In this case, the first predefined parameter λk t1 may be a value obtained by raising a first constant λk, which is greater than 0 and less than 1, to the power of a value obtained by subtracting the predetermined delay constant dij from the time point difference δ δ−dij=t1). In addition, the second predefined parameter μl t2 may be a parameter whose value gradually decreases as the time point difference δ increases, for example. In this case, the second predefined parameter μl t2 may be a value obtained by raising a second constant μl, which is greater than 0 and less than 1, to the power of a negative value of the time point difference δ (−δ=t2).
  • The above-described model 10 according to the present embodiment may be operable to form a Boltzmann machine. That is, the model 10 may be a Boltzmann machine to which time-series data is applied. The model 10 may be a Boltzmann machine that includes hidden layers into which are, input values differing from the time-series data, in addition to the input layers into which the time-series data is input. The learning apparatus 100 according to the embodiment learns the model 10 by adjusting the learning parameters ui, j, k and vi, j, l and the bias parameter bj while sampling and storing the values of the hidden nodes, by using, as training data, input data at one time point that is supplied to the 0-th input layer of the model 10. A learning operation of the learning apparatus 100 is described with reference to FIG. 3.
  • FIG. 3 shows a flow of an operation of the learning apparatus 100 according to the present embodiment. In the present embodiment, the learning apparatus 100 may be operable to learn the model 10 corresponding to time-series input data and determine the learning parameters ui, j, k and vi, j, l and the bias parameter bj, by executing the processing steps of S310 to S360. In the present embodiment, first, an example is described in which the determination of the weight parameters to the hidden nodes and the weight parameters to the input nodes is perforated by the learning apparatus 100 using substantially the same operation.
  • First, the acquiring section 110 may acquire time-series data (S310). The acquiring section 110 may acquire time-series data of a duration equivalent to a total of T layers from the 0-th layer to the (−T+1)-th layer of the model 10. The acquiring section 110 acquires, for example, T pieces of image data in time-series that form the moving image data.
  • Then, the supplying section 120 may supply a plurality of input values corresponding to the input data of the time-series input data at one time point to the plurality of input nodes of the 0-th input layer of the model 10 (S320). Here, x[1,I] [0] (=xj [0], 1≤j≤I) denotes input data supplied to the 0-th input layer.
  • The supplying section 120 supplies, for example, I input values xj [0] corresponding to input data x[1,I] [0] of the time-series input data at the most recent time point to the corresponding nodes j of the 0-th input layer (1≤j≤1). For example, the supplying section 120 supplies I pieces of pixel data included in the last piece of image data of T pieces of image data arranged in time series to form the moving image data to I nodes of the 0-th input layer. The supplying section 120 may supply a value of 1 or 0 as the pixel data to each node of the 0-th input layer. If the duration of the time-series input data is shorter than T, the supplying section 120 may supply the data to a number of layers from the 0-th input layer corresponding to the length of the time series, and may supply a value of 0, for example, to the nodes of the rest of the layers.
  • Then, the supplying section 120 may supply a plurality of input values corresponding to the input data sequence before the one time point to the plurality of nodes included in respective layers from the −1st input layer to the (−T+1)-th input layer of the model 10. Here, let xj (−T, −1] denote input data, supplied to layers from the −1st input layer to the (−T+1)-th input layer (1≤j≤1). The term (−T, −1] indicates layers from the (−T+1)-th layer to the −1st layer. That is, the input data xj (−T, −1] in the tune-series data denotes a history up to the input data xj [0], for example.
  • Next, the storage section 130 samples the values of a plurality of hidden nodes corresponding to the one time point, and respectively stores these values in the corresponding plurality of hidden nodes of the 0-th hidden layer (S330). The storage section 130 may arbitrarily or randomly input values of 1 or 0. The storage section 130 stores sampled values in the corresponding hidden nodes j of the 0-th hidden layer, for example (I+1≤j≤I+H).
  • The storage section 130 may store the values of the hidden nodes before the one time point respectively in a plurality of nodes in each of the corresponding hidden layers from the −1st hidden layer to the (−T+1)-th hidden layer of the model 10. Here, let, let xj (−T, −1] denote the values of the hidden nodes stored in the layers from the −1st hidden layer to the (−T+1)-th hidden layer (I+1≤j≤I+H). That is, the values xj (−T, −1] input to the nodes of each common layer 12 before the one time point denote a history up to the input values xj [0] input to the nodes of the 0-th common layer, for example (1≤j≤I+H).
  • Then, the computing section 140 may compute conditional probabilities of each input value xj [0] (1≤j≤I) of an input node at the one time point, based on the input values xj (−T, −1] (1≤j≤I+H) of the plurality of nodes of the (−T+1) common layer to the −1st common layer and the weight parameter Wij [δ] (S340). The computing section 140 computes a probability <xj [0]>θ of the input value xj [0] (1≤j≤I) of the j-th node of the 0-th input layer being equal to 1 by substituting 1 for xj [0] in the following expression, based on the history x(−T, −1] (1≤j≤I+H) of the plurality of nodes of the common layer 12.
  • X j [ 0 ] θ = p θ , j ( x j [ 0 ] | x ( - T , - 1 ] ) = exp ( τ - 1 E θ , j ( x j [ 0 ] | x ( - T , - 1 ] ) ) 1 + exp ( τ - 1 E θ , j ( x j [ 0 ] | x ( - T , - 1 ] ) ) Expression 2
  • In the present embodiment, an example is described in which the input value xj [0] of each node is binary, i.e. 1 or 0, but the value of the input value xj [0] is not limited to these values. Furthermore, in the step thy computing the conditional probabilities of each input value xj [0] (1≤j≤I) of the 0-th input layer, the computing section 140 may compute the conditional probabilities of the values xj [0] (I+1≤j≤I+H) of the 0-th hidden, layer.
  • Expression 2 is derived as a Boltzmann machine from a known probability formula. For example, θ denotes a set of parameters to be computed, and the formula θ=(bj, ui, j, k, vi, j, l) is established. In addition, τ may be a parameter that is dependent on a known “system temperature” of the Boltzmann machine, and may be preset by a user or the like. Also, Eθ,j(xj [0]|x−(−T,−1]) of Expression 2 is computed by using the following expression.
  • E θ , j ( x j [ 0 ] | x ( - T , - 1 ] ) = - b j x j [ 0 ] - t = - T - 1 ( x [ t ] ) T ( x [ t ] ) W : , j [ - t ] x j [ 0 ] Expression 3
  • Here, “T” denotes a transpose, “:” denotes 1 to n in a case where n (=I+H) denotes the number of nodes, and “:.j” indicates extraction of the j-th column, That is, the second term on the right side of Expression 3 is denoted by the following expression, for example.
  • t = - T - 1 ( x [ t ] ) T W ; , j [ - t ] x j [ 0 ] = ( x 1 [ t ] , x 2 [ t ] , . . . , x n [ t ] ) ( W 1 , j [ - t ] W 2 , j [ - t ] W n , j [ - t ] ) Expresssion 4 = i = 1 N ( k K u i , j , k α i , j , k - l L v i , j , l β i , j , l - l L v j , i , l γ i , l ) x j [ 0 ]
  • Here, αi, j, k, βi, j, l, and γi,1 are denoted by the following expressions..
  • α i , j , k t = - T - d i j λ k - t - d i j x i [ t ] Expression 5 β i , j , l t = - d i j + 1 - 1 μ l t x i [ t ] Expression 6 γ i , l t = - T - 1 μ l - t x i [ t ] Expression 7
  • Accordingly, Pθ,j(1|xj (−T, −1]) obtained by substituting 1 for xj [0] of Expression 2 can be computed from Expression 3 by substituting 1 for xj [0] in expression 5. Note that predetermined initial values (for example, 0) may be substituted for the parameter set θ=(bj, ui, j, k, vi, j, l). In this way, the computing section 140 can compute a conditional probability <xj [0]>θ of each input value xj [0] at the one time point which is denoted by Expression 2.
  • Then, the learning processing section 150 may adjust the parameter set, θ=(bj, ui, j, k, vi, j, l) (S350). When adjusting the bias parameter the learning processing section 150 may determine a direction of the change in the bias parameter b, by using the following expression.
  • b j log P θ ( x [ 1 , l ] [ 0 ] | x ( - T , - 1 ] ) = τ - 1 ( x j [ 0 ] - X j [ 0 ] θ ) Expression 8
  • Here, xj [0] on the right side of Expression 8 denotes an input value supplied as training data by the supplying section 120, and <xj [0]>θ on the right side denotes a probability computed by using Expression 2 (1≤j≤I). The bias parameter bj for each input node (1≤j≤I) may be adjusted and updated as denoted by the following expression by using Expression 8. Note that a coefficient c is a parameter predetermined by the user or the like.
  • b j b j + c 1 τ ( x j [ 0 ] - x j [ 0 ] θ ) Expression 9
  • That is, the learning processing section 150 adjusts the bias parameter bj so as to increase the conditional probability of the input value xj [0] of the node of the 0-th input layer occurring, on a condition that the history x(−T, −1] of the common layer 12 has occurred. The learning processing section 150 may iteratively perform updating of the bias parameter bj denoted by Expression 9 and computing of the probability <xj [0]>θ denoted by Expression 2, to determine the bias parameter bj. The learning processing section 150 stops updating the bias parameter bj and determines the bias parameter b if a difference in the bias parameter bj before and after updating is less than or equal to a predetermined threshold. If a bias parameter bj is also set for a hidden node, the learning processing section 150 may determine the bias parameter bj of the hidden node in the same manner.
  • Alternatively, the learning processing section 150 may decide upon the bias parameter bj by iteratively updating the bias parameter bj a predetermined number of times. If a difference in the bias parameter bj before and after updating is greater than or equal to the predetermined threshold even after the bias parameter bj has been updated the predetermined number of times, the learning processing section 150 may stop updating the bias parameter bj and inform the user that the parameter does not converge.
  • Similarly, when updating the learning parameter ui, j, k, the learning processing section 150 may determine the direction of a change in the learning parameter uu, j, k by using the following expression.
  • u i , j , k log P θ ( x [ 1 , I ] [ 0 ] x ( - T , - 1 ] ) = τ - 1 α i , j , k ( x j [ 0 ] - X j [ 0 ] θ ) Expression 10
  • In addition, when updating the, learning parameter vi, j, l, the learning processing section 150 may determine the direction of a change in the learning parameter ui, j, k using the following expression.
  • v i , j , l log P θ ( x [ 1 , I ] [ 0 ] x ( - T , - 1 ] ) = - τ - 1 β i , j , l ( x j [ 0 ] - X j [ 0 ] θ ) - τ - 1 γ j , l ( x i [ 0 ] - X i [ 0 ] θ ) Expression 11
  • In the same manner as the updating of the bias parameter bj, the learning processing section 150 may iteratively perform updating of the learning parameters ui, j, k and vi, j, l corresponding to the input nodes (1≤j≤I) and computing of the probability <xj [0]>θ to determine the learning parameters ui, j, k and vi, j, l corresponding to the input nodes (1≤j≤I). Alternatively, the learning processing section 150 may iteratively perform an operation for updating the parameter set θ=(bj, ui, j, k, vi, j, l) and then computing the probability <xj [0]>0 denoted by Expression 2 to determine the parameter set θ=(bj, ui, j, k, vi, j, l).
  • As described above, the learning processing section 150 according to the present embodiment can decide upon the learning parameters ui, j, k and vi, j, l and the bias parameter bj through learning. The learning apparatus 100 may then determine whether to continue learning (S360). The learning apparatus 100 may continue learning until it performs the learning process a predetermined number of times, or may continue learning until a stop command is input by the user. Alternatively, the learning apparatus 100 may continue learning until it can no longer acquire time-series data.
  • If the learning apparatus 100 continues learning (S360: YES), the process may return to step S310, in which the acquiring section 110 acquires the next time-series data, and the learning apparatus 100 may then perform leaning of the model 10 based on the next time-series data. For example, the supplying section 120 supplies the 0-th input layer with the next image data in the image data acquired by the acquiring section 110. Furthermore, the storage section 130 samples the values of the hidden layers and stores these values in the 0-th hidden layer. Then, values held in the t-th common layer (−T<t<0) may be supplied to the (t−1)-th common layer. The values held in the (−T+1)-th layer may be deleted. The learning apparatus 100 may perform learning by using image data supplied to the layers from the 0-th input layer to the (−T+1)-th input layer as training data and using the values stored in the layers from the 0-th hidden layer to the (−T+1)-th hidden layer.
  • In this way, the supplying section 120 and the storage section 130 may sequentially acquire new input values xj [0] at the next time point corresponding to;the nodes of the 0-th common layer in the model 10. Then, the computing section 140 may compute a conditional probability <xj [0]>θ of the new input value xj [0] on a condition that the history has occurred for each common layer before the next time point The learning processing section 150 may adjust the weight parameter so as to increase the conditional probability of the new input value occurring on the condition that this history has occurred,
  • If the learning processing section 150 stops learning (S360: NO), the learning processing section 150 may output the learning parameters ui, j, k and vi, j, l and the bias parameter bj that have been determined and store the parameters in the external database 1000 or the like.
  • As described above, the learning apparatus 100 according to the present embodiment may be operable to apply, to time-series input data that is input in time series, a model having a total of T layers by associating one time point with the 0-th common layer and an input data sequence before the one time, point with T−1 layers. The learning apparatus 100 may be operable to apply a model having hidden nodes to each common layer 12. That is, the learning apparatus 100 may be operable to form a time-evolution Boltzmann machine that predicts input data at one time point on the basis of the input data sequence and hidden node values
  • The learning apparatus 100 is able to learn the model by computing a conditional probability of the input value xj [0] at the one time point occurring, based on the input value x(−T, −1], which is a history, for a model that takes time evolution into consideration. Furthermore, since the learning apparatus 100 learns the model using hidden nodes in addition to the time-series input data, the expressive ability and learning ability can be improved.
  • A description has been given of the learning apparatus 100 according to the present embodiment that sequentially acquires new input data from time-series input data and adjusts the weight parameter for each input data acquired. Instead of this configuration, the learning apparatus 100 may acquire time-series input data of a predetermined duration and then adjust the weight parameters. For example, the learning processing section 150 adjusts the weight parameters collectively for a plurality of tune points in response to acquisition of new input data at a plurality of time points corresponding to D) layers.
  • FIG. 4 shows an example of structures of time-series data and training data for use in learning in the present embodiment. In FIG. 4, the horizontal axis denotes time. FIG. 4 shows an example in which the learning apparatus 100 uses time-series data y[1, L] having a duration L that is longer than a duration T of time-series data y[1, T] used as training data by the learning apparatus 100 during learning. In this case, the learning processing section 150 may be operable to adjust weight parameters for a plurality of time points all together, in response to input data at a plurality of time points being newly acquired.
  • The learning apparatus 100 first performs learning using, as first training data, a time-series data segment of the time-series data from a time 1 to a time T. In this case, as described in FIG. 3, the learning apparatus 100 may perform learning by setting the time-series data and corresponding, hidden layer values from the time 1 to the time T as each input value of the common layer 12 at the one time point in order, and incrementally shifting the time points one at a time toward the future. The learning apparatus 100 may use data at a time T as each input value xj [0] at the one time point, and continue learning until the time-series data from the time 1 to a time T−1 becomes the input data sequence x(−T, −1] (i.e. the history).
  • Next, the learning apparatus 100 performs learning using, as second training data, a time-series data segment of the time-series data from a time 2 to a time T+1. The learning apparatus 100 may sequentially use each of D pieces of data in the second training data as the input value xj [0] at the one time point. In this case, the learning apparatus 100 may shift the time point in the interval from the time 2 to the time T+1 one time point at a time toward the future and use, as the history, the corresponding time-series data and hidden nodes of the inter al from the, time 2 to the time T. In this way, the learning apparatus 100 may adjust the parameters D times for the D input values xj [0] and the corresponding D histories. That is, the learning apparatus 100 may use a stochastic gradient technique>in which the learning method described with Expressions 8 to 11 is performed.
  • Alternatively, the learning apparatus 100 may acquire D time-series data sets, generate a plurality of training data sets from time-sequence data segments of a duration of L, and collectively perform learning for D layers. Specifically, the learning apparatus 100 may perform the stochastic gradient technique described using Expressions 8 to 11 collectively for D layers, by using the following expression.
  • θ θ + η x { y ( t , t + T - 1 ] 0 t < D } θ log P θ ( x [ 1 , I ] [ 0 ] x ( - T , - 1 ] ) Expression 12
  • FIG. 5 shows a first modification of the learning apparatus 100 according to the present embodiment. Components of the learning apparatus 100 shown in FIG. 5 that perform substantially the same operations as those of the learning apparatus 100 according to the embodiment illustrated in FIG. 1 are denoted by the same reference numerals, and a description thereof is omitted. In a case where time-series data of a duration L such as described in FIG. 4 is provided, the learning apparatus 100 according to the present modification may be operable to efficiently update parameters by using FIFO memories and learn a model corresponding to the time-series input data. The learning apparatus 100 according to the present modification further includes FIFO memories 160 and an updating section 170.
  • Each of the FIFO memories 160 may sequentially store input data and output the stored data after'a predetermined number of storages have been performed. Each of the FIFO memories 160 may be a memory that first outputs data that has been stored first (FIFO: First in, First Out).
  • Each of the FIFO memories 160 may sequentially store an input value of the common layer 12 and output the input value after a predetermined number of storages have been performed. The learning apparatus 100 may include a plurality of FIFO memories 160, the number of which is greater than or equal to the number of nodes n of the model. The plurality of FIFO memories 160 is desirably provided to have a one-to-one correspondence with the plurality of nodes of the common layer 12. That is, each of the plurality of FIFO memories 160 may be provided in a manner to store a history for a respective node of the common layer 12 or to update the history thereof.
  • The plurality of FIFO memories 160 are connected, to the acquiring section 110 and the storage section 130, and sequentially store input values corresponding to new input data of the common layer 12. The plurality of FIFO memories 160 are also connected to the updating section 170 and sequentially supply the data stored therein to the updating section 170. [008.2] The updating section 170 may he operable to update a plurality of update parameters that are based on the hidden nodes and the input data sequence of the time-series input data before the one time point, from values at a prior time point to values at the one time point, on the basis of values of the update parameters and values of the hidden nodes and input values corresponding to the input data to be reflected next. The updating section 170 may update the update parameters by using values input to the FIFO memories 160 and values output from the FIFO memories 160. The updating section 170 may be connected to the acquiring section 110 and the storage section 130, and may receive values input to the FIFO memories 160. Alternatively, the updating section 170 may receive values input to the FIFO memories 160 from the acquiring section 110 via the supplying section 120.
  • Here, the update parameters are αi, j, k and γi,l shown in Expressions 5 and 7. In this case, the update parameters are based on input values i (1≤i≤I) corresponding to input data of the input data sequence at each time point and the predefined parameters λk t1 and μl t2 of the weight parameter Wij [δ] between this input value i and the target input node j (1≤j≤I) or hidden node j (I+1≤j+I+H), for example. As another example, the update parameters are based on the hidden node i (I+1≤i≤I+H) at each time point and the predefined parameters and λk t1 and μl t2 of the weight parameter Wij [δ] between this hidden node i and the target input node j (1≤j≤I) or hidden node j (I+1≤j≤I+H), for example.
  • The update parameters may be updated every time the acquisition of the time-series input data by the acquiring section 110 and the storage of the sampling values by the storage section 130 are performed sequentially. The above-described learning apparatus 100 according to the present modification may be operable to learn a modification of the model 10. The modification of the model 10 is described with reference to FIG. 6.
  • FIG. 6 shows a modification of the model 10 according to the present embodiment. The model 10 according to the modification needs not have the layered structure including T layers shown in FIG. 2. FIG. 6 shows an example of a model corresponding to one of the FIFO memories 160. Accordingly, the overall configuration of the model 10 according to the present modification includes a storage area that is equivalent to the 0-th common layer in FIG. 2 including the training data, and a number of the configurations illustrated in FIG. 6 equal to the number of nodes n I+H). Neurons i and j and a FIFO sequence 20 of the model 10 according to the present modification are described below.
  • The neuron i may be equivalent to the'input terminal of the FIFO memory 160. An input value yi [t] (1≤i≤I) of each node in the input data of the input data sequence at each time point t and a corresponding value yi [t] the values yi [t] (I+1≤i≤i+H) of the hidden nodes at each time point are sequentially input to the neuron i. The neuron i may set the value yi [t] input thereto as the current input value. Then, at a time point t+1, the neuron i may supply the input value yi [t] input at the time point t to the updating section 170 and to the FIFO sequence 20 as the previous input value and may hold the input value yi [t+1] at the time point t+1 as the current input value.
  • The FIFO sequence 20 may store dij−1 of the latest input values received from the neuron i. The FIFO sequence 20 may supply the dij−1 input values stored therein to the updating section 170. The updating section 170 may be operable to compute the values of the update parameters denoted by Expression 6 by using the input values supplied by the FIFO sequence. If the FIFO sequence 20 holds input values from the time point t−1 to the time point t−dij−1, the FIFO sequence 20 is denoted by the following expression.

  • q i,j≡(y i [t−1], y i [t−d ij +2] , y i [t−d ij +1])   Expression 13:
  • After the input value is input to the neuron i at the time point t1, the FIFO sequence 20 may store the input value yi [t1] up until a time point t3 (=t1+dij−1) which is a predetermined time period dij−1 after the next time point t2 (=t1+1) of the time point t1. At the next time point t4 (=t3+I=t1+dij), the FIFO sequence 20 may supply the input value yi [t1] to the neuron j. The input value yi [t1] supplied to the neuron j at the time point t4 is immediately supplied to the updating section 170 at the time point t4. However, the input value yi [t1] that the neuron j has received from the FIFO sequence 20 at the time point t4 does not serve as an input for the neuron j, and the input value yi [t4] may be input to the neuron j at the time point t4.
  • The neuron j may be equivalent to the output terminal of the FIFO memory 160, and the neuron j may receive the input value yi [t1] input to the neuron i at the time point t1, via the FIFO sequence after the time period dij, i.e. at the time point t1+dij. That is, the model 10 from the neuron i to the neuron j via the FIFO sequence 20 may correspond to the FIFO memory 160 that, stores dij pieces of input data. In addition, the neuron i of the model 10 according to the modification may correspond to, for example, a node for an input data sequence such as a node i of the (−δ)-th common layer of the model 10 shown in FIG. 2, and in this case the neuron j may correspond to, for example, the node j of the 0-th common layer. At the time point t1+dij, the neuron j may supply the received input value y[t1] to the updating section 170.
  • As described above, the model 10 according to the present modification may supply the input values at the time point t−1 and the time point t−dij+1 to the updating section 170 at the time point tin: this way, the updating section 170 can update the update parameters by adding the corresponding, input value in the input data to be reflected next to the update parameters for the time point before the one time point, and then multiplying the resulting sum by a predetermined constant. Note that the update parameters denoted by Expression 8 may be computed in accordance with Expression 8 by using the input values stored in the FIFO sequence 20 that are supplied to the updating section 170.
  • For example, the update parameter γi,l denoted by Expression 7 can be updated by using the input values supplied to the updating section 170 and the second predefined parameter. Specifically, the updating section 170 can compute the update parameter γi,l to be used in the current learning by performing computing at the time point t according to the following expression by using, the prior update parameter γi,l and the input value yi [t−1] received from the neuron i at the time point t.

  • γi,l←μli,l +y i [t−1])   Expression 14:
  • FIG. 7 shows an example of a temporal change in the update parameter γi,l according to the present embodiment. FIG. 7 shows an example in which values greater than 0 (for example, 1) are input to the neuron i as the input value at time points t−5, t−2, and t−1, and these input values are supplied to the updating section 170 at time points t−4, t−1, and t. The second predefined parameter μl is a parameter whose value gradually decreases as the time point difference increases. Accordingly, the update parameter γi,l computed by the updating section 170 tends to decrease as time passes from when the input value of 1 is input to when the next input is given.
  • The update parameter αi, j, k denoted by Expression 5 can be updated by using the input values supplied to the updating section 170 and the first predefined parameter λk. Specifically, the updating section 170 can compute the update parameter αi, j, k to be used in the current learning by performing computing at the time point t according to the following expression, by using the prior update parameter αi, j, k and the input value yi [t−dij] from the neuron j at the time point t.

  • αi, j, k←λki, j, k +y i [t−d di ij ])   Expression 15:
  • FIG. 8 shows an example of a temporal change in the update parameter αi, j, k according to the present embodiment. FIG. 8 shows an example in which values greater than 0 (for example, 1) are supplied to the neuron j as the input value at time points t−3, t−1, and t. The first predefined parameter λk is a parameter whose value gradually decreases as the time point difference increases. Accordingly, the update parameter αi, j, k computed by the updating section 170 tends to decrease as time passes from when the input value of 1 is input to when the next input is given.
  • As described above, the learning apparatus 100 according to the present modification can update the update parameters αi, j, k and γi,l by applying the model 10 shown in FIG. 6 using the FIFO memories 160 and the updating section 170. Note that the updating section 170 can apply the model 10 according to the present modification, for example, by acquiring the input values xi [t−1] at the time point t−1 from the input data input to the FIFO memories 160 and acquiring the input values xi [t−dij] at the time point t−dij from the output of the FIFO memories 160.
  • In addition, the learning apparatus 100 may update the parameter βi,j, l through substantially the same operation as the operation described in FIG. 3. Specifically, the computing section 140 can compute the parameter βi,j, l by determining the sum of products of the second predefined parameter μl and the input value xi (yi in the present modification) for time points from t−1 to t−dij+1 as indicated by Expression 6.
  • In this way, the computing section 140 according to the present modification can compute, by using the plurality of update parameters, conditional probabilities of input data values at one time point on the condition that the hidden node values and input data sequence have occurred. Then, the learning processing section 150 can determine the learning parameters ui, j, k and vi, j, l and the bias parameter bj by performing substantially the same operation as the operation described in FIG. 3
  • In other words, the learning apparatus 100 according to the present embodiment can determine the weight parameter and bias parameters in a manner to increase the probability of predicting the input value to be input to the input layer 14 of the common layer 12, based on the past values that have been input to the common layer 12 of the model 10 before the one time point. Furthermore, the learning apparatus 100 can improve the prediction accuracy, the expressive ability, the learning efficiency, and the like of the input values input to the input layer 14 by having the common layer 12 include the hidden layer 16 in addition to the input layer 14.
  • The learning apparatus 100 according to the present embodiment described above is an example in which a value that is unrelated to the prediction made by the learning apparatus 100 is sampled and input as the hidden node value to be input to the hidden layer 16. Instead of this, the learning apparatus 100 may determine the hidden node value by using a history of the conditional probability of the values of the nodes of the common layer 12. The learning apparatus 100 may determine the weight parameter to a hidden node by using this conditional probability history. The learning apparatus 100 can improve the prediction accuracy by using the conditional probability history of nodes of the common layer 12 to determine the weight parameter to the hidden node and the hidden node value.
  • In this case, the computing section 140 may compute the conditional probability pj,t of the value of a node j of the common layer 12 at one time point t based on the values input to the corresponding node j of the common layer 12 at each time point before the one time point t, and store this conditional probability in the storage section or the like. In addition to the computation of the conditional probability of each input value of the input layer 14 at the one time point described above, the computing section 140 may compute the conditional probability of each hidden node in the layer 16 at the one time point in the same manner. That is, the computing section 140 may use the plurality of update parameters to compute the conditional probability of the value of each hidden node and each input data value at the one time point on the condition that an input data sequence has occurred. Here, the computing section 140 may store the conditional probability pj,t in a FIFO or the like.
  • The computing section 140 may be operable to compute a total likelihood, after the learning by the learning apparatus 100 has continued. The computing section 140 computes the total likelihood pj as shown in the following expression, based on the conditional probabilities pj,t−K+1, pj,t−K+2, . . . , pj,t computed by K instances of learning from the time point t−K+1 to the time point t, for example. The total likelihood pj in Expression 16 indicates a total sum of the conditional probabilities, as an example, but the total likelihood pj may be at least one of a sum, weighted sum, product, or weighted product of the conditional probabilities. Furthermore, K may be an integer greater than or equal to 2, and if the computing section 140 stores the conditional probabilities pj,t in a FIFO or the like, the length of the FIFO sequence may be equal to the value of K.
  • p = s = t - k + 1 t p s Expression 16
  • The computing section 140 may supply the total likelihood pj to the storage section 130. The storage section 130 may sample the values xj [t] of the hidden nodes of the hidden layer 16 at the one time point, based on the most recent likelihood pj,t. That is, the storage section 130 according to the present embodiment may be operable to sample the value of each hidden node at the one time point, by using the conditional probability of the value of each hidden node at the one time point. For example, the storage section 130 samples the values of the hidden nodes based on the history of the conditional probabilities computed by the computing section 140. That is, the storage seal m 130 may sample the values of the hidden nodes after the learning operation of the learning processing section 150 has been performed a plurality of times. The storage section 130 may store a value of 0 in the hidden nodes as the sampling value, until the leaching operation of the learning processing section 150 has been performed a plurality of times.
  • The storage section 130 may store a value of 1 or 0 in the hidden node j as the sampling result, according to the result of a comparison between the value of the total likelihood pj and a threshold value. In this way, when predicting the time series data to be input to the input layer 14, the storage section 130 can store a more preferable value as the hidden node value by performing sampling based on the history of past conditional probabilities.
  • The learning processing section 150 may be operable to determine the weight parameter based on the total likelihood pj. In this case, the learning processing section 150 may compute update amounts and Δui, j, k and Δvi, j, k for the learning parameters ui, j, k and vi, j, k in the weight parameter for one hidden node j at the one time point. For example, the learning processing section 150 may compute these update amounts Δui, j, k [t] and Δvi, j, k [t] as shown in the following expression, based on the value xj [t] of the one hidden node j at the one time point t and on the conditional probability <Xj [t]> of the value of this hidden node j at the one time point t on the condition that the input data'sequence has occurred (I+1≤j≤I+H).

  • Δu i,j,k [t]i,j,k [t−1](x j [t]
    Figure US20180197080A1-20180712-P00001
    X j [t]
    Figure US20180197080A1-20180712-P00002
    )

  • Δv i,j,l (1)|t|i,j,l [t−1](
    Figure US20180197080A1-20180712-P00001
    X j [t]
    Figure US20180197080A1-20180712-P00002
    −x j |t|)

  • Δv i,j,l (2)[t]j,l [t−1](
    Figure US20180197080A1-20180712-P00001
    X i |t|
    Figure US20180197080A1-20180712-P00002
    −x i |t|)   Expression 17:
  • Here, the update amount Δvi, j, k [t] is equal to Δui, j, k (1)[t]+Δui, j, k (2)[t]. The conditional probability <Xj [t]> of the value of the hidden node j may be computed by the computing section 140 using Expression 2. The learning processing section 150 may store the computed update amounts Δvi, j, k [t], Δui, j, k (1)[t], and Δui, j, k (2)[t] in the storage section or the like. The learning processing section 150 may be operable to store the update amounts Δvi, j, k [t], Δui, j, k (1)[t], and Δui, j, k (2)[t] computed for one time point in the FIFO sequence. That is, the learning processing section 150 may be operable to update the learning parameters based on update amounts computed in the past.
  • The learning, processing section 150 changes the ratio by which the update amounts are reflected in the learning parameters, according to the conditional probability of input data occurring at a following time point that is after the one time point t, for example. In this case, the learning processing section 150 may change the ratio by which the update amounts are reflected in the learning parameters according to the conditional probability of a bidden node value occurring at a plurality of following time points that are after the one tune point. In the present embodiment, an example is described in which the learning processing section 150 changes the ratio by which the update amounts are reflected in the learning parameters according to the total likelihood pj after the one time point
  • The learning processing section 150 may update the learning parameters as shown in the following expression, based on the total likelihood pj computed by K instances of teaming from the time point to the time point t and on the update amount at the time point t−K+1, for example. Here, K may be an integer greater than or equal to 2, and if the learning processing section 150 stores the update amounts in the FIFO sequence, the length of the FIFO sequence may be equal to the value of K.

  • ui,j,k←ui,j,k1pjΔui,j,k [t−k+1]

  • vi,j,l←vi,j,l1pj(Δvi,j,l (1)[t−k+1] +Δv i,j,l (2)[t−k+1])   Expression 18:
  • Here, η1 may be a constant for adjusting the update amount. Alternatively, η1 may be a coefficient whose value becomes smaller according to an increase in the number of updates. Yet further, ηl may have a value of substantially 1 at the stage when the learning processing section 150 begins learning, and may be a coefficient whose value becomes smaller according to the amount of learning occurring as time progresses from the time point t. For example, η110/t2. Furthermore, η1 may be a coefficient whose value becomes smaller according to the update amount. For example, η110/(ΣΔui, j, k 1/2). Here, η10 may be a predetermined constant.
  • In the manner described above, the learning processing section 150 may update the learning parameters of a hidden node at one time point according to the conditional probabilities computed at time points before the one time point. In this way, the learning apparatus 100 can more strongly reflect the update amounts at time points before the one time in the learning parameters, in response to the predicted probability of an input value of an input node being large due to the weight parameters at time points before the one time point. That is, the learning apparatus 100 can update the weight parameters of the hidden nodes in a mariner to increase the conditional probabilities.
  • If a FIFO sequence is used to perform an update of such a weight parameter, the learning processing section 150 may extract from the FIFO sequence the update amounts Δvi, j, k [t−K+1, Δui, j, k (1)(t−K+1], and Δui, j, k (2)[t−K+1] of a past time point (e.g. t−K+1) at the following time point (e.g. t) or a time point thereafter. The learning processing section 150 may update the ratio by which the update extracted from the FIFO sequence are reflected in the learning parameters according to the conditional probabilities of hidden node values occurring at the following time point t. For example, the learning processing section 150 may multiply the total likelihood pj respectively by each update amount. In this way, the learning processing section 150 can efficiently perform the update of the weight parameters as described above.
  • As described above, the learning apparatus 100 according to the present embodiment can learn a model 10 in a manner to predict input data at one time point, based on an input data series before the one time point in the time-series input data. Furthermore, the learning apparatus 100 can improve the expressive power, learning power, and the like by including hidden nodes. With such a learning apparatus 100, there are cases where it is relatively difficult to perform the learning.
  • For example, for a hidden node value at one time point, it is difficult to evaluate whether this hidden node value was effective if a certain amount of time points have not yet passed. Accordingly, if hidden nodes, are used, there are cases where the learning accuracy of the learning apparatus 100 for the weight parameter between the input data series and the hidden nodes at: the one time point is lower than the learning accuracy for the weight parameter between the input data series and the input nodes at the one time point. Furthermore, if there is a characteristic data trend in the input data series, there are cases where the learning is affected by this data trend and overfitting is performed, causing unsuitable learning for future data prediction.
  • Therefore, the learning apparatus 100 may be operable to further improve the learning by changing the order in which the time-series input data is input to the model 10 and performing a bidirectional learning process. The following describes such a learning apparatus 100.
  • The present embodiment describes an example in which the learning apparatus 100 is operable to learn a model 10 in a manner to predict input data at one time point, based on the input data series before the one time point in the time-series input data. In this way, the learning apparatus 100 creates a first model that is a forward model that performs a learning process in which the time-series input data is input to the model 10 in order from past data to future data.
  • FIG. 9 shows an exemplary schematic configuration of the first model according to the present embodiment. FIG. 9 shows a schematic configuration of the first model corresponding to the model 10 shown in FIG. 2. In FIG. 9, the value of the node j in the input layer 14 at the one time point t is indicated by xj [t] (1≤j≤I), and the value of the node j in the hidden layer 16 at the one time point t is indicated by hj [t] (I+1≤j≤H). Furthermore, the value of the node i in a plurality of input layers 14 before the one time point t is indicated by xi [−1] (1≤i≤I), and the value of the node i in the hidden layer 16 before the one time point t is indicated by hi [:t−1] (I+1≤i≤H).
  • As described above, the first model may include a plurality of input nodes j that sequentially input a plurality of input values at each time point in a time-series input data sequence, and a weight parameter Wij [δ] between each input node j and each input value xi [1−1] at the time points before the time point: corresponding to the plurality of input nodes j. Here, the weight parameter between a plurality of input layers 14 before the one time point and the input layer 14 at the one time point is WVV. Furthermore, the first model may include a weight parameter Wij [δ] between each input node j and each of a plurality of hidden nodes i corresponding to time points before the time point corresponding to the plurality of input nodes j. Here, the weight parameter between the plurality of hidden layers 16 before the one time point t and the input layer 14 at the one time point t is WHV.
  • The first model may further include a weight parameter Wij [δ] between each input node i corresponding, to a time point before the time point corresponding to the plurality of input nodes j and each hidden node j corresponding to the time point corresponding to the plurality of input nodes j. Here, the weight parameter between the plurality of input layers 14 before the one time point t and the hidden layer 16 at the one time point t is WVH. The first model may further include a weight parameter Wij [δ] between each of the plurality of hidden nodes i corresponding to a time point before the time point corresponding to the plurality of input nodes j and each hidden node j corresponding to the time point corresponding to the plurality of input nodes j. Here, the weight parameter between the plurality of hidden layers 16 before the one time point t and the hidden layer 16 at the one time point t is WHH.
  • With b representing a bias parameter of the first model, the conditional probability p of the, input node j and the hidden node j at the one time point may be calculated based on the following expression. The update of each parameter in the first model is as described above.
  • p ( x j [ t ] , h j [ t ] x i [ : t - 1 ] , h i { ; t - 1 ] ) ~ exp ( b T ( x j [ t ] h j [ t ] ) + δ > 0 ( x i [ t - δ ] , h i { t - δ ] ) T W i j [ δ ] ( x j [ t ] h j [ t ] ) ) Expression 19
  • Here, since the time-series input data for learning is already known, the order in which this time-series input data is input to the model 10 is not limited to the forward order from past data to future data, as in the first model. The learning apparatus 100 may be operable to learns the model 10 in a manner to predict input data at one time point based on the input data series after the one time point i the'time-series input data. In this way, the learning apparatus 100 creates a second model that is a backward model that performs learning process in which the time-series input data is input to the model 10 m order from future data to past data.
  • FIG. 10 shows an exemplary schematic configuration of the second model according to the present embodiment. FIG. 10 shows a schematic configuration of the second model corresponding to the model 10 shown in FIG. 2 In FIG. 10, the value of the node j in the input layer 14 at the one time point t is indicated by xj [t] (1≤j≤I), and the value of the node j in the hidden layer 16 at the one time point t is indicated by hj [t] (I+1≤j≤H). Furthermore, the value of the node i in a plurality of input layers 14 after the one time point t is indicated by xi [t−1:] (1≤i≤I), and the value of the node i in the hidden layer 16 after the one time point t is indicated by hi [t−1:] (I+1≤i≤H). In other words, if the learning operation is performed using the second model, the learning apparatus 100 inputs the data the model 10 from the temporally newest data in the time-series input data.
  • In this case, the second model may include a plurality of input nodes i that sequentially input, in a backwards order, a plurality of input values at each time point in the time-series input data sequence, and a weight parameter Wji [δ] between each input value xj [t] at time points after the tune point corresponding to the plurality of input nodes i and each input node i. Here, the weight parameter between the input layer 14 at the one time point and a plurality of input layers 14 after the one time point is WVV. Furthermore, the second model may include a weight parameter Wji [δ] between each of a plurality of hidden nodes j corresponding to time points after the time point corresponding to the plurality of input nodes i and each input node i. Here, the weight parameter between the input layer 14 at the one time point t and the plurality of hidden layers 16 after the one time point t is WVH.
  • The second model may further include a weight parameter Wji [δ] between each input node j corresponding to a time point after the time point corresponding to the plurality of input nodes i and each hidden node i corresponding to a time point corresponding to the plurality of input nodes i. Here, the weight parameter between the hidden layer 16 at the one time point t and the plurality of input layers 14 after the one time point t is wHV. The second model may further include a weight parameter Wij [δ] between each of the plurality of hidden nodes j corresponding to a time point after the time point corresponding to the plurality of input nodes i and each hidden node i corresponding to the time point corresponding to the plurality of input nodes i. Here, the weight parameter between the hidden layer 16 at the one time point t and the plurality of hidden layers 16 after the one time point tis WHH.
  • With b representing a bias parameter of the second model, >the conditional probability p of the input nodes i and the hidden nodes i at the one time point may be calculated based on the following expression.
  • p ( x j [ t ] , h j [ t ] x i [ t + 1 : ] , h i [ t + 1 : ] ) ~ exp ( b T ( x j [ t ] h j [ t ] ) + δ > 0 ( x i [ t - δ ] , h i { t - δ ] ) T W i j [ δ ] ( x j [ t + δ ] h j [ t + δ ] ) ) ~ exp ( b T ( x j [ t ] h j [ t ] ) + δ > 0 ( x i [ t + δ ] , h i { t + δ ] ) T W i j [ δ ] ( x j [ t ] h j [ t ] ) ) Expression 20
  • As shown in Expression 19 and Expression 20, the learning apparatus 100 may be operable to learn each parameter in common for the first model and the second model. Furthermore, the learning operation of the learning apparatus 100 using the second model may be operable to be performed using the same operation as the learning operation using the first model. For example, the bias parameters for the first model and the second model may be substantially the same parameter. The weight parameters Wij and Wji for the first model and the second model may be operable to be converted according to a transpositional relationship there between. Furthermore, the learning parameters ui,j,k and vi,j,l of the first model ma be operable to be converted to the learning parameters ui,j,k and vj,i,l of the second model by switching i and j.
  • The predefined parameters λk t1 and μl t2 of the first model and the second model may be substantially the same parameter. Since the value of the predefined parameter is determined according to the time point difference, the predefined parameter value may change differently according to each time point in the first model and the second model.
  • in this way, by applying the Conversion operation, the parameters of the first model and the second model are shared. Accordingly, the learning result of the first model and the learning result of the second model based on the same time-series input data sequence will theoretically match. However, there are cases where differences in the ease of learning occur, according to the directional order in which the data is input.
  • For example, the value of the input layer 14 at the one time point t and the plurality of values of the input layer 14 before the one time point t are already known as the time-series input data. Furthermore, the plurality of values of the hidden layer 16 before the one time point t are values obtained after a certain amount of time points have passed from when these values were sampled, for example, and therefore an evaluation result is already obtained for most of these values. Accordingly, the weight parameters WVV and WHV relating to the input layer 14 at the one time point t in the first model are parameters that can be accurately learned.
  • Furthermore, the value of the hidden layer 16 at the one time point t is a sampled value, for example, and therefore an evaluation result of this sampled value becomes known after several instances of learning have been completed. Accordingly, there are cases where the weight parameters WVH and WHH relating to the hidden layer 16 at the one time point t in the first model are difficult to learn, In the present embodiment, the weight parameter WHV that is difficult to learn in the first model is set as a first learning target parameter, and the weight parameter WHV that can be accurately learned is set as a second learning target parameter.
  • In the second model, in the same manner, the weight parameters WVV and WVH relating to the input layer 14 at the one time point t are parameters that can be accurately learned. On the other hand, there are cases where the weight parameters WHV and WHH relating to the hidden layer 16 at the one time point t are difficult to learn. Therefore, the weight parameter WVH that can be accurately learned in the second model is set as a first learning target parameter, and the weight parameter WHV that is difficult to learn is set as a second learning target parameter.
  • Here, the first learning target parameter WVH of the first model may he expressed relatively as the weight parameter between a past input layer 14 and a fixture hidden layer 16. The first learning target parameter WVH of the second model may also be expressed relatively as the weight parameter between a past input layer 14 and a future hidden layer 16. Accordingly, the first learning target parameter WVH of the first model and the first learning target parameter WVH of the second model may be a common parameter, or the first learning target parameter WVH of the first model may be operable to be converted into the first learning target parameter WVH of the second model. Accordingly, by learning the first learning target parameter WVH using the second model that can accurately learn this parameter, the learning apparatus 100 can easily calculate the first learning target parameter WVH that can be difficult to calculate with the first model.
  • Similarly, the second learning target parameter WHV of the first model may be expressed relatively as the weight parameter between a past hidden layer 16 and a future input layer 14. The second learning target parameter WHV of the second model may also be expressed relatively as the weight parameter between a past hidden layer 16 and a future input layer 14. Accordingly, the second learning target parameter WHV of the first model and the second learning target parameter WHV of the second model may be a common parameter, or the second learning target parameter WHV of the first model may be operable to be converted into the second learning target parameter Wuv of the second model. Accordingly, by learning the second learning target parameter WHV using the first model that can accurately learn this parameter, the learning apparatus 100 can easily calculate the second learning target parameter WHV that can be difficult to calculate with the second model,
  • In this way, the learning apparatus 100 according to the present embodiment may be operable to perform efficient learning by using the first model and the second model to learn a parameter that is difficult for one of the models to learn using a parameter that can be accurately learned by the other model. The following describes the learning operation of such a learning apparatus 100.
  • FIG. 11 shows an operational flow of the learning process using the first model and the second model performed by the learning apparatus 100 according to the present embodiment. In the present embodiment, the learning apparatus 100 may be operable to accurately learn a model 10 corresponding to time-series input data by performing the processes from S410 to S470.
  • First, the acquiring section 110 may acquire a time-series input data sequence (S410). In the present embodiment, an example is described in which the acquiring section 110 acquires, as the time-series input data sequence, time series data in an interval L that is longer than an interval T of the time-series data used for one instance of learning. The acquiring section 110 acquires L pieces of image data arranged in time series to form moving image data, for example.
  • Next, the learning apparatus 100 may learn the weight parameter between each input value and each hidden node corresponding to a time point before a time point corresponding to a plurality of input nodes in the first model, using a first model learning process (S420). Specifically, the supplying section 120 may sequentially input the time-series input data sequence into the model 10 in order from the older pieces of input data, and the storage section 130 may sample the values of the plurality of hidden nodes corresponding to the respective time points and store the values respectively in the corresponding plurality of hidden nodes.
  • The learning apparatus 100 may perform this learning using the first model. Specifically, the calculating section 140 may calculate the conditional probability of each input value of the input node at the one time point, and the learning processing section 150 may update the parameter θ. Furthermore, the updating section 170 may update the update parameter. The learning operation of the learning apparatus 100 has already been described above, and therefore is not provided here. The learning apparatus 100 may be operable to perform the learning process a predetermined number of times using the first model.
  • Next, the learning apparatus 100 may convert the parameter learned using the first model into a parameter corresponding to the second model (S430). In this case, the updating section 170 includes a converting section, and the converting section;may be operable to convert the parameter of the first model into the parameter of the second model. The converting section may convert the parameters by performing a transposition, element replacement, and the like.

  • uj,i,k←ui,j,k vj,i,l←vi,j,l

  • Wj,l←Wi,j
  • Next, the learning apparatus 100 may learn the weight parameter between each of the plurality of hidden nodes and each input node corresponding to the time points after the time point corresponding to the plurality of input nodes in the second model, by using the learning process with the second model using the converted parameter (S440). Specifically, the supplying section 120 may sequentially input the time-series input data sequence into the model 10 in order from the newer pieces of input data, and the storage section 130 may sample the values of the plurality of hidden nodes corresponding to each time point and store the sampled values respectively in the corresponding plurality of hidden nodes.
  • The supplying section 120 and the storage section 130 may store each piece of data in a FIFO memory 160 used for learning by the second model, which is different from a FIFO memory 160 used for learning by the first model. The learning, apparatus 100 may be operable to perform the learning process a predetermined number of times using the second model.
  • Next, the learning apparatus 100 may convert the parameter learned using the second model into the parameter corresponding to the first model (S450). In this case, the converting section may be operable to convert the parameter of the second model into the parameter of the first model. The converting section may convert the parameters using a transposition, element replacement, and the like.

  • ui,j,k←uj,i,k vi,j,l←vj,i,l

  • Wi,j←Wj,i
  • The learning apparatus 100 may judge whether the learning is to continue (S460). The learning apparatus 100 may continue the learning until the learning process has been performed a predetermined number of times, or may instead continue the learning until a stop command from the user is input. As another example, the learning apparatus 100 may continue the learning until there is no more data in the time-series input data sequence that can be input.
  • If the learning is to continue (S460: Yes), the learning apparatus 100 returns to step S410, and if there is no more time-series data to be supplied to the model 10, the acquiring section 110 may acquire the next piece of data in the time-series data sequence. The learning apparatus 100 may sequentially perform learning with the first model and the second model based on the next piece of time-series data.
  • If the learning is ended (S460: No), the teaming processing section 150 may perform a final learning process using the first model (S470). The learning processing section 150 may be operable to acquire the learned first model after performing the learning process with the first model one or more times. The learning processing section 150 may be operable to output the learned first model to the database 1000 or the like. The learning processing section 150 may output the first model and end the learning process.
  • In this way, the learning apparatus 100 may output, as the learned model, the model used to perform the last instance of learning corresponding to the model to be output from among the first model and the second model. Specifically, FIG. 11 describes an example in which the learned first model is output, but if the learned second model is output, the learned second model may be output without performing the process of S450 for convening the parameters.
  • The learning apparatus 100 may continue with the learning while focusing more heavily on learning corresponding to the model to be output from among the first model and the second model. For example, the learning apparatus 100 may perform the learning process with the first model a greater number of times than the learning process with the second model. Furthermore, the learning apparatus 100 may be operable to perform the learning using a higher learning rate for the learning process with the first model than for the learning process with the second model, by adjusting a coefficient or the like. In this way, the learning apparatus 100 performs learning in a focused manner corresponding to the model to be output, and can therefore output a more suitable learned parameter model using this model.
  • As described above, the learning process of the learning apparatus 100 according to the present embodiment performs a bidirectional learning process with the first model and the second model, using common parameters. Accordingly, even if parameters having learning accuracies that fluctuate according to the learning direction are included, the learning apparatus 100 can learn these parameters with higher accuracy. Furthermore, by performing, bidirectional learning, the learning apparatus 100 can prevent the occurrence of overfitting caused when model learning is performed using only one of the first model and the second model.
  • The learning process of the learning apparatus 100 according to the present embodiment described above is an example of a bidirectional learning process performed using a forward model and a backward model. In addition to this, the learning apparatus 100 may make a distinction in the parameters being learned according to the model.
  • As described in FIG. 9 and FIG. 10, the first learning target parameter WVH of the first model may be operable to be learned with higher accuracy for learning using the second model than for learning using the first model. Furthermore, the second learning target parameter %Tv of the first model may be operable to be learned with higher accuracy for learning using the first model than for learning using the second model.
  • In this case, the learning using the first model performed by the learning apparatus 100 may be operable to learn the second learning target parameter WHV without changing the first learning target parameter WVH. In other words, when learning with the first model, the learning apparatus 100 may update the second learning target parameter WHV without updating the first learning target parameter WVH. Furthermore, the learning using the second model performed by he learning apparatus 100 may be operable to learn the first learning target parameter WVH without changing the second learning target parameter WHV. In other words, when learning with the second model, the learning apparatus 100 may update the first learning target parameter WVH but need does not need to update the second learning target parameter WHV.
  • Instead of this, the learning apparatus 100 may increase or decrease the update amount, according to the model used for the learning. For example, when learning with the first model, the learning apparatus 100 may perform the update by multiplying the update amount of the first learning target parameter WVH by a coefficient ρ, and multiplying the update amount of the second learning target parameter WHV (1−ρ). Here, the coefficient p may be a positive value that is less than 0.5. In the same manner, when learning with the second model, the learning apparatus 100 may perform the update by multiplying the update amount of the first learning target parameter WVH by (1−ρ), and multiplying, the update amount of the second learning target parameter WHV by the coefficient ρ.
  • In this way, the learning apparatus 100 can perform the learning process more;accurately by encouraging parameter updates that are suitable for the learning, according to the model used for the learning. Accordingly, by including the hidden nodes, the learning apparatus 100 can improve the expressive power, learning power, and the like, and can perform the learning more accurately by performing the bidirectional learning process.
  • The learning apparatus 100 according to the present embodiment described above is an example in which the learning apparatus 100 acquires and learns a time-series input data sequence, but the present invention is not limited to time-series data. For example, for a two-dimensional image that has n pixels vertically and m pixels horizontally, n-pixel column data arranged in a column direction may be'used as the data corresponding to the input data at the one time point. In this case, the plurality of pieces of column data arranged in the row direction may be the data corresponding to the input data before or after the one time point, in this case, the learning apparatus 100 may perform the bidirectional learning for a case where a plurality of pieces of column data are input to the model 10 in one row direction and a case where the plurality of pieces of column data are input to the model 10 in the other row direction. The learning apparatus 100 may perform the bidirectional learning for waveform data, image pattern data, language data and the like, in the same manner.
  • The learning apparatus 100 according to the present embodiment described above is an example in which the learning apparatus 100 acquires one input data sequence and generates training data for the bidirectional learning, based on this one input data sequence, but the data acquired in the present invention is not limited to one input data sequence. The learning apparatus 100 may acquire a plurality of data sequences. In this case, the learning apparatus 100 may use learning with the first model for the first input data sequence and use learning with the second model for a second input data sequence that differs from the first input data sequence.
  • In this case, the first input data sequence and the second input data sequence are preferably at least partially related. Furthermore, if the data sequences are the same type or have substantially the same trend, the first input data sequence and the second input data sequence may be input data sequences for learning that are different from each other and are included in a plurality of input data sequences for learning.
  • In this case, the learning apparatus 100 may be operable to, based on the first model, sequentially input each piece of input data from the first input data sequence, in which are arranged pieces of input data including a plurality of input values, into the input nodes of the model 10, to perform the learning process using the first model. The learning apparatus 100 may be operable to input each piece of input data from the second input data sequence, in which are arranged pieces of input data including a plurality of input values, into the input nodes of the model 10 in a different order than, in the first model, to generate the second model for learning the first learning target parameter included in the first model.
  • The learning apparatus 100 may be operable to input each piece of input data from the second input data sequence in a backwards order, to generate the second model for learning the learning target parameter. The first and second input data sequences may be time-series input data sequences. In this case, the first model may input the first input data sequence in order from the older pieces of input data, and the second model may input the second input data sequence in order from the newer pieces of input data.
  • The learning apparatus 100 may be operable to perform the learning process using both the first model and the second model. The learning apparatus 100 may be operable to output the learned first model after performing the bidirectional learning with the first model and the second model. The learning apparatus 100 may be operable to delete the learned second model and output the learned first model as the predictive model based on the input data sequence. In this way, the learning apparatus 100 may be operable to acquire a plurality of input data sequences for learning and perform the learning process.
  • The learning apparatus 100 according to the present embodiment described above is an example in which the learning apparatus 100 learns by applying substantially the same learning; process to the first learning target parameter and the second learning target parameter. Instead of this, the learning apparatus 100 may learn by applying different learning processes respectively to the first learning target parameter and the second learning target parameter. In other words, the learning apparatus 100 may be operable to perform an update by decreasing the update amount for a parameter that can be difficult to learn.
  • In this case, when learning with the first model and the second model, the teaming apparatus 100 may perform the updates using θv=(bvj, WHV, WVV) and θH=(bHj, WVH, WHH) as the respective update parameters. In other words, the learning apparatus 100 may be operable to independently perform learning with θv as an update parameter than can be accurately updated and with On as an update parameter than can be difficult to learn. Here, bvi indicates the bias parameter corresponding to each node of the input layer 14. Furthermore, bHj indicates the bias parameter corresponding to each node of the hidden layer 16.
  • As an example, a lower bound for the total likelihood can be expressed as shown in the following expression from Jensen's inequality Here, the value {tilde over (h)} of the hidden node may be sampled in correspondence with pθ({tilde over (h)}|x).
  • log p θ ( x ) = log Π t ( h ~ p θ ( x [ t ] | x [ : t - 1 ] , h ~ [ : t - 1 ] ) Π s < t p θ ( h ~ [ s ] | x [ : s - 1 ] , h ~ [ : s - 1 ] ) ) h ~ p θ ( h ~ | x ) t log p θ ( x [ t ] | x [ : t - 1 ] , h ~ [ : t - 1 ] ) Expression 23
  • From Expression 23, the stochastic gradient may be calculated as shown in the following Expression, according to the parameter to be updated.
  • θ V log p θ ( x ) = θ v log p θ ( x [ t ] | x [ : t - 1 ] , h ~ [ : t - 1 ] ) θ H log p θ ( x ) log p θ ( x [ t ] | x [ : t - 1 ] , h ~ [ : t - 1 ] ) ( s t θ H log p θ ( h ^ [ s ] | x [ : s - 1 ] , h ~ [ : s - 1 ] ) ) Expression 24
  • The learning apparatus 100 may be operable to, when learning using the first model, update the parameter θv that can be learned accurately as described above. In this case, the update of the parameter θv can be expressed as shown in the expression below.

  • θV←θV+η∇θV log pθ(x[t]|x[;t−1], {tilde over (h)}[:t−1])   Expression 25:
  • Furthermore, the lemming apparatus 100 may be operable to, when learning using the first model, gradually update the parameter θH that can be difficult to learn, by using a gentle gradient. In tins case, the update of the parameter θH can be expressed as shown in the expression below. The parameter g is a gradient parameter indicating the direction of the gradient. The constant ϵ may be a value less than 1, and may be a value closer to 0, for example. The log pθ operating On ηg in the expression for updating θH need not be included. Furthermore, in the same manner as η1, η may be a constant for adjusting the update amount or may be a coefficient whose value becomes smaller according to an increase in the number of updates.

  • g←ϵg+(1−ϵ)∇θ H log pθ({tilde over (h)}[t]|x[:t−1], {tilde over (h)}[:t−1])

  • θH←θH+log pθ(x[t]|x[:t−1], {tilde over (h)}[t−1])ηg   Expression b 26:
  • In the same manner, the learning apparatus 100 may be operable to, when learning using the second model, update the parameter θv that can be accurately learned, as shown in the following expression.

  • θV←θV+η∇θ V log pθ(x[t]|x[t+1:], {tilde over (h)}[t+1:])   Expression 27:
  • The learning apparatus 100 may be operable to, when learning using the second model, update the parameter θH that can be difficult to learn, as shown in the following expression. ° The log pθ operating on ηg in the expression for updating θH need not be included.

  • g←ϵg+(1−ϵ)∇θ H log pθ({tilde over (h)}[t]|x[t+1:], {tilde over (h)}[t+1:])

  • θH←θH+lot pθ(x[t]|x[t+1:], {tilde over (h)}[t+1:])ηg   Expression 28:
  • As described above, the learning apparatus 100 can perform the learning process more accurately by changing the method used according to the ease of learning the parameter.
  • The bidirectional learning process of the learning apparatus 100 according to the present embodiment described above is an example suitable for a Boltzmann machine, but the present invention is not limited to this. The present invention can be applied to any learning apparatus that is operable to bidirectionally learn a data series and capable of changing the ease of the learning according to the learning direction
  • FIG. 12 shows an example of a computer 800 in which aspects of the present invention may be wholly or partly embodied. A program that is installed in the computer 800 can cause the computer 800 to function as or perform operations associated with apparatuses of the embodiments of the present invention or one or more sections (including modules, components, elements, etc.) thereof, and/or cause the computer 800 to perform processes of the embodiments of the present invention or steps thereof. Such a program may be executed by the CPU 800-12 to cause the computer 800 to perform certain operations associated with some or all of the blocks of flowcharts and block diagrams described herein.
  • The computer 800 according to the present embodiment includes a CPU 800-12, a RAM 800-14, a graphics controller 80016., and a display device 800-18, winch are mutually connected by a host controller 800-10. The computer 800 also includes input/output units such as a communication interface 800-22, a hard disk drive 800-24, a DVD-ROM drive 800-26 and an IC card drive, which are connected to the host controller 800-10 via an input/output controller 800-20. The computer also includes legacy input/output units such as a RUM 800-30 and a keyboard 800-42, which are connected to the input/output controller 800-20 through an input/output chip 800-40.
  • The CPU 800-12 operates according to programs stored in the ROM 800-30 and the RAM 800-14, thereby controlling each unit. The graphics controller 800-16 obtains image data generated by the CPU 800-12 on a frame buffer or the like provided in the RAM 800-14 or in itself, and causes the it tape data to be displayed on the display device 800-18.
  • The communication interface 800-22 communicates with other electronic devices via a network 800-50. The hard disk drive 800-24 stores programs and data used by the CPU 800-12 within the computer 800. The DVD-ROM drive 800-26 reads the programs or the data from the DVD-ROM 800-01, and provides the hard disk drive 800-24 with the programs or the data via the RAM 800-14. The IC card drive reads programs and data from an IC card, and/or writes programs and data into the IC card.
  • The ROM 800-30 stores therein a boot program or the like executed by the computer 800 at the time of activation, and/or a program depending on the hardware of the computer 800. The input/output chip 800-40 may also connect various input/output units via a parallel port, a serial port, a keyboard port, a mouse port, and the like to the input/output controller 800-20.
  • A program is provided by computer readable media such as the DVD-ROM 800-01 or the IC card. The program is read from the computer readable media, installed into the hard disk drive 800-24, RAM 800-14, or ROM 800-30, which are also examples of computer readable media, and executed by the CPU 800-12. The information processing described in these programs is read into the computer 800, resulting in cooperation between a program and the above-mentioned various types of hardware resources. An apparatus or method may be constituted by realizing the operation or processing of information in accordance with the usage of the computer 800-
  • For example, when communication is performed between the computer 800 and an external device, the CPU 800-12 may execute a communication program loaded onto the RAM 800-14 to instruct communication processing to the communication interface 800-22, based on the processing described in the communication program. The communication interface 800-22, under control of the CPU 800-12, reads transmission data stored on a transmission buffering region provided in a recording medium such as the RAM 800-14, the hard disk drive 800-24, the DVD-ROM 800-01, or the IC card, and transmits the read transmission data to network 800-50 or writes reception data received from network 800-50 to a reception buffering region or the like provided on the recording medium.
  • In addition, the CPU 800-12 may cause all or a necessary portion of a file or a database to be read into the RAM 800-14, the file or the database having been stored in an external recording medium such as the hard disk drive 800-24, the DVD-ROM drive 800-26 (DVD-ROM 800-01), the IC card, etc., and perform various types of processing on the data on the RAM 800-14. The CPU 800-12 may then write back the processed data to the external recording medium.
  • Various types of information, such as various types of programs, data, tables, and databases, may be stored in the recording medium to undergo information processing. The CPU 800-12 may perform various types of processing on the, data read from the RAM 800-14, which includes various, types of operations, processing of information, condition judging, conditional branch, unconditional branch, search/replace of information, etc., as described throughout this disclosure and designated by an instruction sequence of programs, and writes the result back to the RAM 800-14. In addition, the CPU 800-12 may search for information in a file, a database, etc., in the recording medium. For example, when a plurality of entries, each having an attribute value of a first attribute is associated with an attribute value of a second attribute, are stored in the recording medium, the CPU 800-12 may search for an entry matching the condition whose attribute value of the first attribute is designated, from among the plurality of entries, and reads the attribute value of the second attribute stored in the entry, thereby obtaining the attribute value of the second attribute associated with the first attribute satisfying the predetermined condition.
  • The above-explained program or software modules may be stored in the computer readable media on or near the computer 800. In addition, a recording medium such as a hard disk or a RAM provided in a server system connected dedicated communication network or the Internet can be used as the computer readable media, thereby providing the program to the computer 800 via the network.
  • A model in accordance with the present invention can used for a myriad of applications including, but not limited to, classification, recognition (e.g., speech recognition, speaker recognition, pattern recognition, etc.), robotics (e.g., robotic control, robotic decision making), machine control (e.g., controlling a movement of a machine (e.g., an assembly line machine), or powering down the machine, or changing the operational state of the machine (e.g., faster, slower, on, off low-power, etc.) based on an output of the model, and so forth. Thus, such model can be used within systems including, but not limited to, classification systems (e.g., speech recognition systems, speaker recognition systems, pattern recognition systems, etc.), machine control systems (or machine controllers), etc. These and other applications to which the present invention can be applied are readily determined by one of ordinary skill in the art, given the teachings of the present invention provided herein, while maintaining the spirit of the present invention.
  • The present invention may be a system, a method, and/or a computer program product. The computer program product may include a computer readable storage medium (or media) having computer readable program instructions thereon for causing a processor to carry out aspects of the present invention.
  • The computer readable storage medium can be a tangible device that can retain and store instructions for use by an instruction execution device. The computer readable storage medium may be, for example, but is not limited to, an electronic storage device, a magnetic storage device, an optical storage device, an electromagnetic storage device, a semiconductor storage device, or any suitable combination of the foregoing. A non-exhaustive list of more specific examples of the computer readable storage medium includes the following: a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a static random access memory (SRAM), a portable compact disc read-only memory (CD-ROM), a digital versatile disk (DVD), a memory stick, a floppy disk, a mechanically encoded device such as punch-cards or raised structures in a groove having instructions recorded thereon, and any suitable combination of the foregoing. A computer readable storage medium, as used herein, is not to be construed as being transitory signals per se, such as radio eaves or other freely propagating electromagnetic waves, electromagnetic waves propagating through a waveguide or other transmission media (e.g., light pulses passing through a fiber-optic cable), or electrical signals transmitted through a wire,
  • Computer readable program instructions described herein can be downloaded to respective computing/processing devices from a computer readable storage medium or to an external computer or external storage device via a network, for example, the Internet, a local area network, a wide area network and/or a wireless network. The network may comprise copper transmission cables, optical transmission fibers, wireless transmission, routers, firewalls, switches, gateway computers and/or edge servers. A network adapter card or network interface in each computing/processing device receives computer readable program instructions from the network and forwards the computer readable program instructions for storage in a computer readable storage medium within the respective computing/processing device.
  • Computer readable program instructions for carrying out operations of the present invention may be assembler instructions, instruction-set-architecture (ISA) instructions, machine instructions, machine dependent instructions, microcode, firmware instructions, state-setting data, or either source code or object code written in any combination of one or more programming languages, including an object oriented programming language such as Smalltalk, C++ or the like, and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer readable program instructions may execute entirely on the user's computer, partly on the user's computer, as a stand-alone software package, partly on the users computer and partly on a remote computer or entirely on the remote computer or server. In the latter scenario, the remote computer may be connected to the user's computer through any type of network, including a local area network (LAN) or a wide area network (WAN), or the connection may be made to an external computer (for example, through the Internet using an Internet Service Provider). In some embodiments, electronic circuitry including, for example, programmable logic circuitry, field-programmable gate arrays (FPGA), or programmable logic arrays (PLA) may execute Me computer readable program instructions by utilizing state information of the computer readable program instructions to individualize the electronic circuitry, in order to perform aspects of the present invention.
  • Aspects of the present invention are described herein with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer readable program instructions.
  • These computer readable program instructions may be provided to a processor of a general purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks. These computer readable program instructions may also be stored in a computer readable storage medium that can direct a computer, a programmable data processing apparatus, and/or other devices to function in a particular manner, such that the computer readable storage medium having instructions stored therein comprises an article of manufacture including instructions which implement aspects of the function/act specified in the flowchart and/or block diagram block or blocks.
  • The computer readable program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other device to cause a series of operational steps to be performed on the computer, other programmable apparatus or other device to produce a computer implemented process, such that the instructions which execute on the computer, other programmable apparatus, or other device implement the functions/acts specified in the flowchart and/or block diagram block or blocks.
  • The flowchart and block diagrams in the Figures illustrate the architecture, functionality, and operation of possible implementations of systems, methods, and computer program products according to various embodiments of the present invention. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executable instructions for implementing the specified logical function(s). In sonic alternative implementations, the functions noted in the block ma occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes, be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems that perform the specified functions or acts or carry out combinations of special purpose hardware and computer instructions.
  • While the embodiments of the present invention have been described, the technical scope of the invention is not limited to the above described embodiments. It is apparent to persons skilled in the art that various alterations and improvements can be added to the above-described embodiments. It is also apparent from the scope of the claims that the embodiments added with such alterations or improvements can be included in the technical scope of the invention.
  • The operations, procedures, steps, and stages of each process performed by an apparatus, system, program, and method shown in the claims, embodiments, or diagrams can, be performed in any order as long as the order is not indicated by “prior to,” “before,” or the like and as long as the output from a previous process is not used in a later process. Even if the process flow is described. using phrases such as “first” or “next” in the claims, embodiments, or diagrams, it does not necessarily mean that the process must he performed in this order.
  • As made dear from the above, with the embodiments of the present invention, it is possible to perform learning while preventing overfitting and a decrease in learning accuracy in a predictive model based on a data sequence, and to it improve the expressive power and learning power of a learning apparatus.

Claims (15)

1. A computer-implemented method for learning a first model, comprising:
generating, by a processor, a second model based on the first model, the first model being configured to perform a learning process based on sequentially inputting each of a plurality of pieces of input data that include a plurality of input values and that are from a first input data sequence, the second model being configured to learn a first learning target parameter included in the first model based on inputting, in an order differing from an order in the first model, each of a plurality of pieces of input data that include a plurality of input values and are from a second input data sequence;
performing, by the processor, a learning process using both the first model and the second model; and
storing, in a memory device, the first model that has been learned.
2. The computer-implemented method of claim 1, wherein the storing the first model that has been learned includes deleting, from the memory device, the second model that has been learned and outputting the first model that has been learned as a predictive model based on an input data sequence.
3. The computer-implemented method of claim 2, wherein the generating the second model includes generating the second model for learning the learning target parameter by inputting, in a backwards order, each of the plurality of pieces of input data from the second input data sequence.
4. The computer-implemented method of claim 3, wherein the first input data sequence and the second input data sequence are time-series input data sequences, wherein the first model inputs the first input data sequence in order from older to newer ones of the plurality of pieces of input data, and wherein the second model inputs the second input data sequence in order from newer to older ones of the plurality of pieces of input data.
5. The computer-implemented method of claim 3, wherein the first model and the second model each include the first learning target parameter and a second learning target parameter, and wherein the performing the learning process includes:
learning the second learning target parameter by using the first model without changing the first learning target parameter, and
learning the first learning target parameter by using the second model without changing the second learning target parameter.
6. The computer-implemented method of claim 5, wherein the first learning target parameter is operable to be learned with higher accuracy by learning using the second model than by learning using the first model, and wherein the second learning target parameter is operable to be learned with higher accuracy by learning using the first model than by learning using the second model.
7. The computer-implemented method of claim 3, wherein the first input data sequence and the second input data sequence are at least partially identical.
8. The computer-implemented method of claim 3, wherein the first input data sequence and the second input data sequence are input data sequences for learning that are different from each other and included in a plurality of input data sequences for learning.
9. The computer-implemented method of claim 3, wherein the performing the learning process includes performing the learning process with the first model a greater number of times than the learning process with the second model.
10. The computer-implemented method of claim 3, wherein the performing the learning process includes performing the learning process with the first model using a higher learning rate than is used for the learning process with the second model.
11. The computer-implemented method of claim 3, wherein the performing the learning process includes obtaining the first model that has been learned by performing the learning process with the first model last.
12. The computer-implemented method of claim 4, wherein
the first model includes a plurality of input nodes that sequentially input a plurality of input values at each time point of the first input data sequence, and a weight parameter between each input node and each input value at a time point before a time point corresponding to the plurality of input nodes, and
the second model includes a plurality of input nodes that input, in a backwards order, a plurality of input values at each time point of the second input data sequence, and a weight parameter between each input node and each input value at a time point after the time point corresponding to the plurality of input nodes.
13. The computer-implemented method of claim 12, wherein
the first model further includes a weight parameter between each input node and each of a plurality of hidden nodes corresponding to the time point before the time point corresponding to the plurality of input nodes, and a weight parameter between each hidden node and each input value corresponding to the time point before the time point corresponding to the plurality of input nodes, and
the second model further includes a weight parameter between each input node and each of a plurality of hidden nodes corresponding to the time point after the time point corresponding to the plurality of input nodes, and a weight parameter between each hidden node and each input value corresponding to the time point after the time point corresponding to the plurality of input nodes.
14. The computer-implemented method of claim 13, wherein the performing the learning process includes:
learning the weight parameter between each hidden node and each input value corresponding to the time point before the time point corresponding to the plurality of input nodes in the first mode, using the learning process with the first model; and
learning the weight parameter between each input node and each of the plurality of hidden nodes corresponding to the time point after the time point corresponding to the plurality of input nodes in the second model, using the learning process with the second model.
15-20. (canceled)
US15/403,986 2017-01-11 2017-01-11 Learning apparatus and method for bidirectional learning of predictive model based on data sequence Abandoned US20180197080A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US15/403,986 US20180197080A1 (en) 2017-01-11 2017-01-11 Learning apparatus and method for bidirectional learning of predictive model based on data sequence
US15/804,702 US20180197082A1 (en) 2017-01-11 2017-11-06 Learning apparatus and method for bidirectional learning of predictive model based on data sequence
PCT/IB2017/057999 WO2018130890A1 (en) 2017-01-11 2017-12-15 Learning apparatus and method for bidirectional learning of predictive model based on data sequence
CN201780083071.4A CN110168577A (en) 2017-01-11 2017-12-15 The learning device and method of the two-way study of prediction model based on data sequence
JP2019535257A JP7123938B2 (en) 2017-01-11 2017-12-15 LEARNING APPARATUS, METHOD AND COMPUTER PROGRAM FOR BIDIRECTIONAL LEARNING OF PREDICTION MODEL BASED ON DATA SEQUENCE

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US15/403,986 US20180197080A1 (en) 2017-01-11 2017-01-11 Learning apparatus and method for bidirectional learning of predictive model based on data sequence

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US15/804,702 Continuation US20180197082A1 (en) 2017-01-11 2017-11-06 Learning apparatus and method for bidirectional learning of predictive model based on data sequence

Publications (1)

Publication Number Publication Date
US20180197080A1 true US20180197080A1 (en) 2018-07-12

Family

ID=62783135

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/403,986 Abandoned US20180197080A1 (en) 2017-01-11 2017-01-11 Learning apparatus and method for bidirectional learning of predictive model based on data sequence
US15/804,702 Pending US20180197082A1 (en) 2017-01-11 2017-11-06 Learning apparatus and method for bidirectional learning of predictive model based on data sequence

Family Applications After (1)

Application Number Title Priority Date Filing Date
US15/804,702 Pending US20180197082A1 (en) 2017-01-11 2017-11-06 Learning apparatus and method for bidirectional learning of predictive model based on data sequence

Country Status (4)

Country Link
US (2) US20180197080A1 (en)
JP (1) JP7123938B2 (en)
CN (1) CN110168577A (en)
WO (1) WO2018130890A1 (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107368476B (en) * 2017-07-25 2020-11-03 深圳市腾讯计算机系统有限公司 Translation method, target information determination method and related device
CN109410198B (en) * 2018-10-25 2022-04-22 北京奇艺世纪科技有限公司 Time sequence action detection method, device and equipment
JP2021140303A (en) * 2020-03-03 2021-09-16 トヨタ自動車株式会社 Learning method, learning system and learning program

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8489529B2 (en) * 2011-03-31 2013-07-16 Microsoft Corporation Deep convex network with joint use of nonlinear random projection, Restricted Boltzmann Machine and batch-based parallelizable optimization
US9875445B2 (en) * 2014-02-25 2018-01-23 Sri International Dynamic hybrid models for multimodal analysis
WO2015193531A1 (en) * 2014-06-16 2015-12-23 Nokia Technologies Oy Data processing
JP5943358B2 (en) * 2014-09-30 2016-07-05 インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation Learning device, processing device, prediction system, learning method, processing method, and program
CN106127806B (en) * 2016-06-17 2018-10-02 淮阴工学院 RGB-D video target tracking methods based on depth Boltzmann machine cross-module formula feature learning

Also Published As

Publication number Publication date
WO2018130890A1 (en) 2018-07-19
JP2020505672A (en) 2020-02-20
JP7123938B2 (en) 2022-08-23
US20180197082A1 (en) 2018-07-12
CN110168577A (en) 2019-08-23

Similar Documents

Publication Publication Date Title
US11842281B2 (en) Reinforcement learning with auxiliary tasks
US11783182B2 (en) Asynchronous deep reinforcement learning
US11100388B2 (en) Learning apparatus and method for learning a model corresponding to real number time-series input data
US11449731B2 (en) Update of attenuation coefficient for a model corresponding to time-series input data
US10783452B2 (en) Learning apparatus and method for learning a model corresponding to a function changing in time series
US10909451B2 (en) Apparatus and method for learning a model corresponding to time-series input data
US9547818B2 (en) Apparatus and method for learning a model corresponding to time-series moving image/video input data
US20200175364A1 (en) Training action selection neural networks using a differentiable credit function
JP2020521205A (en) Multi-task neural network system with task-specific and shared policies
CN111340221B (en) Neural network structure sampling method and device
CN111598253A (en) Training machine learning models using teacher annealing
US20200327450A1 (en) Addressing a loss-metric mismatch with adaptive loss alignment
US20180197082A1 (en) Learning apparatus and method for bidirectional learning of predictive model based on data sequence
US11195116B2 (en) Dynamic boltzmann machine for predicting general distributions of time series datasets
US11755946B2 (en) Cumulative reward predictor training
US11410042B2 (en) Dynamic Boltzmann machine for estimating time-varying second moment
US20240144015A1 (en) Reinforcement learning with auxiliary tasks
US11893480B1 (en) Reinforcement learning with scheduled auxiliary control
CN113469204A (en) Data processing method, device, equipment and computer storage medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KAJINO, HIROSHI;OSOGAMI, TAKAYUKI;REEL/FRAME:040975/0198

Effective date: 20161121

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION