US20230385606A1 - System and method for state estimation in a noisy machine-learning environment - Google Patents

System and method for state estimation in a noisy machine-learning environment Download PDF

Info

Publication number
US20230385606A1
US20230385606A1 US18/187,860 US202318187860A US2023385606A1 US 20230385606 A1 US20230385606 A1 US 20230385606A1 US 202318187860 A US202318187860 A US 202318187860A US 2023385606 A1 US2023385606 A1 US 2023385606A1
Authority
US
United States
Prior art keywords
estimate
measurement
filter
state
prediction
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/187,860
Inventor
Randal Allen
John P. Volpi
Samuel Vasta
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Incucomm Inc
Original Assignee
Incucomm Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Incucomm Inc filed Critical Incucomm Inc
Priority to US18/187,860 priority Critical patent/US20230385606A1/en
Assigned to INCUCOMM, INC. reassignment INCUCOMM, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: VOLPI, JOHN P., ALLEN, RANDAL, Vasta, Samuel
Publication of US20230385606A1 publication Critical patent/US20230385606A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/02Comparing digital values
    • G06F7/023Comparing digital values adaptive, e.g. self learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/30Monitoring
    • G06F11/34Recording or statistical evaluation of computer activity, e.g. of down time, of input/output operation ; Recording or statistical evaluation of user activity, e.g. usability assessment
    • G06F11/3452Performance evaluation by statistical analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/11Complex mathematical operations for solving equations, e.g. nonlinear equations, general mathematical optimization problems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • G06F30/27Design optimisation, verification or simulation using machine learning, e.g. artificial intelligence, neural networks, support vector machines [SVM] or training a model
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F7/00Methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F7/22Arrangements for sorting or merging computer data on continuous record carriers, e.g. tape, drum, disc
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/042Knowledge-based neural networks; Logical representations of neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/10Interfaces, programming languages or software development kits, e.g. for simulating neural networks
    • G06N3/105Shells for specifying net layout
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N5/00Computing arrangements using knowledge-based models
    • G06N5/01Dynamic search techniques; Heuristics; Dynamic trees; Branch-and-bound
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/14Fourier, Walsh or analogous domain transformations, e.g. Laplace, Hilbert, Karhunen-Loeve, transforms
    • G06F17/141Discrete Fourier transforms
    • G06F17/142Fast Fourier transforms, e.g. using a Cooley-Tukey type algorithm
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2111/00Details relating to CAD techniques
    • G06F2111/10Numerical modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2207/00Indexing scheme relating to methods or arrangements for processing data by operating upon the order or content of the data handled
    • G06F2207/38Indexing scheme relating to groups G06F7/38 - G06F7/575
    • G06F2207/48Indexing scheme relating to groups G06F7/48 - G06F7/575
    • G06F2207/4802Special implementations
    • G06F2207/4818Threshold devices
    • G06F2207/4824Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/048Activation functions
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Definitions

  • the present disclosure is directed, in general, to state tracking systems and, more specifically, to a system and method for estimating the state of a system in a noisy measurement environment.
  • Table 1 (below) provides a sampling of the state of the art of machine learning methods, where each method has its own set of implementation requirements; it was presented by Anais Dotis-Georgiou at the Big Data and Artificial Intelligence Conference in 2019 at Addison, TX.
  • supervised learning the designer is required to manually select features, choose the classifier method, and tune the hyperparameters.
  • some algorithms e.g., k-means, k-medoid, and fuzzy c-means
  • principal component analysis requires the data to be scaled, assumes the data is orthogonal, and results in linear correlation
  • nonnegative matrix factorization requires normalization of the data
  • factor analysis is subject to interpretation.
  • Deep learning brings with it its own set of demands. Enormous computing power through high performance graphics processing units (GPUs) is needed to process big data, on the order of 10 5 to 10 6 points. Also, the data must be numerically tagged. Furthermore, it takes a long time to train a model. In the end, because of the depth of complexity, it is virtually impossible to understand how conclusions were reached.
  • GPUs graphics processing units
  • ANN artificial neural network
  • back propagation is an iterative gradient method to reduce the error between the input and output data.
  • humans do not back-propagate when learning, so the analogy is weak in that regard.
  • Other drawbacks include the following:
  • FIG. 1 reveals machine learning forecasting performance is worse than the application of statistical methods, when comparing the vertical axis which represents the symmetric mean absolute percentage error of the various methods.
  • FIG. 1 shows exponential smoothing (Error-Trend-Seasonality, “ETS”) as being superior to all others from a study. Noticeably absent from the field of comparison are optimal estimators, e.g., linear quadratic estimators, commonly called Kalman filters. It is believed that the prior art approaches have not included applying optimal estimation techniques to predictive or prescriptive analytics, let alone combining all three techniques: filtering, smoothing, and predicting. This is reasonable since the Kalman filter (without smoothing or predicting), and its variants, is typically used for state estimation in many control system applications.
  • ETS Error-Trend-Seasonality
  • a system is needed which improves upon machine learning and statistical methods such that the practitioner can perform real-time predictive and prescriptive analytics.
  • the system should avoid the pitfalls of artificial neural networks with their arbitrary hidden layers, iterative feature and method selection, and hyperparameter tuning.
  • the system should not require enormous computing power.
  • such a system will overcome state estimation challenges in a noisy measurement environment.
  • the method includes making a first measurement of a value of a characteristic of a state of a system, and a second measurement of a value of the characteristic of the state of the system after the first measurement. While the second measurement is the last measurement in this example, there may be a plurality of measurements followed by the following steps that apply to the corresponding measurements. After the measurements have been taken, the method carries out a filtering process.
  • the method also includes constructing a first filter measurement estimate after the second measurement coinciding with the first measurement including a first filter measurement covariance matrix describing an accuracy of the first filter measurement estimate, and constructing a first filter time estimate after the first filter measurement estimate including a first filter time covariance matrix describing an accuracy of the first filter time estimate employing a dynamic model of the state of the system.
  • the method also includes constructing a second filter measurement estimate after the first filter time estimate coinciding with the second measurement including a second filter measurement covariance matrix describing an accuracy of the second filter measurement estimate, and constructing a second filter time estimate after the second filter measurement estimate including a second filter time covariance matrix describing an accuracy of the second filter time estimate employing the dynamic model of the state of the system.
  • the method After the filtering process, the method carries out a smoothing process.
  • the method includes constructing a smoothing estimate from the first filter measurement estimate and the second filter measurement estimate.
  • the smoothing estimate may be obtained by sweeping backward recursively from the second filter measurement estimate to the first filter measurement estimate.
  • the method carries out a prediction.
  • the method includes constructing a first prediction estimate after the smoothing estimate that provides a forecast of a value of the characteristic of the state of the system including a first prediction covariance matrix describing an accuracy of the first prediction estimate employing the dynamic model of the state of the system.
  • the method can carry out a plurality of prediction estimates providing corresponding forecasts of a value of the characteristic of the state of the system.
  • FIG. 1 illustrates a graphical representation comparing forecasting performance (Symmetric Mean Absolute Percentage Error) of machine learning and statistical methods for known machine learning processes;
  • FIG. 2 illustrates a process for smoothing over fixed time intervals
  • FIG. 3 illustrates a graphical representation of filtering, smoothing, and prediction performance for analytics
  • FIGS. 4 A and 4 B illustrate performance examples of temperature tracking and vibration filtering in a noisy measurement environment
  • FIG. 5 illustrates a flow diagram of an embodiment of a method of estimating the state of a system
  • FIG. 6 illustrates a block diagram of an embodiment of an apparatus for estimating the state of a system.
  • Intelligent prediction is a system introduced herein that uniquely combines the three forms of optimal estimation (filtering, smoothing, and predicting) to provide utility for predictive and prescriptive analytics with applications to real-time sensor data.
  • the results of this systematic approach outperform the best of the current statistical methods and, as such, outperform machine learning methods.
  • the system architecture introduced herein is based on combining three types of estimation: filtering, smoothing, and predicting, which is an approach new to forecasting, and which it is believed has not been previously considered with regard to machine learning, as illustrated in Table 1.
  • FIG. 1 illustrates on the vertical axis a symmetric mean absolute percentage error for forecasting performance of machine learning and statistical methods for known machine learning processes.
  • the first part of the combined three-part system is an implementation of a discrete-time Kalman filter.
  • the filtering form of optimal estimation is when an estimate coincides with the last measurement point.
  • K k P k ⁇ H k T ( H k P k ⁇ H k T +R k ) ⁇ 1 Filtering gain
  • ⁇ circumflex over (x) ⁇ k ⁇ circumflex over (x) ⁇ k +K k ( z k ⁇ H k ⁇ circumflex over (x) ⁇ k ⁇ )State measurement estimate
  • C. F. van Loan's method is employed to compute ⁇ k and Q k .
  • the Kalman filter has numerous applications for guidance, navigation, and control of aerospace vehicles, e.g., aircraft, spacecraft, rockets, and missiles.
  • the filter will be combined, as introduced herein, with smoothing and predicting with applications to (possibly real-time) predictive and prescriptive analytics.
  • the second part of the three-part system is an implementation of discrete fixed-interval smoothing.
  • the smoothing form of optimal estimation is when an estimate falls within a span of measurement points.
  • the time interval of the measurements is fixed (hence the name) and optimal estimates of the (saved) states ⁇ circumflex over (x) ⁇ k are obtained.
  • x k s ⁇ circumflex over (x) ⁇ k +C k ( x k+1 s ⁇ k ⁇ circumflex over (x) ⁇ k )State smoothing estimate
  • the third part of the three-part system is an implementation of a predictor.
  • the predicting form of optimal estimation is when an estimate falls beyond the last measurement point.
  • the equations for the predictor are identical to the filter with the following three exceptions:
  • the three-step smoother Upon saving the state estimates ⁇ circumflex over (x) ⁇ k , the state covariances P k , and properly initializing x T s and P T s with the last entries of ⁇ circumflex over (x) ⁇ k (T f ) and P k (T f ), the three-step smoother iterates recursively with a backward sweep to an earlier time point, as illustrated in FIG. 2 , showing fixed-intervals smoothing, until all states/covariances are consumed. At this point, the system is prepared for predictive and prescriptive analytics.
  • the last state estimate and state covariance of the smoother is used to initialize the predictor.
  • the predictor runs just like the five-step Kalman filter with two exceptions: (i) the covariance of the measurement noise R k is set to an arbitrarily large value to indicate the measurements are worthless, because there are not any, and (ii) the measurement z k is fixed to its final value, because that is the last piece of information available.
  • FIG. 3 An application of the disclosed three-part system's implementation is shown in FIG. 3 illustrating an example of filtering, smoothing, and predicting for analytics where the future position of a moving object is predicted.
  • the actual position of an object is depicted on the vertical axis with a dashed line
  • the filtered position is represented by dots
  • the smoothed position is denoted by a solid line
  • the predicted position is shown with “x” marking the spots.
  • Measurements are filtered and smoothed up until 50 seconds, at which time, predictions are made 30 seconds into the future.
  • HPS Horizontal Pump System
  • Each data set includes noise which may vary with time.
  • the different sources of data provide a mixture of different characteristics such as seasonality, trends, impulses, and randomness. For example, temperature data is affected by the day/night (diurnal) cycle which creates a (short) seasonal characteristic. Vibration data, however, is not affected by the day/night cycle and is not seasonal but does contain a significant portion of randomness.
  • a missed prediction occurs when an observed measurement exceeds a threshold value, but no forecast was produced which predicted the exception. Any forecast which predicted the exception within twelve periods leading up to the exception was not considered because such a short forecast is not useful. A prediction strategy should produce as few missed predictions as possible.
  • Table 2 illustrated are temperature and vibration filtering after filtering and predicting, showing sensitivities, forecast lengths, and average percent of missed predictions.
  • the filter sensitivity column refers to how closely the signal is being tracked. For instance, temperature changes slowly over time so the filter/prediction combination is set to high sensitivity to track the slowly changing signal; whereas vibration, which contains high frequency noise is set to low sensitivity, as illustrated in FIG. 4 showing performance examples of temperature tracking and vibration filtering in a noisy measurement environment.
  • the Average % Missed Predictions column in Table 2 shows the probability that, given the data has crossed a critical threshold, the corresponding strategy failed to predict the event. A lower percentage indicates better performance for this metric. In each scenario (low noise/high noise), the filter/predictor strategy with a high or medium sensitivity correctly predicted every event. Even in the case of low sensitivity, the filter/predictor strategy outperformed the ETS strategy.
  • the filtering, smoothing, and predicting process introduced herein outperforms the ETS strategy which was the basis of performance assessment over machine learning strategies as shown in Table 1. These results appear to be independent of sensitivity setting (low, medium, or high). Therefore, in general, a practitioner could use the filter/predictor strategy to avoid missed predictions. Furthermore, with the inclusion of smoothing, these results are improved upon as shown in FIG. 3 .
  • FIG. 5 illustrated is a flow diagram of an embodiment of a method 500 of estimating a state of a system.
  • the method 500 may be employable to estimate a state of a system in a machine learning and/or noisy measurement environment.
  • the method 500 is operable on a processor such as a microprocessor coupled to a memory, the memory contained instructions which, when executed by the processor, are operative to perform the functions.
  • the method 500 begins at a start step or module 505 .
  • a first estimate of a state of a system is constructed at a first time including a first covariance matrix describing an accuracy of the first estimate.
  • a second estimate of the state of said system is constructed at a second time, after the first time, including a second covariance matrix describing an accuracy of the second estimate employing a dynamic model of the state of the system; the dynamic model comprises a matrix with coefficients that describes a temporal evolution of the state of the system.
  • a value of a characteristic of the state of the system is measured at the second time. Measuring the value of the characteristic can include making a plurality of independent measurements characterized by a diagonal measurement covariance matrix.
  • the second estimate of the state of the system and the second covariance matrix are adjusted based on the value of the characteristic.
  • a third estimate of the state of the system is constructed at a third time, before the second time, including a third covariance matrix describing an accuracy of the third estimate employing the dynamic model of the state of the system.
  • a fourth estimate of the state of the system is constructed at a fourth time, after the second time, from the second estimate.
  • the fourth time is on a different time scale from the first, second and third times.
  • the dynamic model is altered in response to the value of the characteristic.
  • the state of the system is reported based on the fourth estimate.
  • a fifth estimate of the state of the system is constructed at a fifth time, after the second time, from the second estimate.
  • the dynamic model is a linear dynamic model with constant coefficients.
  • constructing the first estimate and constructing the second estimate are performed by a Kalman filter.
  • the state of the system is altered based on the fourth estimate.
  • the method 500 terminates at end step or module 598 .
  • FIG. 6 illustrated is a block diagram of an embodiment of an apparatus 600 for estimating the state of a system in a machine learning environment.
  • the apparatus 600 is configured to perform functions described hereinabove of constructing the estimate of the state of the system.
  • the apparatus 600 includes a processor (or processing circuitry) 610 , a memory 620 and a communication interface 630 such as a graphical user interface.
  • the functionality of the apparatus 600 may be provided by the processor 610 executing instructions stored on a computer-readable medium, such as the memory 620 shown in FIG. 6 .
  • Alternative embodiments of the apparatus 600 may include additional components (such as the interfaces, devices and circuits) beyond those shown in FIG. 6 that may be responsible for providing certain aspects of the device's functionality, including any of the functionality to support the solution described herein.
  • the processor 610 (or processors), which may be implemented with one or a plurality of processing devices, perform functions associated with its operation including, without limitation, performing the operations of estimating the state of a system, computing covariance matrices, and estimating a future state of the system.
  • the processor 610 may be of any type suitable to the local application environment, and may include one or more of general-purpose computers, special purpose computers, microprocessors, digital signal processors (“DSPs”), field-programmable gate arrays (“FPGAs”), application-specific integrated circuits (“ASICs”), and processors based on a multi-core processor architecture, as non-limiting examples.
  • the processor 610 may include, without limitation, application processing circuitry.
  • the application processing circuitry may be on separate chipsets.
  • part or all of the application processing circuitry may be combined into one chipset, and other application circuitry may be on a separate chipset.
  • part or all of the application processing circuitry may be on the same chipset, and other application processing circuitry may be on a separate chipset.
  • part or all of the application processing circuitry may be combined in the same chipset.
  • the memory 620 may be one or more memories and of any type suitable to the local application environment, and may be implemented using any suitable volatile or nonvolatile data storage technology such as a semiconductor-based memory device, a magnetic memory device and system, an optical memory device and system, fixed memory and removable memory.
  • the programs stored in the memory 620 may include program instructions or computer program code that, when executed by an associated processor, enable the respective apparatus 600 to perform its intended tasks.
  • the memory 620 may form a data buffer for data transmitted to and from the same.
  • Exemplary embodiments of the system, subsystems, and modules as described herein may be implemented, at least in part, by computer software executable by the processor 610 , or by hardware, or by combinations thereof.
  • the communication interface 630 modulates information for transmission by the respective apparatus 600 to another apparatus.
  • the respective communication interface 630 is also configured to receive information from another processor for further processing.
  • the communication interface 630 can support duplex operation for the respective other processor 600 .
  • the inventions disclosed herein combine three techniques of optimal estimation of the state of a system.
  • the three techniques include filtering, smoothing, and predicting processes, and can be performed, without limitation, in a machine learning and/or a noisy measurement environment.
  • the filtering portion of optimal estimation is performed to construct a first estimate of a state vector x k at a time point t k that coincides with a measurement of a value of characteristic of the state of the system at the time point t k .
  • the filtering process employs a covariance matrix that describes the accuracy of the first estimate of the state vector x k at the time point t k .
  • a second estimate of the state vector x k+1 at the time point t k+1 is then constructed by propagating the state of the system forward to a second time point t k+1 , the second time point being after the first time point.
  • the propagating forward employs a dynamic model of the state of the system to produce the estimate of the state vector x k+1 at the second time point t k+1 .
  • the first estimate of the state vector x k and constructing the second estimate of the state vector x k+1 can be performed by employing a Kalman filter.
  • the dynamic model can employ a matrix with coefficients that describes temporal evolution of the state of the system.
  • the dynamic model is a linear dynamic model with constant coefficients.
  • a value of a characteristic of the state of the system x k+1 is measured at the second time point t k+1 .
  • the second estimate of the state of the system and the second covariance matrix are adjusted based on the measured value of the characteristic at the second time point t k+1
  • Measuring the value of the characteristic can include making a plurality of independent measurements characterized by a diagonal measurement covariance matrix.
  • the smoothing portion of optimal estimation is performed by constructing a third state estimate for a time point that is earlier than the time point t k+1 .
  • the earlier time point can fall within or before a span of current measurement points, e.g., between or before the time points t k and t k+1 .
  • the predicting portion then propagates the state estimate forward for a forecast period of interest.
  • the last state estimate and state covariance of the smoother can be used to initialize the predicting.
  • the predictions may be at various time points in the future and over various time scales that are after the second time point.
  • Measurement noise R k can be set to an arbitrarily large value to accommodate the inherent absence of a state measurement at a future time point.
  • the initial conditions for the prediction can be taken as the last a posteriori state estimate and the covariance of the smoother.
  • the exemplary embodiments provide both a method and corresponding apparatus consisting of various modules providing functionality for performing the steps of the method.
  • the modules may be implemented as hardware (embodied in one or more chips including an integrated circuit such as an application specific integrated circuit), or may be implemented as software or firmware for execution by a processor.
  • firmware or software the exemplary embodiments can be provided as a computer program product including a computer readable storage medium embodying computer program code (i.e., software or firmware) thereon for execution by the computer processor.
  • the computer readable storage medium may be non-transitory (e.g., magnetic disks; optical disks; read only memory; flash memory devices; phase-change memory) or transitory (e.g., electrical, optical, acoustical or other forms of propagated signals-such as carrier waves, infrared signals, digital signals, etc.).
  • the coupling of a processor and other components is typically through one or more busses or bridges (also termed bus controllers).
  • the storage device and signals carrying digital traffic respectively represent one or more non-transitory or transitory computer readable storage medium.
  • the storage device of a given electronic device typically stores code and/or data for execution on the set of one or more processors of that electronic device such as a controller.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • Pure & Applied Mathematics (AREA)
  • Computational Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Algebra (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computer Hardware Design (AREA)
  • Medical Informatics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Databases & Information Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Operations Research (AREA)
  • Geometry (AREA)
  • Quality & Reliability (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)
  • Image Analysis (AREA)
  • Complex Calculations (AREA)

Abstract

A system and method for estimating a system state. The method includes making a first measurement and a second measurement of a value of a characteristic of a state of a system. The method includes constructing first filter measurement and time estimates after the second measurement coinciding with the first measurement including corresponding covariance matrices describing an accuracy of the first filter measurement and time estimates. The method includes constructing second filter measurement and time estimates coinciding with the second measurement including corresponding covariance matrices describing an accuracy of the second filter measurement and time estimates. The method includes constructing a smoothing estimate from the first and second filter measurement estimates. The method includes constructing a first prediction estimate that provides a forecast of a value of the characteristic of the state of the system including a first prediction covariance matrix describing an accuracy of the first prediction estimate.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application is a Continuation of U.S. patent application Ser. No. 16/674,848 entitled “System And Method For State Estimation In A Noisy Machine-Learning Environment” filed on Nov. 5, 2019 which claims benefit to U.S. Provisional Patent Application No. 62/756,044, entitled “Hybrid AI,” filed Nov. 5, 2018, which is incorporated herein by reference.
  • This application is related to U.S. application Ser. No. 15/611,476 entitled “PREDICTIVE AND PRESCRIPTIVE ANALYTICS FOR SYSTEMS UNDER VARIABLE OPERATIONS,” filed Jun. 1, 2017, which is incorporated herein by reference. (INC-026)
  • This application is related to U.S. Provisional Application No. 62/627,644 entitled “DIGITAL TWINS, PAIRS, AND PLURALITIES,” filed Feb. 7, 2018, converted to U.S. application Ser. No. 16/270,338 entitled “SYSTEM AND METHOD THAT CHARACTERIZES AN OBJECT EMPLOYING VIRTUAL REPRESENTATIONS THEREOF,” filed Feb. 7, 2019, which are incorporated herein by reference. (INC-030)
  • This application is related to U.S. application Ser. No. 16/674,885 (Attorney Docket No. INC-031B), entitled “SYSTEM AND METHOD FOR ADAPTIVE OPTIMIZATION,” filed Nov. 5, 2019, U.S. application Ser. No. 16/674,942 (Attorney Docket No. INC-031C), entitled “SYSTEM AND METHOD FOR CONSTRUCTING A MATHEMATICAL MODEL OF A SYSTEM IN AN ARTIFICIAL INTELLIGENCE ENVIRONMENT,” filed Nov. 5, 2019, and U.S. application Ser. No. 16/675,000 (Attorney Docket No. INC-031D, entitled “SYSTEM AND METHOD FOR VIGOROUS ARTIFICIAL INTELLIGENCE,” filed Nov. 5, 2019, which are incorporated herein by reference.
  • RELATED REFERENCES
  • Each of the references cited below are incorporated herein by reference.
  • U. S. Patents:
  • Pat. No. Issue Date Patentee
    10,068,170 Sep. 4, 2018 Golovashkin, et al.
  • U.S. Patent Application Publications:
  • Publication Kind Publication
    Number Code Date Applicant
    20190036639 A1 Jan. 31, 2019 Huang; Yan; et al.
  • Nonpatent Literature Documents:
    • Makridakis, S. et al., “Statistical and Machine Learning Forecasting Methods: Concerns and Ways Forward” (2018)
    • Box, G. E. P. et al., “Time Series Analysis Forecasting and Control” (2016)
    • Hyndman, R. J. and Athanasopoulos, G., “Forecasting: Principles and Practice” (2014)
    • Brown, R. G. and Hwang, P. Y. C., “Introduction to Random Signals and Applied Kalman Filtering with MATLAB Exercises” (2012)
    • Grewal, M. S. and Andrews, A. P., “Kalman filtering: Theory and Practice Using MATLAB” (2008)
    • Jategaonkar, R., “Flight Vehicle System Identification: A Time Domain Methodology” (2006)
    • Simon, D., “Optimal State Estimation: Kalman, H-infinity, and Nonlinear Approaches” (2006)
    • Zarchan, P., “Fundamentals of Kalman Filtering: A Practical Approach” (2005)
    • Desat, U. B. et al., “Discrete-Time Complementary Models and Smoothing Algorithms: The Correlated Noise Case” (1983)
    • Van Loan, C. F., “Computing Integrals Involving the Matrix Exponential” (1978)
    • Gelb, A., “Applied Optimal Estimation” (1974)
    • Fraserand, D. C. and Potter, J. E., “The Optimum Linear Smoother as a Combination of Two Linear Filters” (1969)
    • Meditch, J. S., “Stochastic Optimal Linear Estimation and Control” (1969)
    • Raush, H. E. et al., “Maximum Likelihood Estimates of Linear Dynamic Systems” (1965)
    • Kalman, R. E., “A New Approach to Linear Filtering and Prediction Problems” (1960)
    TECHNICAL FIELD
  • The present disclosure is directed, in general, to state tracking systems and, more specifically, to a system and method for estimating the state of a system in a noisy measurement environment.
  • BACKGROUND
  • Niels Bohr is often quoted as saying, “Prediction is very difficult, especially about the future.” Prediction is the practice of extracting information from sets of data to identify patterns and predict future behavior. The Institute for Operations Research and the Management Sciences (INFORMS) defines several types of prediction which have been expanded here for clarity:
      • Descriptive Analytics, in which enormous amounts of historical (big) data are used for describing patterns extracted solely from the data.
      • Predictive Analytics, where, in addition to descriptive analytics, subject matter expert (SME) knowledge is included to capture attributes not reflected in the big data by itself.
      • Decision Analytics, where, in addition to predictive analytics, influence diagrams (mathematical networks or models) are architected to address decision-making strategy.
      • Prescriptive Analytics, where, in addition to decision analytics, advanced mathematical techniques (e.g., optimization) are leveraged to predict missing data. The techniques of prescriptive analytics include data modeling/mining, artificial intelligence (AI), supervised or unsupervised machine learning, and deep learning. The focus of the present disclosure is with machine and deep learning with the intent to train a network or characterize a model for prediction about the future as new data is assimilated.
  • Table 1 (below) provides a sampling of the state of the art of machine learning methods, where each method has its own set of implementation requirements; it was presented by Anais Dotis-Georgiou at the Big Data and Artificial Intelligence Conference in 2019 at Addison, TX.
  • TABLE 1
    Soft Hard
    Regression Classification Clustering Clustering
    Ensemble Decision Fuzzy-C Hierarchical
    methods trees means clustering
    Gaussian Discriminant Gaussian K-
    process analysis mixture means
    General linear K-nearest K-
    model neighbor medoids
    Linear Logistic Self-organizing
    regression regression maps
    Nonlinear naïve
    regression Bayes
    Regression Neural
    tree nets
    Support vector Support vector
    machine machine
  • In the case of supervised learning, the designer is required to manually select features, choose the classifier method, and tune the hyperparameters. In the case of unsupervised learning, some algorithms (e.g., k-means, k-medoid, and fuzzy c-means) require the number of clusters to be selected a priori; principal component analysis requires the data to be scaled, assumes the data is orthogonal, and results in linear correlation; nonnegative matrix factorization requires normalization of the data; and factor analysis is subject to interpretation.
  • Deep learning brings with it its own set of demands. Enormous computing power through high performance graphics processing units (GPUs) is needed to process big data, on the order of 105 to 106 points. Also, the data must be numerically tagged. Furthermore, it takes a long time to train a model. In the end, because of the depth of complexity, it is virtually impossible to understand how conclusions were reached.
  • The artificial neural network (ANN) architecture supporting machine/deep learning is supposedly inspired by the biologic nervous system. The model learns through a process called back propagation which is an iterative gradient method to reduce the error between the input and output data. But humans do not back-propagate when learning, so the analogy is weak in that regard. Other drawbacks include the following:
      • ANNs are shallow. While little innate knowledge is required on behalf of the practitioner, architectural selection becomes an exercise in numerical investigation.
      • Arbitrary numbers of hidden layers and nodes comprise the depth of deep learning. The practitioner is often unable to explain why one architecture is used over another, not to mention the practitioner has no control or influence over what is being learned.
      • ANNs are greedy and brittle. Big data (and big computing) is required to train/test the model which often breaks when presented with new data.
      • ANNs are opaque. There is generally a lack of transparency due to the difficulty in understanding the connection between inputs and outputs—especially for deep learning. This unknown opaqueness leaves the practitioner wondering if the architecture can be trusted.
  • FIG. 1 reveals machine learning forecasting performance is worse than the application of statistical methods, when comparing the vertical axis which represents the symmetric mean absolute percentage error of the various methods. FIG. 1 shows exponential smoothing (Error-Trend-Seasonality, “ETS”) as being superior to all others from a study. Noticeably absent from the field of comparison are optimal estimators, e.g., linear quadratic estimators, commonly called Kalman filters. It is believed that the prior art approaches have not included applying optimal estimation techniques to predictive or prescriptive analytics, let alone combining all three techniques: filtering, smoothing, and predicting. This is reasonable since the Kalman filter (without smoothing or predicting), and its variants, is typically used for state estimation in many control system applications.
  • A system is needed which improves upon machine learning and statistical methods such that the practitioner can perform real-time predictive and prescriptive analytics. The system should avoid the pitfalls of artificial neural networks with their arbitrary hidden layers, iterative feature and method selection, and hyperparameter tuning. Furthermore, the system should not require enormous computing power. Preferably, such a system will overcome state estimation challenges in a noisy measurement environment.
  • SUMMARY
  • Deficiencies of the prior art are generally solved or avoided, and technical advantages are generally achieved, by advantageous embodiments of the present disclosure of a system and method for estimating a state of a system. The method includes making a first measurement of a value of a characteristic of a state of a system, and a second measurement of a value of the characteristic of the state of the system after the first measurement. While the second measurement is the last measurement in this example, there may be a plurality of measurements followed by the following steps that apply to the corresponding measurements. After the measurements have been taken, the method carries out a filtering process. The method also includes constructing a first filter measurement estimate after the second measurement coinciding with the first measurement including a first filter measurement covariance matrix describing an accuracy of the first filter measurement estimate, and constructing a first filter time estimate after the first filter measurement estimate including a first filter time covariance matrix describing an accuracy of the first filter time estimate employing a dynamic model of the state of the system. The method also includes constructing a second filter measurement estimate after the first filter time estimate coinciding with the second measurement including a second filter measurement covariance matrix describing an accuracy of the second filter measurement estimate, and constructing a second filter time estimate after the second filter measurement estimate including a second filter time covariance matrix describing an accuracy of the second filter time estimate employing the dynamic model of the state of the system.
  • After the filtering process, the method carries out a smoothing process. The method includes constructing a smoothing estimate from the first filter measurement estimate and the second filter measurement estimate. The smoothing estimate may be obtained by sweeping backward recursively from the second filter measurement estimate to the first filter measurement estimate. After the smoothing process, the method carries out a prediction. The method includes constructing a first prediction estimate after the smoothing estimate that provides a forecast of a value of the characteristic of the state of the system including a first prediction covariance matrix describing an accuracy of the first prediction estimate employing the dynamic model of the state of the system. Of course, the method can carry out a plurality of prediction estimates providing corresponding forecasts of a value of the characteristic of the state of the system.
  • The foregoing has outlined rather broadly the features and technical advantages of the present disclosure in order that the detailed description of the disclosure that follows may be better understood. Additional features and advantages of the disclosed embodiments will be described hereinafter, which form the subject matter of the claims. It should be appreciated by those skilled in the art that the conception and specific embodiment disclosed may be readily utilized as a basis for modifying or designing other structures or processes for carrying out the same purposes of the present disclosure, and that such equivalent constructions do not depart from the spirit and scope of the disclosure as set forth in the appended claims.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • For a more complete understanding of the present disclosure, reference is now made to the following detailed description taken in conjunction with the accompanying drawings, in which:
  • FIG. 1 illustrates a graphical representation comparing forecasting performance (Symmetric Mean Absolute Percentage Error) of machine learning and statistical methods for known machine learning processes;
  • FIG. 2 illustrates a process for smoothing over fixed time intervals;
  • FIG. 3 illustrates a graphical representation of filtering, smoothing, and prediction performance for analytics;
  • FIGS. 4A and 4B illustrate performance examples of temperature tracking and vibration filtering in a noisy measurement environment;
  • FIG. 5 illustrates a flow diagram of an embodiment of a method of estimating the state of a system; and,
  • FIG. 6 illustrates a block diagram of an embodiment of an apparatus for estimating the state of a system.
  • Corresponding numerals and symbols in the different figures generally refer to corresponding parts unless otherwise indicated and, in the interest of brevity, may not be described after the first instance.
  • DETAILED DESCRIPTION
  • The making and using of exemplary embodiments of the disclosed invention are discussed in detail below. It should be appreciated, however, that the general embodiments are provided to illustrate the inventive concepts that can be embodied in a wide variety of specific contexts, and the specific embodiments are merely illustrative of specific ways to make and use the systems, subsystems, and modules for estimating the state of a system in a real-time, noisy measurement, machine-learning environment. While the principles will be described in the environment of a linear system in a real-time machine-learning environment, any environment such as a nonlinear system, or a non-real-time machine-learning environment, is within the broad scope of the disclosed principles and claims.
  • Intelligent prediction is a system introduced herein that uniquely combines the three forms of optimal estimation (filtering, smoothing, and predicting) to provide utility for predictive and prescriptive analytics with applications to real-time sensor data. The results of this systematic approach outperform the best of the current statistical methods and, as such, outperform machine learning methods.
  • To perform predictive and prescriptive analytics in real-time and avoid the pitfalls of machine learning which utilizes artificial neural networks, the system architecture introduced herein is based on combining three types of estimation: filtering, smoothing, and predicting, which is an approach new to forecasting, and which it is believed has not been previously considered with regard to machine learning, as illustrated in Table 1.
  • FIG. 1 , to which reference is now made, illustrates on the vertical axis a symmetric mean absolute percentage error for forecasting performance of machine learning and statistical methods for known machine learning processes. The first part of the combined three-part system is an implementation of a discrete-time Kalman filter. The filtering form of optimal estimation is when an estimate coincides with the last measurement point.
  • With the following definitions:
      • xk=(n×1), state vector at time tk
      • ϕk=(n×n), state transition matrix
      • wk=(n×1), process white noise, wk˜(0,Qk)
      • Qk is the covariance of the process noise
      • zk=(m×1), measurement vector at time tk
      • Hk=(m×n), measurement matrix
      • vk=(m×1), measurement white noise, vk˜(0, Rk)
      • Rk is the covariance of the measurement noise,
        the dynamic process is described by

  • x k+1k x k +w k,

  • measurements are described by

  • z k =H k x k +v k,
  • and initial conditions given by

  • {circumflex over (x)} 0 =E[x 0]

  • P 0 =E[(x 0 −{circumflex over (x)} 0 )(x 0 −{circumflex over (x)} 0 )T].
  • The discrete-time Kalman filter recursive equations are given by

  • K k =P k H k T(H k P k H k T +R k)−1Filtering gain

  • {circumflex over (x)} k ={circumflex over (x)} k +K k(z k −H k {circumflex over (x)} k )State measurement estimate

  • P k=(I−K k H k)P k (I−K k H k)T +K k R k K k T State measurement covariance

  • {circumflex over (x)} k+1 k {circumflex over (x)} kState time estimate at a next time step t k+1

  • P k+1 k P kϕk T +Q k State time covariance(also at next time step)
  • It is worthy to mention C. F. van Loan's method is employed to compute ϕk and Qk. As previously mentioned, the Kalman filter has numerous applications for guidance, navigation, and control of aerospace vehicles, e.g., aircraft, spacecraft, rockets, and missiles. However, the filter will be combined, as introduced herein, with smoothing and predicting with applications to (possibly real-time) predictive and prescriptive analytics.
  • The second part of the three-part system is an implementation of discrete fixed-interval smoothing. The smoothing form of optimal estimation is when an estimate falls within a span of measurement points. For the proposed system, the time interval of the measurements is fixed (hence the name) and optimal estimates of the (saved) states{circumflex over (x)}k are obtained.
  • With initial conditions given by the last a posteriori estimate and covariance from the filter

  • x T s =x r

  • P T s =P T,
  • the smoother sweeps backward recursively

  • C k =P kϕk T(P k+1 )−1Smoothing gain

  • x k s ={circumflex over (x)} k +C k(x k+1 s−ϕk {circumflex over (x)} k)State smoothing estimate

  • P k s =P k +C k(P k+1 s −P k−1 )C k T State smoothing covariance
  • The third part of the three-part system is an implementation of a predictor. The predicting form of optimal estimation is when an estimate falls beyond the last measurement point. The equations for the predictor are identical to the filter with the following three exceptions:
      • (1) The initial conditions are given by the last a posteriori estimate and covariance of the smoother:

  • {circumflex over (x)} 0 =x T s

  • P 0 =P T s,
      • (2) The covariance of the measurement noise Rk is set to a large value rendering the measurements worthless because there are not any measurements available with prediction.
      • (3) As such, zk is fixed to the value of last measurement.
        The predictor propagates forward for the forecast period of interest. These predictions may be at various points in the future over various time periods. For example, one analytics application might predict temperature one minute into the future and/or five minutes into the future. Also, there may be temperature predictions at hourly or daily rates. These example combinations would be dependent on the application, of course.
  • Those skilled in the art know how to model dynamic process and measurements described by xk+1kxk+wk and zk=Hk xk+vk, respectively. Thus, once initialized with {circumflex over (x)}0 and P0 , the five-step Kalman filter iterates recursively until the set of data to be filtered is exhausted resulting in state estimates {circumflex over (x)}k and state covariances Pk.
  • Upon saving the state estimates {circumflex over (x)}k, the state covariances Pk, and properly initializing xT s and PT s with the last entries of {circumflex over (x)}k (Tf) and Pk (Tf), the three-step smoother iterates recursively with a backward sweep to an earlier time point, as illustrated in FIG. 2 , showing fixed-intervals smoothing, until all states/covariances are consumed. At this point, the system is prepared for predictive and prescriptive analytics.
  • The last state estimate and state covariance of the smoother is used to initialize the predictor. The predictor runs just like the five-step Kalman filter with two exceptions: (i) the covariance of the measurement noise R k is set to an arbitrarily large value to indicate the measurements are worthless, because there are not any, and (ii) the measurement z k is fixed to its final value, because that is the last piece of information available.
  • An application of the disclosed three-part system's implementation is shown in FIG. 3 illustrating an example of filtering, smoothing, and predicting for analytics where the future position of a moving object is predicted. In FIG. 3 , the actual position of an object is depicted on the vertical axis with a dashed line, the filtered position is represented by dots, the smoothed position is denoted by a solid line, and the predicted position is shown with “x” marking the spots. Measurements are filtered and smoothed up until 50 seconds, at which time, predictions are made 30 seconds into the future.
  • Referring again to FIG. 1 , if performance of a three-part system is better than ETS, then its performance is better than the rest. In this section, details are presented showing filtering with prediction is better than ETS. When combined with smoothing, performance improves and forms the basis of the accurate prediction shown in FIG. 3 .
  • Data used to perform the analysis was gathered from a live, operating Horizontal Pump System (HPS). Filtering and predicting was tested with eighteen data sets. Measurements were taken every hour so that one forecasting period represents one hour of elapsed time. The data measures various components of the HPS including bearing and winding temperatures in the motor, pump vibration and suction pressure, overall system health, and other attributes of the system.
  • Each data set includes noise which may vary with time. The different sources of data provide a mixture of different characteristics such as seasonality, trends, impulses, and randomness. For example, temperature data is affected by the day/night (diurnal) cycle which creates a (short) seasonal characteristic. Vibration data, however, is not affected by the day/night cycle and is not seasonal but does contain a significant portion of randomness.
  • A missed prediction occurs when an observed measurement exceeds a threshold value, but no forecast was produced which predicted the exception. Any forecast which predicted the exception within twelve periods leading up to the exception was not considered because such a short forecast is not useful. A prediction strategy should produce as few missed predictions as possible.
  • Turning now to Table 2 (below), illustrated are temperature and vibration filtering after filtering and predicting, showing sensitivities, forecast lengths, and average percent of missed predictions.
  • TABLE 2
    Average
    Forecast % Missed
    Length Predictions
    Sensi- Periods Low High
    Strategy tivity (Days) Noise Noise
    ETS High 24 (1) 18.92% 43.19%
    Filter/Predictor High 24 (1)  0.00%  0.00%
    ETS Medium 24 (1) 16.67% 42.94%
    Filter/Predictor Medium 24 (1)  0.00%  0.00%
    ETS Low 24 (1) 61.17% 62.05%
    Filter/Predictor Low 24 (1) 24.96% 13.89%
    ETS High 336 (14)  0.00%  5.56%
    Filter/Predictor High 336 (14)  0.00%  0.00%
    ETS Medium 336 (14)  0.00%  2.78%
    Filter/Predictor Medium 336 (14)  0.00%  0.00%
    ETS Low 336 (14)  0.00%  5.56%
    Filter/Predictor Low 336 (14)  0.00%  0.00%

    Table 2 compares each strategy (ETS versus filter/prediction) over 24 periods (1 day) and 336 periods (14 days). The filter sensitivity column refers to how closely the signal is being tracked. For instance, temperature changes slowly over time so the filter/prediction combination is set to high sensitivity to track the slowly changing signal; whereas vibration, which contains high frequency noise is set to low sensitivity, as illustrated in FIG. 4 showing performance examples of temperature tracking and vibration filtering in a noisy measurement environment. The Average % Missed Predictions column in Table 2 shows the probability that, given the data has crossed a critical threshold, the corresponding strategy failed to predict the event. A lower percentage indicates better performance for this metric. In each scenario (low noise/high noise), the filter/predictor strategy with a high or medium sensitivity correctly predicted every event. Even in the case of low sensitivity, the filter/predictor strategy outperformed the ETS strategy.
  • The filtering, smoothing, and predicting process introduced herein outperforms the ETS strategy which was the basis of performance assessment over machine learning strategies as shown in Table 1. These results appear to be independent of sensitivity setting (low, medium, or high). Therefore, in general, a practitioner could use the filter/predictor strategy to avoid missed predictions. Furthermore, with the inclusion of smoothing, these results are improved upon as shown in FIG. 3 .
  • Turning now to FIG. 5 , illustrated is a flow diagram of an embodiment of a method 500 of estimating a state of a system. The method 500 may be employable to estimate a state of a system in a machine learning and/or noisy measurement environment. The method 500 is operable on a processor such as a microprocessor coupled to a memory, the memory contained instructions which, when executed by the processor, are operative to perform the functions. The method 500 begins at a start step or module 505.
  • At a step or module 510, a first estimate of a state of a system is constructed at a first time including a first covariance matrix describing an accuracy of the first estimate.
  • At a step or module 520, a second estimate of the state of said system is constructed at a second time, after the first time, including a second covariance matrix describing an accuracy of the second estimate employing a dynamic model of the state of the system; the dynamic model comprises a matrix with coefficients that describes a temporal evolution of the state of the system.
  • At a step or module 530, a value of a characteristic of the state of the system is measured at the second time. Measuring the value of the characteristic can include making a plurality of independent measurements characterized by a diagonal measurement covariance matrix. At a step or module 540, the second estimate of the state of the system and the second covariance matrix are adjusted based on the value of the characteristic.
  • At a step or module 550, a third estimate of the state of the system is constructed at a third time, before the second time, including a third covariance matrix describing an accuracy of the third estimate employing the dynamic model of the state of the system.
  • At a step or module 560, a fourth estimate of the state of the system is constructed at a fourth time, after the second time, from the second estimate. In some embodiments, the fourth time is on a different time scale from the first, second and third times.
  • At a step or module 570, the dynamic model is altered in response to the value of the characteristic.
  • At a step or module 580, the state of the system is reported based on the fourth estimate.
  • At a step or module 590, a fifth estimate of the state of the system is constructed at a fifth time, after the second time, from the second estimate.
  • In certain embodiments, the dynamic model is a linear dynamic model with constant coefficients. In an embodiment, constructing the first estimate and constructing the second estimate are performed by a Kalman filter.
  • At a step or module 595, the state of the system is altered based on the fourth estimate.
  • The method 500 terminates at end step or module 598.
  • The impacts to implementation of predictive analysis of processes introduced herein cannot be understated. Whereas machine learning approaches are directly dependent on a large and fully populated training corpus, purely statistical approaches, such as ETS and the novel filter/predictor strategy introduced herein, learn directly from the real-time signal with additional data or knowledge imposed. Based upon the findings indicated in Table 1, the established ETS approach is already of better performance than the more widely used machine learning techniques. The improvements and advantages of the process introduced herein over ETS (shown in Table 2) only solidifies the merits of the new approach.
  • In short, advantages of the novel filtering, smoothing, and predicting process do not requiring a priori knowledge as it does for machine learning techniques. Because the system combines optimal estimation techniques of filtering, smoothing, and predicting, there are no dependencies on artificial neural nets and their (shallow, greedy, brittle, and opaque) shortcomings.
  • Turning now to FIG. 6 , illustrated is a block diagram of an embodiment of an apparatus 600 for estimating the state of a system in a machine learning environment. The apparatus 600 is configured to perform functions described hereinabove of constructing the estimate of the state of the system. The apparatus 600 includes a processor (or processing circuitry) 610, a memory 620 and a communication interface 630 such as a graphical user interface.
  • The functionality of the apparatus 600 may be provided by the processor 610 executing instructions stored on a computer-readable medium, such as the memory 620 shown in FIG. 6 . Alternative embodiments of the apparatus 600 may include additional components (such as the interfaces, devices and circuits) beyond those shown in FIG. 6 that may be responsible for providing certain aspects of the device's functionality, including any of the functionality to support the solution described herein.
  • The processor 610 (or processors), which may be implemented with one or a plurality of processing devices, perform functions associated with its operation including, without limitation, performing the operations of estimating the state of a system, computing covariance matrices, and estimating a future state of the system. The processor 610 may be of any type suitable to the local application environment, and may include one or more of general-purpose computers, special purpose computers, microprocessors, digital signal processors (“DSPs”), field-programmable gate arrays (“FPGAs”), application-specific integrated circuits (“ASICs”), and processors based on a multi-core processor architecture, as non-limiting examples.
  • The processor 610 may include, without limitation, application processing circuitry. In some embodiments, the application processing circuitry may be on separate chipsets. In alternative embodiments, part or all of the application processing circuitry may be combined into one chipset, and other application circuitry may be on a separate chipset. In still alternative embodiments, part or all of the application processing circuitry may be on the same chipset, and other application processing circuitry may be on a separate chipset. In yet other alternative embodiments, part or all of the application processing circuitry may be combined in the same chipset.
  • The memory 620 (or memories) may be one or more memories and of any type suitable to the local application environment, and may be implemented using any suitable volatile or nonvolatile data storage technology such as a semiconductor-based memory device, a magnetic memory device and system, an optical memory device and system, fixed memory and removable memory. The programs stored in the memory 620 may include program instructions or computer program code that, when executed by an associated processor, enable the respective apparatus 600 to perform its intended tasks. Of course, the memory 620 may form a data buffer for data transmitted to and from the same. Exemplary embodiments of the system, subsystems, and modules as described herein may be implemented, at least in part, by computer software executable by the processor 610, or by hardware, or by combinations thereof.
  • The communication interface 630 modulates information for transmission by the respective apparatus 600 to another apparatus. The respective communication interface 630 is also configured to receive information from another processor for further processing. The communication interface 630 can support duplex operation for the respective other processor 600.
  • In summary, the inventions disclosed herein combine three techniques of optimal estimation of the state of a system. The three techniques include filtering, smoothing, and predicting processes, and can be performed, without limitation, in a machine learning and/or a noisy measurement environment.
  • The filtering portion of optimal estimation is performed to construct a first estimate of a state vector xk at a time point tk that coincides with a measurement of a value of characteristic of the state of the system at the time point tk. The filtering process employs a covariance matrix that describes the accuracy of the first estimate of the state vector xk at the time point tk. A second estimate of the state vector xk+1 at the time point tk+1 is then constructed by propagating the state of the system forward to a second time point tk+1, the second time point being after the first time point. The propagating forward employs a dynamic model of the state of the system to produce the estimate of the state vector xk+1 at the second time point tk+1. The first estimate of the state vector xk and constructing the second estimate of the state vector xk+1 can be performed by employing a Kalman filter.
  • The dynamic model can employ a matrix with coefficients that describes temporal evolution of the state of the system. In certain embodiment, the dynamic model is a linear dynamic model with constant coefficients.
  • A value of a characteristic of the state of the system xk+1 is measured at the second time point tk+1. The second estimate of the state of the system and the second covariance matrix are adjusted based on the measured value of the characteristic at the second time point tk+1
  • Measuring the value of the characteristic can include making a plurality of independent measurements characterized by a diagonal measurement covariance matrix.
  • The smoothing portion of optimal estimation is performed by constructing a third state estimate for a time point that is earlier than the time point tk+1. The earlier time point can fall within or before a span of current measurement points, e.g., between or before the time points tk and tk+1.
  • The predicting portion then propagates the state estimate forward for a forecast period of interest. The last state estimate and state covariance of the smoother can be used to initialize the predicting. The predictions may be at various time points in the future and over various time scales that are after the second time point. Measurement noise Rk can be set to an arbitrarily large value to accommodate the inherent absence of a state measurement at a future time point. The initial conditions for the prediction can be taken as the last a posteriori state estimate and the covariance of the smoother.
  • As described above, the exemplary embodiments provide both a method and corresponding apparatus consisting of various modules providing functionality for performing the steps of the method. The modules may be implemented as hardware (embodied in one or more chips including an integrated circuit such as an application specific integrated circuit), or may be implemented as software or firmware for execution by a processor. In particular, in the case of firmware or software, the exemplary embodiments can be provided as a computer program product including a computer readable storage medium embodying computer program code (i.e., software or firmware) thereon for execution by the computer processor. The computer readable storage medium may be non-transitory (e.g., magnetic disks; optical disks; read only memory; flash memory devices; phase-change memory) or transitory (e.g., electrical, optical, acoustical or other forms of propagated signals-such as carrier waves, infrared signals, digital signals, etc.). The coupling of a processor and other components is typically through one or more busses or bridges (also termed bus controllers). The storage device and signals carrying digital traffic respectively represent one or more non-transitory or transitory computer readable storage medium. Thus, the storage device of a given electronic device typically stores code and/or data for execution on the set of one or more processors of that electronic device such as a controller.
  • Although the embodiments and its advantages have been described in detail, it should be understood that various changes, substitutions, and alterations can be made herein without departing from the spirit and scope thereof as defined by the appended claims. For example, many of the features and functions discussed above can be implemented in software, hardware, or firmware, or a combination thereof. Also, many of the features, functions, and steps of operating the same may be reordered, omitted, added, etc., and still fall within the broad scope of the various embodiments.
  • Moreover, the scope of the various embodiments is not intended to be limited to the particular embodiments of the process, machine, manufacture, composition of matter, means, methods and steps described in the specification. As one of ordinary skill in the art will readily appreciate from the disclosure, processes, machines, manufacture, compositions of matter, means, methods, or steps, presently existing or later to be developed, that perform substantially the same function or achieve substantially the same result as the corresponding embodiments described herein may be utilized as well. Accordingly, the appended claims are intended to include within their scope such processes, machines, manufacture, compositions of matter, means, methods, or steps.

Claims (20)

1. A method, comprising:
making a first measurement of a value of a characteristic of a state of a system;
making a second measurement of a value of said characteristic of said state of said system after said first measurement;
constructing a first filter measurement estimate after said second measurement coinciding with said first measurement including a first filter measurement covariance matrix describing an accuracy of said first filter measurement estimate;
constructing a first filter time estimate after said first filter measurement estimate including a first filter time covariance matrix describing an accuracy of said first filter time estimate employing a dynamic model of said state of said system;
constructing a second filter measurement estimate after said first filter time estimate coinciding with said second measurement including a second filter measurement covariance matrix describing an accuracy of said second filter measurement estimate;
constructing a second filter time estimate after said second filter measurement estimate including a second filter time covariance matrix describing an accuracy of said second filter time estimate employing said dynamic model of said state of said system;
constructing a smoothing estimate from said first filter measurement estimate and said second filter measurement estimate; and
constructing a first prediction estimate after said smoothing estimate that provides a forecast of a value of said characteristic of said state of said system including a first prediction covariance matrix describing an accuracy of said first prediction estimate employing said dynamic model of said state of said system.
2. The method as recited in claim 1 further comprising constructing a second prediction estimate after said first prediction estimate that provides another forecast of a value of said characteristic of said state of said system including a second prediction covariance matrix describing an accuracy of said second prediction estimate employing said dynamic model of said state of said system.
3. The method as recited in claim 1 further comprising constructing a plurality of prediction estimates that provides a corresponding plurality of forecasts of a value of said characteristic of said state of said system including a corresponding plurality of prediction covariance matrices describing an accuracy of said plurality of prediction estimates employing said dynamic model of said state of said system.
4. The method as recited in claim 1 wherein constructing said smoothing estimate comprises sweeping backward recursively from said second filter measurement estimate to said first filter measurement estimate.
5. The method as recited in claim 1 further comprising altering said state of said system based on said first prediction estimate.
6. The method as recited in claim 1 wherein said constructing said first filter measurement estimate, said first filter time estimate, said second filter measurement estimate and said second filter time estimate are performed by a Kalman filter.
7. The method as recited in claim 1 further comprising reporting said state of said system based on said first prediction estimate.
8. The method as recited in claim 1 wherein said first measurement comprises a plurality of independent measurements characterized by a diagonal measurement covariance matrix.
9. The method as recited in claim 1 wherein said dynamic model comprises a linear dynamic model with constant coefficients.
10. The method as recited in claim 1 wherein said dynamic model comprises a matrix with coefficients that describes a temporal evolution of said state of said system.
11. An apparatus operable to construct a state of a system, comprising:
processing circuitry coupled to a memory, configured to:
make a first measurement of a value of a characteristic of said state of said system;
make a second measurement of a value of said characteristic of said state of said system after said first measurement;
construct a first filter measurement estimate after said second measurement coinciding with said first measurement including a first filter measurement covariance matrix describing an accuracy of said first filter measurement estimate;
construct a first filter time estimate after said first filter measurement estimate including a first filter time covariance matrix describing an accuracy of said first filter time estimate employing a dynamic model of said state of said system;
construct a second filter measurement estimate after said first filter time estimate coinciding with said second measurement including a second filter measurement covariance matrix describing an accuracy of said second filter measurement estimate;
construct a second filter time estimate after said second filter measurement estimate including a second filter time covariance matrix describing an accuracy of said second filter time estimate employing said dynamic model of said state of said system;
construct a smoothing estimate from said first filter measurement estimate and said second filter measurement estimate; and
construct a first prediction estimate after said smoothing estimate that provides a forecast of a value of said characteristic of said state of said system including a first prediction covariance matrix describing an accuracy of said first prediction estimate employing said dynamic model of said state of said system.
12. The apparatus as recited in claim 11 wherein said processing circuitry is further configured to construct a second prediction estimate after said first prediction estimate that provides another forecast of a value of said characteristic of said state of said system including a second prediction covariance matrix describing an accuracy of said second prediction estimate employing said dynamic model of said state of said system.
13. The apparatus as recited in claim 11 wherein said processing circuitry is further configured to construct a plurality of prediction estimates that provides a corresponding plurality of forecasts of a value of said characteristic of said state of said system including a corresponding plurality of prediction covariance matrices describing an accuracy of said plurality of prediction estimates employing said dynamic model of said state of said system.
14. The apparatus as recited in claim 11 wherein said processing circuitry is configured to construct said smoothing estimate by sweeping backward recursively from said second filter measurement estimate to said first filter measurement estimate.
15. The apparatus as recited in claim 11 wherein said processing circuitry is further configured to alter said state of said system based on said first prediction estimate.
16. The apparatus as recited in claim 11 wherein said processing circuitry is configured to construct said first filter measurement estimate, said first filter time estimate, said second filter measurement estimate and said second filter time estimate with a Kalman filter.
17. The apparatus as recited in claim 11 wherein said processing circuitry is further configured to report said state of said system based on said first prediction estimate.
18. The apparatus as recited in claim 11 wherein said first measurement comprises a plurality of independent measurements characterized by a diagonal measurement covariance matrix.
19. The apparatus as recited in claim 11 wherein said dynamic model comprises a linear dynamic model with constant coefficients.
20. The apparatus as recited in claim 11 wherein said dynamic model comprises a matrix with coefficients that describes a temporal evolution of said state of said system.
US18/187,860 2018-11-05 2023-03-22 System and method for state estimation in a noisy machine-learning environment Pending US20230385606A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/187,860 US20230385606A1 (en) 2018-11-05 2023-03-22 System and method for state estimation in a noisy machine-learning environment

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US201862756044P 2018-11-05 2018-11-05
US16/674,848 US20200193318A1 (en) 2018-11-05 2019-11-05 System and method for state estimation in a noisy machine-learning environment
US18/187,860 US20230385606A1 (en) 2018-11-05 2023-03-22 System and method for state estimation in a noisy machine-learning environment

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/674,848 Continuation US20200193318A1 (en) 2018-11-05 2019-11-05 System and method for state estimation in a noisy machine-learning environment

Publications (1)

Publication Number Publication Date
US20230385606A1 true US20230385606A1 (en) 2023-11-30

Family

ID=71070932

Family Applications (7)

Application Number Title Priority Date Filing Date
US16/674,885 Abandoned US20200192777A1 (en) 2018-11-05 2019-11-05 System and method for adaptive optimization
US16/675,000 Abandoned US20200193271A1 (en) 2018-11-05 2019-11-05 System and method for vigorous artificial intelligence
US16/674,942 Pending US20200193075A1 (en) 2018-11-05 2019-11-05 System and method for constructing a mathematical model of a system in an artificial intelligence environment
US16/674,848 Abandoned US20200193318A1 (en) 2018-11-05 2019-11-05 System and method for state estimation in a noisy machine-learning environment
US18/050,661 Pending US20230289562A1 (en) 2018-11-05 2022-10-28 System and method for adaptive optimization
US18/187,860 Pending US20230385606A1 (en) 2018-11-05 2023-03-22 System and method for state estimation in a noisy machine-learning environment
US18/449,532 Pending US20240111996A1 (en) 2018-11-05 2023-08-14 Systems and methods for use in operations and maintenance sysyems for controlling the operation of second system

Family Applications Before (5)

Application Number Title Priority Date Filing Date
US16/674,885 Abandoned US20200192777A1 (en) 2018-11-05 2019-11-05 System and method for adaptive optimization
US16/675,000 Abandoned US20200193271A1 (en) 2018-11-05 2019-11-05 System and method for vigorous artificial intelligence
US16/674,942 Pending US20200193075A1 (en) 2018-11-05 2019-11-05 System and method for constructing a mathematical model of a system in an artificial intelligence environment
US16/674,848 Abandoned US20200193318A1 (en) 2018-11-05 2019-11-05 System and method for state estimation in a noisy machine-learning environment
US18/050,661 Pending US20230289562A1 (en) 2018-11-05 2022-10-28 System and method for adaptive optimization

Family Applications After (1)

Application Number Title Priority Date Filing Date
US18/449,532 Pending US20240111996A1 (en) 2018-11-05 2023-08-14 Systems and methods for use in operations and maintenance sysyems for controlling the operation of second system

Country Status (1)

Country Link
US (7) US20200192777A1 (en)

Families Citing this family (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11561983B2 (en) * 2019-03-07 2023-01-24 Throughputer, Inc. Online trained object property estimator
US11560146B2 (en) * 2019-03-26 2023-01-24 Ford Global Technologies, Llc Interpreting data of reinforcement learning agent controller
US11297029B2 (en) * 2019-10-02 2022-04-05 Paypal, Inc. System and method for unified multi-channel messaging with block-based datastore
US11790030B2 (en) * 2020-06-04 2023-10-17 Promoted.ai, Inc. Creating an effective product using an attribute solver
US11511413B2 (en) * 2020-06-12 2022-11-29 Huawei Technologies Co. Ltd. Systems and methods for learning reusable options to transfer knowledge between tasks
WO2022187502A1 (en) * 2021-03-05 2022-09-09 The Nielsen Company (Us), Llc Methods and apparatus to perform computer-based community detection in a network
CN113033102B (en) * 2021-03-30 2023-04-07 西安电子科技大学 Unsupervised learning-based method for evaluating health of cutter head of shield tunneling machine under complex stratum
CN113420847B (en) * 2021-08-24 2021-11-16 平安科技(深圳)有限公司 Target object matching method based on artificial intelligence and related equipment
US20230141928A1 (en) * 2021-10-13 2023-05-11 Oracle International Corporation Adaptive network attack prediction system
US20230169564A1 (en) * 2021-11-29 2023-06-01 Taudata Co., Ltd. Artificial intelligence-based shopping mall purchase prediction device
US11810574B1 (en) * 2022-11-15 2023-11-07 Leslie Helpert Voice-driven internal physiological imaging
CN116494816B (en) * 2023-06-30 2023-09-15 江西驴宝宝通卡科技有限公司 Charging management system and method for charging pile

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7676390B2 (en) * 2003-09-04 2010-03-09 General Electric Company Techniques for performing business analysis based on incomplete and/or stage-based data
BR112012009154A2 (en) * 2009-10-23 2016-08-16 Exxonmobil Upstream Res Co method for improving a geological model of a subsurface region, computer program product, and method for controlling hydrocarbons in a subsurface region
US11074495B2 (en) * 2013-02-28 2021-07-27 Z Advanced Computing, Inc. (Zac) System and method for extremely efficient image and pattern recognition and artificial intelligence platform
US20140278507A1 (en) * 2013-03-15 2014-09-18 Myrtle S. POTTER Methods and systems for growing and retaining the value of brand drugs by computer predictive model
US9454626B2 (en) * 2013-07-30 2016-09-27 Synopsys, Inc. Solving an optimization problem using a constraints solver
US9472188B1 (en) * 2013-11-15 2016-10-18 Noble Systems Corporation Predicting outcomes for events based on voice characteristics and content of a contact center communication
US10366346B2 (en) * 2014-05-23 2019-07-30 DataRobot, Inc. Systems and techniques for determining the predictive value of a feature
CA2953385A1 (en) * 2014-06-30 2016-01-07 Evolving Machine Intelligence Pty Ltd A system and method for modelling system behaviour

Also Published As

Publication number Publication date
US20240111996A1 (en) 2024-04-04
US20200192777A1 (en) 2020-06-18
US20200193075A1 (en) 2020-06-18
US20230289562A1 (en) 2023-09-14
US20200193318A1 (en) 2020-06-18
US20200193271A1 (en) 2020-06-18

Similar Documents

Publication Publication Date Title
US20230385606A1 (en) System and method for state estimation in a noisy machine-learning environment
Liang et al. A novel wind speed prediction strategy based on Bi-LSTM, MOOFADA and transfer learning for centralized control centers
Cannon A flexible nonlinear modelling framework for nonstationary generalized extreme value analysis in hydroclimatology
CN111708876B (en) Method and device for generating information
CN108923983B (en) Method and device for predicting opportunistic network link and readable storage medium
US11100388B2 (en) Learning apparatus and method for learning a model corresponding to real number time-series input data
WO2016003861A1 (en) Solar power forecasting using mixture of probabilistic principal component analyzers
US10783452B2 (en) Learning apparatus and method for learning a model corresponding to a function changing in time series
JP2020525872A (en) Influenza prediction model generation method, device, and computer-readable storage medium
US20200151545A1 (en) Update of attenuation coefficient for a model corresponding to time-series input data
Fernandes et al. Evaluating machine-learning techniques for recruitment forecasting of seven North East Atlantic fish species
US20180197082A1 (en) Learning apparatus and method for bidirectional learning of predictive model based on data sequence
US11197175B2 (en) Forcasting time series data
Rizvi Time series deep learning for robust steady-state load parameter estimation using 1D-CNN
Shuai et al. A kalman filter based approach for outlier detection in sensor networks
Mirikitani et al. Nonlinear maximum likelihood estimation of electricity spot prices using recurrent neural networks
CN117217779A (en) Training method and device of prediction model and information prediction method and device
CN112305915A (en) Label multi-Bernoulli multi-target tracking method and system of TSK iterative regression model
Rivero et al. Short time series prediction: Bayesian Enhanced modified Approach with application to cumulative rainfall series
Du et al. A multi-output two-stage locally regularized model construction method using the extreme learning machine
Baingana et al. Switched dynamic structural equation models for tracking social network topologies
CN115210714A (en) Large scale model simulation by knowledge distillation based NAS
US20230106295A1 (en) System and method for deriving a performance metric of an artificial intelligence (ai) model
CN117808125B (en) Model aggregation method, device, equipment, federal learning system and storage medium
Sun et al. Water level prediction based on polyelement correlation analysis and improved bp neural network

Legal Events

Date Code Title Description
AS Assignment

Owner name: INCUCOMM, INC., TEXAS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ALLEN, RANDAL;VOLPI, JOHN P.;VASTA, SAMUEL;SIGNING DATES FROM 20191106 TO 20191114;REEL/FRAME:063057/0822

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION