US20150227656A1 - Integrated Risk Management System - Google Patents

Integrated Risk Management System Download PDF

Info

Publication number
US20150227656A1
US20150227656A1 US14/693,910 US201514693910A US2015227656A1 US 20150227656 A1 US20150227656 A1 US 20150227656A1 US 201514693910 A US201514693910 A US 201514693910A US 2015227656 A1 US2015227656 A1 US 2015227656A1
Authority
US
United States
Prior art keywords
distribution
module
test
autocorrelation
statistic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US14/693,910
Inventor
Johnathan Mun
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US14/693,910 priority Critical patent/US20150227656A1/en
Publication of US20150227656A1 publication Critical patent/US20150227656A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • G06F17/5009
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F30/00Computer-aided design [CAD]
    • G06F30/20Design optimisation, verification or simulation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/04Forecasting or optimisation specially adapted for administrative or management purposes, e.g. linear programming or "cutting stock problem"
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0633Workflow analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q2220/00Business processing using cryptography
    • G06Q2220/10Usage protection of distributed data files
    • G06Q2220/18Licensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0201Market modelling; Market analysis; Collecting market data
    • G06Q30/0202Market predictions or forecasting for commercial activities

Abstract

A method and system allowing the analysis of risk through the use of Monte Carlo simulation, statistical and data analysis, stochastic forecasting, and optimization. The present invention includes novel methods such as the detailed reporting capabilities coupled with advanced analytical techniques, an integrated risk management process and procedures, adaptive licensing technology, and model profiling and storage procedures.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 12/378,169 filed on Feb. 11, 2009, the entire disclosure of which is incorporated herein by reference.
  • COPYRIGHT AND TRADEMARK NOTICE
  • A portion of the disclosure of this patent document contains materials subject to copyright and trademark protection. The copyright and trademark owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the U.S. Patent and Trademark Office patent files or records, but otherwise reserves all copyrights whatsoever.
  • BACKGROUND OF THE INVENTION
  • The present invention is in the field of finance, economics, math, and business statistics, and relates to the modeling and valuation of risk within all companies, allowing these firms to properly assess, quantify, value, diversify, and hedge their risks.
  • The field of risk analysis is large and complex, and this invention allows a new and novel set of analytics in a integrated and comprehensive manner, where executive reports with detailed explanations, numerical results and charts are generated within one to three mouse clicks. The invention uses Monte Carlo simulation, stochastic forecasting, business statistics, and optimization techniques in a novel way to analyze a user's existing data set to extract valuable and important information. In addition, the invention includes new and novel computer logic to save modeling parameters and inputs within multiple profiles in a single Excel workbook file, and covers a unique method to license software products through the use of a hardware identification algorithm.
  • SUMMARY OF THE INVENTION
  • Risk and uncertainty abound in the business world and impact business decisions and ultimately affects the profitability and survival of the corporation. The present invention's preferred embodiment is encapsulated in the Risk Simulator software, which incorporates a lot of advanced analytical techniques and algorithms and compiles them in such a unique and novel way to facilitate business risk analysis, through an intelligent set of statistical and analytical tests to analyze and extract information that otherwise cannot be obtained manually. That is, instead of requiring the user to understand advanced statistics, financial modeling and mathematics, in order to know what analysis to run on some existing data or the ability to interpret the raw numerical results, this present invention automatically runs the relevant analyses in an integrated fashion, and provides detailed description in its reports, coupled with the numerical results and charts for easy interpretation. The present invention also includes a novel licensing capability that extracts the user's system and hardware information to create a license protection. In addition, an integrated risk management business process method is developed that allows the user to step through the risk analysis methodology step by step in an integrated and comprehensive manner.
  • Monte Carlo simulation refers to a method where risk and uncertainty is quantified, through the use of mathematical algorithms of randomly sampling numbers from a specific distribution. For instance, suppose we need to forecast the revenues of a product the following year but this value is unknown, but nonetheless, we know from past experience that revenues for these types of products has a mean of X and standard deviation of Y, and follows a normal distribution. Further suppose that there are multiple products sold by this company. We can then take advantage of this fact and randomly select data points thousands of times with replacement, from a set of normal distribution with these specifications. The end result is thousands of forecast results, and using these results, we can determine the company's total revenues and the probability that the revenue will exceed some predefined value, and so forth.
  • BRIEF DESCRIPTION OF THE DRAWING
  • FIG. 01 illustrates the Integrated Risk Management Process steps.
  • FIG. 02 illustrates the Risk Simulator menu item with integrated language capabilities.
  • FIG. 03 illustrates the licensing scheme using a computer's unique set of hardware information.
  • FIG. 04 illustrates the simulation profiles.
  • FIG. 05 illustrates a sample report from the integrated system.
  • FIG. 06 illustrates the econometrics modeling capabilities in the system.
  • FIG. 07 illustrates the stochastic forecasting module and report.
  • FIG. 08 illustrates the auto ARIMA module and report.
  • FIG. 09 illustrates the distribution analysis tool and associated probabilities.
  • FIG. 10 illustrates the integrated statistical data analysis module.
  • FIG. 11 illustrates the statistical data analysis report structure.
  • FIG. 12 illustrates the econometric and regression data diagnostic module.
  • FIG. 13 illustrates the econometric and regression diagnostics reporting structure.
  • DETAILED DESCRIPTION OF THE INVENTION
  • FIG. 01 illustrates the integrated risk management process 1 on the process and method undertaken in the preferred embodiment of the present invention in the Risk Simulator software.
  • FIG. 02 illustrates the menu items in the software and that several foreign languages 2 exists in the software and can be changed immediately without having to restart the operating system or having to install a different software version. All language packs are available immediately.
  • FIG. 03 illustrates the licensing schema. The present invention's method allows the software to access the user computer's hardware and software configurations such as the user name on the computer, serial number on the operating system, serial numbers from various hardware devices such as the hard drive, motherboard, wireless and Ethernet card, take these values and apply some proprietary mathematical algorithms to convert them into a 10 to 20 alphanumerical Hardware ID 3. These Hardware IDs are unique to each computer and no two computers have the same identification. The prefix to this Hardware ID indicates the software type while the last letter on the ID indicates the type of hardware configuration on this computer (e.g., the letter “F” indicates that the hard drive and motherboard are properly installed and these serial numbers are used to generate this ID). Other suffix letters indicate various combinations of serial numbers used.
  • FIG. 04 illustrates the profile method in the software 4, where instead of creating multiple versions of the same Excel file, a single Excel file can contain multiple profiles that stores all the relevant information and parameters in the model, the system comprising the ability to create, duplicate, delete, and switch among multiple profiles in a single Excel file and the profile is saved as an encrypted XML code in a hidden sheet in the Excel file.
  • FIG. 05 illustrates a sample report generated from the software, complete with numerical results 5, detailed descriptions of the results 6 and executive dashboards and charts 7.
  • FIG. 06 illustrates the basic econometrics method in the software, where data is loaded directly into the tool 8 and customizable functions and equations can be entered 9 and a sample set of results 10 are available prior to running the entire report.
  • FIG. 07 illustrates another sample report 11 and method called stochastic process forecasting 12, where multiple types of stochastic models are available to forecast future values (e.g., Brownian motion random walk for forecasting stock prices and commodity asset prices, mean-reversion for forecasting interest rates and inflation rates, jump-diffusion for modeling utility, oil and gas prices).
  • FIG. 08 illustrates a novel method called Auto-ARIMA 13 where different combinations of ARIMA (autoregressive integrated moving average) models (see the section on the mathematical details of this modeling approach) are automatically run and analyzed in this report, returning the rankings of the best to the worst forecasting models.
  • FIG. 09 illustrates the distributional analysis tool 14 with 24 statistical distributions 15 where the probability density function (PDF), cumulative distribution function (CDF) and inverse cumulative distribution function (ICDF) 17 are available. The results show the distributional chart and probability tables 18.
  • FIG. 10 illustrates a comprehensive and integrated statistical analysis method where given some data, a set of comprehensive data analysis techniques are available 19 within a single mouse click.
  • FIG. 11 illustrates the results generated using the statistical analysis tool method, where we obtain detailed descriptive statistics 20, distributional fitting 21, hypothesis testing results 22, nonlinear extrapolation 23, normality test 24, stochastic parameter estimations 25, autocorrelation 26, time-series autocorrelation 27, and linear trend 28.
  • FIG. 12 illustrates a comprehensive and integrated forecasting data diagnostic method where given some data, a set of comprehensive data analysis techniques are available 29 within a single mouse click.
  • FIG. 13 illustrates the results of the data diagnostic tool, where the test for heteroskedasticity, micronumerosity, outliers and nonlinearity 30 are tested, followed by tests on normality and sphericity of the errors 31, autocorrelation 32, stochastic parameter estimations 33, multicollinearity 34, and correlation significance analysis 35.
  • Mathematical Probability Distributions
  • This section demonstrates the mathematical models and computations used in creating the Monte Carlo simulations. In order to get started with simulation, one first needs to understand the concept of probability distributions. To begin to understand probability, consider this example: You want to look at the distribution of nonexempt wages within one department of a large company. First, you gather raw data—in this case, the wages of each nonexempt employee in the department. Second, you organize the data into a meaningful format and plot the data as a frequency distribution on a chart. To create a frequency distribution, you divide the wages into group intervals and list these intervals on the chart's horizontal axis. Then you list the number or frequency of employees in each interval on the chart's vertical axis. Now you can easily see the distribution of nonexempt wages within the department. You can chart this data as a probability distribution. A probability distribution shows the number of employees in each interval as a fraction of the total number of employees. To create a probability distribution, you divide the number of employees in each interval by the total number of employees and list the results on the chart's vertical axis.
  • Probability distributions are either discrete or continuous. Discrete probability distributions describe distinct values, usually integers, with no intermediate values and are shown as a series of vertical bars. A discrete distribution, for example, might describe the number of heads in four flips of a coin as 0, 1, 2, 3, or 4. Continuous probability distributions are actually mathematical abstractions because they assume the existence of every possible intermediate value between two numbers; that is, a continuous distribution assumes there is an infinite number of values between any two points in the distribution. However, in many situations, you can effectively use a continuous distribution to approximate a discrete distribution even though the continuous model does not necessarily describe the situation exactly.
  • SProbability Density Functions, Cumulative Distribution Functions, and Probability Mass Functions
  • In mathematics and Monte Carlo simulation, a probability density function (PDF) represents a continuous probability distribution in terms of integrals. If a probability distribution has a density of f(x), then intuitively the infinitesimal interval of [x, x+dx] has a probability of f(x) dx. The PDF therefore can be seen as a smoothed version of a probability histogram; that is, by providing an empirically large sample of a continuous random variable repeatedly, the histogram using very narrow ranges will resemble the random variable's PDF. The probability of the interval between [a, b] is given by
  • a b f ( x ) x ,
  • which means that the total integral of the function f must be 1.0. It is a common mistake to think of f(a) as the probability of a. This is incorrect. In fact, f(a) can sometimes be larger than 1—consider a uniform distribution between 0.0 and 0.5. The random variable x within this distribution will have f(x) greater than 1. The probability in reality is the function f(x)dx discussed previously, where dx is an infinitesimal amount.
  • The cumulative distribution function (CDF) is denoted as F(x)=P(X≦x) indicating the probability of X taking on a less than or equal value to x. Every CDF is monotonically increasing, is continuous from the right, and at the limits, have the following properties:
  • lim x - F ( x ) = 0 and lim x + F ( x ) = 1.
  • Further, the CDF is related to the PDF by
  • F ( b ) - F ( a ) = P ( a X b ) = a b f ( x ) x ,
  • where the PDF function f is the derivative of the CDF function F.
  • In probability theory, a probability mass function or PMF gives the probability that a discrete random variable is exactly equal to some value. The PMF differs from the PDF in that the values of the latter, defined only for continuous random variables, are not probabilities; rather, its integral over a set of possible values of the random variable is a probability. A random variable is discrete if its probability distribution is discrete and can be characterized by a PMF. Therefore, X is a discrete random variable if
  • u P ( X = u ) = 1
  • as u runs through all possible values of the random variable X.
    Discrete Distributions Following is a detailed listing of the different types of probability distributions that can be used in Monte Carlo simulation.
  • Bernoulli or Yes/No Distribution
  • The Bernoulli distribution is a discrete distribution with two outcomes (e.g., head or tails, success or failure, 0 or 1). The Bernoulli distribution is the binomial distribution with one trial and can be used to simulate Yes/No or Success/Failure conditions. This distribution is the fundamental building block of other more complex distributions. For instance:
      • Binomial distribution: Bernoulli distribution with higher number of n total trials and computes the probability of x successes within this total number of trials.
      • Geometric distribution: Bernoulli distribution with higher number of trials and computes the number of failures required before the first success occurs.
      • Negative binomial distribution: Bernoulli distribution with higher number of trials and computes the number of failures before the xth success occurs.
  • The mathematical constructs for the Bernoulli distribution are as follows:
  • P ( x ) = { 1 - p for x = 0 p for x = 1 or P ( x ) = p x ( 1 - p ) 1 - x mean = p standard deviation = p ( 1 - p ) skewness = 1 - 2 p p ( 1 - p ) excess kurtosis = 6 p 2 - 6 p + 1 p ( 1 - p )
  • The probability of success (p) is the only distributional parameter. Also, it is important to note that there is only one trial in the Bernoulli distribution, and the resulting simulated value is either 0 or 1. The input requirements are such that Probability of Success >0 and <1 (that is, 0.0001≦p≦0.9999).
  • Binomial Distribution
  • The binomial distribution describes the number of times a particular event occurs in a fixed number of trials, such as the number of heads in 10 flips of a coin or the number of defective items out of 50 items chosen.
  • The three conditions underlying the binomial distribution are:
      • For each trial, only two outcomes are possible that are mutually exclusive.
      • The trials are independent—what happens in the first trial does not affect the next trial.
      • The probability of an event occurring remains the same from trial to trial.
  • The mathematical constructs for the binomial distribution are as follows:
  • P ( x ) = n ! x ! ( n - x ) ! p x ( 1 - p ) ( n - x ) for n > 0 ; x = 0 , 1 , 2 , n ; and 0 < p < 1 mean = np standard deviation = np ( 1 - p ) skewness = 1 - 2 p np ( 1 - p ) excess kurtosis = 6 p 2 - 6 p + 1 np ( 1 - p )
  • The probability of success (p) and the integer number of total trials (n) are the distributional parameters. The number of successful trials is denoted x. It is important to note that probability of success (p) of 0 or 1 are trivial conditions and do not require any simulations, and hence, are not allowed in the software. The input requirements are such that Probability of Success >0 and <1 (that is, 0.0001≦p≦0.9999), the Number of Trials ≧1 or positive integers and ≦1000 (for larger trials, use the normal distribution with the relevant computed binomial mean and standard deviation as the normal distribution's parameters).
  • Discrete Uniform
  • The discrete uniform distribution is also known as the equally likely outcomes distribution, where the distribution has a set of N elements, and each element has the same probability. This distribution is related to the uniform distribution but its elements are discrete and not continuous. The mathematical constructs for the discrete uniform distribution are as follows:
  • P ( x ) = 1 N mean = N + 1 2 ranked value standard deviation = ( N - 1 ) ( N + 1 ) 12 ranked value skewness = 0 ( that is , the distribution is perfectly symmetrical ) excess kurtosis = - 6 ( N 2 + 1 ) 5 ( N - 1 ) ( N + 1 ) ranked value
  • The input requirements are such that Minimum<Maximum and both must be integers (negative integers and zero are allowed).
  • Geometric Distribution
  • The geometric distribution describes the number of trials until the first successful occurrence, such as the number of times you need to spin a roulette wheel before you win.
  • The three conditions underlying the geometric distribution are:
      • The number of trials is not fixed.
      • The trials continue until the first success.
      • The probability of success is the same from trial to trial.
  • The mathematical constructs for the geometric distribution are as follows:
  • P ( x ) = p ( 1 - p ) x - 1 for 0 < p < 1 and x = 1 , 2 , , n mean = 1 p - 1 standard deviation = 1 - p p 2 skewness = 2 - p 1 - p excess kurtosis = p 2 - 6 p + 6 1 - p
  • The probability of success (p) is the only distributional parameter. The number of successful trials simulated is denoted x, which can only take on positive integers. The input requirements are such that Probability of success >0 and <1 (that is, 0.0001≦p≦0.9999). It is important to note that probability of success (p) of 0 or 1 are trivial conditions and do not require any simulations, and hence, are not allowed in the software.
  • Hypergeometric Distribution
  • The hypergeometric distribution is similar to the binomial distribution in that both describe the number of times a particular event occurs in a fixed number of trials. The difference is that binomial distribution trials are independent, whereas hypergeometric distribution trials change the probability for each subsequent trial and are called trials without replacement. For example, suppose a box of manufactured parts is known to contain some defective parts. You choose a part from the box, find it is defective, and remove the part from the box. If you choose another part from the box, the probability that it is defective is somewhat lower than for the first part because you have removed a defective part. If you had replaced the defective part, the probabilities would have remained the same, and the process would have satisfied the conditions for a binomial distribution.
  • The three conditions underlying the hypergeometric distribution are:
      • The total number of items or elements (the population size) is a fixed number, a finite population. The population size must be less than or equal to 1,750.
      • The sample size (the number of trials) represents a portion of the population.
      • The known initial probability of success in the population changes after each trial.
  • The mathematical constructs for the hypergeometric distribution are as follows:
  • P ( x ) = ( N x ) ! x ! ( N x - x ) ! ( N - N x ) ! ( n - x ) ! ( N - N x - n + x ) ! N n ! ( N - n ) ! for x = Max ( n - ( N - N x ) , 0 ) , , Min ( n , N x ) mean = N x n N standard deviation = ( N - N x ) N x n ( N - n ) N 2 ( N - 1 ) skewness = ( N - 2 N x ) ( N - 2 n ) N - 2 N - 1 ( N - N x ) N x n ( N - n ) excess kurtosis = V ( N , N x , n ) ( N - N x ) N x n ( - 3 + N ) ( - 2 + N ) ( - N + n ) where V ( N , N x , n ) = ( N - N x ) 3 - ( N - N x ) 5 + 3 ( N - N x ) 2 N x - 6 ( N - N x ) 3 N x + ( N - N x ) 4 N x + 3 ( N - N x ) N x 2 - 12 ( N - N x ) 2 N x 2 + 8 ( N - N x ) 3 N x 2 + N x 3 - 6 ( N - N x ) N x 3 + 8 ( N - N x ) 2 N x 3 + ( N - N x ) N x 4 - N x 5 - 6 ( N - N x ) 3 N x + 6 ( N - N x ) 4 N x + 18 ( N - N x ) 2 N x n - 6 ( N - N x ) 3 N x n + 18 ( N - N x ) N x 2 n - 24 ( N - N x ) 2 N x 2 n - 6 ( N - N x ) 3 n - 6 ( N - N x ) N x 3 n + 6 N x 4 n + 6 ( N - N x ) 2 n 2 - 6 ( N - N x ) 3 n 2 - 24 ( N - N x ) N x n 2 + 12 ( N - N x ) 2 N x n 2 + 6 N x 2 n 2 + 12 ( N - N x ) N x 2 n 2 - 6 N x 3 n 2
  • The number of items in the population (N), trials sampled (n), and number of items in the population that have the successful trait (Nx) are the distributional parameters. The number of successful trials is denoted x. The input requirements are such that Population ≧2 and integer, Trials >0 and integer
  • Successes >0 and integer, Population>Successes
  • Trials<Population and Population <1750.
  • Negative Binomial Distribution
  • The negative binomial distribution is useful for modeling the distribution of the number of trials until the rth successful occurrence, such as the number of sales calls you need to make to close a total of 10 orders. It is essentially a superdistribution of the geometric distribution. This distribution shows the probabilities of each number of trials in excess of r to produce the required success r.
  • Conditions
  • The three conditions underlying the negative binomial distribution are:
      • The number of trials is not fixed.
      • The trials continue until the rth success.
      • The probability of success is the same from trial to trial.
  • The mathematical constructs for the negative binomial distribution are as follows:
  • P ( x ) = ( x + r - 1 ) ! ( r - 1 ) ! x ! p x ( 1 - p ) ( n - x ) for x = r , r + 1 , ; and 0 < p < 1 mean = r ( 1 - p ) p standard deviation = r ( 1 - p ) p 2 skewness = 2 - p r ( 1 - p ) excess kurtosis = p 2 - 6 p + 6 r ( 1 - p )
  • Probability of success (p) and required successes (r) are the distributional parameters. Where the input requirements are such that Successes required must be positive integers >0 and <8000, Probability of success >0 and <1 (that is, 0.0001≦p≦0.9999). It is important to note that probability of success (p) of 0 or 1 are trivial conditions and do not require any simulations, and hence, are not allowed in the software.
  • Poisson Distribution
  • The Poisson distribution describes the number of times an event occurs in a given interval, such as the number of telephone calls per minute or the number of errors per page in a document.
  • Conditions
  • The three conditions underlying the Poisson distribution are:
      • The number of possible occurrences in any interval is unlimited.
      • The occurrences are independent. The number of occurrences in one interval does not affect the number of occurrences in other intervals.
      • The average number of occurrences must remain the same from interval to interval.
  • The mathematical constructs for the Poisson are as follows:
  • P ( x ) = - λ λ x x ! for x and λ > 0 mean = λ standard deviation = λ skewness = 1 λ excess kurtosis = 1 λ
  • Rate (λ) is the only distributional parameter and the input requirements are such that Rate >0 and ≦1000 (that is, 0.0001≦rate ≦1000).
  • Continuous Distributions Beta Distribution
  • The beta distribution is very flexible and is commonly used to represent variability over a fixed range. One of the more important applications of the beta distribution is its use as a conjugate distribution for the parameter of a Bernoulli distribution. In this application, the beta distribution is used to represent the uncertainty in the probability of occurrence of an event. It is also used to describe empirical data and predict the random behavior of percentages and fractions, as the range of outcomes is typically between 0 and 1. The value of the beta distribution lies in the wide variety of shapes it can assume when you vary the two parameters, alpha and beta. If the parameters are equal, the distribution is symmetrical. If either parameter is 1 and the other parameter is greater than 1, the distribution is J-shaped. If alpha is less than beta, the distribution is said to be positively skewed (most of the values are near the minimum value). If alpha is greater than beta, the distribution is negatively skewed (most of the values are near the maximum value). The mathematical constructs for the beta distribution are as follows:
  • f ( x ) = ( x ) ( α - 1 ) ( 1 - x ) ( β - 1 ) [ Γ ( α ) Γ ( β ) Γ ( α + β ) ] for α > 0 ; β > 0 ; x > 0 mean = α α + β standard deviation = αβ ( α + β ) 2 ( 1 + α + β ) skewness = 2 ( β - α ) 1 + α + β ( 2 + α + β ) αβ excess kurtosis = 3 ( α + β + 1 ) [ αβ ( α + β - 6 ) + 2 ( α + β ) 2 ] αβ ( α + β + 2 ) ( α + β + 3 ) - 3
  • Alpha (α) and beta (β) are the two distributional shape parameters, and Γ is the gamma function. The two conditions underlying the beta distribution are:
      • The uncertain variable is a random value between 0 and a positive value.
      • The shape of the distribution can be specified using two positive values.
  • Input requirements:
  • Alpha and beta >0 and can be any positive value
  • Cauchy Distribution or Lorentzian Distribution or Breit-Wigner Distribution
  • The Cauchy distribution, also called the Lorentzian distribution or Breit-Wigner distribution, is a continuous distribution describing resonance behavior. It also describes the distribution of horizontal distances at which a line segment tilted at a random angle cuts the x-axis.
  • The mathematical constructs for the cauchy or Lorentzian distribution are as follows:
  • f ( x ) = 1 π γ / 2 ( x - m ) 2 + γ 2 / 4
  • The cauchy distribution is a special case where it does not have any theoretical moments (mean, standard deviation, skewness, and kurtosis) as they are all undefined. Mode location (m) and scale (γ) are the only two parameters in this distribution. The location parameter specifies the peak or mode of the distribution while the scale parameter specifies the half-width at half-maximum of the distribution. In addition, the mean and variance of a cauchy or Lorentzian distribution are undefined. In addition, the cauchy distribution is the Student's t distribution with only 1 degree of freedom. This distribution is also constructed by taking the ratio of two standard normal distributions (normal distributions with a mean of zero and a variance of one) that are independent of one another. The input requirements are such that Location can be any value whereas Scale >0 and can be any positive value.
  • Chi-Square Distribution
  • The chi-square distribution is a probability distribution used predominatly in hypothesis testing, and is related to the gamma distribution and the standard normal distribution. For instance, the sums of independent normal distributions are distributed as a chi-square (χ2) with k degrees of freedom:
  • Z 1 2 + Z 2 2 + + Z k 2 d χ k 2
  • The mathematical constructs for the chi-square distribution are as follows:
  • f ( x ) = 2 - k / 2 Γ ( k / 2 ) x k / 2 - 1 - x / 2 for all x > 0
  • Γ is the gamma function. Degrees of freedom k is the only distributional parameter.
  • The chi-square distribution can also be modeled using a gamma distribution by setting the
  • shape parameter = k 2
  • and scale=2S2 where S is the scale. The input requirements are such that
  • Degrees of freedom >1 and must be an integer <1000.
  • Exponential Distribution
  • The exponential distribution is widely used to describe events recurring at random points in time, such as the time between failures of electronic equipment or the time between arrivals at a service booth. It is related to the Poisson distribution, which describes the number of occurrences of an event in a given interval of time. An important characteristic of the exponential distribution is the “memoryless” property, which means that the future lifetime of a given object has the same distribution, regardless of the time it existed. In other words, time has no effect on future outcomes. The mathematical constructs for the exponential distribution are as follows:
  • f ( x ) = λ - λ x for x 0 ; λ > 0 mean = 1 λ standard deviation = 1 λ skewness = 2 ( this value applies to all success rate λ inputs ) excess kurtosis = 6 ( this value applies to all success rate λ inputs )
  • Success rate (λ) is the only distributional parameter. The number of successful trials is denoted x.
  • The condition underlying the exponential distribution is:
      • The exponential distribution describes the amount of time between occurrences.
  • Input requirements: Rate >0 and ≦300
  • Extreme Value Distribution or Gumbel Distribution
  • The extreme value distribution (Type 1) is commonly used to describe the largest value of a response over a period of time, for example, in flood flows, rainfall, and earthquakes. Other applications include the breaking strengths of materials, construction design, and aircraft loads and tolerances. The extreme value distribution is also known as the Gumbel distribution.
  • The mathematical constructs for the extreme value distribution are as follows:
  • f ( x ) = 1 β z - Z where z = x - m β for β > 0 ; and any value of x and m mean = m + 0.577215 β standard deviation = 1 6 π 2 β 2 skewness = 12 6 ( 1.2020569 ) π 3 = 1.13955 ( this applies for all values of mode and scale ) excess kurtosis = 5.4 ( this applies for all values of mode and scale )
  • Mode (m) and scale (β) are the distributional parameters. There are two standard parameters for the extreme value distribution: mode and scale. The mode parameter is the most likely value for the variable (the highest point on the probability distribution). The scale parameter is a number greater than 0. The larger the scale parameter, the greater the variance. The input requirements are such that Mode can be any value and Scale >0.
  • F Distribution or Fisher-Snedecor Distribution
  • The F distribution, also known as the Fisher-Snedecor distribution, is another continuous distribution used most frequently for hypothesis testing. Specifically, it is used to test the statistical difference between two variances in analysis of variance tests and likelihood ratio tests. The F distribution with the numerator degree of freedom n and denominator degree of freedom m is related to the chi-square distribution in that:
  • χ n 2 / n d χ m 2 / m ~ F n , m or f ( x ) = Γ ( n + m 2 ) ( n m ) n / 2 x n / 2 - 1 Γ ( n 2 ) Γ ( m 2 ) [ x ( n m ) + 1 ] ( n + m ) / 2 mean = m m - 2 standard deviation = 2 m 2 ( m + n - 2 ) n ( m - 2 ) 2 ( m - 4 ) for all m > 4 skewness = 2 ( m + 2 n - 2 ) m - 6 2 ( m - 4 ) n ( m + n - 2 ) excess kurtosis = 12 ( - 16 + 20 m - 8 m 2 + m 3 + 44 n - 32 mn + 5 m 2 n - 22 n 2 + 5 mn 2 n ( m - 6 ) ( m - 8 ) ( n + m - 2 )
  • The numerator degree of freedom n and denominator degree of freedom m are the only distributional parameters. The input requirements are such that Degrees of freedom numerator and degrees of freedom denominator both >0 integers.
  • Gamma Distribution (Erlang Distribution)
  • The gamma distribution applies to a wide range of physical quantities and is related to other distributions: lognormal, exponential, Pascal, Erlang, Poisson, and Chi-Square. It is used in meteorological processes to represent pollutant concentrations and precipitation quantities. The gamma distribution is also used to measure the time between the occurrence of events when the event process is not completely random. Other applications of the gamma distribution include inventory control, economic theory, and insurance risk theory.
  • The gamma distribution is most often used as the distribution of the amount of time until the rth occurrence of an event in a Poisson process. When used in this fashion, the three conditions underlying the gamma distribution are:
      • The number of possible occurrences in any unit of measurement is not limited to a fixed number.
      • The occurrences are independent. The number of occurrences in one unit of measurement does not affect the number of occurrences in other units.
      • The average number of occurrences must remain the same from unit to unit.
  • The mathematical constructs for the gamma distribution are as follows:
  • f ( x ) = ( x β ) α - 1 - x β Γ ( α ) β with any value of α > 0 and β > 0 mean = αβ standard deviation = αβ 2 skewness = 2 α excess kurtosis = 6 α
  • Shape parameter alpha (α) and scale parameter beta (β) are the distributional parameters, and Γ is the gamma function. When the alpha parameter is a positive integer, the gamma distribution is called the Erlang distribution, used to predict waiting times in queuing systems, where the Erlang distribution is the sum of independent and identically distributed random variables each having a memoryless exponential distribution. Setting n as the number of these random variables, the mathematical construct of the Erlang distribution is:
  • f ( x ) = x n - 1 - x ( n - 1 ) ! for all x > 0
  • and all positive integers of n, where the input requirements are such that Scale Beta >0 and can be any positive value, Shape Alpha ≧0.05 and any positive value, and Location can be any value.
  • Logistic Distribution
  • The logistic distribution is commonly used to describe growth, that is, the size of a population expressed as a function of a time variable. It also can be used to describe chemical reactions and the course of growth for a population or individual.
  • The mathematical constructs for the logistic distribution are as follows:
  • f ( x ) = μ - x α α [ 1 + μ - x α ] for any value of α and μ mean = μ standard deviation = 1 3 π 2 α 2 skewness = 0 ( this applies to all mean and scale inputs ) excess kurtosis = 1.2 ( this applies to all mean and scale inputs )
  • Mean (μ) and scale (α) are the distributional parameters. There are two standard parameters for the logistic distribution: mean and scale. The mean parameter is the average value, which for this distribution is the same as the mode, because this distribution is symmetrical. The scale parameter is a number greater than 0. The larger the scale parameter, the greater the variance.
  • Input requirements:
  • Scale >0 and can be any positive value
  • Mean can be any value
  • Lognormal Distribution
  • The lognormal distribution is widely used in situations where values are positively skewed, for example, in financial analysis for security valuation or in real estate for property valuation, and where values cannot fall below zero. Stock prices are usually positively skewed rather than normally (symmetrically) distributed. Stock prices exhibit this trend because they cannot fall below the lower limit of zero but might increase to any price without limit. Similarly, real estate prices illustrate positive skewness and are lognormally distributed as property values cannot become negative.
  • The three conditions underlying the lognormal distribution are:
      • The uncertain variable can increase without limits but cannot fall below zero.
      • The uncertain variable is positively skewed, with most of the values near the lower limit.
      • The natural logarithm of the uncertain variable yields a normal distribution.
  • Generally, if the coefficient of variability is greater than 30 percent, use a lognormal distribution. Otherwise, use the normal distribution.
  • The mathematical constructs for the lognormal distribution are as follows:
  • f ( x ) = 1 x 2 π ln ( σ ) - [ ln ( x ) - ln ( μ ) ] 2 2 [ ln ( σ ) ] 2 for x > 0 ; μ > 0 and σ > 0 mean = exp ( μ + σ 2 2 ) standard deviation = exp ( σ 2 + 2 μ ) [ exp ( σ 2 ) - 1 ] skewness = exp ( σ 2 ) - 1 ( 2 + exp ( σ 2 ) ) excess kurtosis = exp ( 4 σ 2 ) + 2 exp ( 3 σ 2 ) + 3 exp ( 2 σ 2 ) - 6
  • Mean (μ) and standard deviation (σ) are the distributional parameters. The input requirements are such that Mean and Standard deviation are both >0 and can be any positive value. By default, the lognormal distribution uses the arithmetic mean and standard deviation. For applications for which historical data are available, it is more appropriate to use either the logarithmic mean and standard deviation, or the geometric mean and standard deviation.
  • Normal Distribution
  • The normal distribution is the most important distribution in probability theory because it describes many natural phenomena, such as people's IQs or heights. Decision makers can use the normal distribution to describe uncertain variables such as the inflation rate or the future price of gasoline.
  • Conditions
  • The three conditions underlying the normal distribution are:
      • Some value of the uncertain variable is the most likely (the mean of the distribution).
      • The uncertain variable could as likely be above the mean as it could be below the mean (symmetrical about the mean).
      • The uncertain variable is more likely to be in the vicinity of the mean than further away.
  • The mathematical constructs for the normal distribution are as follows:)
  • f ( x ) = 1 2 π σ - ( x - μ ) 2 2 σ 2 for all values of x and μ ; while σ > 0 mean = μ standard deviation = σ skewness = 0 ( this applies to all inputs of mean and standard deviation ) excess kurtosis = 0 ( this applies to all inputs of mean and standard deviation )
  • Mean (μ) and standard deviation (σ) are the distributional parameters. The input requirements are such that Standard deviation >0 and can be any positive value and Mean can be any value.
  • Pareto Distribution
  • The Pareto distribution is widely used for the investigation of distributions associated with such empirical phenomena as city population sizes, the occurrence of natural resources, the size of companies, personal incomes, stock price fluctuations, and error clustering in communication circuits.
  • The mathematical constructs for the pareto are as follows:
  • f ( x ) = β L β x ( 1 + β ) for x > L mean = β L β - 1 standard deviation = β L 2 ( β - 1 ) 2 ( β - 2 ) skewness = β - 2 β [ 2 ( β + 1 ) β - 3 ] excess kurtosis = 6 ( β 3 + β 2 - 6 β - 2 ) β ( β - 3 ) ( β - 4 )
  • Location (L) and shape (β) are the distributional parameters.
  • There are two standard parameters for the Pareto distribution: location and shape. The location parameter is the lower bound for the variable. After you select the location parameter, you can estimate the shape parameter. The shape parameter is a number greater than 0, usually greater than 1. The larger the shape parameter, the smaller the variance and the thicker the right tail of the distribution. The input requirements are such that Location >0 and can be any positive value while Shape >0.05.
  • Student's t Distribution
  • The Student's t distribution is the most widely used distribution in hypothesis test. This distribution is used to estimate the mean of a normally distributed population when the sample size is small, and is used to test the statistical significance of the difference between two sample means or confidence intervals for small sample sizes.
  • The mathematical constructs for the t-distribution are as follows:
  • f ( t ) = Γ [ ( r + 1 ) / 2 ] r π Γ [ r / 2 ] ( 1 + t 2 / r ) - ( r + 1 ) / 2 mean = 0 ( this applies to all degrees of freedom r except if the distribution is shifted to another nonzero central location ) standard deviation = r r - 2 skewness = 0 excess kurtosis = 6 r - 4 for all r > 4 where t = x - x _ s and Γ is the gamma function .
  • Degree of freedom r is the only distributional parameter. The t-distribution is related to the F-distribution as follows: the square of a value of t with r degrees of freedom is distributed as F with 1 and r degrees of freedom. The overall shape of the probability density function of the t-distribution also resembles the bell shape of a normally distributed variable with mean 0 and variance 1, except that it is a bit lower and wider or is leptokurtic (fat tails at the ends and peaked center). As the number of degrees of freedom grows (say, above 30), the t-distribution approaches the normal distribution with mean 0 and variance 1. The input requirements are such that Degrees of freedom ≧1 and must be an integer.
  • Triangular Distribution
  • The triangular distribution describes a situation where you know the minimum, maximum, and most likely values to occur. For example, you could describe the number of cars sold per week when past sales show the minimum, maximum, and usual number of cars sold.
  • Conditions
  • The three conditions underlying the triangular distribution are:
      • The minimum number of items is fixed.
      • The maximum number of items is fixed.
      • The most likely number of items falls between the minimum and maximum values, forming a triangular-shaped distribution, which shows that values near the minimum and maximum are less likely to occur than those near the most-likely value.
  • The mathematical constructs for the triangular distribution are as follows:
  • f ( x ) = { 2 ( x - Min ) ( Max - Min ) ( Likely - min ) for Min < x < Likely 2 ( Max - x ) ( Max - Min ) ( Max - Likely ) for Likely < x < Max mean = 1 3 ( Min + Likely + Max ) standard deviation = 1 18 ( Min 2 + Likely 2 + Max 2 - MinMax - MinLikely - MaxLikely ) skewness = 2 ( Min + Max - 2 Likely ) ( 2 Min - Max - Likely ) ( Min - 2 Max + Likely ) 5 ( Min 2 + Max 2 + Likely 2 - MinMax - MinLikely - MaxLikely ) 3 / 2 excess kurtosis = - 0.6
  • Minimum (Min), most likely (Likely) and maximum (Max) are the distributional parameters and the input requirements are such that Min≦Most Likely≦Max and can take any value, Min<Max and can take any value.
  • Uniform Distribution
  • With the uniform distribution, all values fall between the minimum and maximum and occur with equal likelihood.
  • The three conditions underlying the uniform distribution are:
      • The minimum value is fixed.
      • The maximum value is fixed.
      • All values between the minimum and maximum occur with equal likelihood.
  • The mathematical constructs for the uniform distribution are as follows:
  • f ( x ) = 1 Max - Min for all values such that Min < Max mean = Min + Max 2 standard deviation = ( Max - Min ) 2 12 skewness = 0 excess kurtosis = - 1.2 ( this applies to all inputs of Min and Max )
  • Maximum value (Max) and minimum value (Min) are the distributional parameters. The input requirements are such that Min<Max and can take any value.
  • Weibull Distribution (Rayleigh Distribution)
  • The Weibull distribution describes data resulting from life and fatigue tests. It is commonly used to describe failure time in reliability studies as well as the breaking strengths of materials in reliability and quality control tests. Weibull distributions are also used to represent various physical quantities, such as wind speed. The Weibull distribution is a family of distributions that can assume the properties of several other distributions. For example, depending on the shape parameter you define, the Weibull distribution can be used to model the exponential and Rayleigh distributions, among others. The Weibull distribution is very flexible. When the Weibull shape parameter is equal to 1.0, the Weibull distribution is identical to the exponential distribution. The Weibull location parameter lets you set up an exponential distribution to start at a location other than 0.0. When the shape parameter is less than 1.0, the Weibull distribution becomes a steeply declining curve. A manufacturer might find this effect useful in describing part failures during a burn-in period.
  • The mathematical constructs for the Weibull distribution are as follows:
  • f ( x ) = α β [ x β ] α - 1 - ( x β ) α mean = βΓ ( 1 + α - 1 ) standard deviation = β 2 [ Γ ( 1 + 2 α - 1 ) - Γ 2 ( 1 + α - 1 ) ] skewness = 2 Γ 3 ( 1 + β - 1 ) - 3 Γ ( 1 + β - 1 ) Γ ( 1 + 2 β - 1 ) + Γ ( 1 + 3 β - 1 ) [ Γ ( 1 + 2 β - 1 ) - Γ 2 ( 1 + β - 1 ) ] 3 / 2 excess kurtosis = - 6 Γ 4 ( 1 + β - 1 ) + 12 Γ 2 ( 1 + β - 1 ) Γ ( 1 + 2 β - 1 ) - 3 Γ 2 ( 1 + 2 β - 1 ) - 4 Γ ( 1 + β - 1 ) Γ ( 1 + 3 β - 1 ) + Γ ( 1 + 4 β - 1 ) [ Γ ( 1 + 2 β - 1 ) - Γ 2 ( 1 + β - 1 ) ] 2
  • Location (L), shape (α) and scale (β) are the distributional parameters, and Γ is the Gamma function. The input requirements are such that Scale >0 and can be any positive value, Shape ≧0.05 and
  • Location can take on any value.
  • Multiple Regression Analysis and Econometric Data Analysis
  • This section demonstrates the mathematical models and computations used in creating the general regression equations, which take the form of Y=β01X12X2+ . . . +βnXn+ε where β0 is the intercept, βi are the slope coefficients, and ε is the error term. The Y term is the dependent variable and the X terms are the independent variables, where these X variables are also known as the regressors. The dependent variable is named as such as it depends on the independent variable, for example, sales revenue depends on the amount of marketing costs expended on a product's advertising and promotion, making the dependent variable sales and the independent variable marketing costs. An example of a bivariate regression where there is only a single Y and a single X variable, is seen as simply inserting the best-fitting line through a set of data points in a two-dimensional plane. In other cases, a multivariate regression can be performed, where there are multiple or k number of independent X variables or regressors where in this case, the best-fitting line will be within a k+1 dimensional plane.
  • Fitting a line through a set of data points in a multidimensional scatter plot may result in numerous possible lines. The best-fitting line is defined as the single unique line that minimizes the total vertical errors, that is, the sum of the absolute distances between the actual data points (Yi) and the estimated line (Ŷ). To find the best-fitting unique line that minimizes the errors, a more sophisticated approach is applied, using multivariate regression analysis. Regression analysis therefore finds the unique best-fitting line by requiring that the total errors be minimized, or by calculating
  • Min i = 1 n ( Y i - Y ^ i ) 2
  • Only one unique line will minimize this sum of squared errors as shown in the equation above. The errors (vertical distances between the actual data and the predicted line) are squared to avoid the negative errors from canceling out the positive errors. Solving this minimization problem with respect to the slope and intercept requires calculating first derivatives and setting them equal to zero:
  • β 0 i = 1 n ( Y i - Y ^ i ) 2 = 0 and i = 1 n ( Y i - Y ^ i ) 2 = 0
  • Which yields the simple bivariate regression's set of least squares equations:
  • β 1 = i = 1 n ( X i - X _ ) ( Y i - Y _ ) i = 1 n ( X i - X _ ) 2 = i = 1 n X i Y i - i = 1 n X i i = 1 n Y i n i = 1 n X i 2 - ( i = 1 n X i ) 2 n β 0 = Y _ - β 1 X _
  • For multivariate regression, the analogy is expanded to account for multiple independent variables, where Yi12 X2,i3X3,ii and the estimated slopes can be calculated by:
  • β ^ 2 = Y i X 2 , i X 3 , i 2 - Y i X 3 , i X 2 , i X 3 , i X 2 , i 2 X 3 , i 2 - ( X 2 , i X 3 , i ) 2 β ^ 3 = Y i X 3 , i X 2 , i 2 - Y i X 2 , i X 2 , i X 3 , i X 2 , i 2 X 3 , i 2 - ( X 2 , i X 3 , i ) 2
  • This set of results can be summarized using matrix notations: [X′ X]−1[X′ Y].
  • In running multivariate regressions, great care must be taken to set up and interpret the results. For instance, a good understanding of econometric modeling is required (e.g., identifying regression pitfalls such as structural breaks, multicollinearity, heteroskedasticity, autocorrelation, specification tests, nonlinearities, and so forth) before a proper model can be constructed. Therefore the present invention includes some advanced econometrics approaches that are based on the principles of multiple regression outlined above.
  • One approach used is that of an Auto-ARIMA, which is based on the fundamental concepts of ARIMA theory or Autoregressive Integrated Moving Average models. ARIMA(p,d,q) models are the extension of the AR model that uses three components for modeling the serial correlation in the time series data. The first component is the autoregressive (AR) term. The AR(p) model uses the p lags of the time series in the equation. An AR(p) model has the form: yt=a1yt-1+ . . . +apyt-p+et. The second component is the integration (d) order term. Each integration order corresponds to differencing the time series. I(1) means differencing the data once. I (d) means differencing the data d times. The third component is the moving average (MA) term. The MA(q) model uses the q lags of the forecast errors to improve the forecast. An MA(q) model has the form: yt=et+b1et-1+ . . . +bet-q. Finally, an ARMA(p,q) model has the combined form: yt=a1yt-1+ . . . +apyt-p+e1+b1et-1+ . . . +bq et-q. Using this ARIMA concept, various combinations of p, d, q integers are tested in an automated and systematic fashion to determine the best-fitting model for the user's data.
  • In order to determine the best fitting model, we apply several goodness-of-fit statistics to provide a glimpse into the accuracy and reliability of the estimated regression model. They usually take the form of a t-statistic, F-statistic, R-squared statistic, adjusted R-squared statistic, Durbin-Watson statistic, Akaike Criterion, Schwarz Criterion, and their respective probabilities.
  • The R-squared (R2), or coefficient of determination, is an error measurement that looks at the percent variation of the dependent variable that can be explained by the variation in the independent variable for a regression analysis. The coefficient of determination can be calculated by:
  • R 2 = 1 - i - 1 n ( Y i - Y ^ i ) 2 i = 1 n ( Y i - Y _ ) 2 = 1 - S S E T S S
  • Where the coefficient of determination is one less the ratio of the sums of squares of the errors (SSE) to the total sums of squares (TSS). In other words, the ratio of SSE to TSS is the unexplained portion of the analysis, thus, one less the ratio of SSE to TSS is the explained portion of the regression analysis.
  • The estimated regression line is characterized by a series of predicted values (Ŷ) the average value of the dependent variable's data points is denoted Y; and the individual data points are characterized by Yi. Therefore, the total sum of squares, that is, the total variation in the data or the total variation about the average dependent value, is the total of the difference between the individual dependent values and its average (the total squared distance of YiY). The explained sum of squares, the portion that is captured by the regression analysis, is the total of the difference between the regression's predicted value and the average dependent variable's data set (seen as the total squared distance of Ŷ− Y). The difference between the total variation (TSS) and the explained variation (ESS) is the unexplained sums of squares, also known as the sums of squares of the errors (SSE).
  • Another related statistic, the adjusted coefficient of determination, or the adjusted R-squared ( R 2), corrects for the number of independent variables (k) in a multivariate regression through a degrees of freedom correction to provide a more conservative estimate:
  • R _ 2 = 1 - i = 1 n ( Y i - Y ^ i ) 2 / ( k - 2 ) i = 1 n ( Y i - Y _ ) 2 / ( k - 1 ) = 1 - S S E / ( k - 1 ) T S S / ( k - 1 )
  • The adjusted R-squared should be used instead of the regular R-squared in multivariate regressions because every time an independent variable is added into the regression analysis, the R-squared will increase; indicating that the percent variation explained has increased. This increase occurs even when nonsensical regressors are added. The adjusted R-squared takes the added regressors into account and penalizes the regression accordingly, providing a much better estimate of a regression model's goodness-of-fit.
  • Other goodness-of-fit statistics include the t-statistic and the F-statistic. The former is used to test if each of the estimated slope and intercept(s) is statistically significant, that is, if it is statistically significantly different from zero (therefore making sure that the intercept and slope estimates are statistically valid). The latter applies the same concepts but simultaneously for the entire regression equation including the intercept and slopes. Using the previous example, the following illustrates how the t-statistic and F-statistic can be used in a regression analysis.
  • When running the Autoeconometrics methodology, multiple regression issues and errors are first tested for. These include items such as heteroskedasticity, multicollinearity, micronumerosity, lags, leads, autocorrelation and others. For instance, several tests exist to test for the presence of heteroskedasticity. These tests also are applicable for testing misspecifications and nonlinearities. The simplest approach is to graphically represent each independent variable against the dependent variable as illustrated earlier. Another approach is to apply one of the most widely used model, the White's test, where the test is based on the null hypothesis of no heteroskedasticity against an alternate hypothesis of heteroskedasticity of some unknown general form. The test statistic is computed by an auxiliary or secondary regression, where the squared residuals or errors from the first regression are regressed on all possible (and nonredundant) cross products of the regressors. For example, suppose the following regression is estimated:

  • Y=β 01 X+β 2 Z+ε t
  • The test statistic is then based on the auxiliary regression of the errors (ε):

  • εt 201 X+α 2 Z+a 3 X 24 Z 25 XZ+v t
  • The nR2 statistic is the White's test statistic, computed as the number of observations (n) times the centered R-squared from the test regression. White's test statistic is asymptotically distributed as a χ2; with degrees of freedom equal to the number of independent variables (excluding the constant) in the test regression.
  • The White's test is also a general test for model misspecification, because the null hypothesis underlying the test assumes that the errors are both homoskedastic and independent of the regressors, and that the linear specification of the model is correct. Failure of any one of these conditions could lead to a significant test statistic. Conversely, a nonsignificant test statistic implies that none of the three conditions is violated. For instance, the resulting F-statistic is an omitted variable test for the joint significance of all cross products, excluding the constant.
  • One method to fix heteroskedasticity is to make it homoskedastic by using a weighted least squares (WLS) approach. For instance, suppose the following is the original regression equation:

  • Y=β 01 X 12 X 23 X 3
  • Further suppose that X2 is heteroskedastic. Then transform the data used in the regression into:
  • Y = β 0 X 2 + β 1 X 1 X 2 + β 2 + β 3 X 3 X 2 + ɛ X 2
  • The model can be redefined as the following WLS regression:

  • Y WLS0 WLS1 WLS X 12 WLS X 23 WLS X 3 +v
  • Alternatively, the Park's test can be applied to test for heteroskedasticity and to fix it. The Park's test model is based on the original regression equation, uses its errors, and creates an auxiliary regression that takes the form of:

  • ln e i 212 ln X k,j
  • Suppose β2 is found to be statistically significant based on a t-test, then heteroskedasticity is found to be present in the variable Xk,i. The remedy therefore is to use the following regression specification:
  • Y X k β 2 = β 1 X k β 2 + β 2 X 2 X k β 2 + β 3 X 3 X k β 2 + ɛ .
  • Multicollinearity exists when there is a linear relationship between the independent variables. When this occurs, the regression equation cannot be estimated at all. In near collinearity situations, the estimated regression equation will be biased and provide inaccurate results. This situation is especially true when a step-wise regression approach is used, where the statistically significant independent variables will be thrown out of the regression mix earlier than expected, resulting in a regression equation that is neither efficient nor accurate.
  • As an example, suppose the following multiple regression analysis exists, where Yi12X2,i3X3,ii
  • The estimated slopes can be calculated through
  • β ^ 2 = Y i X 2 , i X 3 , i 2 - Y i X 3 , i X 2 , i X 3 , i X 2 , i 2 X 3 , i 2 - ( X 2 , i X 3 , i ) 2 β ^ 3 = Y i X 3 , i X 2 , i 2 - Y i X 2 , i X 2 , i X 3 , i X 2 , i 2 X 3 , i 2 - ( X 2 , i X 3 , i ) 2
  • Now suppose that there is perfect multicollinearity, that is, there exists a perfect linear relationship between X2 and X3, such that X3,i=λX2,i for all positive values of λ. Substituting this linear relationship into the slope calculations for β2, the result is indeterminate. In other words, we have
  • β ^ 2 = Y i X 2 , i λ 2 X 2 , i 2 - Y i λ X 2 , i λ X 2 , i 2 X 2 , i 2 λ 2 X 2 , i 2 - ( λ X 2 , i 2 ) 2 = 0 0
  • The same calculation and results apply to β3, which means that the multiple regression analysis breaks down and cannot be estimated given a perfect collinearity condition.
  • One quick test of the presence of multicollinearity in a multiple regression equation is that the R-squared value is relatively high while the t-statistics are relatively low. Another quick test is to create a correlation matrix between the independent variables. A high cross correlation indicates a potential for multicollinearity. The rule of thumb is that a correlation with an absolute value greater than 0.75 is indicative of severe multicollinearity.
  • Another test for multicollinearity is the use of the variance inflation factor (VIF), obtained by regressing each independent variable to all the other independent variables, obtaining the R-squared value and calculating the VIF of that variable by estimating:
  • V I F i = 1 ( 1 - R i 2 )
  • A high VIF value indicates a high R-squared near unity. As a rule of thumb, a VIF value greater than 10 is usually indicative of destructive multicollinearity. The Autoeconometrics method computes for multicollinearity and corrects the data before running the next iteration when enumerating through the entire set of possible combinations and permutations of models.
  • One very simple approach to test for autocorrelation is to graph the time series of a regression equation's residuals. If these residuals exhibit some cyclicality, then autocorrelation exists. Another more robust approach to detect autocorrelation is the use of the Durbin-Watson statistic, which estimates the potential for a first-order autocorrelation. The Durbin-Watson test also identifies model misspecification. That is, if a particular time-series variable is correlated to itself one period prior. Many time-series data tend to be autocorrelated to their historical occurrences. This relationship can be due to multiple reasons, including the variables' spatial relationships (similar time and space), prolonged economic shocks and events, psychological inertia, smoothing, seasonal adjustments of the data, and so forth.
  • The Durbin-Watson statistic is estimated by the sum of the squares of the regression errors for one period prior, to the sum of the current period's errors:
  • D W = ( ɛ t - ɛ t - 1 ) 2 ɛ t 2
  • There is a Durbin-Watson critical statistic table at the end of the book that provides a guide as to whether a statistic implies any autocorrelation.
  • Another test for autocorrelation is the Breusch-Godfrey test, where for a regression function in the form of:

  • Y=f(X 1 , X 2 , . . . , X k)
  • Estimate this regression equation and obtain its errors εt. Then, run the secondary regression function in the form of:

  • Y=f(X 1 ,X 2 , . . . , X k, εt-1, εt-2, εt-p)
  • Obtain the R-squared value and test it against a null hypothesis of no autocorrelation versus an alternate hypothesis of autocorrelation, where the test statistic follows a Chi-Square distribution of p degrees of freedom:

  • R 2(n−p)˜χdf-p 2
  • Fixing autocorrelation requires the application of advanced econometric models including the applications of ARIMA (as described above) or ECM (Error Correction Models). However, one simple fix is to take the lags of the dependent variable for the appropriate periods, add them into the regression function, and test for their significance, for instance:

  • Y t =f(Y t-1 , Y t-2 , . . . , Y t-p , X 1 , X 2 , . . . , X k)
  • In interpreting the results of an Autoeconometrics model, most of the specifications are identical to the multivariate regression analysis. However, there are several additional sets of results specific to the econometric analysis. The first is the addition of Akaike Information Criterion (AIC) and Schwarz Criterion (SC), which are often used in ARIMA model selection and identification. That is, AIC and SC are used to determine if a particular model with a specific set of p, d, and q parameters is a good statistical fit. SC imposes a greater penalty for additional coefficients than the AIC but generally, the model with the lowest AIC and SC values should be chosen. Finally, an additional set of results called the autocorrelation (AC) and partial autocorrelation (PAC) statistics are provided in the ARIMA report.
  • For instance, if autocorrelation AC(1) is nonzero, it means that the series is first order serially correlated. If AC dies off more or less geometrically with increasing lags, it implies that the series follows a low-order autoregressive process. If AC drops to zero after a small number of lags, it implies that the series follows a low-order moving-average process. In contrast, PAC measures the correlation of values that are k periods apart after removing the correlation from the intervening lags. If the pattern of autocorrelation can be captured by an autoregression of order less than k, then the partial autocorrelation at lag k will be close to zero. The Ljung-Box Q-statistics and their p-values at lag k are also provided, where the null hypothesis being tested is such that there is no autocorrelation up to order k. The dotted lines in the plots of the autocorrelations are the approximate two standard error bounds. If the autocorrelation is within these bounds, it is not significantly different from zero at approximately the 5% significance level. Finding the right ARIMA model takes practice and experience. These AC, PAC, SC, and AIC are highly useful diagnostic tools to help identify the correct model specification. Finally, the ARIMA parameter results are obtained using sophisticated optimization and iterative algorithms, which means that although the functional forms look like those of a multivariate regression, they are not the same. ARIMA is a much more computationally intensive and advanced econometric approach.

Claims (18)

I claim:
1. A system for analyzing business risk comprising:
a computing device comprising a processor communicatively connected to a storage medium, a motherboard, and an Ethernet card;
an operating system stored in a memory of said device configured to provide instructions to said processor;
computer readable instructions residing in a memory of said device, wherein said computer readable instructions comprise a risk simulation module comprising:
a plurality of forecasting and risk simulation models and methods, and a plurality of basic econometric models;
a basic econometrics module configured to run one or more basic econometric models by (1) identifying input variables from said user provided data and designating at least one independent variable and at least one dependent variable from among said input variables, (2) calculating at least one of the following metrics: R-Squared, Adjusted R-Squared, Multiple R, Standard Error of the Estimates, ANOVA F Statistic, and ANOVA p-Value; (3) testing for regression errors including at least one of: heteroskedasticity, multicollinearity, micronumerosity, lags, leads, and autocorrelation, and (4) adjusting the data to fix any identified regression errors;
an autoregressive integrated moving average (ARIMA) module configured to analyze and rank said forecasting and risk simulation models and methods from best to worst based on said adjusted user provided data, by testing various combinations of p, d, and q integers to determine the best-fitting model for the user provided data, wherein one or more of said ranked forecasting and risk simulation models may be selected for use in a simulation;
a simulation selection module configured to allow a user to select an active simulation defined by one or more forecasting and/or risk simulation models being applied to a set of input variables derived from the user provided data;
a stochastic process forecasting module configured to forecast future values for at least one of equities, assets, interest rates, inflation rates, and commodities using at least one of Brownian motion random walk, mean-reversion, and jump-diffusion;
a distribution analysis module configured to generate the probability density function (PDF), cumulative distribution function (CDF), and the inverse cumulative distribution function (ICDF) of distributions calculated in the Risk Simulator; and
a Statistical Analyses module comprising a Descriptive Statistics sub-module which includes descriptive statistics functions, a Distributional Fitting sub-module which includes distributional fitting functions, a Histogram and Charts sub-module which includes histogram and chart generating functions, a Hypothesis Testing sub-module which includes hypothesis testing functions to determine the probability that a given hypothesis is true, a Nonlinear Extrapolation sub-module which includes extrapolation functions that extrapolates or extends non-linear data into the future, a Normality Test sub-module which includes functions for determining whether the user provided data set is well-modeled by a normal distribution and how likely it is for a variable underlying the data set to be normally distributed, a Stochastic Process Parameter Estimation sub-module which includes functions for estimating parameters to achieve a best fit regarding characteristics of the user provided data set, a Time-series Autocorrelation sub-module which includes functions for identifying auto correlation as a function of time, and a Trend Line Projection sub-module which includes trend line projection functions.
2. The system of claim 1, wherein the best-fitting model for the user-provided data is determined by applying several goodness-of-fit statistics comprising a t-statistic, F-statistic, R-squared statistic, adjusted R-squared statistic, Durbin-Watson statistic, Akaike Criterion, Schwartz Criterion, and their respective probabilities.
3. The system of claim 1, wherein said testing for regression errors is accomplished by graphically representing each independent variable against the at least one dependent variable.
4. The system of claim 1, wherein said testing for regression errors is accomplished by applying White's test.
5. The system of claim 1, wherein when heteroskedasticity is detected by said testing, said heteroskedasticity is made homoskedastic by using a weighted least squares (WLS) approach.
6. The system of claim 1, wherein Park's test is used to test for and fix heteroskedasticity.
7. The system of claim 1, wherein the test for multicollinearity comprises gauging whether the R-squared value too high and the t-statistics are too low.
8. The system of claim 1, wherein the test for multicollinearity comprises a correlation matrix between the independent variables, wherein when correlation values in said matrix are high the system identifies a potential for multicollinearity.
9. The system of claim 8, wherein the system determines that multicollinearity is severe when the cross correlation between the independent variables has an absolute value that is greater than 0.75.
10. The system of claim 1, wherein the test for multicollinearity comprises the use of a variance inflation factor (VIF).
11. The system of claim 1, wherein the test for autocorrelation comprises graphing the time series of a regression equation's residuals.
12. The system of claim 11, wherein the system determines that autocorrelation exists when said residuals exhibit some cyclicality.
13. The system of claim 1, wherein the test for autocorrelation comprises using the Durbin-Watson statistic.
14. The system of claim 13, wherein the Durbin-Watson statistic is used to identify model misspecification.
15. The system of claim 1, wherein the test for autocorrelation comprises the Breusch-Godfrey test.
16. The system of claim 1, wherein any autocorrelation is fixed by taking the lags of the at least one dependent variable for a relevant period, adding them into the regression function, and testing for their significance.
17. The system of claim 1, wherein the ARIMA module uses an Akaike Information Criterion (AIC) and Schwartz Criterion (SC) to analyze and rank the forecasting and risk simulation models.
18. The system of claim 17, wherein said ARIMA module generates a report comprising autocorrelation (AC) and partial autocorrelation statistics (PAC), wherein said AC, PAC, SC, and AIC are used to help identify the best model.
US14/693,910 2009-02-11 2015-04-23 Integrated Risk Management System Abandoned US20150227656A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US14/693,910 US20150227656A1 (en) 2009-02-11 2015-04-23 Integrated Risk Management System

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US12/378,169 US9020857B2 (en) 2009-02-11 2009-02-11 Integrated risk management process
US14/693,910 US20150227656A1 (en) 2009-02-11 2015-04-23 Integrated Risk Management System

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/378,169 Continuation US9020857B2 (en) 2009-02-11 2009-02-11 Integrated risk management process

Publications (1)

Publication Number Publication Date
US20150227656A1 true US20150227656A1 (en) 2015-08-13

Family

ID=42541161

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/378,169 Active 2032-05-02 US9020857B2 (en) 2009-02-11 2009-02-11 Integrated risk management process
US14/693,910 Abandoned US20150227656A1 (en) 2009-02-11 2015-04-23 Integrated Risk Management System

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/378,169 Active 2032-05-02 US9020857B2 (en) 2009-02-11 2009-02-11 Integrated risk management process

Country Status (1)

Country Link
US (2) US9020857B2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018065942A1 (en) * 2016-10-07 2018-04-12 KPMG Australia IP Holdings Pty Ltd Method and system for collecting, visualising and analysing risk data
US20180268336A1 (en) * 2017-03-15 2018-09-20 Katerra, Inc. Generating Construction Metrics Using Probabilistic Methods
US10496948B1 (en) * 2015-12-18 2019-12-03 Xactly Corporation Computer trend visualization using quadratic simplified closed form linear regression
US10963292B2 (en) * 2013-05-06 2021-03-30 Sas Institute Inc. Techniques to manage virtual classes for statistical tests
US11586705B2 (en) 2019-12-02 2023-02-21 International Business Machines Corporation Deep contour-correlated forecasting

Families Citing this family (43)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110054982A1 (en) * 2009-09-01 2011-03-03 Edward Kim Methods and systems for randomizing starting retail store inventory when determining distribution center and warehouse demand forecasts
US8332250B2 (en) * 2010-03-15 2012-12-11 Xerox Corporation Systems and methods for determining print revenue behavior
US20110307300A1 (en) * 2010-05-11 2011-12-15 Npa Vision, Inc. Method and system for identifying and then probabalistically projecting and aggregating difficult-to-predict industry operating statistics
JP5548048B2 (en) * 2010-06-30 2014-07-16 アズビル株式会社 Signal determination apparatus and signal determination method
US20120004946A1 (en) * 2010-07-01 2012-01-05 Bank Of America Corporation Integrated Operational Risk Management
US10805331B2 (en) 2010-09-24 2020-10-13 BitSight Technologies, Inc. Information technology security assessment system
US20120179421A1 (en) * 2010-12-07 2012-07-12 Gautam Dasgupta Emergency Response Management Apparatuses, Methods and Systems
JP5896668B2 (en) * 2011-09-26 2016-03-30 三菱重工業株式会社 Demand forecasting device, demand forecasting method, and demand forecasting program
CN103930912A (en) * 2011-11-08 2014-07-16 国际商业机器公司 Time-series data analysis method, system and computer program
US9299051B2 (en) * 2012-09-12 2016-03-29 Siemens Industry, Inc. Methods and systems for evaluating the performance of building processes
US9881339B2 (en) * 2012-12-18 2018-01-30 Johnathan Mun Project economics analysis tool
US20140180755A1 (en) * 2012-12-21 2014-06-26 Fluor Technologies Corporation Identifying, Assessing, And Tracking Black Swan Risks For An Engineering And Construction Program
US9438615B2 (en) 2013-09-09 2016-09-06 BitSight Technologies, Inc. Security risk management
CN103632197B (en) * 2013-11-01 2016-08-31 国家电网公司 A kind of electricity transaction information issues Risk Identification Method
US20220027928A1 (en) * 2014-09-17 2022-01-27 Newmark & Company Real Estate, Inc. Industrial Momentum Index
US10453014B2 (en) 2015-07-14 2019-10-22 Bank Of America Corporation Systems and methods for utilizing compliance drivers to conserve system resources and reduce compliance violations
US20170076216A1 (en) * 2015-09-10 2017-03-16 Siemens Aktiengesellschaft Maintenance event planning using adaptive predictive methodologies
CN105512466A (en) * 2015-11-30 2016-04-20 华北电力大学 Power grid project implementation phase risk early warning method based on extreme value theory
US11182720B2 (en) 2016-02-16 2021-11-23 BitSight Technologies, Inc. Relationships among technology assets and services and the entities responsible for them
CN106934254B (en) * 2017-02-15 2020-05-26 中国银联股份有限公司 Analysis method and device for open source license
US10425380B2 (en) 2017-06-22 2019-09-24 BitSight Technologies, Inc. Methods for mapping IP addresses and domains to organizations using user activity data
CN108229071B (en) * 2018-02-09 2021-06-22 杭州安脉盛智能技术有限公司 Cutter performance degradation evaluation method and system based on AR model and SVDD algorithm
US10257219B1 (en) 2018-03-12 2019-04-09 BitSight Technologies, Inc. Correlated risk in cybersecurity
US10812520B2 (en) 2018-04-17 2020-10-20 BitSight Technologies, Inc. Systems and methods for external detection of misconfigured systems
CN109215781B (en) * 2018-09-14 2021-11-12 道之精准医学科技(上海)有限公司 Method and system for constructing risk assessment model of Kawasaki disease based on logistic algorithm
US11200323B2 (en) * 2018-10-17 2021-12-14 BitSight Technologies, Inc. Systems and methods for forecasting cybersecurity ratings based on event-rate scenarios
US10521583B1 (en) 2018-10-25 2019-12-31 BitSight Technologies, Inc. Systems and methods for remote detection of software through browser webinjects
US11699094B2 (en) * 2018-10-31 2023-07-11 Salesforce, Inc. Automatic feature selection and model generation for linear models
US10726136B1 (en) 2019-07-17 2020-07-28 BitSight Technologies, Inc. Systems and methods for generating security improvement plans for entities
US11956265B2 (en) 2019-08-23 2024-04-09 BitSight Technologies, Inc. Systems and methods for inferring entity relationships via network communications of users or user devices
CN110597875B (en) * 2019-08-28 2022-02-25 上海汇纳数据科技有限公司 Method, system, medium and equipment for judging store operation condition
US20220188322A1 (en) * 2019-09-04 2022-06-16 Michael E. Adel Method and system of database analysis and compression
US10848382B1 (en) 2019-09-26 2020-11-24 BitSight Technologies, Inc. Systems and methods for network asset discovery and association thereof with entities
US11032244B2 (en) 2019-09-30 2021-06-08 BitSight Technologies, Inc. Systems and methods for determining asset importance in security risk management
CN111080502B (en) * 2019-12-17 2023-09-08 清华苏州环境创新研究院 Big data identification method for regional enterprise data abnormal behaviors
US10791140B1 (en) 2020-01-29 2020-09-29 BitSight Technologies, Inc. Systems and methods for assessing cybersecurity state of entities based on computer network characterization
US10893067B1 (en) 2020-01-31 2021-01-12 BitSight Technologies, Inc. Systems and methods for rapidly generating security ratings
US10764298B1 (en) 2020-02-26 2020-09-01 BitSight Technologies, Inc. Systems and methods for improving a security profile of an entity based on peer security profiles
US11023585B1 (en) 2020-05-27 2021-06-01 BitSight Technologies, Inc. Systems and methods for managing cybersecurity alerts
US11122073B1 (en) 2020-12-11 2021-09-14 BitSight Technologies, Inc. Systems and methods for cybersecurity risk mitigation and management
EP4272139A1 (en) * 2020-12-31 2023-11-08 Schneider Electric Systems USA, Inc. Systems and methods for addressing gaps in an industrial operation due to operator variability
CN113793689B (en) * 2021-08-06 2023-10-24 兰州理工大学 Epidemic disease monitoring method based on accumulation and control diagram
CN116227961B (en) * 2022-12-05 2024-04-09 贝壳找房(北京)科技有限公司 Resource allocation method, device, equipment and computer readable storage medium

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7212978B2 (en) * 1998-06-01 2007-05-01 Harrah's Operating Company, Inc. Customer valuation in a resource price manager
US7742972B2 (en) * 1999-07-21 2010-06-22 Longitude Llc Enhanced parimutuel wagering
US7313531B2 (en) * 2001-11-29 2007-12-25 Perot Systems Corporation Method and system for quantitatively assessing project risk and effectiveness
US7742940B1 (en) * 2002-12-17 2010-06-22 Hewlett-Packard Development Company, L.P. Method and system for predicting revenue based on historical pattern indentification and modeling
JP4756675B2 (en) * 2004-07-08 2011-08-24 インターナショナル・ビジネス・マシーンズ・コーポレーション System, method and program for predicting computer resource capacity
US7797184B2 (en) * 2004-10-06 2010-09-14 Hewlett-Packard Development Company, L.P. Methods and systems for cumulative attribute forecasting using a PDF of a current-to-future value ratio
US7664671B2 (en) * 2004-12-01 2010-02-16 Hewlett-Packard Development Company, L.P. Methods and systems for profile-based forecasting with dynamic profile selection
US7836111B1 (en) * 2005-01-31 2010-11-16 Hewlett-Packard Development Company, L.P. Detecting change in data

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10963292B2 (en) * 2013-05-06 2021-03-30 Sas Institute Inc. Techniques to manage virtual classes for statistical tests
US10496948B1 (en) * 2015-12-18 2019-12-03 Xactly Corporation Computer trend visualization using quadratic simplified closed form linear regression
WO2018065942A1 (en) * 2016-10-07 2018-04-12 KPMG Australia IP Holdings Pty Ltd Method and system for collecting, visualising and analysing risk data
GB2569763A (en) * 2016-10-07 2019-06-26 KPMG Australia IP Holdings Pty Ltd Method and system for collecting, visualising and analysing risk data
US20180268336A1 (en) * 2017-03-15 2018-09-20 Katerra, Inc. Generating Construction Metrics Using Probabilistic Methods
US11586705B2 (en) 2019-12-02 2023-02-21 International Business Machines Corporation Deep contour-correlated forecasting

Also Published As

Publication number Publication date
US20100205042A1 (en) 2010-08-12
US9020857B2 (en) 2015-04-28

Similar Documents

Publication Publication Date Title
US9020857B2 (en) Integrated risk management process
US11062274B2 (en) Maintenance planning apparatus and maintenance planning method
Snyder et al. Forecasting the intermittent demand for slow-moving inventories: A modelling approach
US9547477B2 (en) Autoeconometrics modeling method
US9811794B2 (en) Qualitative and quantitative modeling of enterprise risk management and risk registers
US20150088783A1 (en) System and method for modeling and quantifying regulatory capital, key risk indicators, probability of default, exposure at default, loss given default, liquidity ratios, and value at risk, within the areas of asset liability management, credit risk, market risk, operational risk, and liquidity risk for banks
US8290802B2 (en) System and method for product deployment and in-service product risk simulation
US6799154B1 (en) System and method for predicting the timing of future service events of a product
US20140324521A1 (en) Qualitative and quantitative analytical modeling of sales performance and sales goals
US20130103615A1 (en) Project economics analysis tool
US20100205108A1 (en) Credit and market risk evaluation method
US20040236676A1 (en) Disaster risk assessment system, disaster risk assessment support method, disaster risk assessment service providing system, disaster risk assessment method, and disaster risk assessment service providing method
WO2015094545A1 (en) System and method for modeling and quantifying regulatory capital, key risk indicators, probability of default, exposure at default, loss given default, liquidity ratios, and value at risk, within the areas of asset liability management, credit risk, market risk, operational risk, and liquidity risk for banks
Mun Understanding and choosing the right probability distributions
Horbenko et al. Robust estimation of operational risk
Owadally et al. An agent-based system with temporal data mining for monitoring financial stability on insurance markets
WO2017163259A2 (en) Service churn model
Covey et al. Testing for Granger's full causality
Kulk et al. Quantifying IT estimation risks
Lee A semi-parametric approach for estimating critical fractiles under autocorrelated demand
Haris et al. Warranty Claim Quantity Forecasting via Multiple Linear Regressions
Cook Estimating required contingency funds for construction projects using multiple linear regression
Jadraque A phenomenological approach to predicting tuberculosis cases with an assessment of measurement errors
US11854087B1 (en) Systems and methods for water loss mitigation messaging
Meyers et al. Thinking outside the triangle

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION