US20210027183A1 - System and method for performance evaluation of probability forecast - Google Patents
System and method for performance evaluation of probability forecast Download PDFInfo
- Publication number
- US20210027183A1 US20210027183A1 US17/066,695 US202017066695A US2021027183A1 US 20210027183 A1 US20210027183 A1 US 20210027183A1 US 202017066695 A US202017066695 A US 202017066695A US 2021027183 A1 US2021027183 A1 US 2021027183A1
- Authority
- US
- United States
- Prior art keywords
- probability distribution
- forecast
- score
- prior knowledge
- distribution forecast
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000011156 evaluation Methods 0.000 title claims abstract description 30
- 238000009826 distribution Methods 0.000 claims abstract description 187
- 238000004364 calculation method Methods 0.000 description 14
- 230000006870 function Effects 0.000 description 9
- 230000015654 memory Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 4
- 238000000638 solvent extraction Methods 0.000 description 4
- 230000008901 benefit Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000013209 evaluation strategy Methods 0.000 description 2
- 238000005192 partition Methods 0.000 description 2
- 239000013598 vector Substances 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003467 diminishing effect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003203 everyday effect Effects 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000009827 uniform distribution Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/04—Inference or reasoning models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/18—Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N5/00—Computing arrangements using knowledge-based models
- G06N5/02—Knowledge representation; Symbolic representation
-
- G06N7/005—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N7/00—Computing arrangements based on specific mathematical models
- G06N7/01—Probabilistic graphical models, e.g. probabilistic networks
Definitions
- the present application relates in general to evaluating probability distribution forecasts. More specifically, the present application refers to a system and method for evaluating probability distribution forecasts by computing an accuracy score and an information score and further computing a performance score based on the accuracy score and information score.
- the probability distribution forecast evaluation system comprises at least one processor and at least one memory device.
- the at least one memory device may store a plurality of instructions which, when executed by the at least one processor, cause the at least one processor to operate with the at least one memory device to acquire one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an accuracy score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an information score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; and compute a performance score based on the accuracy score and the information score.
- the accuracy score may be computed based on the probability distribution forecast and the probability distribution realization.
- the accuracy score may further be computed by calculating a dissimilarity score between the probability distribution forecast and the probability distribution realization.
- the probability distribution forecast evaluation system may calculate a dissimilarity score that is either (1) the Kullback-Leibler (KL) divergence between the probability distribution forecast and the probability distribution realization, or (2) a quadratic approximation of the KL divergence between the probability distribution forecast and the probability distribution realization.
- the probability distribution forecast evaluation system may compute the information score based on the probability distribution realization and the prior knowledge of the probability distribution forecast.
- the information score may be computed by calculating a dissimilarity score between the probability distribution realization and the prior knowledge of the probability distribuion forecast.
- the probability distribution forecast evaluation system may calculate a dissimilarity score that is either (1) the Kullback-Leibler (KL) divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast, or (2) a quadratic approximation of the KL divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast.
- KL Kullback-Leibler
- the probability distribution forecast evaluation system may compute the performance score by subtracting the accuracy score from the information score. Further, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast are computed based on samples. Still further, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast may be partitioned into discrete probability bins. In some embodiments, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast contain sample errors and the performance score may be normalized to account for the sample errors. Lastly, a relative performance score may be further computed based on the computed performance score and the entropy of the prior knowledge of the probability distribution forecast.
- FIG. 1 is a schematic diagram of an example implementation of the presently described probability distribution forecast evaluation system.
- FIG. 2 is a flowchart of an example method for implementing the presently described probability distribution forecast evaluation system.
- the systems and method disclosed herein rely in different capacities on scoring and manipulating probability distributions.
- These probability distributions may correspond to any set of financial, business, weather, or other data.
- the probability distributions may be forecasts or predictions of a given probability distribution, actual realizations of a probability distributions, or prior knowledge of a given probability distribution, which may also comprise a reference probability distribution.
- the probability distributions may be stored either as continuous probability density functions, as discretized probability mass functions stored in vectors, or as samples that correspond to the probability distribution. It should be understood that the probability distributions may also be stored in other manners without departing from the scope of the present disclosure.
- FIG. 1 is a schematic diagram of an example implementation 100 of the system described herein.
- the example implementation 100 includes a probability distribution forecast evaluation system 110 .
- the probability distribution forecast evaluation system 110 includes an analysis module 120 comprising a processor 130 coupled to a memory 140 .
- the analysis module 120 is coupled to a storage device 150 via link 142 and a communication device via link 144 .
- the storage device 150 is further connected to the communication device 160 via link 152 .
- the communication device connects the probability distribution forecast evaluation system 110 to the network 170 via link 162 .
- the probability distribution forecast evaluation system 110 further includes a user interface device 112 , which includes a display 114 and a user entry device 116 .
- the user interface device 112 may consist of a display 114 implemented as a computer monitor and a user entry device 116 implemented as one or more of a computer mouse, keyboard, voice recognition system, touch screen device, or other similar computer input device.
- the user interface device 112 may be implemented as a physically separate device that connects to the probability distribution system 110 via the network 170 and link 162 .
- the user interface 112 could be a separate computing device such as a laptop, desktop computer, smartphone, or tablet.
- the display 114 may include a computer display, smartphone display, or tablet display and the user entry device may include a smartphone or tablet touchscreen, voice recognition system, or keyboard.
- the probability distribution forecast evaluation system 110 may not include any user interface.
- the network 170 may be implemented as a local, closed network, or may include one or more connections to the Internet.
- the link 162 may be implemented as a wired connection, or as a wireless connection such as Wi-Fi, Bluetooth, 4G/LTE, or any other wireless protocol.
- the storage device 150 may be implemented as any for of data storage device.
- the storage device 150 may be implemented as one or more of a hard disk drive (HDD), solid state drive (SSD), flash-based storage, read-only memory (ROM).
- the storage device 150 may be coupled to the communication device 160 via link 152 to receive information or data from the network 170 .
- the storage device may store one or more of a probability distribution forecast, a realization of a probability distribution forecast, and a prior knowledge of the of the probability distribution forecast. This information may be stored either as a continuous forecast and may also or alternatively be stored as a set of empirical samples.
- the processor 130 may be configured to perform a series of instructions that are stored in the memory 140 .
- the memory 140 may be implemented as one or more random access memories (RAMs). Although depicted in the singular, the processor 130 may be implemented as one or more computer processing units (CPUs).
- the memory 140 may contain a series of instructions that, when executed by the processor 130 , cause the processor 130 to acquire one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an accuracy score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an information score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; and compute a performance score based on the accuracy score and the information score.
- FIG. 2 depicts a flowchart of an example method 200 for implementing the system described herein.
- the method 200 is an implementation of the probability distribution forecast evaluation system 110 .
- the method 200 may be implemented as a series of instructions stored on the memory 140 that cause the processor 130 to perform the method 200 when executed.
- the method 200 acquires a probability distribution forecast (“forecast”), a probability distribution realization corresponding to the probability distribution forecast (“realization”), and a prior knowledge of the probability distribution forecast (“prior knowledge”).
- the method 200 may not acquire all three of the forecast, realization, and prior knowledge. Instead, it may only acquire a subset, such as just the forecast or both of the forecast and the prior knowledge.
- the forecast, realization, and prior knowledge may be user entered, such as through the user interface 112 ; may be stored, such as in storage device 150 ; or may be looked up, downloaded, or acquired over a network, such as network 170 .
- the system disclosed herein contemplates that each of the forecast, realization, and prior knowledge are acquired by different means. For example, the forecast may be user entered, the realization may be downloaded from the Internet, and the prior knowledge may be stored on a storage device.
- the forecast may be a stored version of a previously-recorded forecast of a statistical probability distribution forecast.
- the forecast may predict metrics in one or more areas such as financial metrics, economic metrics, business performance metrics, weather metrics, or any other metric that may be useful to predict.
- the realization may be a stored version of the realized probability distribution of the metrics that the forecast predicted.
- the prior knowledge may be a stored version of prior knowledge about the probability distribution that the forecast predicted. For example, this may be a reference probability distribution for the forecast.
- Each of the forecast, realization, and prior knowledge may be stored as a probability density function, a probability mass function, or a collection of samples. Further, the forecast, realization, and prior knowledge may be partitioned into one or more bins. Each of the forecast, realization, and prior knowledge may have a different number of bins. Further, each of the forecast, realization, and prior knowledge may be stored in a different function. For example, the forecast and prior knowledge may be stored as a probability density function and the realization may be stored as a collection of samples.
- the method 200 progresses to calculate an accuracy score for the forecast.
- the accuracy score shows how reliable a forecast is, measuring how reliably the forecast predicts the probability distribution. This measure is important to evaluating a probability distribution forecast because accuracy is an important component of a useful forecast.
- the accuracy score may be based on the forecast and the realization. In some embodiments, the accuracy score is calculated by computing a dissimilarity score between the forecast and the realization. A dissimilarity score may be a measurement of how different two sets of data are. For example, the accuracy score may be calculated by calculating the Kullback-Leibler (KL) divergence between the forecast and the realization, as shown below:
- KL Kullback-Leibler
- This implementation of the accuracy score will be lower when a forecast is more accurate.
- the accuracy score may be calculated using a quadratic approximation of the KL divergence between the forecast and the realization.
- the accuracy score is calculated based on samples of one or more of the forecast and the realization, it may be useful to calculate the accuracy score using the quadratic approximation of the KL divergence between the samples of the forecast and the realization.
- the KL divergence between the samples of the forecast and the realization can be approximated with chi-squared statistics using a quadratic approximation as shown below:
- the other may be converted to samples by partitioning it into bins.
- the forecast may be partitioned into bins.
- the system of the present disclosure may convert both distributions by partitioning them into bins. As described above, it may be beneficial to partition both distributions into the same number of bins.
- the method 200 calculates an information score for the forecast.
- the information score measures the amount of information a forecast contains compared to the prior knowledge about the forecasted metrics. This prior knowledge may include reference probability distributions for the forecasted metrics.
- the information score is an important part of evaluating forecasts. Conventional techniques, such as the Brier score, treat more certainty (or a better accuracy score) as the better forecast. Such systems fail to properly account for the amount of information in a forecast. For example, even if a forecast is less certain than the prior knowledge, it is still useful to know that the future is more uncertain than the prior knowledge. As a further example, an area may get rain on 30% of the days over the course of the year.
- the information score may be based on the forecast and the prior knowledge.
- the information score is calculated by computing a dissimilarity score between the forecast and the realization.
- the information score may be calculated by calculating the KL divergence between the realization and the prior knowledge, as shown below:
- This implementation of the information score will be higher when a forecast provides more information.
- the forecast may be a conditional distribution.
- both the realization and the forecast may be conditional on the same condition.
- the prior knowledge is generally either an unconditional distribution, or is conditional on a different condition from the forecast and the realization.
- the forecast may be used as a proxy for the realization in the information score calculation.
- the information score may be calculated as shown below:
- the information score may be calculated using a quadratic approximation of the KL divergence between the realization and the prior knowledge. For example, if the accuracy score is calculated based on samples of one or more of the realization and the prior knowledge, it may be useful to calculate the information score using the quadratic approximation of the KL divergence between the samples of the realization and the prior knowledge.
- the KL divergence between the samples of the realization and the prior knowledge can be approximated with chi-squared statistics using a quadratic approximation as shown below:
- the forecast may be used as a proxy for the realization in the information score calculation.
- the information score can be calculated with the quadratic approximation of the KL divergence between the samples of the forecast and the prior knowledge as shown below:
- the other may be converted to samples by partitioning it into bins.
- the forecast may be partitioned into bins.
- the realization distribution is stored as a set of samples in 5 bins
- he prior knowledge may be partitioned into 5 bins.
- the disclosed system may convert both distributions by partitioning them into bins. As described above, it may be beneficial to partition both distributions into the same number of bins.
- the method 200 calculates a performance score based on the accuracy score and the information score.
- the performance score is calculated by subtracting the accuracy score from the information score, as shown below:
- the accuracy score and information score can be calculated independently. This means that, the number of bins used in the accuracy score calculation can differ from the number of bins used in the information score calculation. This improves the simplicity and numerical stability of the calculation.
- the above implementation of the performance score calculation does not depend on how each of the information score and the accuracy score are calculated. Accordingly, this implementation may be used even if the accuracy score and information score are calculated based on samples or if the information score is calculated by using the forecast as a proxy for the realization as discussed above. In fact, in some embodiments it may be preferable to use the forecast as a proxy for the realization in the information score calculation. Doing this may result in a simplified, and therefore faster, calculation. Calculating the performance score in this manner may also be more robust if one or more of the forecast, prior knowledge, and realization is stored as samples. An example implementation of this is shown below:
- the method 200 will also calculate a relative performance score (not depicted in FIG. 2 ).
- the relative performance score may be calculated based on the entropy of the prior knowledge.
- the entropy of the prior knowledge may be calculated as defined below:
- the method 200 will also calculate a confidence interval for the performance score.
- the confidence interval for the performance score may be based on the probability distribution of the performance score. For example, in some embodiments, those with skill in the art will note that one or more of the accuracy score and the information score may have a probability distribution if calculated based on samples. Accordingly, a probability distribution of the performance score may be calculated based on one or more of the probability distribution of the accuracy score and the probability distribution of the information score. The confidence interval for the performance score may then be calculated based on the probability distribution of the performance score. In practice, the confidence interval of the performance score may be valuable because it provides more information on the range of values that the performance score may take. This provides users with a greater understanding of the performance of the forecast.
- each of the multiple forecasts may differ from one another.
- one or more of the prior knowledge and realization could be different for one or more of the multiple forecasts as well.
- each of the forecasts may use different bin divisions and the prior knowledge and realization that correspond to each of the forecasts may also use different bin divisions. For example, if the multiple forecasts use different bin divisions, the prior knowledge and realization that correspond to each of the multiple forecasts may use the same division as their corresponding forecast.
- the method 200 may select bin divisions for each distribution such that the probability bins of the multiple forecasts are the same.
- the accuracy and information scores may be calculated as defined below:
- the method 200 may use the forecast as a proxy for the realization in the information score calculation as described above.
- the information score may be calculated as shown below:
- the method 200 may select equal quantiles for both the prior knowledge and the realization.
- the performance score may be calculated as defined below:
- q ) 2 - K ⁇ ⁇ k 1 K ⁇ ( f k - o _ k
- q ) 2 - ⁇ k 1 K ⁇ ( 1 K - o _ k
- the method 200 may use the forecast as a proxy for the realization in the information score calculation as described above.
- the performance score may be calculated as shown below:
- f - f k ) 2 - K ⁇ ⁇ k 1 K ⁇ ( f k - o _ k
- f ) 2 - ⁇ k 1 K ⁇ ( 1 K - o _ k
- the method 200 may calculate the information and accuracy scores as shown below:
- the method 200 may use the forecast as a proxy for the realization in the information score calculation as described above.
- the information score may be calculated as shown below:
- the method 200 may substitute the denominator in the above calculations for 1/K, resulting in the simplified calculations shown below:
- the method 200 may use the forecast as a proxy for the realization in the information score calculation as described above.
- the information score may be calculated as shown below:
- the method 200 may also calculate a relative performance score for the multiple forecasts.
- the relative performance score for the multiple forecasts may be calculated based on the entropy of the prior knowledge corresponding to the multiple forecasts.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- General Engineering & Computer Science (AREA)
- Computational Mathematics (AREA)
- Mathematical Analysis (AREA)
- Mathematical Optimization (AREA)
- Pure & Applied Mathematics (AREA)
- Computing Systems (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Algebra (AREA)
- Probability & Statistics with Applications (AREA)
- Computational Linguistics (AREA)
- Operations Research (AREA)
- Databases & Information Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
A method and system for probability distribution forecast evaluation are disclosed. The present disclosure is directed to embodiments of a system that evaluates probability distribution forecasts by acquiring one or more of a probability distribution forecast, a probability distribution realization, and a prior knowledge of the probability distribution forecast. The system disclosed herein may then compute an accuracy score and an information score based on the acquired forecast, realization, and prior knowledge. In evaluating the forecast, a performance score may also be computed based on the accuracy score and the information score.
Description
- This application claims priority to and the benefit of U.S. Provisional Patent Application Ser. No. 62/317,774, filed on Apr. 4, 2016, and U.S. patent application Ser. No. 15/478,910, filed on Apr. 4, 2017, both of which are incorporated herein by reference in their entirety.
- The present application relates in general to evaluating probability distribution forecasts. More specifically, the present application refers to a system and method for evaluating probability distribution forecasts by computing an accuracy score and an information score and further computing a performance score based on the accuracy score and information score.
- Modern forecasting techniques and technologies have resulted in a large number of forecasts and predictions across a variety of industries and applications. The widespread availability of such forecasts enables businesses, groups, and individuals to better plan their behavior and generally prepare for the future. However, these forecasts are only useful to the extent they are accurate. It is therefore important for users of these forecasts to have a way to evaluate these forecasts.
- The large number of available forecasts and large number of predictions inherent in each forecast means that any forecast evaluation must be systematic and repeatable between multiple forecasts. Existing forecast evaluation strategies focus primarily on the forecasts' accuracy, or how far the forecast's predictions tend to be from what actually happens. However, focusing on accuracy alone fails to measure whether the forecast provides more information than what is already available. A forecast that provides more information may be more useful even if it is less accurate. For example, a forecast that tells you it rains on 30% of the days in July may be highly accurate. However, a forecast that predicts an 80% chance of rain tomorrow may be much more useful, even if it is inaccurate and there is actually a 90% chance of rain tomorrow. Therefore, a forecast evaluation strategy that takes into account both a forecast's accuracy and the additional information it provides is needed.
- A system and method to enable probability distribution forecast evaluation are disclosed. The probability distribution forecast evaluation system comprises at least one processor and at least one memory device. The at least one memory device may store a plurality of instructions which, when executed by the at least one processor, cause the at least one processor to operate with the at least one memory device to acquire one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an accuracy score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an information score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; and compute a performance score based on the accuracy score and the information score.
- The accuracy score may be computed based on the probability distribution forecast and the probability distribution realization. The accuracy score may further be computed by calculating a dissimilarity score between the probability distribution forecast and the probability distribution realization. The probability distribution forecast evaluation system may calculate a dissimilarity score that is either (1) the Kullback-Leibler (KL) divergence between the probability distribution forecast and the probability distribution realization, or (2) a quadratic approximation of the KL divergence between the probability distribution forecast and the probability distribution realization.
- The probability distribution forecast evaluation system may compute the information score based on the probability distribution realization and the prior knowledge of the probability distribution forecast. The information score may be computed by calculating a dissimilarity score between the probability distribution realization and the prior knowledge of the probability distribuion forecast. The probability distribution forecast evaluation system may calculate a dissimilarity score that is either (1) the Kullback-Leibler (KL) divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast, or (2) a quadratic approximation of the KL divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast.
- The probability distribution forecast evaluation system may compute the performance score by subtracting the accuracy score from the information score. Further, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast are computed based on samples. Still further, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast may be partitioned into discrete probability bins. In some embodiments, one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast contain sample errors and the performance score may be normalized to account for the sample errors. Lastly, a relative performance score may be further computed based on the computed performance score and the entropy of the prior knowledge of the probability distribution forecast.
-
FIG. 1 is a schematic diagram of an example implementation of the presently described probability distribution forecast evaluation system. -
FIG. 2 is a flowchart of an example method for implementing the presently described probability distribution forecast evaluation system. - The systems and method disclosed herein rely in different capacities on scoring and manipulating probability distributions. These probability distributions may correspond to any set of financial, business, weather, or other data. The probability distributions may be forecasts or predictions of a given probability distribution, actual realizations of a probability distributions, or prior knowledge of a given probability distribution, which may also comprise a reference probability distribution. The probability distributions may be stored either as continuous probability density functions, as discretized probability mass functions stored in vectors, or as samples that correspond to the probability distribution. It should be understood that the probability distributions may also be stored in other manners without departing from the scope of the present disclosure.
- Throughout this detailed description, multiple equations are used to illustrate potential embodiments of the system disclosed herein. To aid in understanding these exemplary equations, Table 1 below sets out definitions for terms used in the equations.
-
TABLE 1 Term Definition K Number of bins in a given probability distribution, or a number of possible outcomes N Total number of realized samples or total number of forecasts Mk Total number of forecasts for the k-th bin. fk Forecast probability (frequency) for k-th bin fnk Forecast probability (frequency) for the k-th bin at the n-th sample pk Realized probability (frequency) for the k-th bin ōk Observed frequency for the k-th bin; i.e., an empirical ex post estimate of pk onk For the n-th sample, if it falls in the k-th bin, onk = 1; otherwise, onk = 0 ōk|p Observed frequency for the k-th bin, according to a bin division of p qk Prior knowledge for the k-th bin, or a reference probability distribution used as the prior knowledge f, p, ō, q The probability density functions with the definitions as described above (i.e., f-forecast, p-realization, ō-observed realization estimate, and q-prior knowledge) f, p, ō, q Vectors representing the probability mass functions with the definitions as described above (i.e., the discretized version of the probability density functions f, p, ō, q with a certain bin division) - Those skilled in the art will understand that these exemplary equations are not the only way to implement embodiments of the system disclosed herein, and various changes and modifications to the preferred embodiments will be apparent to them. Such changes and modifications can be made without departing from the spirit and scope of the present subject matter and without diminishing its intended advantages. It is therefore intended that such changes and modifications be covered by the appended claims.
-
FIG. 1 is a schematic diagram of anexample implementation 100 of the system described herein. Theexample implementation 100 includes a probability distributionforecast evaluation system 110. The probability distributionforecast evaluation system 110 includes ananalysis module 120 comprising aprocessor 130 coupled to amemory 140. Theanalysis module 120 is coupled to astorage device 150 vialink 142 and a communication device vialink 144. Thestorage device 150 is further connected to thecommunication device 160 vialink 152. The communication device connects the probability distributionforecast evaluation system 110 to thenetwork 170 via link 162. The probability distributionforecast evaluation system 110 further includes auser interface device 112, which includes adisplay 114 and auser entry device 116. - The
user interface device 112 may consist of adisplay 114 implemented as a computer monitor and auser entry device 116 implemented as one or more of a computer mouse, keyboard, voice recognition system, touch screen device, or other similar computer input device. In an alternative embodiment not depicted inFIG. 1 , theuser interface device 112 may be implemented as a physically separate device that connects to theprobability distribution system 110 via thenetwork 170 and link 162. For example, theuser interface 112 could be a separate computing device such as a laptop, desktop computer, smartphone, or tablet. In this case, thedisplay 114 may include a computer display, smartphone display, or tablet display and the user entry device may include a smartphone or tablet touchscreen, voice recognition system, or keyboard. In still other embodiments, the probability distributionforecast evaluation system 110 may not include any user interface. - The
network 170 may be implemented as a local, closed network, or may include one or more connections to the Internet. The link 162 may be implemented as a wired connection, or as a wireless connection such as Wi-Fi, Bluetooth, 4G/LTE, or any other wireless protocol. Thestorage device 150 may be implemented as any for of data storage device. For example, thestorage device 150 may be implemented as one or more of a hard disk drive (HDD), solid state drive (SSD), flash-based storage, read-only memory (ROM). Thestorage device 150 may be coupled to thecommunication device 160 vialink 152 to receive information or data from thenetwork 170. The storage device may store one or more of a probability distribution forecast, a realization of a probability distribution forecast, and a prior knowledge of the of the probability distribution forecast. This information may be stored either as a continuous forecast and may also or alternatively be stored as a set of empirical samples. - The
processor 130 may be configured to perform a series of instructions that are stored in thememory 140. Thememory 140 may be implemented as one or more random access memories (RAMs). Although depicted in the singular, theprocessor 130 may be implemented as one or more computer processing units (CPUs). As discussed in greater detail below, thememory 140 may contain a series of instructions that, when executed by theprocessor 130, cause theprocessor 130 to acquire one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an accuracy score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; compute an information score based on the one or more of a probability distribution forecast, a probability distribution realization corresponding to the probability distribution forecast, and a prior knowledge of the probability distribution forecast; and compute a performance score based on the accuracy score and the information score. -
FIG. 2 depicts a flowchart of anexample method 200 for implementing the system described herein. In some embodiments, themethod 200 is an implementation of the probability distributionforecast evaluation system 110. For example, themethod 200 may be implemented as a series of instructions stored on thememory 140 that cause theprocessor 130 to perform themethod 200 when executed. - Starting at
step 210, themethod 200 acquires a probability distribution forecast (“forecast”), a probability distribution realization corresponding to the probability distribution forecast (“realization”), and a prior knowledge of the probability distribution forecast (“prior knowledge”). In some embodiments, themethod 200 may not acquire all three of the forecast, realization, and prior knowledge. Instead, it may only acquire a subset, such as just the forecast or both of the forecast and the prior knowledge. The forecast, realization, and prior knowledge may be user entered, such as through theuser interface 112; may be stored, such as instorage device 150; or may be looked up, downloaded, or acquired over a network, such asnetwork 170. The system disclosed herein contemplates that each of the forecast, realization, and prior knowledge are acquired by different means. For example, the forecast may be user entered, the realization may be downloaded from the Internet, and the prior knowledge may be stored on a storage device. - The forecast may be a stored version of a previously-recorded forecast of a statistical probability distribution forecast. The forecast may predict metrics in one or more areas such as financial metrics, economic metrics, business performance metrics, weather metrics, or any other metric that may be useful to predict. The realization may be a stored version of the realized probability distribution of the metrics that the forecast predicted. The prior knowledge may be a stored version of prior knowledge about the probability distribution that the forecast predicted. For example, this may be a reference probability distribution for the forecast.
- Each of the forecast, realization, and prior knowledge may be stored as a probability density function, a probability mass function, or a collection of samples. Further, the forecast, realization, and prior knowledge may be partitioned into one or more bins. Each of the forecast, realization, and prior knowledge may have a different number of bins. Further, each of the forecast, realization, and prior knowledge may be stored in a different function. For example, the forecast and prior knowledge may be stored as a probability density function and the realization may be stored as a collection of samples.
- Once the forecast, realization, and prior knowledge are acquired, at
step 220 themethod 200 progresses to calculate an accuracy score for the forecast. The accuracy score shows how reliable a forecast is, measuring how reliably the forecast predicts the probability distribution. This measure is important to evaluating a probability distribution forecast because accuracy is an important component of a useful forecast. - The accuracy score may be based on the forecast and the realization. In some embodiments, the accuracy score is calculated by computing a dissimilarity score between the forecast and the realization. A dissimilarity score may be a measurement of how different two sets of data are. For example, the accuracy score may be calculated by calculating the Kullback-Leibler (KL) divergence between the forecast and the realization, as shown below:
-
- This implementation of the accuracy score will be lower when a forecast is more accurate.
- In other embodiments, the accuracy score may be calculated using a quadratic approximation of the KL divergence between the forecast and the realization.
- For example, if the accuracy score is calculated based on samples of one or more of the forecast and the realization, it may be useful to calculate the accuracy score using the quadratic approximation of the KL divergence between the samples of the forecast and the realization. Those skilled in the art will recognize that the KL divergence between the samples of the forecast and the realization can be approximated with chi-squared statistics using a quadratic approximation as shown below:
-
- In this instance, if only one of the forecast and the realization is stored as samples, the other may be converted to samples by partitioning it into bins. For example, if the realization distribution is stored as samples, the forecast may be partitioned into bins. In some embodiments, it may be preferable to convert the distribution that was not sampled into the same number of bins as the sampled distribution. In the preceding situation, for example, if the realization distribution is stored as a set of samples in 5 bins, the forecast may be partitioned into 5 bins. In other cases, if neither the forecast nor the realization is stored as samples, the system of the present disclosure may convert both distributions by partitioning them into bins. As described above, it may be beneficial to partition both distributions into the same number of bins.
- At
step 230, themethod 200 calculates an information score for the forecast. The information score measures the amount of information a forecast contains compared to the prior knowledge about the forecasted metrics. This prior knowledge may include reference probability distributions for the forecasted metrics. The information score is an important part of evaluating forecasts. Conventional techniques, such as the Brier score, treat more certainty (or a better accuracy score) as the better forecast. Such systems fail to properly account for the amount of information in a forecast. For example, even if a forecast is less certain than the prior knowledge, it is still useful to know that the future is more uncertain than the prior knowledge. As a further example, an area may get rain on 30% of the days over the course of the year. However, a forecast that predicted a 30% chance of rain for every day might be very accurate over the course of the year, but it does not contain any information and is therefore not very useful. On the other hand, a forecast that predicts a 90% chance of rain on days when it only rains 80% of the time might not be that accurate, but it does contain information about which days are more likely to receive rain and is therefore more useful. Accordingly, embodiments of the presently disclosed system incorporate an information score into its evaluation. - The information score may be based on the forecast and the prior knowledge. In some embodiments, the information score is calculated by computing a dissimilarity score between the forecast and the realization. For example, the information score may be calculated by calculating the KL divergence between the realization and the prior knowledge, as shown below:
-
- This implementation of the information score will be higher when a forecast provides more information.
- In some embodiments, the forecast may be a conditional distribution. In these cases, both the realization and the forecast may be conditional on the same condition. Similarly, the prior knowledge is generally either an unconditional distribution, or is conditional on a different condition from the forecast and the realization. In this instance, when the accuracy score is high, the forecast may be used as a proxy for the realization in the information score calculation. For example, the information score may be calculated as shown below:
-
- In still further embodiments, the information score may be calculated using a quadratic approximation of the KL divergence between the realization and the prior knowledge. For example, if the accuracy score is calculated based on samples of one or more of the realization and the prior knowledge, it may be useful to calculate the information score using the quadratic approximation of the KL divergence between the samples of the realization and the prior knowledge. Those skilled in the art will recognize that the KL divergence between the samples of the realization and the prior knowledge can be approximated with chi-squared statistics using a quadratic approximation as shown below:
-
- In similar embodiments to those discussed above, if the forecast is accurate (i.e., the accuracy score is small), the forecast may be used as a proxy for the realization in the information score calculation. In this instance, the information score can be calculated with the quadratic approximation of the KL divergence between the samples of the forecast and the prior knowledge as shown below:
-
- In this instance, if only one of the prior knowledge and the realization is stored as samples, the other may be converted to samples by partitioning it into bins. For example, if the realization distribution is stored as samples, the forecast may be partitioned into bins. In some embodiments, it may be preferable to convert the distribution that was not sampled into the same number of bins as the sampled distribution. In the preceding situation, for example, if the realization distribution is stored as a set of samples in 5 bins, he prior knowledge may be partitioned into 5 bins. In other cases, if neither the prior knowledge nor the realization is stored as samples, the disclosed system may convert both distributions by partitioning them into bins. As described above, it may be beneficial to partition both distributions into the same number of bins.
- After calculating the accuracy and information scores, at
step 240, themethod 200 calculates a performance score based on the accuracy score and the information score. In some embodiments, the performance score is calculated by subtracting the accuracy score from the information score, as shown below: -
PS=INFO−ACC - When defined like this, the larger the performance score is, the better the forecast is. One benefit of this implementation is that the accuracy score and information score can be calculated independently. This means that, the number of bins used in the accuracy score calculation can differ from the number of bins used in the information score calculation. This improves the simplicity and numerical stability of the calculation.
- Similarly, the above implementation of the performance score calculation does not depend on how each of the information score and the accuracy score are calculated. Accordingly, this implementation may be used even if the accuracy score and information score are calculated based on samples or if the information score is calculated by using the forecast as a proxy for the realization as discussed above. In fact, in some embodiments it may be preferable to use the forecast as a proxy for the realization in the information score calculation. Doing this may result in a simplified, and therefore faster, calculation. Calculating the performance score in this manner may also be more robust if one or more of the forecast, prior knowledge, and realization is stored as samples. An example implementation of this is shown below:
-
- In some embodiments, after calculating the performance score, the
method 200 will also calculate a relative performance score (not depicted inFIG. 2 ). The relative performance score may be calculated based on the entropy of the prior knowledge. In some embodiments, the entropy of the prior knowledge may be calculated as defined below: -
- where u represents a uniform distribution, i.e., uk=1/K.
- In still further embodiments, after calculating the performance score, the
method 200 will also calculate a confidence interval for the performance score. The confidence interval for the performance score may be based on the probability distribution of the performance score. For example, in some embodiments, those with skill in the art will note that one or more of the accuracy score and the information score may have a probability distribution if calculated based on samples. Accordingly, a probability distribution of the performance score may be calculated based on one or more of the probability distribution of the accuracy score and the probability distribution of the information score. The confidence interval for the performance score may then be calculated based on the probability distribution of the performance score. In practice, the confidence interval of the performance score may be valuable because it provides more information on the range of values that the performance score may take. This provides users with a greater understanding of the performance of the forecast. - The previously-discussed example embodiments of the
method 200 focus primarily on analyzing a single forecast at a time. However, the present disclosure also contemplates that themethod 200 may also analyze multiple forecasts, where each of the multiple forecasts may differ from one another. In addition, one or more of the prior knowledge and realization could be different for one or more of the multiple forecasts as well. In some embodiments, each of the forecasts may use different bin divisions and the prior knowledge and realization that correspond to each of the forecasts may also use different bin divisions. For example, if the multiple forecasts use different bin divisions, the prior knowledge and realization that correspond to each of the multiple forecasts may use the same division as their corresponding forecast. - In some embodiments, the
method 200 may select bin divisions for each distribution such that the probability bins of the multiple forecasts are the same. In such embodiments, the accuracy and information scores may be calculated as defined below: -
- In the above embodiment, the
method 200 may use the forecast as a proxy for the realization in the information score calculation as described above. For example, the information score may be calculated as shown below: -
- In still further embodiments, the
method 200 may select equal quantiles for both the prior knowledge and the realization. In this case, the performance score may be calculated as defined below: -
- In the above embodiment, the
method 200 may use the forecast as a proxy for the realization in the information score calculation as described above. In such an embodiment, the performance score may be calculated as shown below: -
- In some cases, it may impractical or impossible to divide the multiple forecasts and their corresponding realizations and prior knowledge into bins with a constant size. For example, it may be impossible to keep bins of a constant size if the outcome is binary. Thus, in some embodiments, the
method 200 may calculate the information and accuracy scores as shown below: -
- In the above embodiment, the
method 200 may use the forecast as a proxy for the realization in the information score calculation as described above. In such an embodiment, the information score may be calculated as shown below: -
- In similar embodiments, the
method 200 may substitute the denominator in the above calculations for 1/K, resulting in the simplified calculations shown below: -
- In the above embodiment, the
method 200 may use the forecast as a proxy for the realization in the information score calculation as described above. In such an embodiment, the information score may be calculated as shown below: -
- In some embodiments, after calculating the performance score for the multiple forecasts, the
method 200 may also calculate a relative performance score for the multiple forecasts. The relative performance score for the multiple forecasts may be calculated based on the entropy of the prior knowledge corresponding to the multiple forecasts.
Claims (20)
1. A probability distribution forecast evaluation system comprising:
at least one processor;
at least one memory device that stores a plurality of instructions which, when executed by the at least one processor, cause the at least one processor to operate with the at least one memory device to:
acquire a probability distribution forecast and a prior knowledge of the probability distribution forecast;
compute an information score based on the probability distribution forecast and the prior knowledge of the probability distribution forecast; and
compute a performance score based on the information score.
2. The probability distribution forecast evaluation system of claim 1 , further comprising instructions that, when executed by the at least one processor, cause the at least one processor to operate with the at least one memory device to:
acquire a probability distribution realization corresponding to the probability distribution forecast;
compute an accuracy score based on one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast; and
compute the performance score based on the accuracy score and the information score.
3. The probability distribution forecast evaluation system of claim 2 , wherein the accuracy score is computed based on the probability distribution forecast and the probability distribution realization.
4. The probability distribution forecast evaluation system of claim 3 , wherein the accuracy score is computed by calculating a dissimilarity score between the probability distribution forecast and the probability distribution realization.
5. The probability distribution forecast evaluation system of claim 4 , wherein the dissimilarity score is either (1) the Kullback-Leibler (KL) divergence between the probability distribution forecast and the probability distribution realization, or (2) a quadratic approximation of the KL divergence between the probability distribution forecast and the probability distribution realization.
6. The probability distribution forecast evaluation system of claim 2 , wherein the performance score is computed by subtracting the accuracy score from the information score.
7. The probability distribution forecast evaluation system of claim 2 , wherein a relative performance score is further computed based on the computed performance score and an entropy of the prior knowledge of the probability distribution forecast.
8. The probability distribution forecast evaluation system of claim 1 , wherein the information score is computed by calculating a dissimilarity score between the probability distribution realization and the prior knowledge of the probability distribution forecast.
9. The probability distribution forecast evaluation system of claim 8 , wherein the dissimilarity score is either (1) the Kullback-Leibler (KL) divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast, or (2) a quadratic approximation of the KL divergence between the probability distribution realization and the prior knowledge of the probability distribution forecast.
10. The probability distribution forecast evaluation system of claim 1 , wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast are computed based on samples.
11. The probability distribution forecast evaluation system of claim 10 , wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast are partitioned into discrete probability bins.
12. The probability distribution forecast evaluation system of claim 10 , wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast contain sample errors and the performance score is normalized to account for the sample errors.
13. A method comprising:
acquiring a probability distribution forecast and a prior knowledge of the probability distribution forecast;
computing an information score based on the probability distribution forecast and the prior knowledge of the probability distribution forecast; and
computing a performance score based on the information score.
14. The method of claim 13 , further comprising:
acquiring a probability distribution realization corresponding to the probability distribution forecast;
computing an accuracy score based on one or more of the probability distribution forecast, the probability distribution realization corresponding to the probability distribution forecast, and the prior knowledge of the probability distribution forecast; and
computing the performance score based on the accuracy score and the information score.
15. The method of claim 14 , wherein the accuracy score is computed based on the probability distribution forecast and the probability distribution realization.
16. The method of claim 14 , wherein the performance score is computed by subtracting the accuracy score from the information score.
17. The method of claim 14 , wherein a relative performance score is further computed based on the computed performance score and an entropy of the prior knowledge of the probability distribution forecast.
18. The method of claim 13 , wherein the information score is computed by calculating a dissimilarity score between the probability distribution realization and the prior knowledge of the probability distribution forecast.
19. The method of claim 13 , wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast are computed based on samples, and wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast are partitioned into discrete probability bins.
20. The method of claim 13 , wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast are computed based on samples, and wherein one or more of the probability distribution forecast and the prior knowledge of the probability distribution forecast contain sample errors and the performance score is normalized to account for the sample errors.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/066,695 US20210027183A1 (en) | 2016-04-04 | 2020-10-09 | System and method for performance evaluation of probability forecast |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201662317774P | 2016-04-04 | 2016-04-04 | |
US15/478,910 US10803393B2 (en) | 2016-04-04 | 2017-04-04 | System and method for performance evaluation of probability forecast |
US17/066,695 US20210027183A1 (en) | 2016-04-04 | 2020-10-09 | System and method for performance evaluation of probability forecast |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/478,910 Continuation US10803393B2 (en) | 2016-04-04 | 2017-04-04 | System and method for performance evaluation of probability forecast |
Publications (1)
Publication Number | Publication Date |
---|---|
US20210027183A1 true US20210027183A1 (en) | 2021-01-28 |
Family
ID=59959396
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/478,910 Active 2039-07-22 US10803393B2 (en) | 2016-04-04 | 2017-04-04 | System and method for performance evaluation of probability forecast |
US17/066,695 Abandoned US20210027183A1 (en) | 2016-04-04 | 2020-10-09 | System and method for performance evaluation of probability forecast |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/478,910 Active 2039-07-22 US10803393B2 (en) | 2016-04-04 | 2017-04-04 | System and method for performance evaluation of probability forecast |
Country Status (1)
Country | Link |
---|---|
US (2) | US10803393B2 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109840308B (en) * | 2017-11-29 | 2023-12-19 | 中国电力科学研究院有限公司 | Regional wind power probability forecasting method and system |
US11650551B2 (en) * | 2019-10-04 | 2023-05-16 | Mitsubishi Electric Research Laboratories, Inc. | System and method for policy optimization using quasi-Newton trust region method |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8386401B2 (en) * | 2008-09-10 | 2013-02-26 | Digital Infuzion, Inc. | Machine learning methods and systems for identifying patterns in data using a plurality of learning machines wherein the learning machine that optimizes a performance function is selected |
US9594983B2 (en) * | 2013-08-02 | 2017-03-14 | Digimarc Corporation | Learning systems and methods |
US10113910B2 (en) * | 2014-08-26 | 2018-10-30 | Digimarc Corporation | Sensor-synchronized spectrally-structured-light imaging |
-
2017
- 2017-04-04 US US15/478,910 patent/US10803393B2/en active Active
-
2020
- 2020-10-09 US US17/066,695 patent/US20210027183A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
US20170286840A1 (en) | 2017-10-05 |
US10803393B2 (en) | 2020-10-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11423263B2 (en) | Comparison method and comparison apparatus | |
CN110443657B (en) | Client flow data processing method and device, electronic equipment and readable medium | |
Meller et al. | Ensemble distribution models in conservation prioritization: from consensus predictions to consensus reserve networks | |
US10877012B2 (en) | Automatic siting for air quality monitoring stations | |
Papalexiou et al. | How extreme is extreme? An assessment of daily rainfall distribution tails | |
Sohnesen et al. | Is random forest a superior methodology for predicting poverty? An empirical assessment | |
US20210027183A1 (en) | System and method for performance evaluation of probability forecast | |
US11869021B2 (en) | Segment valuation in a digital medium environment | |
Dragomir et al. | Does the galaxy–halo connection vary with environment? | |
Luckstead et al. | The size distributions of all Indian cities | |
Monteiro et al. | Derivation of fragility functions for seismic assessment of RC bridge portfolios using different intensity measures | |
JP5833817B2 (en) | A method for approximating user task representation by document usage clustering | |
Herla et al. | Snow profile alignment and similarity assessment for aggregating, clustering, and evaluating of snowpack model output for avalanche forecasting | |
CN113159934A (en) | Method and system for predicting passenger flow of network, electronic equipment and storage medium | |
Gowtham Sethupathi et al. | Efficient rainfall prediction and analysis using machine learning techniques | |
US11886230B2 (en) | Method and system of automatically predicting anomalies in online forms | |
Kulinich et al. | A Markov chain method for weighting climate model ensembles | |
Wolf et al. | Robust detection of plant species distribution shifts under biased sampling regimes | |
CN110704614A (en) | Information processing method and device for predicting user group type in application | |
Pang et al. | The clustering of spatially associated species unravels patterns in tropical tree species distributions | |
CN111325350B (en) | Suspicious tissue discovery system and method | |
CN111988813B (en) | Method, device and computer equipment for determining weak coverage cell in mobile communication network | |
Ashton et al. | Calibrating gravitational-wave search algorithms with conformal prediction | |
Xie | Deep Learning Architectures for PM2. 5 and Visibility Predictions | |
Santillán et al. | Modelling uncertainty of flood quantile estimations at ungauged sites by Bayesian networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |