US20220318711A1 - Automated supply chain demand forecasting - Google Patents

Automated supply chain demand forecasting Download PDF

Info

Publication number
US20220318711A1
US20220318711A1 US17/708,985 US202217708985A US2022318711A1 US 20220318711 A1 US20220318711 A1 US 20220318711A1 US 202217708985 A US202217708985 A US 202217708985A US 2022318711 A1 US2022318711 A1 US 2022318711A1
Authority
US
United States
Prior art keywords
data
demand
products
product demand
product
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/708,985
Inventor
Javier Recasens
Prasanna Kumar Ragavan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Coupa Software Inc
Original Assignee
Coupa Software Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Coupa Software Inc filed Critical Coupa Software Inc
Priority to US17/708,985 priority Critical patent/US20220318711A1/en
Assigned to Coupa Software Incorporated reassignment Coupa Software Incorporated ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RAGAVAN, PRASANNA KUMAR, RECASENS, JAVIER
Publication of US20220318711A1 publication Critical patent/US20220318711A1/en
Assigned to SSLP LENDING, LLC reassignment SSLP LENDING, LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Coupa Software Incorporated, YAPTA, INC.
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0201Market modelling; Market analysis; Collecting market data
    • G06Q30/0202Market predictions or forecasting for commercial activities
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/063Operations research, analysis or management
    • G06Q10/0631Resource planning, allocation, distributing or scheduling for enterprises or organisations
    • G06Q10/06315Needs-based resource requirements planning or analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/06Resources, workflows, human or project management; Enterprise or organisation planning; Enterprise or organisation modelling
    • G06Q10/067Enterprise or organisation modelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/08Logistics, e.g. warehousing, loading or distribution; Inventory or stock management

Definitions

  • One technical field of the present disclosure is computer-assisted forecasting of demand of goods or materials in complex supply chains. Another technical field is predictive modeling, including time series analysis. Another technical field is supply chain management. Another technical field is logistics as applied to disruptive conditions.
  • Demand forecasting is a field of predictive data analytics directed to optimizing supply chain decisions by predicting customer demand using computer-implemented algorithms.
  • Demand forecasting may be a part of production planning, inventory management, marketing strategy development, and various other aspects of corporate decision-making.
  • Demand forecasting methods may involve qualitative or quantitative assessment of pertinent data, including historical sales data.
  • Demand forecasting may be accomplished by building a model and testing that model. A variety of techniques may be used to validate a model through testing.
  • Time series analysis involves forecasting future behavior by analyzing past behavior. Time series analysis may be used to predict future demand for certain products based, at least in part, on past sales of those products. Time series analysis may identify data features such as trends, seasonality, cyclicity, and irregularity.
  • Supply chain networks may involve distribution centers, suppliers, vendors, and manufacturers, to utilize to meet consumer demand for a particular finished good.
  • Supply chain network techniques often employ multiple levels of interdependence and connectivity between sites within the supply chain network. Multiple models or techniques may be utilized to predict the behavior and interactions between these sites to optimally deliver goods and services to various points or locations along the supply chain network.
  • An effective solution requires computer implementation to manage issues of scale and real-time response timing that accounts for factors both internal and external to particular sites in a supply chain network.
  • buyer computers or buyer accounts may interoperate with dozens to hundreds of different supply chains, each with dozens to hundreds of nodes, in association with thousands to tens of thousands of products or components.
  • Each node in all these complex supply chains may be associated with different production requirements that impact demand factors.
  • vast data about external events can rapidly impact the accuracy of demand forecasting. Disruptive events such as obstruction of critical canals (Suez, Panama), pandemic or epidemic, or natural disasters can occur with little warning and rapidly render forecasts based on past data invalid.
  • buyer computers require real-time responses to queries about demand.
  • Buyer accounts need the ability to add, delete, or rearrange data describing internal or external demand factors while receiving a real-time response to an updated query for a demand forecast.
  • FIG. 1 illustrates a distributed computer system showing the context of use and principal functional elements with which one embodiment could be implemented.
  • FIG. 2 illustrates an example process for preparing a training data set.
  • FIG. 3 illustrates an example process for disruption-resistant demand forecasting.
  • FIG. 4 illustrates an example of time series cluster visualization.
  • FIG. 5A , FIG. 5B , and FIG. 5C illustrate examples of generalized processes for forming forecasting models and using the forecasting models in a production computing environment that may be used to implement certain embodiments.
  • FIG. 6 illustrates a computer system with which one embodiment could be implemented.
  • a novel computer-implemented method is presented for implementing technical machine learning solutions to the technical problem of machine learning model development, validation, and deployment in the domain of predictive modeling.
  • the disclosure presents solutions implemented via client-server Software as a Service (SaaS) techniques or distributed computer systems generally.
  • SaaS Software as a Service
  • a variety of novel systems are presented for predicting consumer demand.
  • a diverse array of systems may be used to implement the novel methods presented in this disclosure.
  • a distributed computer system is programmed to receive training data representing historic consumer demand for products, to detect changepoints in that data that may be associated with disruptive events, to identify relevant data for modeling, to perform clustering, to process configuration information, to train one or more machine learning models that are capable of evaluating other received data more accurately, and to output results to a user display device.
  • an online distributed computer system or platform for predictive modeling provides a system for generating product demand models based on historical product demand data and using those models to predict future customer demand for products.
  • the platform comprises functionality for importing data, preparing training data sets, training, validating, and executing models, and visualizing results.
  • FIG. 1 illustrates a distributed computer system showing the context of use and principal functional elements with which one embodiment could be implemented.
  • FIG. 1 and the other drawing figures and all of the description and claims in this disclosure, are intended to present, disclose and claim, among other things, a technical system and technical methods in which specially programmed computers, using a special-purpose distributed computer system design, execute functions that have not been available before to provide a practical application of computing technology to the problem of machine learning model development, validation, and deployment.
  • the disclosure presents a technical solution to a technical problem, and any interpretation of the disclosure or claims to cover any judicial exception to patent eligibility, such as an abstract idea, mental process, method of organizing human activity or mathematical algorithm, has no support in this disclosure and is erroneous.
  • a distributed computer system 100 comprises components that are implemented at least partially by hardware at one or more computing devices, such as one or more hardware processors executing stored program instructions stored in one or more memories for performing the functions that are described herein.
  • computing devices such as one or more hardware processors executing stored program instructions stored in one or more memories for performing the functions that are described herein.
  • all functions described herein are intended to indicate operations that are performed using programming in a special-purpose computer or general-purpose computer, in various embodiments.
  • FIG. 1 illustrates only one of many possible arrangements of components configured to execute the programming described herein. Other arrangements may include fewer or different components, and the division of work between the components may vary depending on the arrangement.
  • distributed computer system 100 comprises data acquisition logic 102 coupled to project management logic 104 that is programed to store project data in persistent storage, for example in organizational units such as projects 130 .
  • the project management logic 104 may also be coupled to data storage logic 108 which is programmed to manage the storage of data across the distributed computer system, for example in database 110 .
  • data acquisition logic 102 is programmed to receive input signals specifying network locations of data sources or data files, to access and read the specified data sources or data files, and to import records from the specified data sources or data files into memory comprising database 110 or other memory, including, but not limited to, networked memory or cloud storage memory.
  • projects 130 may be stored in database 110 of distributed computer system 100 , which may be a relational database. In other embodiments, projects 130 may be stored in other memory accessible by distributed computer system 100 .
  • database 110 stores a variety of data comprising product demand data 120 , third-party data 122 , training data 124 , testing data 126 , and output data 128 .
  • distributed computer system 100 further comprises data processing logic which is coupled to the project management logic 104 and the data storage logic 108 , and which is programmed to process data accessible by distributed computer system 100 .
  • the data processing logic 106 may be programmed to initialize, train, execute, and validate machine learning models or statistical models. In an embodiment, machine learning models take as input training data 124 and third-party data 122 .
  • the data processing logic may first transform the product demand data 120 into one or more training data 124 sets. In an embodiment, a subset of the product demand data 120 may be used as training data 124 while a disjoint subset of the product demand data 120 is used as testing data 126 to test, validate, or otherwise assess the efficacy of a machine learning model or statistical model.
  • the data processing logic 106 is configured to visualize and display the output data 128 to a user display device 140 .
  • FIG. 2 illustrates a flow diagram of a computer-implemented system preparing a training data set for product demand forecasting.
  • FIG. 2 and each other flow diagram herein is intended as an illustration at the functional level at which skilled persons, in the art to which this disclosure pertains, communicate with one another to describe and implement algorithms using programming.
  • the flow diagrams are not intended to illustrate every instruction, method object or sub-step that would be needed to program every aspect of a working program, but are provided at the same functional level of illustration that is normally used at the high level of skill in this art to communicate the basis of developing working programs.
  • the present disclosure provides techniques to improve short-term demand forecasts for tactical response through novel time series analysis methodologies and with the help of external data.
  • the disclosed methods and systems may allow a user to identify demand patterns unfolding across an entire product portfolio.
  • product portfolios or subsets thereof selected for analysis may correspond to projects 130 as illustrated in FIG. 1 .
  • Management of projects 130 and the coordination of processes required for training data set preparation may be controlled by project management logic 104 as further illustrated in FIG. 1 .
  • product demand data may comprise data representing historical customer demand for a single product, a product line of a company, a portfolio of related products, all products sold or offered for sale by a company or one of its subdivisions or subsidiaries, or products sold or offered for sale by any number of related or unrelated companies or individuals.
  • the product demand data may comprise any other data representing historical demand for products, or even services.
  • product demand data corresponds to product demand data 120 ( FIG. 1 ) acquired by data acquisition logic 102 and stored in database 110 of distributed computer system 100 according to data storage logic 108 , as illustrated in FIG. 1 .
  • database 220 FIG. 2
  • database 110 FIG. 1 ).
  • product demand data may be associated with a geographical location.
  • product demand data may be global data, comprising historical demand data for one or more products in a plurality of countries around the world.
  • product demand data may be limited to a particular geographic region, country, state, province, city, or any other geographic or political subdivision.
  • the product demand data comprises a dataset representing downstream consumption data rather than data from purchase orders or shipments.
  • downstream consumption data e.g., Point of Sale (POS) data
  • POS Point of Sale
  • downstream consumption data may be a more accurate representation of actual customer demand than other demand data, because many demand transactions (e.g., consumer store visits) with zero sales of a given item may reflect a lack of sufficient stock as opposed to a lack of demand.
  • a variety of other data sets such as data concerning purchase orders or shipment data, may be used for a variety of reasons, including an inability to obtain downstream consumption data.
  • a true demand may be estimated by imputing sales with a variety of imputation algorithms or substitution methods.
  • imputation and substitution methods include, but are not limited to: (1) methods that account for substitution effects to estimate lost sales due to stock outs, (2) methods that use a demand rate for imputation, (2) if visits per store data is unavailable, estimating the basket of a visitor and the approximating the potential demand, or (4) if an item is primarily sold online and the item is out of stock, approximating lost sales by utilizing historical site visits of a specific product to a finalized sales ratio in conjunction with current site visit data.
  • other methods of estimating true demand using imputation or substitution may be used. The following is a non-limiting example of a method that accounts for substitution effects to estimate lost sales due to sock outs.
  • toilet paper of brand “A” is out of stock and toilet paper of brand “B′”s demand is higher by 40%, this percentage can be split by the historical sales of brand “A” and brand “B” to estimate brand “A” and “B′”s true demand.
  • POS data is not available.
  • product demand data may be data related to sales or demand from a previous year or series of years.
  • product demand data may be data related to historical sales or demand tabulated with respect to any other unit of time, or even contain estimates, interpolations, or projections from any period of time.
  • product demand data may be tagged, labeled, or otherwise associated with not only a level of demand or sales and time data, but also one or more of a geographic location or data concerning a source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, stock keeping unit (SKU), or any data or metadata commonly used in the art.
  • SKU stock keeping unit
  • the product demand data set may be transformed using one of the described imputation or substitution methods before or after being loaded into memory as seen in block 204 .
  • the product demand data may be cleaned or deduplicated before or after being loaded into memory as seen in block 204 .
  • the product demand data may be processed in other ways before or after being loaded into memory as seen in block 204 such as by removing outliers from the product demand data, cleaning the product demand data, removing unreadable data, or correcting errors in the data.
  • the data processing at block 204 or other data processing illustrated in FIG. 2 is actuated by data processing logic 106 as illustrated in FIG. 1 .
  • the product demand data is formatted for processing at block 208 .
  • Formatting product demand data for processing may comprise transforming the data into a convenient format to be used as input in a machine learning model or statistical model.
  • Data transformations effectuated at block 208 may comprise (1) resizing inputs to a particular fixed size (2) converting non-numeric data features into numeric features ones, (3) normalizing numeric data features, (4) lower-casing or tokenizing metadata text features, or (5) other data transformations used to process data prior to machine learning or statistical analysis.
  • distributed computer system 100 is programmed to receive user input to select a data source for the product demand data.
  • the user input may be visual or graphical, in the form of selection of graphical user interface widgets.
  • block 202 illustrates examples of data sources that may be selected.
  • a data set from the specified source 202 is written into memory, as seen in block 204 .
  • the memory referenced in block 204 is main memory of a virtual machine instance of a cloud computing center that implements elements of a computer system, for networked access using client computers or workstations.
  • the computer system may be implemented using a dynamic plurality of virtual machine instances that client computers of end users access using Software as a Service (SaaS) techniques and correspond to distributed computer system 100 ( FIG. 1 ).
  • the memory block referenced in block 204 is a block of memory in the memory 606 of the physical computing device illustrated in FIG. 6 .
  • a disruption-resistant demand prediction workflow may begin at step 310 ( FIG. 3 ) where a training data set is received for analysis.
  • the training data set comprises the entire data set loaded into memory at block 204 .
  • the training data set comprises a subset of the data set loaded into memory at block 204 .
  • the subset of the data set is selected by user input.
  • the subset of the data set is selected so that the remaining data in the data set may be used for testing, evaluation, or validation of modeling results.
  • the subset of the data set may be selected at least partially based on any of the various data or metadata associated with the product demand data, including but not limited to information about geographic location, source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, or SKU.
  • the data is not partitioned at block 204 , but is instead partitioned at block 210 .
  • product demand data comprising a training data set is stored in database 220 after it is cleaned, transformed, processed, or formatted.
  • FIG. 3 illustrates a flow diagram overview of a computer-implemented method for processing a training data set for product demand forecasting.
  • a computer system receives a training data set.
  • the training data set may have been prepared and loaded into memory as explained in Section 2.2 of this disclosure.
  • the training data set may comprise product demand data as explained in Section 2.2 of this disclosure.
  • the computer system receives product demand data according to other specifications.
  • the training data set may comprise any kind of time series data set.
  • the method illustrated in FIG. 3 . may be implemented on the distributed computer system illustrated in FIG. 1 and the coordination of processes required for disruption-resistant demand prediction may be controlled by project management logic 104 .
  • the method illustrated in FIG. 3 . may be implemented on computer system 600 illustrated in FIG. 6 .
  • the method illustrated in FIG. 3 may be implemented by another type of system.
  • a distributed computer system 100 calculates any break points in the training data set.
  • Break points may correspond to the occurrence of real-life disruptive events that cause sudden level or trend changes in historical product demand data. These disruptive events may be epidemics like COVID-19, other disease outbreaks, other events that cause increased portions of a population to stay at home or restrict movement, or a variety of other disruptive events. Break points may correspond to a shift from one type of behavior in a consumer populace, such as panic, buying, stabilization, or normalcy to another such phase of behavior in a consumer populace. Break points may correspond directly with disruptive events or they may be time-lagged. Break points may occur as a result of responses to a disruptive event, including governmental responses. Governmental responses may comprise financial responses such as changes to fiscal or monetary responses or legal responses such as restrictions, regulation, or curfews. Break points may also be known in the art of predictive modeling or other arts as changepoints.
  • a distributed computer system 100 calculates zero or more changepoints in the training data set by executing a changepoint detection algorithm.
  • changepoints may be detected within the entire training data.
  • the training data set may be federated, or hierarchical and changepoints may be detected within subsets of the training data set.
  • Changepoints may be detected in subsets corresponding to any of the various data or metadata associated with product demand data, including but not limited to information about geographic location, source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, or SKU.
  • the training data set may comprise time series data that can be clustered to identify and group time series (e.g., products and locations) that are experiencing similar shifts in demand patterns.
  • distributed computer system 100 may be programmed to cluster the training data set, and then execute a changepoint detection algorithm to detect changepoints in a plurality of representative time series corresponding to a plurality of clusters.
  • a distributed computer system 100 receives input from a user which specifies whether to cluster the training data set and calculate changepoints in representative time series corresponding to detected clusters using a changepoint detection algorithm. Such clustering may save time and computing resources when the training data set contains time series data for a large number of products, in some instances for thousands or millions of products.
  • a computer system calculates changepoints using machine learning methods.
  • Machine learning methods used for changepoint detection may be supervised or unsupervised.
  • Supervised methods comprise methods that may use (1) multi class classifiers, such as decision tree, nearest neighbor, support vector machine (SVM), na ⁇ ve Bayes, Bayesian net, hidden Markov model (HMM), conditional random field (CRF), Gaussian mixture model (GMM) methods, (2) binary class classifiers, such as support vector machine (SVM), na ⁇ ve Bayes, or logistic regression methods, or (3) virtual classifiers.
  • Unsupervised methods may use a likelihood ratio, a subspace model, probabilistic methods, kernel-based methods, graph-based methods, or clustering. Changepoint detection may be aided by an indicator akin to a moving average convergence/divergence (MACD) indicator.
  • MCD moving average convergence/divergence
  • a distributed computer system 100 identifies relevant data for modeling.
  • the entire training data set may be modeled, or a subset of the training data set may be modeled.
  • step 330 comprises distributed computer system 100 calculating or retrieving a baseline forecast for one or more products to determine if any of those products are currently being, or will be in the future, impacted by a disruptive event.
  • a baseline forecast is retrieved which was calculated at a time prior to a time value detected to be a changepoint in one or more products at step 320 .
  • demand statistics comprising mean demand level, median demand level, standard deviation of demand level, or other major demand statistics are calculated for selected periods before and after a detected changepoint in time series data for one or more products or clusters.
  • distributed computer system 100 may be programmed to flag that time series as significantly impacted by a disruptive event.
  • a large deviation, for example, of demand statistics or a forecast error prior to and after the disruptive event may be quantified as a deviation exceeding 1.5*Inter Quartile Range (IQR), above and below the 75th and 25th percentiles of the historical data.
  • IQR Inter Quartile Range
  • distributed computer system 100 may determine which representative time series of one or more clusters should be flagged as significantly impacted by the disruptive event using a discussed method.
  • a distributed computer system 100 processes configuration information which partially determines one or more of (1) which data to model, (2) which of one or more types of models to use, (3) which third-party data, if any, to feed into the one or models, (4) how those one or more models are to be initialized, configured, or run, and (5) other configuration information.
  • Configuration information may comprise a cross-learning preference, a preference for accuracy or control, input narrowing specifications comprising a selection of a specific product, region, or cluster to model, a dampening factor, a future inflection point prediction, a third-party data selection, or a short-term-forecasting preference.
  • one or more pieces of configuration information is received as input from a user.
  • all configuration information is hard-coded into memory accessible by distributed computer system 100 .
  • a distributed computer system 100 executes instructions to create one or more models to predict future product demand data based on the training data set.
  • one or more models are run based on the configuration information processed at step 340 .
  • third party data may be used to improve the predictive power of one or more models. For example, when dealing with a disruptive event such as a disease outbreak, like the COVID-19 epidemic, external metrics around restrictions on societal activities or compliance may provide insight on future product demand. Examples of such third-party data include, but are not limited to, mobility data from technology companies that provide maps, GPS tracking, or navigational directions as a service. Such data may capture the impact of local restrictions and how much compliance is observed in various localities, for example, by tracking a percent change in visits to places like grocery stores and parks within a geographic area. Location-based mobility metrics for one or more calendar days may be compared to a baseline value for that day in the immediately preceding calendar year.
  • third-party projections of effects derived from a disruptive event may be correlated with future product demand.
  • Other indicators impacting demand may include a social distance index, case counts, school closures data, unemployment claims data, consumer sentiment, hospital utilization data, or other indicators.
  • one or more of these third-party projections or indicators provide additional data sources with which one or more models may be trained.
  • third party data is selected for processing in response to user input.
  • third party data is selected for processing based on the configuration instructions processed at step 340 .
  • step 350 in an embodiment, data engineering may be used to process related regressors such as external metrics, location-based mobility data, third party projections, or other indicators impacting demand before they are used to train models. Data engineering may be usefully employed because many related regressors may not be forward-looking, meaning they are not available at prediction time and cannot be used directly to train models.
  • step 350 may comprise generating lags (lead indicators) and window statistics with a starting point limited by the forecast horizon. Next, correlation analysis may be performed on lagged regressors to determine optimal lags.
  • step 350 may comprise using forecasted regressors, which may have high accuracy validation metrics, at prediction time.
  • one or more machine learning models are trained and executed after third-party data has been engineered for processing.
  • Training may comprise programmatically supplying a training dataset to a machine learning model that is executing in the computer system and programmatically activating a training function of the model with a reference to or identification of the training dataset to be used.
  • third party data may be cleaned, deduplicated, or otherwise modified before it used for training models.
  • third-party data is used without being engineered or modified.
  • one or more models are trained and executed without the use of third-party data.
  • Models may be trained or executed at step 350 , including, but not limited to, Classical Models (e.g., Statistical Models or State Space Models) or Machine Learning & AI based Models (e.g., Machine Learning Models or Neural Networks).
  • Statistical Models comprise autoregressive (AR) models, moving average (MA) models, autoregressive integrated moving average (ARIMA) models, vector autoregression (VAR) models, hierarchical models, or others.
  • State Space Models comprise exponential smoothing (ETS) models, Hidden Markov models, Bayesian structural time-series models, or others.
  • Machine Learning models comprise support vector machines (SVMs), tree-based models, kNN's, or others.
  • Neural Networks comprise temporal convolutional neural (TCN) networks, multi-layer perceptron networks (FF neural networks), recurrent neural networks (e.g., long short-term memory (LSTM) or gated recurrent units (GRUs)), convolutional neural networks (CNN), or others.
  • TCN temporal convolutional neural
  • FF neural networks multi-layer perceptron networks
  • recurrent neural networks e.g., long short-term memory (LSTM) or gated recurrent units (GRUs)
  • CNN convolutional neural networks
  • one or more models executed at step 350 comprise lift-adjusted seasonal na ⁇ ve models, quantile regression models, stockout prediction models.
  • a distributed computer system 100 only executes new models for those time series flagged as significantly impacted by the disruptive event at step 330 .
  • changepoint detection may be implemented for similar (e.g., clustered) products. Calculated lag may represent time required to change phase, enabling a distributed computer system 100 to compute the expected point in time where changes should be expected.
  • the aforementioned process may be understood as using predicted changepoint detection and assigning binary regressors in the forecasting period.
  • One or more models executed at step 350 may be modified with this technique to potentially improve forecasting accuracy.
  • HTS forecasting may be implemented using a bottom-up approach, a top-down approach, a middle-out approach, or an optimal reconciliation approach. HTS forecasting may be most useful when product demand data is highly federated and the type of HTS approach may be determined by the level of hierarchy in the product demand data. Such approaches may increase coherency in model output.
  • a distributed computer system 100 executes instructions to create one or more models to predict future product demand data based on the training data set.
  • one or more models are run based on the configuration information processed at step 340 .
  • a distributed computer system 100 executes “back testing,” or time series validation, to assess the forecasting accuracy of one or more models.
  • Back testing may be preferable to simpler cross-validation methodologies sometimes used in machine learning model validation. For example, simple random splitting of samples into train and test sets might introduce data leakage in the time series context where there is a temporal dependency between observations.
  • validating one or more models comprises executing “walk-forward cross-validation.” In other embodiments, other validation techniques may be used which may incorporate folds sizes and gaps between the sets.
  • Back testing may be used for a Hyperparameter Tuning Set, an Ensemble Set, an Estimator Selection Set, or an Unbiased Test Set, independently and sequentially.
  • a distributed computer system 100 outputs the results of the models run at step 350 , including to, for example, user display device 140 ( FIG. 1 ).
  • modeling results are output in a visualization pane of a graphical user interface (GUI).
  • GUI graphical user interface
  • Modeling results may be output in the form of graphs, tables, charts, or other forms.
  • Modeling results may be written to a file, such as a CSV, Excel, or txt file, or to another type of file. Modeling results may be stored in a database or transmitted as output to be input for another system.
  • FIG. 5A , FIG. 5B , FIG. 5C illustrate examples of generalized processes for forming forecasting models and using the forecasting models in a production computing environment that may be used to implement certain embodiments.
  • Model Building begins with a data analysis and data engineering step in which historical datasets are assessed for the presence of features useful in forecasting, missing fields are identified for interpolation or enrichment from other data sources, and normalization is conducted to adjust the format of otherwise incompatible data fields.
  • a Model Specification and Evaluation step follows in which a forecasting model is selected or specified in mathematical terms, implemented in program code, and executed to evaluate or process the historical data to product output in the form of forecasting results. The model is deemed adequate if the output forecast is found usable. If not, the model may be re-specified by adjusting mathematical terms or changing input data features, and re-evaluated.
  • the Operationalization stage performs forecasting in a production environment.
  • the model that was validated in the prior stage is instantiated in one or more executable instances, for example, using computing instances of a cloud computing center. These instances are executed to generate forecasts that support the decision-making process, normally using newly created data having new timestamps and possibly new features.
  • the output is evaluated to check stability and continued validity. Over time, model drift or degradation is expected to occur, which requires retraining of the selected models, so control may transfer back to the model specification and evaluation stage. If the model is in a stable state, then a forecast update may be output based on evaluation of the new timestamps and features and used for decision-making.
  • FIG. 5B illustrates an embodiment of the Model Building stage of FIG. 5A in greater detail.
  • the Model Building stage begins with a Business Understanding step comprising identifying one or more relevant metrics for forecasting, determining how to measure the metrics, and determining how to measure successful forecasting via one or more success metrics.
  • the Data Preparation Phase comprises a data ingestion step that involves identifying one or more different data sources, locating datasets in one or more data analytics environments, and reading one or more files containing the datasets. Relational database tables, spreadsheets, flat files, or other sources may be read.
  • the data ingestion step is followed by a data exploration & understanding step that involves exploring the characteristics and meaning of a dataset and determining if the data quality is adequate to answer one or more questions about the data. These steps may be looped through or iteratively repeated until data quality is sufficient.
  • a final step in the Data Preparation Phase comprises a data pre-processing & feature engineering step that involves cleaning missing values and outliers from the data and creating data features from the raw data to facilitate model training.
  • the Model Building stage then continues, in an embodiment, with a Data Modeling Phase.
  • the Data Modeling Phase comprises a model building & selection step that involves using algorithms to scan historical data and extract patterns, building one or more models to predict future values, and selecting one or more best models based on project criteria. These steps may be looped through or iteratively repeated to build new models as more historical data is ingested or project criteria changes.
  • the model building and selection step is followed by a model deployment step that involves deploying one or more models and data pipelines to a production environment for application consumption.
  • forecasting solution acceptance comprises confirming that the one or more pipelines and the one or more models satisfy success criteria and validating that a final solution adds value to a business and its processes.
  • FIG. 5C illustrates an embodiment of the Operationalization stage of FIG. 5A in greater detail.
  • data collection continues, and raw data is fed into a cloud-based system for generating forecasts in a production environment.
  • the Operationalization stage continues with supplemental data exploration and aggregation, wherein the aggregated data may be explored to assess the characteristics of the aggregated data set and reevaluate data quality.
  • the data exploration and aggregation step is followed by a feature engineering step in the Operationalization stage. Feature engineering may include cleaning missing values and outliers from the data and creating data features from the data to facilitate model training.
  • the Operationalization stage of FIG. 5A involves feeding processed data which is the resulting output of feature engineering into a Re-training Web Service or a Scoring Web Service.
  • a Re-training Web Service retrains one or more machine learning or statistical models previously described and may comprise a web service instance of a cloud computing system denoted “Re-training Web Service.”
  • a Scoring Web Service executes forecast models or processes previously described and may comprise a web service instance of a cloud computing system denoted “Scoring Web Service.”
  • the output is reevaluated to check stability and continued validity. If the model is in a stable state, then a forecast update may be output based on evaluation of new timestamps and features and used for decision-making.
  • the Operationalization stage of FIG. 5A involves transmitting forecasted results from the cloud-based system to a Consumption Client, which may be any client configured to receive, or consume, the output of one or more models executed in the production environment.
  • the Consumption Client may be a server accessible via an API call, a relational database or other file system, or any application layer configured to receive forecasted results as input.
  • a computer system identifies relevant data for modeling.
  • relevant data may be identified for modeling is: (1) by calculating a baseline forecast, (2) by calculating relevant demand statistics or forecast errors for a period of time before or after a changepoint, and (3) by identifying data corresponding to segments of time series after changepoints with high forecast error as data which is particularly relevant for modeling.
  • Another helpful technique which may be used to identify relevant data for modeling is clustering.
  • clustering may provide insights into product velocity and demand as a disruptive event takes firm hold in a region.
  • Clustering may be used to identify and group time series (e.g. items and locations) that are experiencing similar shifts in demand patterns. Since impact from a disruptive event may change rapidly over time, it may be advantageous to condense the amount of information that needs to be processed, while providing detailed information on the types of patterns that are unfolding across an entire portfolio of products.
  • groups of products in the product portfolio exhibiting signs of normal behavior may be identified.
  • groups of products deviating away from normalcy may be identified.
  • data for one or more product clusters identified as deviating from normalcy may be then be used for modeling at step 350 . The described process may allow for scaling to millions of products quickly and at lower computational cost.
  • a K-Means clustering approach may be used to scale the clustering process to millions of time series.
  • the K-Means clustering approach may comprise using a demand pattern observed over a window of time since the onset of the disruption as provided features. These features (demand over a window) may be normalized and smoothed to ensure some randomness from day to-day fluctuations does not have a large impact on the clustering process.
  • the normalization may be interpreted as the number of folds in the average historical demand at a certain period (wherein values >1 indicate increased demand, and values ⁇ 1 indicate less demand).
  • FIG. 4 shows example output from a clustering, where, for each cluster, the features (normalized time series over a recent horizon) are plotted against the time periods.
  • Clusters 3 and 4 may be deemed clusters where the demand has been showing an increasing trend without any indication of return to normalcy.
  • Cluster 3 may be distinguished from Cluster 4 by a level of increase in demand.
  • Demand in Cluster 3 appears to grow unabated, experiencing up to five times the normal volume of demand.
  • Cluster 3 may represent, for example products such as toilet paper, hand sanitizers, and other disinfectants during early stages of the COVID-19 epidemic.
  • demand in Cluster 4 appears to rise to two times historical values.
  • FIG. 4 shows Clusters 1, 2, and 6 as groups products that have experienced a growth in demand followed by signs of returning to normalcy or settling at a new normal. However, what distinguishes each one of these clusters is the degree of the growth, rate of the growth, and the demand level the clusters appear to stabilize at. Clusters 1 and 6 appear to have experienced a gradual increase in demand, while demand in Cluster 1 appeared to stop growing, and began returning to normalcy. Demand for products in Cluster 6 appeared to continue to grow before settling at a new normal.
  • Cluster 6 appears to group together products that experienced a negative effect on demand due to COVID-19, wherein demand stayed flat for some time, and then, as the severity of the disease progressed, evidenced a steep decline in demand before settling at a new normal (e.g., products such as camping equipment).
  • a computer system processes configuration information which partially determines one or more of (1) which data to model, (2) which of one or more types of models to use, (3) which third-party data, if any, to feed into the one or models, (4) how those one or more models are to be initialized, configured, or run, and (5) other configuration information.
  • a computer system processes configuration information comprising data which indicates a preference for accuracy or for control.
  • configuration information comprising data which indicates a preference for accuracy or for control.
  • the data indicates a preference for accuracy
  • one or more of an LSTM or another Neural Network type model, or a Bayesian structural time series model is subsequently executed at step 350 .
  • the data indicates a preference for control
  • one or both of an ETS or lift-adjusted seasonal na ⁇ ve model is subsequently run at step 350 .
  • configuration information processed at step 340 may also indicate that a high-demand scenario is present.
  • one or both of quantile regression and a stockout prediction model are executed at step 350 when configuration information processed at step 340 indicates a high-demand scenario.
  • Quantile regression may be used to forecast ranges of high demand estimates to help meet acceptable service levels. Quantile regression forecasting involves modeling an entire demand distribution; and, it may be particularly effective when costs associated with over-forecasting and under-forecasting are unequal.
  • Stockout prediction models may help determine the probability of a stockout over a given timeframe.
  • This model may effectively identify items with a high likelihood of running out of stock and can be implemented as a classification algorithm with the target variable as either a stock-out indicator ( 1 or 0 ) or an inventory level range (e.g., low/medium/high). Rate-of-consumption metrics, inventory levels, and demand forecasts may serve as inputs to a stockout prediction model.
  • a quantile regression or stockout prediction model may be executed at step 350 even when a high-demand scenario is not indicated or present.
  • configuration information processed at step 340 may comprise data indicating a cross-learning preference.
  • Cross-learning a characteristic of global methods, may be used by a variety of models to learn patterns between similar time series, particularly when a large number of time series are being dealt with.
  • cross-learning is implemented in one or more models executed at step 350 when configuration information processed at step 340 indicates a cross-learning preference.
  • cross-learning may be implemented in one or more models executed at step 350 even when configuration information does not indicate a cross-learning preference.
  • configuration information processed at step 340 may comprise data indicating a short-term-forecasting preference.
  • more recent product demand data is weighted more heavily in one or more models executed at step 350 when configuration information processed at step 340 indicates a short-term-forecasting preference.
  • data weighting may be implemented in one or more models executed at step 350 even when configuration information does not indicate a short-term-forecasting preference.
  • configuration information processed at step 340 may comprise data indicating a third-party data selection.
  • third party data is incorporated in training one or more models executed at step 350 when configuration information processed at step 340 indicates a third-party data selection.
  • configuration information processed at step 340 may comprise data indicating a dampening factor. Notwithstanding seasonality and noise, most SKU's may be associated with a generally increasing trend which may be evident in a previously calculated or received baseline model. A dampening factor reduces the changes presented by a new model in comparison to a baseline model calculated before a disruptive event. A dampening factor may be thought of as a “Social Restrictions Lever” with regards to the COVID-19 epidemic. A higher dampening factor may model a prediction of fewer social restrictions and a tendency for consumer demand to emulate a baseline forecast more closely than newly predicted. In an embodiment, a dampening factor is incorporated in executing one or more models at step 350 when configuration information processed at step 340 indicates a dampening factor.
  • configuration information processed at step 340 may comprise data indicating a future inflection point prediction.
  • Data about a disruptive event (such as third-party projections concerning COVID-19 death rates) may indicate a future changepoint is likely.
  • a future changepoint may be implemented in executing one or more models at step 350 via a regressor or structural break.
  • the future change point indicates a likelihood of, for example, an inflection point, a change in level, or a change in trend for product demand.
  • a future inflection point prediction may comprise data indicating that models executed at 350 should be adjusted to reflect a “Return to Normalcy” by a certain “Recovery Date” and that demand predictions should revert to a new baseline prediction after that date.
  • the “Recovery Date”, may be, for example, when a death rate from an epidemic (e.g., COVID-19) is expected to fall below 6 deaths per day or 10% of a peak death rate for one day in a location, or another date when normalcy is predicted to return.
  • a future inflection point prediction is incorporated in executing one or more models at step 350 when configuration information processed at step 340 indicates a future inflection point prediction.
  • the techniques described herein are implemented by at least one computing device.
  • the techniques may be implemented in whole or in part using a combination of at least one server computer and/or other computing devices that are coupled using a network, such as a packet data network.
  • the computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as at least one application-specific integrated circuit (ASIC) or field programmable gate array (FPGA) that is persistently programmed to perform the techniques, or may include at least one general purpose hardware processor programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination.
  • ASIC application-specific integrated circuit
  • FPGA field programmable gate array
  • Such computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the described techniques.
  • the computing devices may be server computers, workstations, personal computers, portable computer systems, handheld devices, mobile computing devices, wearable devices, body mounted or implantable devices, smartphones, smart appliances, internetworking devices, autonomous or semi-autonomous devices such as robots or unmanned ground or aerial vehicles, any other electronic device that incorporates hard-wired and/or program logic to implement the described techniques, one or more virtual computing machines or instances in a data center, and/or a network of server computers and/or personal computers.
  • FIG. 6 is a block diagram that illustrates an example computer system with which an embodiment may be implemented.
  • a computer system 600 and instructions for implementing the disclosed technologies in hardware, software, or a combination of hardware and software are represented schematically, for example as boxes and circles, at the same level of detail that is commonly used by persons of ordinary skill in the art to which this disclosure pertains for communicating about computer architecture and computer systems implementations.
  • Computer system 600 includes an input/output (I/O) subsystem 602 which may include a bus and/or other communication mechanism(s) for communicating information and/or instructions between the components of the computer system 600 over electronic signal paths.
  • the I/O subsystem 602 may include an I/O controller, a memory controller and at least one I/O port.
  • the electronic signal paths are represented schematically in the drawings, for example as lines, unidirectional arrows, or bidirectional arrows.
  • At least one hardware processor 604 is coupled to I/O subsystem 602 for processing information and instructions.
  • Hardware processor 604 may include, for example, a general-purpose microprocessor or microcontroller and/or a special-purpose microprocessor such as an embedded system or a graphics processing unit (GPU) or a digital signal processor or ARM processor.
  • Processor 604 may comprise an integrated arithmetic logic unit (ALU) or may be coupled to a separate ALU.
  • ALU arithmetic logic unit
  • Computer system 600 includes one or more units of memory 606 , such as a main memory, which is coupled to I/O subsystem 602 for electronically digitally storing data and instructions to be executed by processor 604 .
  • Memory 606 may include volatile memory such as various forms of random-access memory (RAM) or other dynamic storage device.
  • RAM random-access memory
  • Memory 606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 604 .
  • Such instructions when stored in non-transitory computer-readable storage media accessible to processor 604 , can render computer system 600 into a special-purpose machine that is customized to perform the operations specified in the instructions.
  • Computer system 600 further includes non-volatile memory such as read only memory (ROM) 608 or other static storage device coupled to I/O subsystem 602 for storing information and instructions for processor 604 .
  • the ROM 608 may include various forms of programmable ROM (PROM) such as erasable PROM (EPROM) or electrically erasable PROM (EEPROM).
  • a unit of persistent storage 610 may include various forms of non-volatile RAM (NVRAM), such as FLASH memory, or solid-state storage, magnetic disk or optical disk such as CD-ROM or DVD-ROM and may be coupled to I/O subsystem 602 for storing information and instructions.
  • Storage 610 is an example of a non-transitory computer-readable medium that may be used to store instructions and data which when executed by the processor 604 cause performing computer-implemented methods to execute the techniques herein.
  • the instructions in memory 606 , ROM 608 or storage 610 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls.
  • the instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps.
  • the instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications.
  • the instructions may implement a web server, web application server or web client.
  • the instructions may be organized as a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
  • SQL structured query language
  • Computer system 600 may be coupled via I/O subsystem 602 to at least one output device 612 .
  • output device 612 is a digital computer display. Examples of a display that may be used in various embodiments include a touch screen display or a light-emitting diode (LED) display or a liquid crystal display (LCD) or an e-paper display.
  • Computer system 600 may include other type(s) of output devices 612 , alternatively or in addition to a display device. Examples of other output devices 612 include printers, ticket printers, plotters, projectors, sound cards or video cards, speakers, buzzers or piezoelectric devices or other audible devices, lamps or LED or LCD indicators, haptic devices, actuators or servos.
  • At least one input device 614 is coupled to I/O subsystem 602 for communicating signals, data, command selections or gestures to processor 604 .
  • input devices 614 include touch screens, microphones, still and video digital cameras, alphanumeric and other keys, keypads, keyboards, graphics tablets, image scanners, joysticks, clocks, switches, buttons, dials, slides, and/or various types of sensors such as force sensors, motion sensors, heat sensors, accelerometers, gyroscopes, and inertial measurement unit (IMU) sensors and/or various types of transceivers such as wireless, such as cellular or Wi-Fi, radio frequency (RF) or infrared (IR) transceivers and Global Positioning System (GPS) transceivers.
  • RF radio frequency
  • IR infrared
  • GPS Global Positioning System
  • control device 616 may perform cursor control or other automated control functions such as navigation in a graphical interface on a display screen, alternatively or in addition to input functions.
  • Control device 616 may be a touchpad, a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 604 and for controlling cursor movement on display 612 .
  • the input device may have at least two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane.
  • An input device 614 may include a combination of multiple different input devices, such as a video camera and a depth sensor.
  • computer system 600 may comprise an internet of things (IoT) device in which one or more of the output device 612 , input device 614 , and control device 616 are omitted.
  • the input device 614 may comprise one or more cameras, motion detectors, thermometers, microphones, seismic detectors, other sensors or detectors, measurement devices or encoders and the output device 612 may comprise a special-purpose display such as a single-line LED or LCD display, one or more indicators, a display panel, a meter, a valve, a solenoid, an actuator or a servo.
  • IoT internet of things
  • input device 614 may comprise a global positioning system (GPS) receiver coupled to a GPS module that is capable of triangulating to a plurality of GPS satellites, determining and generating geo-location or position data such as latitude-longitude values for a geophysical location of the computer system 600 .
  • Output device 612 may include hardware, software, firmware and interfaces for generating position reporting packets, notifications, pulse or heartbeat signals, or other recurring data transmissions that specify a position of the computer system 600 , alone or in combination with other application-specific data, directed toward host 624 or server 630 .
  • Computer system 600 may implement the techniques described herein using customized hard-wired logic, at least one ASIC or FPGA, firmware and/or program instructions or logic which when loaded and used or executed in combination with the computer system causes or programs the computer system to operate as a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 600 in response to processor 604 executing at least one sequence of at least one instruction contained in main memory 606 . Such instructions may be read into main memory 606 from another storage medium, such as storage 610 . Execution of the sequences of instructions contained in main memory 606 causes processor 604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
  • Non-volatile media includes, for example, optical or magnetic disks, such as storage 610 .
  • Volatile media includes dynamic memory, such as memory 606 .
  • Common forms of storage media include, for example, a hard disk, solid state drive, flash drive, magnetic data storage medium, any optical or physical data storage medium, memory chip, or the like.
  • Storage media is distinct from but may be used in conjunction with transmission media.
  • Transmission media participates in transferring information between storage media.
  • transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise a bus of I/O subsystem 602 .
  • Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
  • Various forms of media may be involved in carrying at least one sequence of at least one instruction to processor 604 for execution.
  • the instructions may initially be carried on a magnetic disk or solid-state drive of a remote computer.
  • the remote computer can load the instructions into its dynamic memory and send the instructions over a communication link such as a fiber optic or coaxial cable or telephone line using a modem.
  • a modem or router local to computer system 600 can receive the data on the communication link and convert the data to a format that can be read by computer system 600 .
  • a receiver such as a radio frequency antenna or an infrared detector can receive the data carried in a wireless or optical signal and appropriate circuitry can provide the data to I/O subsystem 602 such as place the data on a bus.
  • I/O subsystem 602 carries the data to memory 606 , from which processor 604 retrieves and executes the instructions.
  • the instructions received by memory 606 may optionally be stored on storage 610 either before or after execution by processor 604 .
  • Computer system 600 also includes a communication interface 618 coupled to bus 602 .
  • Communication interface 618 provides a two-way data communication coupling to network link(s) 620 that are directly or indirectly connected to at least one communication networks, such as a network 622 or a public or private cloud on the Internet.
  • network 622 may be an Ethernet networking interface, integrated-services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of communications line, for example an Ethernet cable or a metal cable of any kind or a fiber-optic line or a telephone line.
  • Network 622 broadly represents a local area network (LAN), wide-area network (WAN), campus network, internetwork or any combination thereof.
  • Communication interface 618 may comprise a LAN card to provide a data communication connection to a compatible LAN, or a cellular radiotelephone interface that is wired to send or receive cellular data according to cellular radiotelephone wireless networking standards, or a satellite radio interface that is wired to send or receive digital data according to satellite wireless networking standards.
  • communication interface 618 sends and receives electrical, electromagnetic or optical signals over signal paths that carry digital data streams representing various types of information.
  • Network link 620 typically provides electrical, electromagnetic, or optical data communication directly or through at least one network to other data devices, using, for example, satellite, cellular, Wi-Fi, or BLUETOOTH technology.
  • network link 620 may provide a connection through a network 622 to a host computer 624 .
  • network link 620 may provide a connection through network 622 or to other computing devices via internetworking devices and/or computers that are operated by an Internet Service Provider (ISP) 626 .
  • ISP 626 provides data communication services through a world-wide packet data communication network represented as internet 628 .
  • a server computer 630 may be coupled to internet 628 .
  • Server 630 broadly represents any computer, data center, virtual machine or virtual computing instance with or without a hypervisor, or computer executing a containerized program system such as DOCKER or KUBERNETES.
  • Server 630 may represent an electronic digital service that is implemented using more than one computer or instance and that is accessed and used by transmitting web services requests, uniform resource locator (URL) strings with parameters in HTTP payloads, API calls, app services calls, or other service calls.
  • URL uniform resource locator
  • Computer system 600 and server 630 may form elements of a distributed computing system that includes other computers, a processing cluster, server farm or other organization of computers that cooperate to perform tasks or execute applications or services.
  • Server 630 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps.
  • the instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications.
  • Server 630 may comprise a web application server that hosts a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
  • SQL structured query language
  • Computer system 600 can send messages and receive data and instructions, including program code, through the network(s), network link 620 and communication interface 618 .
  • a server 630 might transmit a requested code for an application program through Internet 628 , ISP 626 , local network 622 and communication interface 618 .
  • the received code may be executed by processor 604 as it is received, and/or stored in storage 610 , or other non-volatile storage for later execution.
  • the execution of instructions as described in this section may implement a process in the form of an instance of a computer program that is being executed, and consisting of program code and its current activity.
  • a process may be made up of multiple threads of execution that execute instructions concurrently.
  • a computer program is a passive collection of instructions, while a process may be the actual execution of those instructions.
  • Several processes may be associated with the same program; for example, opening up several instances of the same program often means more than one process is being executed. Multitasking may be implemented to allow multiple processes to share processor 604 .
  • computer system 600 may be programmed to implement multitasking to allow each processor to switch between tasks that are being executed without having to wait for each task to finish.
  • switches may be performed when tasks perform input/output operations, when a task indicates that it can be switched, or on hardware interrupts.
  • Time-sharing may be implemented to allow fast response for interactive user applications by rapidly performing context switches to provide the appearance of concurrent execution of multiple processes simultaneously.
  • an operating system may prevent direct communication between independent processes, providing strictly mediated and controlled inter-process communication functionality.

Landscapes

  • Business, Economics & Management (AREA)
  • Engineering & Computer Science (AREA)
  • Human Resources & Organizations (AREA)
  • Strategic Management (AREA)
  • Economics (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Development Economics (AREA)
  • General Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Tourism & Hospitality (AREA)
  • Quality & Reliability (AREA)
  • Operations Research (AREA)
  • Accounting & Taxation (AREA)
  • Finance (AREA)
  • Game Theory and Decision Science (AREA)
  • Educational Administration (AREA)
  • Data Mining & Analysis (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

In an embodiment, a method includes receiving training data representing historic consumer demand for products, detecting changepoints in that data that may be associated with disruptive events, identifying relevant data for modeling, performing clustering, processing configuration information, training one or more machine learning models that are capable of evaluating other received data more accurately, and outputting results to a user display device.

Description

    BENEFIT CLAIM
  • This application claims the benefit under 35 U.S.C. § 119(e) of provisional application 63/169,017, filed Mar. 31, 2021, the entire contents of which are hereby incorporated by reference for all purposes as if fully set forth herein.
  • COPYRIGHT NOTICE
  • A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright or rights whatsoever. © 2020-2022 Coupa Software, Inc.
  • TECHNICAL FIELD
  • One technical field of the present disclosure is computer-assisted forecasting of demand of goods or materials in complex supply chains. Another technical field is predictive modeling, including time series analysis. Another technical field is supply chain management. Another technical field is logistics as applied to disruptive conditions.
  • BACKGROUND
  • The approaches described in this section are approaches that could be pursued, but not necessarily approaches that have been previously conceived or pursued. Therefore, unless otherwise indicated, it should not be assumed that any of the approaches described in this section qualify as prior art merely by virtue of their inclusion in this section.
  • Demand forecasting is a field of predictive data analytics directed to optimizing supply chain decisions by predicting customer demand using computer-implemented algorithms. Demand forecasting may be a part of production planning, inventory management, marketing strategy development, and various other aspects of corporate decision-making. Demand forecasting methods may involve qualitative or quantitative assessment of pertinent data, including historical sales data. Demand forecasting may be accomplished by building a model and testing that model. A variety of techniques may be used to validate a model through testing.
  • A statistical technique which may be used in demand forecasting is time series analysis. Time series analysis involves forecasting future behavior by analyzing past behavior. Time series analysis may be used to predict future demand for certain products based, at least in part, on past sales of those products. Time series analysis may identify data features such as trends, seasonality, cyclicity, and irregularity.
  • Supply chain networks may involve distribution centers, suppliers, vendors, and manufacturers, to utilize to meet consumer demand for a particular finished good. Supply chain network techniques often employ multiple levels of interdependence and connectivity between sites within the supply chain network. Multiple models or techniques may be utilized to predict the behavior and interactions between these sites to optimally deliver goods and services to various points or locations along the supply chain network.
  • Individual sites within a supply chain network, for example manufacturing or production facilities, often feature complex interdependence and connectivity within the site due to multiple finished goods that may be manufactured at the site. Accurate forecasting of demand may require baseline data describing interactions between existing inventory, demand for finished goods, supply of raw materials, production processes each having a plurality of production process steps, production periods, and on-site equipment that must be managed at a particular site. Thus, an entity may incur excess use of materials, power, chemical resources, machine time, or other physical effects at individual sites along a typical supply chain if demand forecasting fails to consider internal and external factors that impact a particular site.
  • An effective solution requires computer implementation to manage issues of scale and real-time response timing that accounts for factors both internal and external to particular sites in a supply chain network. In some environments, buyer computers or buyer accounts may interoperate with dozens to hundreds of different supply chains, each with dozens to hundreds of nodes, in association with thousands to tens of thousands of products or components. Each node in all these complex supply chains may be associated with different production requirements that impact demand factors. Furthermore, vast data about external events can rapidly impact the accuracy of demand forecasting. Disruptive events such as obstruction of critical canals (Suez, Panama), pandemic or epidemic, or natural disasters can occur with little warning and rapidly render forecasts based on past data invalid. Even given this level of complexity, buyer computers require real-time responses to queries about demand. Buyer accounts need the ability to add, delete, or rearrange data describing internal or external demand factors while receiving a real-time response to an updated query for a demand forecast.
  • Fulfilling these requirements with a human-based solution has become impractical. If a solution could provide automated means of managing millions of data items, while still supporting real-time response, it would represent a practical application of machine-based computing technology that should gain widespread use across industry.
  • SUMMARY
  • The appended claims may serve as a summary of the invention.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • In the drawings:
  • FIG. 1 illustrates a distributed computer system showing the context of use and principal functional elements with which one embodiment could be implemented.
  • FIG. 2 illustrates an example process for preparing a training data set.
  • FIG. 3 illustrates an example process for disruption-resistant demand forecasting.
  • FIG. 4 illustrates an example of time series cluster visualization.
  • FIG. 5A, FIG. 5B, and FIG. 5C illustrate examples of generalized processes for forming forecasting models and using the forecasting models in a production computing environment that may be used to implement certain embodiments.
  • FIG. 6 illustrates a computer system with which one embodiment could be implemented.
  • DETAILED DESCRIPTION
  • In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide a thorough understanding of the present invention. It will be apparent, however, that the present invention may be practiced without these specific details. In other instances, well-known structures and devices are shown in block diagram form in order to avoid unnecessarily obscuring the present invention.
  • The text of this disclosure, in combination with the drawing figures, is intended to state in prose the algorithms that are necessary to program a computer to implement the claimed inventions, at the same level of detail that is used by people of skill in the arts to which this disclosure pertains to communicate with one another concerning functions to be programmed, inputs, transformations, outputs and other aspects of programming. That is, the level of detail set forth in this disclosure is the same level of detail that persons of skill in the art normally use to communicate with one another to express algorithms to be programmed or the structure and function of programs to implement the inventions claimed herein.
  • Embodiments are described in sections below according to the following outline:
      • 1. GENERAL OVERVIEW
      • 2. EXAMPLE IMPLEMENTATIONS OF COMPUTER SYSTEM AND PROGRAM FLOW
        • 2.1 COMPUTER SYSTEM OVERVIEW
        • 2.2 TRAINING DATA PREPARATION
        • 2.3 DISRUPTION-RESISTANT DEMAND PREDICTION OVERVIEW
        • 2.4 IDENTIFYING RELEVANT DATA & CLUSTERING
        • 2.5 CONFIGURATION INSTRUCTIONS & FLEXIBILITY
      • 3. IMPLEMENTATION EXAMPLE—HARDWARE OVERVIEW
  • 1. General Overview
  • When a highly disruptive event occurs, historical data may be insufficient for computer-implemented modeling of the future. One example of an incredibly disruptive event is the COVID-19 pandemic of 2019-2021 which has radically impacted consumer behavior, greatly changing demand for a variety of products and services and thereby creating enormous disruption and uncertainty in many industries' operations; other disruptive events could similarly affect demand and this disclosure is not limited to the context of pandemic. Companies have seen demand surge, plummet, or sometimes both, and the resulting confusion makes it hard for their supply chain teams to react. In an embodiment, this disclosure presents, among other things, novel time series forecasting methods and systems for predicting consumer demand for products under disruptive conditions.
  • In an embodiment, a novel computer-implemented method is presented for implementing technical machine learning solutions to the technical problem of machine learning model development, validation, and deployment in the domain of predictive modeling. In an embodiment, the disclosure presents solutions implemented via client-server Software as a Service (SaaS) techniques or distributed computer systems generally. In other embodiments, a variety of novel systems are presented for predicting consumer demand. In other embodiments, a diverse array of systems may be used to implement the novel methods presented in this disclosure.
  • In an embodiment, a distributed computer system is programmed to receive training data representing historic consumer demand for products, to detect changepoints in that data that may be associated with disruptive events, to identify relevant data for modeling, to perform clustering, to process configuration information, to train one or more machine learning models that are capable of evaluating other received data more accurately, and to output results to a user display device.
  • 2. Example Implementations of Computer System and Program Flow
  • 2.1 Computer System Overview
  • In an embodiment, an online distributed computer system or platform for predictive modeling provides a system for generating product demand models based on historical product demand data and using those models to predict future customer demand for products. In an embodiment, the platform comprises functionality for importing data, preparing training data sets, training, validating, and executing models, and visualizing results.
  • FIG. 1 illustrates a distributed computer system showing the context of use and principal functional elements with which one embodiment could be implemented.
  • FIG. 1, and the other drawing figures and all of the description and claims in this disclosure, are intended to present, disclose and claim, among other things, a technical system and technical methods in which specially programmed computers, using a special-purpose distributed computer system design, execute functions that have not been available before to provide a practical application of computing technology to the problem of machine learning model development, validation, and deployment. In this manner, the disclosure presents a technical solution to a technical problem, and any interpretation of the disclosure or claims to cover any judicial exception to patent eligibility, such as an abstract idea, mental process, method of organizing human activity or mathematical algorithm, has no support in this disclosure and is erroneous.
  • In an embodiment, a distributed computer system 100 comprises components that are implemented at least partially by hardware at one or more computing devices, such as one or more hardware processors executing stored program instructions stored in one or more memories for performing the functions that are described herein. In other words, all functions described herein are intended to indicate operations that are performed using programming in a special-purpose computer or general-purpose computer, in various embodiments. FIG. 1 illustrates only one of many possible arrangements of components configured to execute the programming described herein. Other arrangements may include fewer or different components, and the division of work between the components may vary depending on the arrangement.
  • In an embodiment, distributed computer system 100 comprises data acquisition logic 102 coupled to project management logic 104 that is programed to store project data in persistent storage, for example in organizational units such as projects 130. The project management logic 104 may also be coupled to data storage logic 108 which is programmed to manage the storage of data across the distributed computer system, for example in database 110. In an embodiment, data acquisition logic 102 is programmed to receive input signals specifying network locations of data sources or data files, to access and read the specified data sources or data files, and to import records from the specified data sources or data files into memory comprising database 110 or other memory, including, but not limited to, networked memory or cloud storage memory.
  • In an embodiment, projects 130 may be stored in database 110 of distributed computer system 100, which may be a relational database. In other embodiments, projects 130 may be stored in other memory accessible by distributed computer system 100. In an embodiment, database 110 stores a variety of data comprising product demand data 120, third-party data 122, training data 124, testing data 126, and output data 128.
  • In an embodiment, distributed computer system 100 further comprises data processing logic which is coupled to the project management logic 104 and the data storage logic 108, and which is programmed to process data accessible by distributed computer system 100. The data processing logic 106 may be programmed to initialize, train, execute, and validate machine learning models or statistical models. In an embodiment, machine learning models take as input training data 124 and third-party data 122. The data processing logic may first transform the product demand data 120 into one or more training data 124 sets. In an embodiment, a subset of the product demand data 120 may be used as training data 124 while a disjoint subset of the product demand data 120 is used as testing data 126 to test, validate, or otherwise assess the efficacy of a machine learning model or statistical model. In an embodiment, the data processing logic 106 is configured to visualize and display the output data 128 to a user display device 140.
  • 2.2 Training Data Preparation
  • FIG. 2 illustrates a flow diagram of a computer-implemented system preparing a training data set for product demand forecasting. FIG. 2 and each other flow diagram herein is intended as an illustration at the functional level at which skilled persons, in the art to which this disclosure pertains, communicate with one another to describe and implement algorithms using programming. The flow diagrams are not intended to illustrate every instruction, method object or sub-step that would be needed to program every aspect of a working program, but are provided at the same functional level of illustration that is normally used at the high level of skill in this art to communicate the basis of developing working programs.
  • In one embodiment, the present disclosure provides techniques to improve short-term demand forecasts for tactical response through novel time series analysis methodologies and with the help of external data. In addition to improving short-term demand forecasts, the disclosed methods and systems may allow a user to identify demand patterns unfolding across an entire product portfolio. In an embodiment, product portfolios or subsets thereof selected for analysis may correspond to projects 130 as illustrated in FIG. 1. Management of projects 130 and the coordination of processes required for training data set preparation may be controlled by project management logic 104 as further illustrated in FIG. 1.
  • In an embodiment, product demand data may comprise data representing historical customer demand for a single product, a product line of a company, a portfolio of related products, all products sold or offered for sale by a company or one of its subdivisions or subsidiaries, or products sold or offered for sale by any number of related or unrelated companies or individuals. In other embodiments, the product demand data may comprise any other data representing historical demand for products, or even services. In an embodiment, product demand data corresponds to product demand data 120 (FIG. 1) acquired by data acquisition logic 102 and stored in database 110 of distributed computer system 100 according to data storage logic 108, as illustrated in FIG. 1. In an embodiment, database 220 (FIG. 2) corresponds to database 110 (FIG. 1).
  • In an embodiment, product demand data may be associated with a geographical location. In these embodiments, product demand data may be global data, comprising historical demand data for one or more products in a plurality of countries around the world. In other embodiments, product demand data may be limited to a particular geographic region, country, state, province, city, or any other geographic or political subdivision.
  • In an embodiment, the product demand data comprises a dataset representing downstream consumption data rather than data from purchase orders or shipments. In times of disruption, such as during a global pandemic like COVID-19, downstream consumption data (e.g., Point of Sale (POS) data) may be a more accurate representation of actual customer demand than other demand data, because many demand transactions (e.g., consumer store visits) with zero sales of a given item may reflect a lack of sufficient stock as opposed to a lack of demand.
  • In other embodiments, a variety of other data sets, such as data concerning purchase orders or shipment data, may be used for a variety of reasons, including an inability to obtain downstream consumption data. In an embodiment, at block 206, a true demand may be estimated by imputing sales with a variety of imputation algorithms or substitution methods. In embodiments, imputation and substitution methods include, but are not limited to: (1) methods that account for substitution effects to estimate lost sales due to stock outs, (2) methods that use a demand rate for imputation, (2) if visits per store data is unavailable, estimating the basket of a visitor and the approximating the potential demand, or (4) if an item is primarily sold online and the item is out of stock, approximating lost sales by utilizing historical site visits of a specific product to a finalized sales ratio in conjunction with current site visit data. In other embodiments, other methods of estimating true demand using imputation or substitution may be used. The following is a non-limiting example of a method that accounts for substitution effects to estimate lost sales due to sock outs. If toilet paper of brand “A” is out of stock and toilet paper of brand “B′”s demand is higher by 40%, this percentage can be split by the historical sales of brand “A” and brand “B” to estimate brand “A” and “B′”s true demand. One of ordinary skill in the art, will recognize that a variety of other imputation or substitution methods may be used to estimate true demand when, for example, POS data is not available.
  • In an embodiment, product demand data may be data related to sales or demand from a previous year or series of years. In other embodiments product demand data may be data related to historical sales or demand tabulated with respect to any other unit of time, or even contain estimates, interpolations, or projections from any period of time.
  • In an embodiment, product demand data, may be tagged, labeled, or otherwise associated with not only a level of demand or sales and time data, but also one or more of a geographic location or data concerning a source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, stock keeping unit (SKU), or any data or metadata commonly used in the art.
  • In an embodiment, the product demand data set may be transformed using one of the described imputation or substitution methods before or after being loaded into memory as seen in block 204. In other embodiments, the product demand data may be cleaned or deduplicated before or after being loaded into memory as seen in block 204. In other embodiments, the product demand data may be processed in other ways before or after being loaded into memory as seen in block 204 such as by removing outliers from the product demand data, cleaning the product demand data, removing unreadable data, or correcting errors in the data. In an embodiment, the data processing at block 204 or other data processing illustrated in FIG. 2 is actuated by data processing logic 106 as illustrated in FIG. 1.
  • In an embodiment, the product demand data is formatted for processing at block 208. Formatting product demand data for processing may comprise transforming the data into a convenient format to be used as input in a machine learning model or statistical model. Data transformations effectuated at block 208 may comprise (1) resizing inputs to a particular fixed size (2) converting non-numeric data features into numeric features ones, (3) normalizing numeric data features, (4) lower-casing or tokenizing metadata text features, or (5) other data transformations used to process data prior to machine learning or statistical analysis.
  • In an embodiment, distributed computer system 100 is programmed to receive user input to select a data source for the product demand data. The user input may be visual or graphical, in the form of selection of graphical user interface widgets. In an embodiment, block 202 illustrates examples of data sources that may be selected.
  • In an embodiment, in response to a selection, a data set from the specified source 202 is written into memory, as seen in block 204. In an embodiment, the memory referenced in block 204 is main memory of a virtual machine instance of a cloud computing center that implements elements of a computer system, for networked access using client computers or workstations. For example, the computer system may be implemented using a dynamic plurality of virtual machine instances that client computers of end users access using Software as a Service (SaaS) techniques and correspond to distributed computer system 100 (FIG. 1). In another embodiment, the memory block referenced in block 204 is a block of memory in the memory 606 of the physical computing device illustrated in FIG. 6.
  • In an embodiment, once data is in memory at block 204, a disruption-resistant demand prediction workflow may begin at step 310 (FIG. 3) where a training data set is received for analysis. In an embodiment, the training data set comprises the entire data set loaded into memory at block 204. In another embodiment, the training data set comprises a subset of the data set loaded into memory at block 204. In an embodiment, the subset of the data set is selected by user input. In an embodiment, the subset of the data set is selected so that the remaining data in the data set may be used for testing, evaluation, or validation of modeling results. In an embodiment the subset of the data set may be selected at least partially based on any of the various data or metadata associated with the product demand data, including but not limited to information about geographic location, source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, or SKU. In an embodiment, the data is not partitioned at block 204, but is instead partitioned at block 210. In an embodiment, product demand data comprising a training data set is stored in database 220 after it is cleaned, transformed, processed, or formatted.
  • 2.3 Disruption-Resistant Demand Prediction Overview
  • FIG. 3 illustrates a flow diagram overview of a computer-implemented method for processing a training data set for product demand forecasting.
  • In an embodiment, at step 310, a computer system receives a training data set. In an embodiment, the training data set may have been prepared and loaded into memory as explained in Section 2.2 of this disclosure. In an embodiment, the training data set may comprise product demand data as explained in Section 2.2 of this disclosure. In other embodiments, the computer system receives product demand data according to other specifications. In embodiments, the training data set may comprise any kind of time series data set. In an embodiment, the method illustrated in FIG. 3. may be implemented on the distributed computer system illustrated in FIG. 1 and the coordination of processes required for disruption-resistant demand prediction may be controlled by project management logic 104. In an embodiment, the method illustrated in FIG. 3. may be implemented on computer system 600 illustrated in FIG. 6. In other embodiments, the method illustrated in FIG. 3 may be implemented by another type of system.
  • At step 320 of FIG. 3, in an embodiment, a distributed computer system 100 calculates any break points in the training data set. Break points may correspond to the occurrence of real-life disruptive events that cause sudden level or trend changes in historical product demand data. These disruptive events may be epidemics like COVID-19, other disease outbreaks, other events that cause increased portions of a population to stay at home or restrict movement, or a variety of other disruptive events. Break points may correspond to a shift from one type of behavior in a consumer populace, such as panic, buying, stabilization, or normalcy to another such phase of behavior in a consumer populace. Break points may correspond directly with disruptive events or they may be time-lagged. Break points may occur as a result of responses to a disruptive event, including governmental responses. Governmental responses may comprise financial responses such as changes to fiscal or monetary responses or legal responses such as restrictions, regulation, or curfews. Break points may also be known in the art of predictive modeling or other arts as changepoints.
  • At step 320, in an embodiment, a distributed computer system 100 calculates zero or more changepoints in the training data set by executing a changepoint detection algorithm. In an embodiment, changepoints may be detected within the entire training data. In other embodiments, the training data set may be federated, or hierarchical and changepoints may be detected within subsets of the training data set. Changepoints may be detected in subsets corresponding to any of the various data or metadata associated with product demand data, including but not limited to information about geographic location, source, manufacturer, buyer, seller, shipper, contract terms, warranty terms, description, material, size, color, packaging, part number, or SKU. In some embodiments, the training data set may comprise time series data that can be clustered to identify and group time series (e.g., products and locations) that are experiencing similar shifts in demand patterns. In those embodiments, distributed computer system 100 may be programmed to cluster the training data set, and then execute a changepoint detection algorithm to detect changepoints in a plurality of representative time series corresponding to a plurality of clusters. In an embodiment, a distributed computer system 100 receives input from a user which specifies whether to cluster the training data set and calculate changepoints in representative time series corresponding to detected clusters using a changepoint detection algorithm. Such clustering may save time and computing resources when the training data set contains time series data for a large number of products, in some instances for thousands or millions of products.
  • At step 320 of FIG. 3, in an embodiment, a computer system calculates changepoints using machine learning methods. Machine learning methods used for changepoint detection may be supervised or unsupervised. Supervised methods comprise methods that may use (1) multi class classifiers, such as decision tree, nearest neighbor, support vector machine (SVM), naïve Bayes, Bayesian net, hidden Markov model (HMM), conditional random field (CRF), Gaussian mixture model (GMM) methods, (2) binary class classifiers, such as support vector machine (SVM), naïve Bayes, or logistic regression methods, or (3) virtual classifiers. Unsupervised methods may use a likelihood ratio, a subspace model, probabilistic methods, kernel-based methods, graph-based methods, or clustering. Changepoint detection may be aided by an indicator akin to a moving average convergence/divergence (MACD) indicator.
  • At step 330 of FIG. 3, in an embodiment, a distributed computer system 100 identifies relevant data for modeling. The entire training data set may be modeled, or a subset of the training data set may be modeled. In an embodiment, step 330 comprises distributed computer system 100 calculating or retrieving a baseline forecast for one or more products to determine if any of those products are currently being, or will be in the future, impacted by a disruptive event. In an embodiment, a baseline forecast is retrieved which was calculated at a time prior to a time value detected to be a changepoint in one or more products at step 320. In an embodiment, demand statistics comprising mean demand level, median demand level, standard deviation of demand level, or other major demand statistics are calculated for selected periods before and after a detected changepoint in time series data for one or more products or clusters. In an embodiment, if there is a large deviance in major demand statistics before or after a changepoint in a product time series, then distributed computer system 100 may be programmed to flag that time series as significantly impacted by a disruptive event. A large deviation, for example, of demand statistics or a forecast error prior to and after the disruptive event may be quantified as a deviation exceeding 1.5*Inter Quartile Range (IQR), above and below the 75th and 25th percentiles of the historical data. One of ordinary skill in the art will recognize that other metrics may be employed to quantify a significant forecast error or deviation for flagging. In embodiments where distributed computer system 100 clusters time series at step 320, distributed computer system 100 may determine which representative time series of one or more clusters should be flagged as significantly impacted by the disruptive event using a discussed method.
  • At step 340 of FIG. 3, in an embodiment, a distributed computer system 100 processes configuration information which partially determines one or more of (1) which data to model, (2) which of one or more types of models to use, (3) which third-party data, if any, to feed into the one or models, (4) how those one or more models are to be initialized, configured, or run, and (5) other configuration information. Configuration information may comprise a cross-learning preference, a preference for accuracy or control, input narrowing specifications comprising a selection of a specific product, region, or cluster to model, a dampening factor, a future inflection point prediction, a third-party data selection, or a short-term-forecasting preference. In an embodiment, one or more pieces of configuration information is received as input from a user. In other embodiments, all configuration information is hard-coded into memory accessible by distributed computer system 100.
  • At step 350 of FIG. 3, in an embodiment, a distributed computer system 100 executes instructions to create one or more models to predict future product demand data based on the training data set. In an embodiment, one or more models are run based on the configuration information processed at step 340.
  • At step 350, in an embodiment, third party data may be used to improve the predictive power of one or more models. For example, when dealing with a disruptive event such as a disease outbreak, like the COVID-19 epidemic, external metrics around restrictions on societal activities or compliance may provide insight on future product demand. Examples of such third-party data include, but are not limited to, mobility data from technology companies that provide maps, GPS tracking, or navigational directions as a service. Such data may capture the impact of local restrictions and how much compliance is observed in various localities, for example, by tracking a percent change in visits to places like grocery stores and parks within a geographic area. Location-based mobility metrics for one or more calendar days may be compared to a baseline value for that day in the immediately preceding calendar year. In another example, third-party projections of effects derived from a disruptive event, such as predicted death rate curves for an outbreak like COVID-19 may be correlated with future product demand. Other indicators impacting demand may include a social distance index, case counts, school closures data, unemployment claims data, consumer sentiment, hospital utilization data, or other indicators. In embodiments, one or more of these third-party projections or indicators provide additional data sources with which one or more models may be trained. In an embodiment, third party data is selected for processing in response to user input. In an embodiment, third party data is selected for processing based on the configuration instructions processed at step 340.
  • At step 350, in an embodiment, data engineering may be used to process related regressors such as external metrics, location-based mobility data, third party projections, or other indicators impacting demand before they are used to train models. Data engineering may be usefully employed because many related regressors may not be forward-looking, meaning they are not available at prediction time and cannot be used directly to train models. In embodiments, step 350 may comprise generating lags (lead indicators) and window statistics with a starting point limited by the forecast horizon. Next, correlation analysis may be performed on lagged regressors to determine optimal lags. In other embodiments, step 350 may comprise using forecasted regressors, which may have high accuracy validation metrics, at prediction time.
  • At step 350, in an embodiment one or more machine learning models are trained and executed after third-party data has been engineered for processing. Training may comprise programmatically supplying a training dataset to a machine learning model that is executing in the computer system and programmatically activating a training function of the model with a reference to or identification of the training dataset to be used. In an embodiment, third party data may be cleaned, deduplicated, or otherwise modified before it used for training models. In other embodiments, third-party data is used without being engineered or modified. In other embodiments, one or more models are trained and executed without the use of third-party data. Various models may be trained or executed at step 350, including, but not limited to, Classical Models (e.g., Statistical Models or State Space Models) or Machine Learning & AI based Models (e.g., Machine Learning Models or Neural Networks). Statistical Models comprise autoregressive (AR) models, moving average (MA) models, autoregressive integrated moving average (ARIMA) models, vector autoregression (VAR) models, hierarchical models, or others. State Space Models comprise exponential smoothing (ETS) models, Hidden Markov models, Bayesian structural time-series models, or others. Machine Learning models comprise support vector machines (SVMs), tree-based models, kNN's, or others. Neural Networks comprise temporal convolutional neural (TCN) networks, multi-layer perceptron networks (FF neural networks), recurrent neural networks (e.g., long short-term memory (LSTM) or gated recurrent units (GRUs)), convolutional neural networks (CNN), or others. In an embodiment, one or more models executed at step 350 comprise lift-adjusted seasonal naïve models, quantile regression models, stockout prediction models.
  • At step 350, in an embodiment, a distributed computer system 100 only executes new models for those time series flagged as significantly impacted by the disruptive event at step 330.
  • Different geographic regions may be in different phases of the lifecycle of a disruptive event such as a disease outbreak (e.g., COVID-19). As some cities, state, or countries may be further along an epidemic curve, lagged features may be used to predict patterns in other cities, states, or countries respectively. At step 350, in an embodiment, to determine lag between cities/states/countries, changepoint detection may be implemented for similar (e.g., clustered) products. Calculated lag may represent time required to change phase, enabling a distributed computer system 100 to compute the expected point in time where changes should be expected. The aforementioned process may be understood as using predicted changepoint detection and assigning binary regressors in the forecasting period. One or more models executed at step 350 may be modified with this technique to potentially improve forecasting accuracy.
  • For certain disruptive events such as viral outbreaks (e.g., COVID-19) causals may not be applicable at a SKU level, but rather at higher levels of the hierarchy. Furthermore, when modeling aggregated data, noise is removed, and substitution effects are minimized. Thus, it may be more effective to model at higher levels of the product or location hierarchy to better capture level shifts. Once such shifts are detected, they may be propagated to lower levels of a hierarchy via Hierarchical Time Series (HTS) forecasting or in post-processing. At step 350, in an embodiment, one or more models executed may employ HTS forecasting. HTS forecasting may be implemented using a bottom-up approach, a top-down approach, a middle-out approach, or an optimal reconciliation approach. HTS forecasting may be most useful when product demand data is highly federated and the type of HTS approach may be determined by the level of hierarchy in the product demand data. Such approaches may increase coherency in model output.
  • At step 350 of FIG. 3, in an embodiment, a distributed computer system 100 executes instructions to create one or more models to predict future product demand data based on the training data set. In an embodiment, one or more models are run based on the configuration information processed at step 340.
  • At step 350 of FIG. 3, in an embodiment, a distributed computer system 100 executes “back testing,” or time series validation, to assess the forecasting accuracy of one or more models. Back testing may be preferable to simpler cross-validation methodologies sometimes used in machine learning model validation. For example, simple random splitting of samples into train and test sets might introduce data leakage in the time series context where there is a temporal dependency between observations. In an embodiment, validating one or more models comprises executing “walk-forward cross-validation.” In other embodiments, other validation techniques may be used which may incorporate folds sizes and gaps between the sets. Back testing may be used for a Hyperparameter Tuning Set, an Ensemble Set, an Estimator Selection Set, or an Unbiased Test Set, independently and sequentially.
  • At step 360 of FIG. 3, in an embodiment, a distributed computer system 100 outputs the results of the models run at step 350, including to, for example, user display device 140 (FIG. 1). In an embodiment, modeling results are output in a visualization pane of a graphical user interface (GUI). Modeling results may be output in the form of graphs, tables, charts, or other forms. Modeling results may be written to a file, such as a CSV, Excel, or txt file, or to another type of file. Modeling results may be stored in a database or transmitted as output to be input for another system.
  • FIG. 5A, FIG. 5B, FIG. 5C illustrate examples of generalized processes for forming forecasting models and using the forecasting models in a production computing environment that may be used to implement certain embodiments.
  • Referring first to FIG. 5A, in an embodiment, development and deployment of forecasting models involves a two-stage approach. A first stage, termed a Model Building stage, is part of the development process which comprises an analysis of the data and the evaluation of forecasting models. Model Building begins with a data analysis and data engineering step in which historical datasets are assessed for the presence of features useful in forecasting, missing fields are identified for interpolation or enrichment from other data sources, and normalization is conducted to adjust the format of otherwise incompatible data fields. A Model Specification and Evaluation step follows in which a forecasting model is selected or specified in mathematical terms, implemented in program code, and executed to evaluate or process the historical data to product output in the form of forecasting results. The model is deemed adequate if the output forecast is found usable. If not, the model may be re-specified by adjusting mathematical terms or changing input data features, and re-evaluated.
  • When the model is adequate, control transfers to an Operationalization stage. The Operationalization stage performs forecasting in a production environment. In an embodiment, the model that was validated in the prior stage is instantiated in one or more executable instances, for example, using computing instances of a cloud computing center. These instances are executed to generate forecasts that support the decision-making process, normally using newly created data having new timestamps and possibly new features. As forecast output is produced, the output is evaluated to check stability and continued validity. Over time, model drift or degradation is expected to occur, which requires retraining of the selected models, so control may transfer back to the model specification and evaluation stage. If the model is in a stable state, then a forecast update may be output based on evaluation of the new timestamps and features and used for decision-making.
  • FIG. 5B illustrates an embodiment of the Model Building stage of FIG. 5A in greater detail. In an embodiment, the Model Building stage begins with a Business Understanding step comprising identifying one or more relevant metrics for forecasting, determining how to measure the metrics, and determining how to measure successful forecasting via one or more success metrics.
  • The Model Building stage then continues, in an embodiment, with a Data Preparation Phase. In an embodiment, the Data Preparation Phase comprises a data ingestion step that involves identifying one or more different data sources, locating datasets in one or more data analytics environments, and reading one or more files containing the datasets. Relational database tables, spreadsheets, flat files, or other sources may be read. In an embodiment, the data ingestion step is followed by a data exploration & understanding step that involves exploring the characteristics and meaning of a dataset and determining if the data quality is adequate to answer one or more questions about the data. These steps may be looped through or iteratively repeated until data quality is sufficient. In an embodiment, once data quality is sufficient, a final step in the Data Preparation Phase comprises a data pre-processing & feature engineering step that involves cleaning missing values and outliers from the data and creating data features from the raw data to facilitate model training.
  • The Model Building stage then continues, in an embodiment, with a Data Modeling Phase. In an embodiment, the Data Modeling Phase comprises a model building & selection step that involves using algorithms to scan historical data and extract patterns, building one or more models to predict future values, and selecting one or more best models based on project criteria. These steps may be looped through or iteratively repeated to build new models as more historical data is ingested or project criteria changes. In an embodiment, the model building and selection step is followed by a model deployment step that involves deploying one or more models and data pipelines to a production environment for application consumption.
  • The Model Building stage then concludes, in an embodiment, with a Business Validation step that involves forecasting solution acceptance. In an embodiment, forecasting solution acceptance comprises confirming that the one or more pipelines and the one or more models satisfy success criteria and validating that a final solution adds value to a business and its processes.
  • FIG. 5C illustrates an embodiment of the Operationalization stage of FIG. 5A in greater detail. In an embodiment, data collection continues, and raw data is fed into a cloud-based system for generating forecasts in a production environment. In an embodiment, once new raw data is ingested into the cloud-based system, the Operationalization stage continues with supplemental data exploration and aggregation, wherein the aggregated data may be explored to assess the characteristics of the aggregated data set and reevaluate data quality. In an embodiment, the data exploration and aggregation step is followed by a feature engineering step in the Operationalization stage. Feature engineering may include cleaning missing values and outliers from the data and creating data features from the data to facilitate model training.
  • In an embodiment, the Operationalization stage of FIG. 5A involves feeding processed data which is the resulting output of feature engineering into a Re-training Web Service or a Scoring Web Service. A Re-training Web Service retrains one or more machine learning or statistical models previously described and may comprise a web service instance of a cloud computing system denoted “Re-training Web Service.” A Scoring Web Service executes forecast models or processes previously described and may comprise a web service instance of a cloud computing system denoted “Scoring Web Service.” In an embodiment, as new forecast output is produced in the production environment, the output is reevaluated to check stability and continued validity. If the model is in a stable state, then a forecast update may be output based on evaluation of new timestamps and features and used for decision-making.
  • In an embodiment, the Operationalization stage of FIG. 5A involves transmitting forecasted results from the cloud-based system to a Consumption Client, which may be any client configured to receive, or consume, the output of one or more models executed in the production environment. The Consumption Client may be a server accessible via an API call, a relational database or other file system, or any application layer configured to receive forecasted results as input.
  • 2.4 Identifying Relevant Data & Clustering
  • At step 330 of FIG. 3, in an embodiment, a computer system identifies relevant data for modeling. As explained in greater detail in Section 2.3 of this disclosure, one way that relevant data may be identified for modeling is: (1) by calculating a baseline forecast, (2) by calculating relevant demand statistics or forecast errors for a period of time before or after a changepoint, and (3) by identifying data corresponding to segments of time series after changepoints with high forecast error as data which is particularly relevant for modeling. Another helpful technique which may be used to identify relevant data for modeling is clustering.
  • At step 330 of FIG. 3, in an embodiment, clustering may provide insights into product velocity and demand as a disruptive event takes firm hold in a region. Clustering may be used to identify and group time series (e.g. items and locations) that are experiencing similar shifts in demand patterns. Since impact from a disruptive event may change rapidly over time, it may be advantageous to condense the amount of information that needs to be processed, while providing detailed information on the types of patterns that are unfolding across an entire portfolio of products. By clustering and focusing on a representative time series for each pattern of demand unraveling within each clustering, groups of products in the product portfolio exhibiting signs of normal behavior may be identified. Likewise, groups of products deviating away from normalcy may be identified. In an embodiment, data for one or more product clusters identified as deviating from normalcy may be then be used for modeling at step 350. The described process may allow for scaling to millions of products quickly and at lower computational cost.
  • At step 330, in an embodiment, a K-Means clustering approach may be used to scale the clustering process to millions of time series. The K-Means clustering approach may comprise using a demand pattern observed over a window of time since the onset of the disruption as provided features. These features (demand over a window) may be normalized and smoothed to ensure some randomness from day to-day fluctuations does not have a large impact on the clustering process. The normalization may be interpreted as the number of folds in the average historical demand at a certain period (wherein values >1 indicate increased demand, and values <1 indicate less demand).
  • FIG. 4 shows example output from a clustering, where, for each cluster, the features (normalized time series over a recent horizon) are plotted against the time periods. In FIG. 4, Clusters 3 and 4 may be deemed clusters where the demand has been showing an increasing trend without any indication of return to normalcy. Cluster 3 may be distinguished from Cluster 4 by a level of increase in demand. Demand in Cluster 3 appears to grow unabated, experiencing up to five times the normal volume of demand. Cluster 3 may represent, for example products such as toilet paper, hand sanitizers, and other disinfectants during early stages of the COVID-19 epidemic. On the other hand, demand in Cluster 4 appears to rise to two times historical values.
  • FIG. 4 shows Clusters 1, 2, and 6 as groups products that have experienced a growth in demand followed by signs of returning to normalcy or settling at a new normal. However, what distinguishes each one of these clusters is the degree of the growth, rate of the growth, and the demand level the clusters appear to stabilize at. Clusters 1 and 6 appear to have experienced a gradual increase in demand, while demand in Cluster 1 appeared to stop growing, and began returning to normalcy. Demand for products in Cluster 6 appeared to continue to grow before settling at a new normal. Finally, Cluster 6 appears to group together products that experienced a negative effect on demand due to COVID-19, wherein demand stayed flat for some time, and then, as the severity of the disease progressed, evidenced a steep decline in demand before settling at a new normal (e.g., products such as camping equipment).
  • 2.5 Configuration Instructions & Flexibility
  • At step 340 of FIG. 3, in an embodiment, a computer system processes configuration information which partially determines one or more of (1) which data to model, (2) which of one or more types of models to use, (3) which third-party data, if any, to feed into the one or models, (4) how those one or more models are to be initialized, configured, or run, and (5) other configuration information.
  • At step 340, in an embodiment, a computer system processes configuration information comprising data which indicates a preference for accuracy or for control. In an embodiment, if the data indicates a preference for accuracy, then one or more of an LSTM or another Neural Network type model, or a Bayesian structural time series model is subsequently executed at step 350. In an embodiment, if the data indicates a preference for control, then one or both of an ETS or lift-adjusted seasonal naïve model is subsequently run at step 350.
  • In an embodiment, configuration information processed at step 340 may also indicate that a high-demand scenario is present. In an embodiment, one or both of quantile regression and a stockout prediction model are executed at step 350 when configuration information processed at step 340 indicates a high-demand scenario. Quantile regression may be used to forecast ranges of high demand estimates to help meet acceptable service levels. Quantile regression forecasting involves modeling an entire demand distribution; and, it may be particularly effective when costs associated with over-forecasting and under-forecasting are unequal. Stockout prediction models may help determine the probability of a stockout over a given timeframe. This model may effectively identify items with a high likelihood of running out of stock and can be implemented as a classification algorithm with the target variable as either a stock-out indicator (1 or 0) or an inventory level range (e.g., low/medium/high). Rate-of-consumption metrics, inventory levels, and demand forecasts may serve as inputs to a stockout prediction model. In some embodiments, one or both of a quantile regression or stockout prediction model may be executed at step 350 even when a high-demand scenario is not indicated or present.
  • In an embodiment, configuration information processed at step 340 may comprise data indicating a cross-learning preference. Cross-learning, a characteristic of global methods, may be used by a variety of models to learn patterns between similar time series, particularly when a large number of time series are being dealt with. In an embodiment, cross-learning is implemented in one or more models executed at step 350 when configuration information processed at step 340 indicates a cross-learning preference. In other embodiments, cross-learning may be implemented in one or more models executed at step 350 even when configuration information does not indicate a cross-learning preference.
  • In an embodiment, configuration information processed at step 340 may comprise data indicating a short-term-forecasting preference. In an embodiment, more recent product demand data is weighted more heavily in one or more models executed at step 350 when configuration information processed at step 340 indicates a short-term-forecasting preference. In other embodiments, data weighting may be implemented in one or more models executed at step 350 even when configuration information does not indicate a short-term-forecasting preference.
  • In an embodiment, configuration information processed at step 340 may comprise data indicating a third-party data selection. In an embodiment, third party data is incorporated in training one or more models executed at step 350 when configuration information processed at step 340 indicates a third-party data selection.
  • In an embodiment, configuration information processed at step 340 may comprise data indicating a dampening factor. Notwithstanding seasonality and noise, most SKU's may be associated with a generally increasing trend which may be evident in a previously calculated or received baseline model. A dampening factor reduces the changes presented by a new model in comparison to a baseline model calculated before a disruptive event. A dampening factor may be thought of as a “Social Restrictions Lever” with regards to the COVID-19 epidemic. A higher dampening factor may model a prediction of fewer social restrictions and a tendency for consumer demand to emulate a baseline forecast more closely than newly predicted. In an embodiment, a dampening factor is incorporated in executing one or more models at step 350 when configuration information processed at step 340 indicates a dampening factor.
  • In an embodiment, configuration information processed at step 340 may comprise data indicating a future inflection point prediction. Data about a disruptive event (such as third-party projections concerning COVID-19 death rates) may indicate a future changepoint is likely. In such cases, a future changepoint may be implemented in executing one or more models at step 350 via a regressor or structural break. The future change point indicates a likelihood of, for example, an inflection point, a change in level, or a change in trend for product demand. In an embodiment, a future inflection point prediction may comprise data indicating that models executed at 350 should be adjusted to reflect a “Return to Normalcy” by a certain “Recovery Date” and that demand predictions should revert to a new baseline prediction after that date. The “Recovery Date”, may be, for example, when a death rate from an epidemic (e.g., COVID-19) is expected to fall below 6 deaths per day or 10% of a peak death rate for one day in a location, or another date when normalcy is predicted to return. In an embodiment, a future inflection point prediction is incorporated in executing one or more models at step 350 when configuration information processed at step 340 indicates a future inflection point prediction.
  • 3. Implementation Example—Hardware Overview
  • According to one embodiment, the techniques described herein are implemented by at least one computing device. The techniques may be implemented in whole or in part using a combination of at least one server computer and/or other computing devices that are coupled using a network, such as a packet data network. The computing devices may be hard-wired to perform the techniques, or may include digital electronic devices such as at least one application-specific integrated circuit (ASIC) or field programmable gate array (FPGA) that is persistently programmed to perform the techniques, or may include at least one general purpose hardware processor programmed to perform the techniques pursuant to program instructions in firmware, memory, other storage, or a combination. Such computing devices may also combine custom hard-wired logic, ASICs, or FPGAs with custom programming to accomplish the described techniques. The computing devices may be server computers, workstations, personal computers, portable computer systems, handheld devices, mobile computing devices, wearable devices, body mounted or implantable devices, smartphones, smart appliances, internetworking devices, autonomous or semi-autonomous devices such as robots or unmanned ground or aerial vehicles, any other electronic device that incorporates hard-wired and/or program logic to implement the described techniques, one or more virtual computing machines or instances in a data center, and/or a network of server computers and/or personal computers.
  • FIG. 6 is a block diagram that illustrates an example computer system with which an embodiment may be implemented. In the example of FIG. 6, a computer system 600 and instructions for implementing the disclosed technologies in hardware, software, or a combination of hardware and software, are represented schematically, for example as boxes and circles, at the same level of detail that is commonly used by persons of ordinary skill in the art to which this disclosure pertains for communicating about computer architecture and computer systems implementations.
  • Computer system 600 includes an input/output (I/O) subsystem 602 which may include a bus and/or other communication mechanism(s) for communicating information and/or instructions between the components of the computer system 600 over electronic signal paths. The I/O subsystem 602 may include an I/O controller, a memory controller and at least one I/O port. The electronic signal paths are represented schematically in the drawings, for example as lines, unidirectional arrows, or bidirectional arrows.
  • At least one hardware processor 604 is coupled to I/O subsystem 602 for processing information and instructions. Hardware processor 604 may include, for example, a general-purpose microprocessor or microcontroller and/or a special-purpose microprocessor such as an embedded system or a graphics processing unit (GPU) or a digital signal processor or ARM processor. Processor 604 may comprise an integrated arithmetic logic unit (ALU) or may be coupled to a separate ALU.
  • Computer system 600 includes one or more units of memory 606, such as a main memory, which is coupled to I/O subsystem 602 for electronically digitally storing data and instructions to be executed by processor 604. Memory 606 may include volatile memory such as various forms of random-access memory (RAM) or other dynamic storage device. Memory 606 also may be used for storing temporary variables or other intermediate information during execution of instructions to be executed by processor 604. Such instructions, when stored in non-transitory computer-readable storage media accessible to processor 604, can render computer system 600 into a special-purpose machine that is customized to perform the operations specified in the instructions.
  • Computer system 600 further includes non-volatile memory such as read only memory (ROM) 608 or other static storage device coupled to I/O subsystem 602 for storing information and instructions for processor 604. The ROM 608 may include various forms of programmable ROM (PROM) such as erasable PROM (EPROM) or electrically erasable PROM (EEPROM). A unit of persistent storage 610 may include various forms of non-volatile RAM (NVRAM), such as FLASH memory, or solid-state storage, magnetic disk or optical disk such as CD-ROM or DVD-ROM and may be coupled to I/O subsystem 602 for storing information and instructions. Storage 610 is an example of a non-transitory computer-readable medium that may be used to store instructions and data which when executed by the processor 604 cause performing computer-implemented methods to execute the techniques herein.
  • The instructions in memory 606, ROM 608 or storage 610 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps. The instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications. The instructions may implement a web server, web application server or web client. The instructions may be organized as a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
  • Computer system 600 may be coupled via I/O subsystem 602 to at least one output device 612. In one embodiment, output device 612 is a digital computer display. Examples of a display that may be used in various embodiments include a touch screen display or a light-emitting diode (LED) display or a liquid crystal display (LCD) or an e-paper display. Computer system 600 may include other type(s) of output devices 612, alternatively or in addition to a display device. Examples of other output devices 612 include printers, ticket printers, plotters, projectors, sound cards or video cards, speakers, buzzers or piezoelectric devices or other audible devices, lamps or LED or LCD indicators, haptic devices, actuators or servos.
  • At least one input device 614 is coupled to I/O subsystem 602 for communicating signals, data, command selections or gestures to processor 604. Examples of input devices 614 include touch screens, microphones, still and video digital cameras, alphanumeric and other keys, keypads, keyboards, graphics tablets, image scanners, joysticks, clocks, switches, buttons, dials, slides, and/or various types of sensors such as force sensors, motion sensors, heat sensors, accelerometers, gyroscopes, and inertial measurement unit (IMU) sensors and/or various types of transceivers such as wireless, such as cellular or Wi-Fi, radio frequency (RF) or infrared (IR) transceivers and Global Positioning System (GPS) transceivers.
  • Another type of input device is a control device 616, which may perform cursor control or other automated control functions such as navigation in a graphical interface on a display screen, alternatively or in addition to input functions. Control device 616 may be a touchpad, a mouse, a trackball, or cursor direction keys for communicating direction information and command selections to processor 604 and for controlling cursor movement on display 612. The input device may have at least two degrees of freedom in two axes, a first axis (e.g., x) and a second axis (e.g., y), that allows the device to specify positions in a plane. Another type of input device is a wired, wireless, or optical control device such as a joystick, wand, console, steering wheel, pedal, gearshift mechanism or other type of control device. An input device 614 may include a combination of multiple different input devices, such as a video camera and a depth sensor.
  • In another embodiment, computer system 600 may comprise an internet of things (IoT) device in which one or more of the output device 612, input device 614, and control device 616 are omitted. Or, in such an embodiment, the input device 614 may comprise one or more cameras, motion detectors, thermometers, microphones, seismic detectors, other sensors or detectors, measurement devices or encoders and the output device 612 may comprise a special-purpose display such as a single-line LED or LCD display, one or more indicators, a display panel, a meter, a valve, a solenoid, an actuator or a servo.
  • When computer system 600 is a mobile computing device, input device 614 may comprise a global positioning system (GPS) receiver coupled to a GPS module that is capable of triangulating to a plurality of GPS satellites, determining and generating geo-location or position data such as latitude-longitude values for a geophysical location of the computer system 600. Output device 612 may include hardware, software, firmware and interfaces for generating position reporting packets, notifications, pulse or heartbeat signals, or other recurring data transmissions that specify a position of the computer system 600, alone or in combination with other application-specific data, directed toward host 624 or server 630.
  • Computer system 600 may implement the techniques described herein using customized hard-wired logic, at least one ASIC or FPGA, firmware and/or program instructions or logic which when loaded and used or executed in combination with the computer system causes or programs the computer system to operate as a special-purpose machine. According to one embodiment, the techniques herein are performed by computer system 600 in response to processor 604 executing at least one sequence of at least one instruction contained in main memory 606. Such instructions may be read into main memory 606 from another storage medium, such as storage 610. Execution of the sequences of instructions contained in main memory 606 causes processor 604 to perform the process steps described herein. In alternative embodiments, hard-wired circuitry may be used in place of or in combination with software instructions.
  • The term “storage media” as used herein refers to any non-transitory media that store data and/or instructions that cause a machine to operation in a specific fashion. Such storage media may comprise non-volatile media and/or volatile media. Non-volatile media includes, for example, optical or magnetic disks, such as storage 610. Volatile media includes dynamic memory, such as memory 606. Common forms of storage media include, for example, a hard disk, solid state drive, flash drive, magnetic data storage medium, any optical or physical data storage medium, memory chip, or the like.
  • Storage media is distinct from but may be used in conjunction with transmission media. Transmission media participates in transferring information between storage media. For example, transmission media includes coaxial cables, copper wire and fiber optics, including the wires that comprise a bus of I/O subsystem 602. Transmission media can also take the form of acoustic or light waves, such as those generated during radio-wave and infra-red data communications.
  • Various forms of media may be involved in carrying at least one sequence of at least one instruction to processor 604 for execution. For example, the instructions may initially be carried on a magnetic disk or solid-state drive of a remote computer. The remote computer can load the instructions into its dynamic memory and send the instructions over a communication link such as a fiber optic or coaxial cable or telephone line using a modem. A modem or router local to computer system 600 can receive the data on the communication link and convert the data to a format that can be read by computer system 600. For instance, a receiver such as a radio frequency antenna or an infrared detector can receive the data carried in a wireless or optical signal and appropriate circuitry can provide the data to I/O subsystem 602 such as place the data on a bus. I/O subsystem 602 carries the data to memory 606, from which processor 604 retrieves and executes the instructions. The instructions received by memory 606 may optionally be stored on storage 610 either before or after execution by processor 604.
  • Computer system 600 also includes a communication interface 618 coupled to bus 602. Communication interface 618 provides a two-way data communication coupling to network link(s) 620 that are directly or indirectly connected to at least one communication networks, such as a network 622 or a public or private cloud on the Internet. For example, communication interface 618 may be an Ethernet networking interface, integrated-services digital network (ISDN) card, cable modem, satellite modem, or a modem to provide a data communication connection to a corresponding type of communications line, for example an Ethernet cable or a metal cable of any kind or a fiber-optic line or a telephone line. Network 622 broadly represents a local area network (LAN), wide-area network (WAN), campus network, internetwork or any combination thereof. Communication interface 618 may comprise a LAN card to provide a data communication connection to a compatible LAN, or a cellular radiotelephone interface that is wired to send or receive cellular data according to cellular radiotelephone wireless networking standards, or a satellite radio interface that is wired to send or receive digital data according to satellite wireless networking standards. In any such implementation, communication interface 618 sends and receives electrical, electromagnetic or optical signals over signal paths that carry digital data streams representing various types of information.
  • Network link 620 typically provides electrical, electromagnetic, or optical data communication directly or through at least one network to other data devices, using, for example, satellite, cellular, Wi-Fi, or BLUETOOTH technology. For example, network link 620 may provide a connection through a network 622 to a host computer 624.
  • Furthermore, network link 620 may provide a connection through network 622 or to other computing devices via internetworking devices and/or computers that are operated by an Internet Service Provider (ISP) 626. ISP 626 provides data communication services through a world-wide packet data communication network represented as internet 628. A server computer 630 may be coupled to internet 628. Server 630 broadly represents any computer, data center, virtual machine or virtual computing instance with or without a hypervisor, or computer executing a containerized program system such as DOCKER or KUBERNETES. Server 630 may represent an electronic digital service that is implemented using more than one computer or instance and that is accessed and used by transmitting web services requests, uniform resource locator (URL) strings with parameters in HTTP payloads, API calls, app services calls, or other service calls. Computer system 600 and server 630 may form elements of a distributed computing system that includes other computers, a processing cluster, server farm or other organization of computers that cooperate to perform tasks or execute applications or services. Server 630 may comprise one or more sets of instructions that are organized as modules, methods, objects, functions, routines, or calls. The instructions may be organized as one or more computer programs, operating system services, or application programs including mobile apps. The instructions may comprise an operating system and/or system software; one or more libraries to support multimedia, programming or other functions; data protocol instructions or stacks to implement TCP/IP, HTTP or other communication protocols; file format processing instructions to parse or render files coded using HTML, XML, JPEG, MPEG or PNG; user interface instructions to render or interpret commands for a graphical user interface (GUI), command-line interface or text user interface; application software such as an office suite, internet access applications, design and manufacturing applications, graphics applications, audio applications, software engineering applications, educational applications, games or miscellaneous applications. Server 630 may comprise a web application server that hosts a presentation layer, application layer and data storage layer such as a relational database system using structured query language (SQL) or no SQL, an object store, a graph database, a flat file system or other data storage.
  • Computer system 600 can send messages and receive data and instructions, including program code, through the network(s), network link 620 and communication interface 618. In the Internet example, a server 630 might transmit a requested code for an application program through Internet 628, ISP 626, local network 622 and communication interface 618. The received code may be executed by processor 604 as it is received, and/or stored in storage 610, or other non-volatile storage for later execution.
  • The execution of instructions as described in this section may implement a process in the form of an instance of a computer program that is being executed, and consisting of program code and its current activity. Depending on the operating system (OS), a process may be made up of multiple threads of execution that execute instructions concurrently. In this context, a computer program is a passive collection of instructions, while a process may be the actual execution of those instructions. Several processes may be associated with the same program; for example, opening up several instances of the same program often means more than one process is being executed. Multitasking may be implemented to allow multiple processes to share processor 604. While each processor 604 or core of the processor executes a single task at a time, computer system 600 may be programmed to implement multitasking to allow each processor to switch between tasks that are being executed without having to wait for each task to finish. In an embodiment, switches may be performed when tasks perform input/output operations, when a task indicates that it can be switched, or on hardware interrupts. Time-sharing may be implemented to allow fast response for interactive user applications by rapidly performing context switches to provide the appearance of concurrent execution of multiple processes simultaneously. In an embodiment, for security and reliability, an operating system may prevent direct communication between independent processes, providing strictly mediated and controlled inter-process communication functionality.
  • In the foregoing specification, embodiments of the invention have been described with reference to numerous specific details that may vary from implementation to implementation. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense. The sole and exclusive indicator of the scope of the invention, and what is intended by the applicants to be the scope of the invention, is the literal and equivalent scope of the set of claims that issue from this application, in the specific form in which such claims issue, including any subsequent correction.

Claims (18)

What is claimed is:
1. A computer-implemented method of forecasting supply chain demand of products of goods or services, executed by a computing system associated with a supply chain network, the method comprising:
retrieving a training data set comprising product demand data that has been acquired by data acquisition logic of the computing system, the product demand data indicating consumer demand for millions of products at a plurality of time points;
clustering the training data set into a plurality of time series clusters;
calculating one or more break points in one or more of the time series clusters of the training data set, each of the break points corresponding to a disruptive event, by executing a supervised multi class machine learning classifier;
retrieving a baseline forecast for one or more of the products and calculating one or more of mean demand level, median demand level, or standard deviation of demand level for selected periods of the training data set that are before and after one or more of the break points;
identifying a deviation between the baseline forecast and a particular time series cluster among the plurality of the time series clusters, the deviation exceeding 1.5*Inter Quartile Range (IQR), above and below the 75th and 25th percentiles of the baseline forecast and, in response thereto, flagging the particular time series cluster as impacted by a disruptive event;
creating one or more machine learning models to predict future product demand data based on the training data set;
processing configuration information that specifies third-party data for training the one or more machine learning models and, in response thereto, accessing one or more of mobility tracking data specifying a percent change in visits to stores within a geographic area, a social distance index, school closures data, case count data, unemployment claims data, consumer sentiment data, hospital utilization data, as additional data source for which the one or more machine learning models may be trained;
programmatically activating a training function of the one or more machine learning models with a reference to or identification of the training dataset and the third-party data.
2. The computer-implemented method of claim 1, at least a portion of the product demand data comprising upstream consumption data that corresponds to one or more other computing systems that are upstream from the computing system associated with the supply chain network; the method further comprising updating the product demand data for the one or more products by imputing sales values based on the upstream consumption data.
3. The computer-implemented method of claim 1, further comprising clustering the training data set into a plurality of time series clusters based on moving average convergence/divergence (MACD) indicators that are associated with the product demand data.
4. A computer-implemented method of forecasting supply chain demand of products of goods or services, executed by a computing system associated with a supply chain network, the method comprising:
retrieving product demand data that has been acquired by data acquisition logic of the computing system, the product demand data indicating consumer demand for one or more products at a plurality of time points;
determining that at least a portion of the product demand data comprises upstream consumption data that corresponds to one or more other computing systems that are upstream from the computing system associated with the supply chain network;
updating the product demand data for the one or more products by imputing sales values based on the upstream consumption data;
determining a plurality of changepoints for the one or more products represented in the updated product demand data based on evaluation of the product demand data using a trained machine learning model, each changepoint corresponding to a time point of the plurality of time points;
clustering the plurality of changepoints into groups and identifying demand patterns associated with the one or more products;
determining, based on the demand patterns, a baseline model of expected consumer demand for each of the one or more products;
determining, for a particular product of the one or more products, a probability that the particular product will experience a disruptive event based on a deviation from the baseline model of the expected consumer demand.
5. The computer-implemented method of claim 4, the product demand data further comprising downstream consumption data obtained from Point of Sale (POS) computer systems.
6. The computer-implemented method of claim 4, further comprising clustering the plurality of changepoints into the groups based on moving average convergence/divergence (MACD) indicators that are associated with the updated product demand data.
7. The computer-implemented method of claim 4, further comprising updating the product demand data for the one or more products by transforming the product demand data by one or more of: formatting the product demand data, deduplicating the product demand data, or correcting errors associated with the product demand data.
8. The computer-implemented method of claim 4, further comprising determining the baseline model of the expected consumer demand for each of the one or more products based on one or more of: GPS location tracking, social-related data, school closures data, unemployment claims data, consumer sentiment data, hospital utilization data, school closure data, unemployment data, or consumer sentiment data.
9. A computer system comprising:
one or more processors; and
one or more computer-readable non-transitory storage media in communication with the one or more processors, the one or more computer-readable non-transitory storage media storing one or more sequences of instructions that when executed by the one or more processors, cause the one or more processors to:
retrieve product demand data that has been acquired by data acquisition logic of the computing system, the product demand data indicating consumer demand for one or more products at a plurality of time points;
determine that at least a portion of the product demand data comprises upstream consumption data that corresponds to one or more other computing systems that are upstream from the computing system associated with the supply chain network;
update the product demand data for the one or more products by imputing sales values based on the upstream consumption data;
determine a plurality of changepoints for the one or more products represented in the updated product demand data based on evaluation of the product demand data using a trained machine learning model, each changepoint corresponding to a time point of the plurality of time points;
cluster the plurality of changepoints into groups and identifying demand patterns associated with the one or more products;
determine, based on the demand patterns, a baseline model of expected consumer demand for each of the one or more products;
determine, for a particular product of the one or more products, a probability that the particular product will experience a disruptive event based on a deviation from the baseline model of the expected consumer demand.
10. The system of claim 9, the product demand data further comprising downstream consumption data obtained from Point of Sale (POS) computer systems.
11. The system of claim 9, the one or more sequences of instructions when executed by the one or more processors further cause the one or more processors to:
cluster the plurality of changepoints into the groups based on moving average convergence/divergence (MACD) indicators that are associated with the updated product demand data.
12. The system of claim 9, the one or more sequences of instructions when executed by the one or more processors further cause the one or more processors to:
update the product demand data for the one or more products by transforming the product demand data by one or more of: formatting the product demand data, deduplicating the product demand data, or correcting errors associated with the product demand data.
13. The system of claim 9, the one or more sequences of instructions when executed by the one or more processors further cause the one or more processors to:
determine the baseline model of the expected consumer demand for each of the one or more products based on one or more of: GPS location tracking, social-related data, school closures data, unemployment claims data, consumer sentiment data, hospital utilization data, school closure data, unemployment data, or consumer sentiment data.
14. One or more computer-readable non-transitory storage media including instructions that, when executed by one or more processors, are configured to cause the one or more processors to:
retrieve product demand data that has been acquired by data acquisition logic of the computing system, the product demand data indicating consumer demand for one or more products at a plurality of time points;
determine that at least a portion of the product demand data comprises upstream consumption data that corresponds to one or more other computing systems that are upstream from the computing system associated with the supply chain network;
update the product demand data for the one or more products by imputing sales values based on the upstream consumption data;
determine a plurality of changepoints for the one or more products represented in the updated product demand data based on evaluation of the product demand data using a trained machine learning model, each changepoint corresponding to a time point of the plurality of time points;
cluster the plurality of changepoints into groups and identifying demand patterns associated with the one or more products;
determine, based on the demand patterns, a baseline model of expected consumer demand for each of the one or more products;
determine, for a particular product of the one or more products, a probability that the particular product will experience a disruptive event based on a deviation from the baseline model of the expected consumer demand.
15. The one or more computer-readable non-transitory storage media of claim 15, the product demand data further comprising downstream consumption data obtained from Point of Sale (POS) computer systems.
16. The one or more computer-readable non-transitory storage media of claim 15, the instructions being configured to further cause the one or more processors to:
cluster the plurality of changepoints into the groups based on moving average convergence/divergence (MACD) indicators that are associated with the updated product demand data.
17. The one or more computer-readable non-transitory storage media of claim 15, the instructions being configured to further cause the one or more processors to:
update the product demand data for the one or more products by transforming the product demand data by one or more of: formatting the product demand data, deduplicating the product demand data, or correcting errors associated with the product demand data.
18. The one or more computer-readable non-transitory storage media of claim 15, the instructions being configured to further cause the one or more processors to:
determine the baseline model of the expected consumer demand for each of the one or more products based on one or more of: GPS location tracking, social-related data, school closures data, unemployment claims data, consumer sentiment data, hospital utilization data, school closure data, unemployment data, or consumer sentiment data.
US17/708,985 2021-03-31 2022-03-30 Automated supply chain demand forecasting Pending US20220318711A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/708,985 US20220318711A1 (en) 2021-03-31 2022-03-30 Automated supply chain demand forecasting

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202163169017P 2021-03-31 2021-03-31
US17/708,985 US20220318711A1 (en) 2021-03-31 2022-03-30 Automated supply chain demand forecasting

Publications (1)

Publication Number Publication Date
US20220318711A1 true US20220318711A1 (en) 2022-10-06

Family

ID=83449468

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/708,985 Pending US20220318711A1 (en) 2021-03-31 2022-03-30 Automated supply chain demand forecasting

Country Status (1)

Country Link
US (1) US20220318711A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230410134A1 (en) * 2019-10-11 2023-12-21 Kinaxis Inc. Systems and methods for dynamic demand sensing and forecast adjustment
US20240144139A1 (en) * 2019-11-05 2024-05-02 Strong Force Vcn Portfolio 2019, Llc Systems, methods, kits, and apparatuses for automated intelligent procurement in value chain networks

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20230410134A1 (en) * 2019-10-11 2023-12-21 Kinaxis Inc. Systems and methods for dynamic demand sensing and forecast adjustment
US20240144139A1 (en) * 2019-11-05 2024-05-02 Strong Force Vcn Portfolio 2019, Llc Systems, methods, kits, and apparatuses for automated intelligent procurement in value chain networks

Similar Documents

Publication Publication Date Title
Fong et al. Finding an accurate early forecasting model from small dataset: A case of 2019-ncov novel coronavirus outbreak
US11126635B2 (en) Systems and methods for data processing and enterprise AI applications
US11138376B2 (en) Techniques for information ranking and retrieval
US11599752B2 (en) Distributed and redundant machine learning quality management
US11531943B1 (en) Intelligence driven method and system for multi-factor optimization of schedules and resource recommendations for smart construction
US20240046001A1 (en) Automated standardized location digital twin and location digital twin method factoring in dynamic data at different construction levels, and system thereof
US11693917B2 (en) Computational model optimizations
US20070156479A1 (en) Multivariate statistical forecasting system, method and software
US20220318711A1 (en) Automated supply chain demand forecasting
US10578730B2 (en) Method, apparatus and system for location detection and object aggregation
US11875408B2 (en) Techniques for accurate evaluation of a financial portfolio
JP2023514465A (en) machine learning platform
US11803793B2 (en) Automated data forecasting using machine learning
TWI821019B (en) Intelligence driven method and system for multi-factor optimization of schedules and resource recommendations for smart construction
US20220027400A1 (en) Techniques for information ranking and retrieval
US20240168860A1 (en) Apparatus and method for computer-implemented modeling of multievent processes
US20240070899A1 (en) Systems and methods for managing assets
US20230083762A1 (en) Adversarial bandit control learning framework for system and process optimization, segmentation, diagnostics and anomaly tracking
US11971806B2 (en) System and method for dynamic monitoring of changes in coding data
Kettunen Anomaly detection in business metric monitoring
Omolo Using an Ensemble of Machine Learning Algorithms to Predict Economic Recession
Malibari Developing “Smartness” in Emerging Environments and Applications using Artificial Intelligence
Gauthier Melançon Quantifying Uncertainty in Systems-Two Practical Use Cases Using Machine Learning to Predict and Explain Systems Failures
Melançon Quantifying Uncertainty in Systems-Two Practical Use Cases Using Machine Learning to Predict and Explain Systems Failures
Rodríguez Martínez Design of a machine learning application for anomaly detection in real-time series

Legal Events

Date Code Title Description
AS Assignment

Owner name: COUPA SOFTWARE INCORPORATED, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RECASENS, JAVIER;RAGAVAN, PRASANNA KUMAR;REEL/FRAME:059448/0250

Effective date: 20210331

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: SSLP LENDING, LLC, TEXAS

Free format text: SECURITY INTEREST;ASSIGNORS:COUPA SOFTWARE INCORPORATED;YAPTA, INC.;REEL/FRAME:062887/0181

Effective date: 20230228

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED