WO2000055790A2  Gradient criterion method for neural networks and application to targeted marketing  Google Patents
Gradient criterion method for neural networks and application to targeted marketingInfo
 Publication number
 WO2000055790A2 WO2000055790A2 PCT/US2000/006735 US0006735W WO2000055790A2 WO 2000055790 A2 WO2000055790 A2 WO 2000055790A2 US 0006735 W US0006735 W US 0006735W WO 2000055790 A2 WO2000055790 A2 WO 2000055790A2
 Authority
 WO
 Grant status
 Application
 Patent type
 Prior art keywords
 model
 neural
 invention
 training
 method
 Prior art date
Links
Classifications

 G—PHYSICS
 G06—COMPUTING; CALCULATING; COUNTING
 G06Q—DATA PROCESSING SYSTEMS OR METHODS, SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL, SUPERVISORY OR FORECASTING PURPOSES, NOT OTHERWISE PROVIDED FOR
 G06Q30/00—Commerce, e.g. shopping or ecommerce
 G06Q30/02—Marketing, e.g. market research and analysis, surveying, promotions, advertising, buyer profiling, customer management or rewards; Price estimation or determination

 G—PHYSICS
 G06—COMPUTING; CALCULATING; COUNTING
 G06N—COMPUTER SYSTEMS BASED ON SPECIFIC COMPUTATIONAL MODELS
 G06N3/00—Computer systems based on biological models
 G06N3/02—Computer systems based on biological models using neural network models
 G06N3/08—Learning methods
Abstract
Description
TITLE OF THE INVENTION: Gradient Cπteπon Method for Neural Networks and Application to Targeted Marketing FIELD OF THE INVENTION: This invention relates generally to the development of neural network models to optimize the effects of targeted marketing programs. More specifically, this invention is an improvement on the Maximum Likelihood method of training neural networks using a gradient criterion, and is specially designed for binary output havmg strongly uneven proportion, which is typical for direct marketing problems.
BACKGROUND OF THE INVENTION: The goal of most modeling procedures is to minimize the discrepancy between real results and model outputs. If the discrepancy, or error, can be accumulated on a record by record basis, it is suitable for gradient algorithms like Maximum Likelihood.
The goal of target marketmg modelmg is typically to find a method to calculate the probability of any prospect in the list to respond to an offer. The neural network model is built based on the experimental data (test mailing), and the traditional approach to this problem is to choose a model and compute model parameters with a model fitting procedure The topology of model — for example, number of nodes, input and transfer functions — defines the formula that expresses the probability of response as a function of attributes In a special model fitting procedure, the output of the model is tested against actual output (from the results of a test mailing) and discrepancy is accumulated in a special error function. Different types of error functions can be used (e.g.. mean square, absolute error); model parameters are determined to minimize the error function. The best fitting of model parameters is an implicit indication that the model is good (not necessarily the best) in terms of its original objective. Thus the model building process is defined by two entities: the type of model and the error (or utility) function. The type of model defines the ability of the model to discern various patterns in the data. For example, increasing the number of nodes results in more complicated formulae, so a model can more accurately discern complicated patterns. The "goodness" of the model is ultimately defined by the choice of an error function, since it is the error function that is minimized during the model training process. To reach the goal of modeling, one wants to use a utility function that assigns probabilities that are most in compliance with the results of the experiment (the test mailing). The Maximum Likelihood criterion is the explicit measure of this compliance. However, the modeling process as it exists today has a significant drawback: it uses conventional utility functions (least mean square, cross entropy) only because there is a mathematical apparatus developed for these utility functions. What would really be useful is a process that builds a response model that directly maximizes Maximum Likelihood. For example, a random variable X exists with the distribution p(X, A), where A is an unknown vector of parameters to be estimated based on the independent observations of X: (xi, x_{2}, ... , XN) The goal is to find such a vector A that makes a probability of the output p(xι,A)*p( x_{2},A)* ... *p( X_{N},A) maximally possible. Note that the function p(X, A) should be a known function of two variables. The Maximum Likelihood technique provides the mathematical apparatus to solve this optimization problem. In general, the Maximum Likelihood method can be applied to neural networks as follows. Let the neural network calculate a value of the output variable y based on the input vector X. The observed values (y ι , y , • .. , YN) represent the actual output with some error e. Assuming that this error has, for example, a normal distribution, the method can find weights W of the neural network that makes a probability of the output p(yι,W)*p( y_{2}, W)* ... *p( VN,W) maximally possible. In the case of a normal probability function, the Maximum Likelihood criterion is equivalent to the Least Mean Square criterionwhich is, in fact, most widely used for neural network training. In the case of target marketing, the observed output X is a binary variable that is equal to 1 if a customer responded to the offer, and is 0 otherwise. The normality assumption is too rough, and leads to a suboptimal set of neural network weights if used in neural network training. This is a typical direct marketing scenario.
SUMMARY OF THE INVENTION: The present invention represents a unique application of the Maximum Likelihood statistical method to commercial neural network technologies. The present invention utilizes the specific nature of the output in target marketing problems and makes it possible to produce more accurate and predictive results. It is best used on "noisy" data and when one is interested in determining a distribution's overall accuracy, or best general description of reality. The present invention provides a competitive advantage over offtheshelf modeling packages in that it greatly enhances the application of Maximum Likelihood to quantitative marketing applications such as customer acquisition, crossselling/up selling, predictive customer profitability modeling, and channel optimization. Specifically, the superior predictive modeling capability provided by using the present invention means that marketing analysts will be better able to: • Predict the propensity of individual prospects to respond to an offer, thus enabling marketers to better identify target markets. • Identify customers and prospects who are most likely to default on loans, so that remedial action can be taken, or so that those prospects can be excluded from certain offers. • Identify customers or prospects who are most likely to prepay loans, so a better estimate can be made of revenues. • Identify customers who are most amenable to crosssell and upsell opportunities. • Predict claims experience, so that insurers can better establish risk and set premiums appropriately. • Identify instances of creditcard fraud.
BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 shows the dataflow of the method of training the model of the present invention.
Figure 2 illustrates a preferred system architecture for employing the present invention.
DETAILED DESCRIPTION OF THE INVENTION The present invention uses the neural network to calculate a propensity score g(X, W), where Wϊs a set of weights of the neural network, Nis a vector of customer attributes (input vector). The probability to respond to an offer for a customer with attributes X can be calculated by a formula: _{P =} s(x,w) l + g(X,W)
If there are N independent samples and among them n are responders, the probability of such output is:
γ
Using the logarithm of L as a training criterion (training error) in the form of:
Err = ln£ = ∑ln(l + .g )  ∑ln(g,)  ∑ln(l g,) teresp lenonresp The neural network training procedure finds the optimal weights PFthat minimize Err and thus maximize likelihood of the observed output L. One can use back propagation or a similar method to perform training. The gradient criterion that is required by a training procedure is computed as follows:
£ ^{■•} = (£& ∑l /g, + ∑ T^^{~} )g
1=1 A ^{"} " g_{t} teresp iznonresp 1 & _{l} In order for the training procedure be robust and stable the output of the neural network should be in the middle of the working interval [0, 1]. To ensure that, the present invention introduces the normalized propensity score/which is related to g as:
g(X,W) = f^{h} (X,W) Now, let/ be the output of the neural network and choose the parameter x in such a way that/may be of the order of 0.5. Let R be an average response rate in the sample. The above condition is satisfied if:
τ = l/ln
R While training the model, the criterion is optimized so the calculation is based on the output of the neural network using the formula:
Err = \nP = ∑^{]}n(l + f '^{τ} )   ∑Hf)  ∑hι(l  /;^{1/τ} ) =1 ^ it≡resp tenon _resp The gradient criterion is computed as follows: ι N fllτ l , N  1/τ l
^ 1=1 1 "r^{"} J i * teresp tenon _resp 1 J _{t} The method was tested on a variety of business cases against both Least Mean Square and CrossEntropy criteria. In all cases the method gave 20%  50% improvement in the lift on top 20% of the target marketing sample customer pools. As shown in figure 1, the method inputs data from modeling database 11 into a selected model 12 to calculate scores 13. The error 14 is calculated from comparison with the known responses from modeling database 11 and checked for convergence 15 below a desired level. When convergence occurs, a new model 16 is the result to be used for targeted marketing 17. Otherwise, the process minimizes the error and solves for a new set of weights at 18 and begins a new iteration.
The present invention operates on a computer system and is used for targeted marketing purposes. In a preferred embodiment as shown in figure 2, the system runs on a threetier architecture that supports CORBA as an intercommunications protocol. The desktop client software on targeted marketing workstations 20 supports JAVA. The central application server 22 and multithreaded calculation engines 24, 25 run on Windows NT or UNIX. Modeling database 26 is used for training new models to be applied for targeted marketing related to customer database 28. The recommended minimum system requirements for application server 22 and multithreaded calculation engines 24, 25 are as follows:
*Approximately 100 MB/1 million records in customer database. The above assumes the user client is installed on a PC with the recommended configuration found below.
The recommended minimum requirements for the targeted marketing workstations 20 are as follows:
Using the present invention in conjunction with a neural network, the present invention provides a user with data indicating the individuals or classes of individuals who are most likely to respond to direct marketing.
Claims
Priority Applications (2)
Application Number  Priority Date  Filing Date  Title 

US12421799 true  19990315  19990315  
US60/124,217  19990315 
Applications Claiming Priority (1)
Application Number  Priority Date  Filing Date  Title 

CA 2403249 CA2403249A1 (en)  19990315  20000315  Gradient criterion method for neural networks and application to targeted marketing 
Publications (3)
Publication Number  Publication Date 

WO2000055790A2 true true WO2000055790A2 (en)  20000921 
WO2000055790A3 true WO2000055790A3 (en)  20001214 
WO2000055790B1 true WO2000055790B1 (en)  20010222 
Family
ID=22413524
Family Applications (1)
Application Number  Title  Priority Date  Filing Date 

PCT/US2000/006735 WO2000055790B1 (en)  19990315  20000315  Gradient criterion method for neural networks and application to targeted marketing 
Country Status (2)
Country  Link 

CA (1)  CA2403249A1 (en) 
WO (1)  WO2000055790B1 (en) 
Cited By (7)
Publication number  Priority date  Publication date  Assignee  Title 

WO2002037302A1 (en) *  20001031  20020510  Granlund Goesta  Training of associative networks 
US6993493B1 (en) *  19990806  20060131  Marketswitch Corporation  Method for optimizing net present value of a crossselling marketing campaign 
US8027871B2 (en)  20061103  20110927  Experian Marketing Solutions, Inc.  Systems and methods for scoring sales leads 
US8732004B1 (en)  20040922  20140520  Experian Information Solutions, Inc.  Automated analysis of data to generate prospect notifications based on trigger events 
US9058340B1 (en)  20071119  20150616  Experian Marketing Solutions, Inc.  Service for associating network users with profiles 
US9767309B1 (en)  20151123  20170919  Experian Information Solutions, Inc.  Access control system for implementing access restrictions of regulated database records while identifying and providing indicators of regulated database records matching validation criteria 
US9916596B1 (en)  20070131  20180313  Experian Information Solutions, Inc.  Systems and methods for providing a direct marketing campaign planning environment 
Families Citing this family (5)
Publication number  Priority date  Publication date  Assignee  Title 

US8036979B1 (en)  20061005  20111011  Experian Information Solutions, Inc.  System and method for generating a finance attribute from tradeline data 
WO2008147918A3 (en)  20070525  20090122  Experian Information Solutions  System and method for automated detection of neverpay data sets 
WO2010132492A3 (en)  20090511  20140320  Experian Marketing Solutions, Inc.  Systems and methods for providing anonymized user profile data 
US9152727B1 (en)  20100823  20151006  Experian Marketing Solutions, Inc.  Systems and methods for processing consumer information for targeted marketing applications 
US9576030B1 (en)  20140507  20170221  Consumerinfo.Com, Inc.  Keeping up with the joneses 
Citations (3)
Publication number  Priority date  Publication date  Assignee  Title 

EP0554083A2 (en) *  19920130  19930804  Ricoh Company, Ltd  Neural network learning system 
US5504675A (en) *  19941222  19960402  International Business Machines Corporation  Method and apparatus for automatic selection and presentation of sales promotion programs 
US5774868A (en) *  19941223  19980630  International Business And Machines Corporation  Automatic sales promotion selection system and method 
Patent Citations (3)
Publication number  Priority date  Publication date  Assignee  Title 

EP0554083A2 (en) *  19920130  19930804  Ricoh Company, Ltd  Neural network learning system 
US5504675A (en) *  19941222  19960402  International Business Machines Corporation  Method and apparatus for automatic selection and presentation of sales promotion programs 
US5774868A (en) *  19941223  19980630  International Business And Machines Corporation  Automatic sales promotion selection system and method 
NonPatent Citations (1)
Title 

SURENDRAN A C ET AL: "Unsupervised, smooth training of feedforward neural networks for mismatch compensation" 1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING PROCEEDINGS (CAT. NO.97TH8241), 1997 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING PROCEEDINGS, SANTA BARBARA, CA, USA, 1417 DEC. 1997, pages 482489, XP002148180 1997, New York, NY, USA, IEEE, USA ISBN: 0780336984 * 
Cited By (10)
Publication number  Priority date  Publication date  Assignee  Title 

US6993493B1 (en) *  19990806  20060131  Marketswitch Corporation  Method for optimizing net present value of a crossselling marketing campaign 
US7499868B2 (en)  19990806  20090303  Marketswitch Corporation  Method for optimizing net present value of a crossselling marketing campaign 
US8015045B2 (en)  19990806  20110906  Experian Information Solutions, Inc.  Method for optimizing net present value of a crossselling marketing campaign 
US8285577B1 (en)  19990806  20121009  Experian Information Solutions, Inc.  Method for optimizing net present value of a crossselling marketing campaign 
WO2002037302A1 (en) *  20001031  20020510  Granlund Goesta  Training of associative networks 
US8732004B1 (en)  20040922  20140520  Experian Information Solutions, Inc.  Automated analysis of data to generate prospect notifications based on trigger events 
US8027871B2 (en)  20061103  20110927  Experian Marketing Solutions, Inc.  Systems and methods for scoring sales leads 
US9916596B1 (en)  20070131  20180313  Experian Information Solutions, Inc.  Systems and methods for providing a direct marketing campaign planning environment 
US9058340B1 (en)  20071119  20150616  Experian Marketing Solutions, Inc.  Service for associating network users with profiles 
US9767309B1 (en)  20151123  20170919  Experian Information Solutions, Inc.  Access control system for implementing access restrictions of regulated database records while identifying and providing indicators of regulated database records matching validation criteria 
Also Published As
Publication number  Publication date  Type 

WO2000055790A3 (en)  20001214  application 
CA2403249A1 (en)  20000921  application 
WO2000055790B1 (en)  20010222  application 
Similar Documents
Publication  Publication Date  Title 

Baumol et al.  Rules of thumb and optimally imperfect decisions  
Plutowski et al.  Selecting concise training sets from clean data  
Palmquist  Property value models  
Thompson et al.  Prevalence odds ratio or prevalence ratio in the analysis of cross sectional data: what is to be done?  
Reisinger et al.  Structural equation modeling: Critical issues and new developments  
Teachman  Methodological issues in the analysis of family formation and dissolution  
Silk et al.  Pretestmarket evaluation of new packaged goods: A model and measurement methodology  
Ward et al.  A diffusion approximation for a Markovian queue with reneging  
Brown et al.  Family boundary ambiguity and the measurement of family structure: The significance of cohabitation  
US6321179B1 (en)  System and method for using noisy collaborative filtering to rank and present items  
Daraio et al.  Introducing environmental variables in nonparametric frontier models: a probabilistic approach  
US7158959B1 (en)  Automated webbased targeted advertising with quotas  
US7809601B2 (en)  Intelligent performancebased product recommendation system  
US6763354B2 (en)  Mining emergent weighted association rules utilizing backlinking reinforcement analysis  
US7143075B1 (en)  Automated webbased targeted advertising with quotas  
Rieskamp  The probabilistic nature of preferential choice.  
Kenny et al.  A general procedure for the estimation of interdependence.  
US7437308B2 (en)  Methods for estimating the seasonality of groups of similar items of commerce data sets based on historical sales date values and associated error information  
Reich et al.  Evaluating machine learning models for engineering problems  
Pascual et al.  From individuals to population densities: searching for the intermediate scale of nontrivial determinism  
Ghosh et al.  Predicting influential users in online social networks  
Yamaguchi  Tasks and heterogeneous human capital  
Chan et al.  Bayesian poisson regression for crowd counting  
US20050278139A1 (en)  Automatic match tuning  
US20080208652A1 (en)  Method and system utilizing online analytical processing (olap) for making predictions about business locations 
Legal Events
Date  Code  Title  Description 

AK  Designated states 
Kind code of ref document: A2 Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW 

AL  Designated countries for regional patents 
Kind code of ref document: A2 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG 

121  Ep: the epo has been informed by wipo that ep was designated in this application  
AK  Designated states 
Kind code of ref document: A3 Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW 

AL  Designated countries for regional patents 
Kind code of ref document: A3 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG 

DFPE  Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)  
AK  Designated states 
Kind code of ref document: B1 Designated state(s): AE AL AM AT AU AZ BA BB BG BR BY CA CH CN CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW 

AL  Designated countries for regional patents 
Kind code of ref document: B1 Designated state(s): GH GM KE LS MW SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG 

B  Later publication of amended claims  
REG  Reference to national code 
Ref country code: DE Ref legal event code: 8642 

WWE  Wipo information: entry into national phase 
Ref document number: 2403249 Country of ref document: CA 

122  Ep: pct application nonentry in european phase 