CA2389222A1 - Method for modeling market response rates - Google Patents

Method for modeling market response rates Download PDF

Info

Publication number
CA2389222A1
CA2389222A1 CA002389222A CA2389222A CA2389222A1 CA 2389222 A1 CA2389222 A1 CA 2389222A1 CA 002389222 A CA002389222 A CA 002389222A CA 2389222 A CA2389222 A CA 2389222A CA 2389222 A1 CA2389222 A1 CA 2389222A1
Authority
CA
Canada
Prior art keywords
prospects
selecting
list
group
classifier
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002389222A
Other languages
French (fr)
Inventor
Yu-To Chen
Piero Patrone Bonissone
Margaret Stewart Trench
Jeremiah Francis Donoghue
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
GE Financial Assurance Holdings Inc
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CA2389222A1 publication Critical patent/CA2389222A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising

Abstract

Published without an

Description

METHOD FOR MODELING MARKET RESPONSE RATES
FIELD OF THE INVENTION
The present invention generally relates to direct marketing, and more specifically to modeling market response rates to direct solicitations.
BACKGROUND OF THE INVENTION
Direct marketing usually involves directly contacting persons or entities, such as for example by mail, with a specific message or solicitation. The persons to be contacted are usually identified by a mailing list. Today's direct marketer, however, faces a variety of problems, such as for example rising postal and printing costs, which effect the cost of doing business. The success of a direct mailing is dependent on the number of responses created by the direct mailing (i.e., the response rate). As a result, blindly mailing a direct mail piece to everyone on a mailing list (e.g., mass mailings) can be costly and inefficient because the response rate will in all likelihood be low.
Cost-conscious direct marketers use their knowledge about persons identified on a mailing list (i.e., a prospect) to determine the best prospects to mail to.
Usually, a marketer will use a set of descriptor variables about each prospect, such as for example demographics and credit card ownership, to target good prospects (i.e., prospects which will find the mailing interesting). For example, the Rao and Steckel model includes acquiring a set of descriptor variables and conducting a knowledge engineering session to screen the variables. In this regard, a marketing committee may be appointed and prior experience and intuition may be used to pick out the demographic variables most relevant to the response rate. After that, the probability of a response is modeled as a beta-logistic distribution and its parameters are estimated by maximum likelihood and a response score, R(i) for each prospect i, and a profit score, P(i) for each prospect i, are generated. Next, each prospect is assigned a value of R(i) x P(i), and each prospect is ranked high to low based on the assigned value of R(i) x P(i). The ranking of prospects is intended to account for both responsiveness and profitability of the direct mailing proposal.

There are at least two drawbacks to the Rao and Steckel model. First, the variable screening process based on opinion is subjective and error-prone. Second, using a simple distribution to describe the response probability in a high-dimensional (e.g., hundreds of attributes to a prospect), noisy environment (i.e., incomplete or missing data) is inadequate. In this regard, the drawback of using a simple distribution to describe the response probability is that a simple distribution assumes that the behaviors of various people will follow more or less a magic distribution and is governed by pure randomness.
This assumption ignores the fact that there may be some reasons behind a person's response to a solicitation.
BRIEF SUMMARY OF THE INVENTION
Thus there is a particular need for a consistent and sustainable process for building response models which predict the likelihood of a prospect responding to a marketing solicitation. The present invention is a method for modeling market response rates. The method is used to evaluate and filter large contact lists with the aim of accomplishing two goals. The tactical goal is to improve a market response rate to cut costs associated with mailing, phone and electronic mail campaigns and produce more leads. The strategic goal is to assess the incremental risk of non-responsiveness associated with the incremental volume derived from growing a market in different directions (i.e., the tradeoffs between growing business (e.g., more solicitation) and risk of loss).
The method of the present invention uses an internal experience database and an external demographic database, coupled with variable screening schemes and non-parametric modeling techniques. The method comprises a number of steps. First, a data acquisition step associates descriptor variables with prospects. Second, a variable selection step identifies the descriptor variables in order to identify prospects most likely to respond to a direct marketing solicitation. Third, a model selection step examines and assesses a number of competitive algorithms, and selects the algorithm that will best predict the response rate. Fourth, a parameter estimation step ensures the best fit of data once an algorithm is chosen. Finally, a validation step ensures the robustness of the modeling process. Robustness means the model will work even though the data in the future is likely to be different from the data used to build the model.
BRIEF DESCRIPTION OF THE DRAWINGS
Fig. 1 is a flow diagram illustrating the steps of one embodiment of a method for modeling market response rates in accordance with the present invention;
Fig. 2 is a block diagram illustrating an example of cross-referencing of mailing records with multiple universal files in the data acquisition step of the method illustrated in Fig. 1;
Fig. 3 is a flow diagram illustrating steps included in the variable selection step of the method illustrated in Fig. 1;
Fig. 4 is a table showing an exemplary subset of Census Tract and Block Group variables selected in the variable selection step of the method illustrated in Fig.
1;
Fig. S is a table showing the results of testing of models constructed using training data;
Fig. 6 is a table showing the performance of ZIPS classifiers;
Fig. 7 is a table showing the performance of Census Tract and Block Group classifiers;
Fig. 8 is a table showing the performance of Donnelley classifiers; and Fig. 9 is a graph summarizing the results shown in Figs. 6-8.
DETAILED DESCRIPTION OF THE INVENTION
The first embodiment of the invention is a method for estimating response rates for a direct marketing campaign. For ease of explanation, this invention will be described with reference to a direct marketing campaign that uses the mail, however, other processes can be used with this invention such as direct marketing with a phone, the Internet (electronic mail), fax machines, etc. A direct marketing campaign through the mail can provide a variety of information to the intended recipients. One possible example is to mail pieces advertising long term health care insurance. This invention should not be limited to advertising long term health care insurance and can be use for a variety of other insurance applications as well other areas that do not relate to insurance. The method of this invention associates demographic variables to prospects and uses non-parametric modeling techniques to predict mailing response rates for the prospects. The method is operable in two modes -- a training mode and a testing mode - for cross-validation purposes. Specifically, the data set is divided into two sets, a training data set that is used to build the model, and a test data set that is used to test the robustness of the model. In the training mode, historical mailing data is analyzed off line and a decision logic (i.e., a model) is formulated to estimate the mailing response rates.
In the testing mode, the decision logic analyzes prospects on the fly and predicts the response rates for prospects.
Referring now to Fig. 1, the method first comprises the step of acquiring data.
This step generally comprises attaching household or area level demographics to a prospect (e.g., a mailing record), randomly sampling the prospects, and splitting the randomly sampled prospects into a training set and a testing set. First, mailing records on a mailing list are cross referenced with a universal file (i.e., the entire data set) so that information regarding demographic variables associated with a mailing record are attached to the mailing record. Preferably, the mailing records are cross referenced with multiple universal files. For example, as shown in Fig. 2, the mailing records can be broken down into groups using universal files available from various vendors, including for example Donnelley, Census Tract and Block Group ("CTBG") and ZIPS. If multiple universal files are used, preferably the mailing records are broken down in to subgroups.
Continuing the example, four groups of data can be created -- Group 1: Matched with Donnelley household demographic key; Group 2: Not matched with Donnelley, but matched with CTBG demographic key; Group 3: Not matched with either Donnelley or CTBG, but matched with ZIPS demographic key; and Group 4: Not matched with Donnelley, CTBG or ZIPS. Preferably, the mailing record is attached with individual, household and area level demographic information which are useful for identifying segments having the strongest relationship to the mailing response rate.
Preferably, for each universe file used for cross referencing, equal number of responders and non-responders are included in the group. In this regard, all of the responders and a sub-sampling (i.e., random drawing of non-responders) of the non-responders are typically included because non-responders are much greater in size than that of responders. Each group is next randomly split into two sets - a training set and a testing set.
For example, the training set may be about 2/3 of the size of the group, and the testing set may be about 1/3 of the group.
Referring again to Fig. l, the method next comprises the step of variably selecting descriptor variables. Generally, it is desirable to use as few variables as possible in the presence of noise. This is often referred to as the "principle of parsi.monious." There may be combinations (linear or nonlinear) of variables that are irrelevant to the underlying process, that due to noise in data appear to increase the prediction accuracy. Preferably, variables with the greater discrimination power in response prediction are selected. Generally, descriptor variables are selected using the misclassification rate as a measure of the discrimination power of each input variable given the same size of tree it constructs. In this regard, there are two types of misclassification, "wasted-mail" and "missed-opportunity." A model takes as input a list of prospects attached with demographic variables (Xs) and known response (Y's) and will produce as output four numbers -- first, the number of known responders being classified as responders (sensitivity of the classifier); second, the number of known responders being classified as non-responders (missed-opportunity); third, the number of known non-responders being classified as responders (wasted-mail); and fourth, the number of known non-responders being classified as non-responders (specificity of the classifier). It is preferred that both missed-opportunity and wasted-mail be minimized.
However, tradeoffs may be necessary. For example, one performance evaluation criterion would be to minimize the misclassification cost, that is to minimize the sum of misclassification rates weighted by cost: objective = minimize (# of wasted-mail x cost per wasted-mail + # of missed-opportunity x cost per missed-opportunity).
This step involves two parts. First, it is preferred that CART, a commercially available statistical algorithm for classification be used for variable selection. Assuming there are N input variables and one output variable, and that there are equal number of responders and non-responders in the training data set, a tree model is constructed for each input variable and the output variable is used to measure how good the variable is, as shown in Fig. 3. Next, the tree is allowed to grow until the size of each terminal node is preferably no smaller than 1/100 of the original data set. Next, the tree is pruned until the number of terminal nodes is preferably around 10, which provides a balance between robustness and accuracy. Next, the misclassification rate of the tree model is computed.
(At this point, there are N tree models. Each tree has about 10 terminal nodes.) Next, each (N) tree model ranked in ascending order of their misclassification rates. Finally, the top 20 trees and their input variables are selected. For example, a subset of CTBG
variables selected by CART are shown in Fig. 4. This step secondly involves selecting variables out of the available Donnelley, CTBG and ZIPS variables. In this regard, each input variable is grouped into two samples: responders and non-responders. The mean difference of the two groups for the particular input variable is next tested.
In addition, the variance difference of the two groups is tested. If both the mean and variance values are significant different, then the input variable is selected. The selection criteria in this case is that both P-values from two-sample T-test and F-test are significant at 0.01 level.
The variables that are common to the two groups of variables are preferably used.
Referring again to Fig. l, the method next comprises the step of selecting a classifier that will best model the mailing response rate. First, available classifiers that are to be considered as possible models are selected. For example, commercially available classifiers, including METROMAIL, Multivariate Adaptive Regression Splines, Logistic Regression, Neural Networks with Back-Propagation, CART and No Data Optimal Classifier (e.g., human intuition), can be selected. Next, the selected classifiers are compared using available universal files, such as for example, the ZIPS
universal file. In this regard, the selected classifiers are constructed using the selected universal file that is split into a training set, to construct the classifier, and a test set, to test the robustness of the constructed classifier. Preferably, multiple universal files are used. For example, the ZIPS universal file is split into a training set and a testing set, and the training set is used to construct a number of different classifiers such as METROMAIL, Multivariate Adaptive Regression ("MARS"), Splines ("C4.5"), Logistic Regression ("LR"), Neural Networks with Back-Propagation ("NN-BP"), CART and No Data Optimal Classifier ("NDOC"). Next, the ZIPS testing set is used to test the constructed classifiers for robustness. In this example, the ZIPS universe file contained 8,407 responders and 454,732 non-responders.
With reference to Fig. 5, the results of the testing of each model constructed using the training data are shown using the test data to validate. This comprises the validation step shown in Fig. 1. In order to make a cost evaluation, the cost per missed-opportunity (i.e., no mailing was made to prospect which would have responded) is estimated to be $17.85, while the cost per wasted-mail is estimated to be $.33 (i.e., the cost of postage). Based on the results of the example, the CART classifier was determined to be the best. The results of the test, and the best classifier will vary according to the classifiers used, the universal file used, and the assumptions made.
Refernng again to Fig. 1, the method further comprises the step of estimating the parameters. In this regard, it is noted that NODAC (no-data optimal classifier) is in essence a "brain-dead" approach. It does not utilize any data/information to reach its conclusion. It only depends on the gut feeling of prior probabilities and estimation of misclassification cost. Theoretically speaking, it assigns any observation to a class j that minimizes ~'n(i)C(j~i), for all i, where n(i) is the prior probability of class i and C(j~i) is the cost of misclassifying class i as class j. Assuming class 1 to be responders, and class 0 to be non-responders, there are two costs-missed-opportunity, C(0~1), and wasted-mail, C(1 ~ 0). The latter is to misclassify class 0 as class 1, i.e., mistaken non-responders as responders. The former is vice versa, i.e., mistaken responders as non-responders. In this example, NODAC has only two choices. It either blindly mails out to all prospects or does not mail at all. The decision is based on the minimum of the two numbers: ~z(1)C(0~1) or ~z(0)C(1 ~0). The latter was the total cost of wasted-mail (mail to all and see how much it costs): ~c(0)C(1 ~ 0) = 0.98 x $0.33 = 0.3240. The former was the total cost of missed-opportunity (do not mail at all): n(1)C(0~ 1) = 0.018 x $17.85 =
0.3240. As can be seen by this example, the two costs were almost the same.
The cost estimates in this case are called the break-even costs for the prior estimates.
The parameters that can be set in tree structured classification include the priors, a~(i), and variable misclassification costs, C(j~i), for class i and j. The priors are more or less fixed -- not much can be done about the 1.56% response rate on the average.
Consequently, instead of bumping up the 1.56% response rate, it is preferred that the classifier's prediction accuracy is improved by using better estimates of misclassification costs.
As discussed above, break-even costs for the ZIPS priors are $17.85 and $0.33 for missed-opportunity and wasted-mail, respectively. In this example, a high confidence on the estimate of wasted-mail is presumed to be within 10%. In contrast, the missed-opportunity depends on how the profit is modeled. Nevertheless, the lower bound of the figure is the break-even cost of missed-opportunity: $17.85 for the ZIPS. In this regard, it is not worth doing any business if a lead's value is lower than that. On the other hand, if the estimate of missed-opportunity increased in the positive direction, we will be tempted to mail out to all prospects due to the fact that the cost of missed-opportunity would be too high. In other words, as the cost of missed-opportunity increases, the NODAC (no-data optimal classifier) will become more and more dominate.
By way of further examples, the following shows the relative performance of METROMAIL, NODAC and CART under three different cost estimates for the three demographics. Fig. 6 illustrates the performance of ZIPS classifiers. Figs. 7 and 8 show the performance of CTBG and Donnelley, respectively.
There are two numbers in each cell of Fig 6. One is the total misclassification cost. The other is the percentage improvement over METROMAIL. If the missed-WO 01/20512 PCT/(JS00/24414 opportunity cost is estimated as $20, then CART is clearly the better classifier. If it is estimated as $60, then both CART and NODAC give similar performance. If the estimate of missed-opportunity is further increased to $100, then CART is acting exactly like NODAC (i.e., mailing out to all prospects). Note that the performance of NODAC
is the same throughout the three different estimates of missed-opportunity because the break-even cost is around $20. Consequently, NODAC just mails out to all prospects as long as the cost of missed opportunity is estimated greater than $20.
A similar trend as ZIPS is observed for CTBG, as shown in Fig. 7. CART is the better classifier if the missed-opportunity cost is around $20. If the missed-opportunity is estimated beyond $60, then both CART and NODAC behave in the same way. The CART's dominance over NODAC decreases as the missed-opportunity cost is increasing from $20 to $60.
From Fig. 8, it is clear that CART is the better classifier throughout various levels of cost estimates of missed-opportunity. CART has potential savings over METROMAIL by $2.4 MM if the cost of per missed-opportunity is $100.
The results shown in Figs. 6-8 are summarized in Fig. 9. The X-axis is the cost estimates of missed-opportunity, while the Y-axis is the dollars saved over METROMAIL classifier summed across the three demographics. If a lead's value is less than $20, then it is not worth doing any business. Note that a lead's value is the same as the miss-opportunity cost. At the break-even cost, $20, CART can save $318,000 over the current METROMAIL classifier. If a lead's value is $60, then either CART
or NODAC can save up to $1.BMM. If a lead is valued at $100, then either CART or NODAC can save over $4.4MM over the METROMAIL.
It is noted that CART is the better classifier if a lead is valued less than $60. If a lead's value is greater than $60, then there is not much gained using CART
over NODAC. Assuming the missed-opportunity and the wasted-mail costs are $100 and $0.33, respectively, (i.e., the missed-opportunity cost is 303 times greater than that of the wasted-mail cost), it would break even if and only if at least one responder in 303 mailings was obtained. Knowing that the prior probability of response rate is 1.56%, the NODAC approach way would be used (i.e., mail out to all prospects).
It should be apparent that the method of the present invention provides a consistent and sustainable process for building response models which can be used in a variety of direct marketing scenarios that use the phone, Internet (electronic mail), fax machines, etc.
It is therefore apparent that there has been provided in accordance with the present invention, a method that fully satisfies the aims and advantages and objectives set forth herein. The invention has been described with reference to several embodiments, however, it will be appreciated that variations and modifications can be effected by a person of ordinary skill in the art without departing from the scope of the invention.

Claims (17)

1. A method for modeling market response rates comprising the steps of:
acquiring a list of prospects and attaching descriptor variables to said list of prospects;
variably selecting descriptor variables;
selecting a model by examining and assessing at least one algorithm; and validating the model to ensure the robustness of the modeling process.
2. The method of claim 1, wherein said step of acquiring a list of prospects further comprises:
attaching demographics to a prospect;
randomly sampling the prospects; and splitting the randomly sampled prospects into a training set and a testing set.
3. The method of claim 1, wherein said step of acquiring a list of prospects further comprises:
cross referencing said list of prospects with a universal file to cause variables associated with a mailing record to be attached to said prospect.
4. The method of claim 3, wherein said list of prospects is cross referenced with multiple universal files.
5. The method of claim 4, wherein said list of prospects are cross referenced with at least two of the group of universal files consisting of Donnelley, Census Tract and Block Group and ZIP5.
6. The method of claim 1, wherein said step of acquiring a list of prospects further comprises selecting an equal number of responders and non-responders for each universe file used for cross referencing.
7. The method of claim 1, wherein said step of variably selecting descriptor variables further comprises using the misclassification rate as a measure of the discrimination power of each input variable.
8. The method of claim 1, wherein said step of variably selecting descriptor variables further comprises:

selecting CART, and using CART to performing the steps of:

constructing a tree model for each input variable available for selection;

growing the tree until the size of each terminal node is preferably no smaller than 1/100 of the original data set;
pruning the tree until the number of terminal nodes reaches a predetermined number;
computing the misclassification rate of each tree model;
ranking each tree model in ascending order of their misclassification rates;
and selecting at least one input variable based on said ranking.
9. The method of claim 1, wherein said step of variably selecting descriptor variables further comprises:
selecting at least one variable out of the available Donnelley, Census Tract and Block Group and ZIP5 variables;
grouping each input variable into a responder group and a nonresponder group;

determining the mean difference of the responder group and nonresponder group for the input variable selected;
determining the variance difference of the responder group and nonresponder group for the input variable selected;
selecting the input variable if both the mean and variance values are significant different.
10. The method of claim 1, wherein said step of selecting a model by examining and assessing at least one algorithm further comprises:
selecting multiple classifiers for examination and testing; and comparing the selected classifiers using a universal file.
11. The method of claim 10, wherein said step of selecting an algorithm comprises selecting an algorithm from group consisting of METROMAIL, Multivariate Adaptive Regression Splines, Logistic Regression, Neural Networks with Back-Propagation, CART and No Data Optimal Classifier.
12. The method of claim 10, wherein said universal file is the ZIP5 universal file.
13. The method of claim 10, wherein said step of comparing the selected classifiers using a universal file further comprises:
splitting the universal file that is split into a training set and a testing set; and constructing the classifier using the training set.
14. The method of claim 13, where said step of validating the model to ensure the robustness of the modeling process further comprises:

testing each classifier using the testing set.
15. The method of claim 1, further comprising the step of estimating the parameters to ensure that the best fit of data once an algorithm is chosen.
16. The method of claim 1, wherein said method is applied to modeling response rates in the insurance sector.
17. A method for modeling market response rates comprising the steps of:
acquiring a list of prospects;
attaching demographics to said prospect;
randomly sampling the prospects;
splitting the randomly sampled prospects into a training set and a testing set;
variably selecting descriptor variables using the misclassification rate;
selecting multiple classifiers;
selecting a universal file;
comparing the selected classifiers using the selected universal file by constructing the classifier using said training set;
validating the selected classifier by testing each classifier using the testing set.
CA002389222A 1999-09-15 2000-09-06 Method for modeling market response rates Abandoned CA2389222A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US39659999A 1999-09-15 1999-09-15
US09/396,599 1999-09-15
PCT/US2000/024414 WO2001020512A2 (en) 1999-09-15 2000-09-06 Method for modeling market response rates

Publications (1)

Publication Number Publication Date
CA2389222A1 true CA2389222A1 (en) 2001-03-22

Family

ID=23567909

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002389222A Abandoned CA2389222A1 (en) 1999-09-15 2000-09-06 Method for modeling market response rates

Country Status (4)

Country Link
EP (1) EP1224590A2 (en)
AU (1) AU7351300A (en)
CA (1) CA2389222A1 (en)
WO (1) WO2001020512A2 (en)

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8370242B2 (en) 2001-06-08 2013-02-05 Genworth Financial, Inc. Systems and methods for providing a benefit product with periodic guaranteed minimum income
US8433634B1 (en) 2001-06-08 2013-04-30 Genworth Financial, Inc. Systems and methods for providing a benefit product with periodic guaranteed income
US7899688B2 (en) 2001-12-31 2011-03-01 Genworth Financial, Inc. Process for optimization of insurance underwriting suitable for use by an automated system
US7844477B2 (en) 2001-12-31 2010-11-30 Genworth Financial, Inc. Process for rule-based insurance underwriting suitable for use by an automated system
US8005693B2 (en) 2001-12-31 2011-08-23 Genworth Financial, Inc. Process for determining a confidence factor for insurance underwriting suitable for use by an automated system
US7844476B2 (en) 2001-12-31 2010-11-30 Genworth Financial, Inc. Process for case-based insurance underwriting suitable for use by an automated system
US8793146B2 (en) 2001-12-31 2014-07-29 Genworth Holdings, Inc. System for rule-based insurance underwriting suitable for use by an automated system
US7895062B2 (en) 2001-12-31 2011-02-22 Genworth Financial, Inc. System for optimization of insurance underwriting suitable for use by an automated system
US7818186B2 (en) 2001-12-31 2010-10-19 Genworth Financial, Inc. System for determining a confidence factor for insurance underwriting suitable for use by an automated system
US7801748B2 (en) 2003-04-30 2010-09-21 Genworth Financial, Inc. System and process for detecting outliers for insurance underwriting suitable for use by an automated system
US7813945B2 (en) 2003-04-30 2010-10-12 Genworth Financial, Inc. System and process for multivariate adaptive regression splines classification for insurance underwriting suitable for use by an automated system
US7698159B2 (en) 2004-02-13 2010-04-13 Genworth Financial Inc. Systems and methods for performing data collection
CN111047343A (en) * 2018-10-15 2020-04-21 京东数字科技控股有限公司 Method, device, system and medium for information push

Also Published As

Publication number Publication date
WO2001020512A8 (en) 2002-04-11
AU7351300A (en) 2001-04-17
WO2001020512A2 (en) 2001-03-22
EP1224590A2 (en) 2002-07-24

Similar Documents

Publication Publication Date Title
Simester et al. Targeting prospective customers: Robustness of machine-learning methods to typical data challenges
JP4529058B2 (en) Distribution system
Wittink et al. Forecasting with conjoint analysis
US8417560B2 (en) Systems, methods, and apparatus for analyzing the influence of marketing assets
JP4600392B2 (en) How to select relevant campaign messages to send to recipients
Lo et al. WMR--A graph-based algorithm for friend recommendation
US8688518B2 (en) Method, algorithm, and computer program for targeting messages including advertisements in an interactive measurable medium
US20070011224A1 (en) Real-time Internet data mining system and method for aggregating, routing, enhancing, preparing, and analyzing web databases
US20030195793A1 (en) Automated online design and analysis of marketing research activity and data
US20060009994A1 (en) System and method for reputation rating
US20150161661A1 (en) Preservation of scores of the quality of traffic to network sites across clients and over time
CA2389222A1 (en) Method for modeling market response rates
US20220156795A1 (en) Segment content optimization delivery system and method
Bloemer et al. Comparing complete and partial classification for identifying customers at risk
Curtis et al. The citizen versus consumer hypothesis: evidence from a contingent valuation survey
Safa et al. An artificial neural network classification approach for improving accuracy of customer identification in e-commerce
EP3545484A1 (en) System and method for measuring the performance of a brand and predicting its future sales
Augustin et al. Bayesian aggregation of categorical distributions with applications in crowdsourcing
JP2009223423A (en) User state prediction system, user state estimation method, and user state prediction program
Qabbaah et al. Decision tree analysis to improve e-mail marketing campaigns
AU2014204115A1 (en) Using a graph database to match entities by evaluating Boolean expressions
Theerthaana et al. A study to improve the response in email campaigning by comparing data mining segmentation Approaches in Aditi Technologies
KR100727555B1 (en) Creating method for decision tree using time-weighted entropy and recording medium thereof
JP4538934B2 (en) Distribution system
Kelly Mining data to discover customer segments

Legal Events

Date Code Title Description
FZDE Discontinued