CA3165582A1 - Data processing method and system based on similarity model - Google Patents

Data processing method and system based on similarity model

Info

Publication number
CA3165582A1
CA3165582A1 CA3165582A CA3165582A CA3165582A1 CA 3165582 A1 CA3165582 A1 CA 3165582A1 CA 3165582 A CA3165582 A CA 3165582A CA 3165582 A CA3165582 A CA 3165582A CA 3165582 A1 CA3165582 A1 CA 3165582A1
Authority
CA
Canada
Prior art keywords
discrete
data
factors
label
algorithm
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CA3165582A
Other languages
French (fr)
Inventor
Xiang QIAN
Chengcheng XIA
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
10353744 Canada Ltd
Original Assignee
10353744 Canada Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 10353744 Canada Ltd filed Critical 10353744 Canada Ltd
Publication of CA3165582A1 publication Critical patent/CA3165582A1/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q30/00Commerce
    • G06Q30/02Marketing; Price estimation or determination; Fundraising
    • G06Q30/0201Market modelling; Market analysis; Collecting market data

Landscapes

  • Business, Economics & Management (AREA)
  • Strategic Management (AREA)
  • Engineering & Computer Science (AREA)
  • Accounting & Taxation (AREA)
  • Development Economics (AREA)
  • Finance (AREA)
  • Entrepreneurship & Innovation (AREA)
  • Game Theory and Decision Science (AREA)
  • Data Mining & Analysis (AREA)
  • Economics (AREA)
  • Marketing (AREA)
  • Physics & Mathematics (AREA)
  • General Business, Economics & Management (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)
  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

A similarity model-based data processing method and system, which may effectively improve the conversion rate of customers at reduced costs by using similarity model-based data processing technical means. The method comprises: collecting a plurality of customer data; extracting continuous label data from each piece of customer data, and obtaining multiple groups of discrete label data after binning conversion; calculating the similarity distance for discrete factors in each group of discrete label data, while screening out multiple groups of new discrete label data consisting of discrete factors which contribute significantly; calculating the weight for the discrete factors in the new discrete label data respectively by using the random forest algorithm and the gradient boosting decision tree algorithm, and obtaining weighted results of multiple groups of discrete factors after weighted summation; and calculating the final similarity distance between each piece of customer data and positive sample data respectively by using the Manhattan distance algorithm according to the weighted result of each group of discrete factors and the similarity distance of each discrete factor.

Description

DATA PROCESSING METHOD AND SYSTEM BASED ON SIMILARITY MODEL
BACKGROUND OF THE INVENTION
Technical Field [0001] The present invention relates to the field of big data analysis technology, and more particularly to a data processing method and a data processing system based on a similarity model.
Description of Related Art
[0002] Being one of the kernel viewpoints in the network marketing conception with attitude, precision marketing is based on precision positioning, and relies on the modern information technical means, big data technology in particular, to create a customized customer communication service system, to enhance efficiencies of communication and service by the enterprise with respect to customers, and to reduce operational cost.
Winning over and converting are two principal processes of internet operation, of which winning over means to promulgate internet products, expose brands, and develop new users of the products. Converting means to convert low consumption value users of internet products to high value users, namely to promote consumption behaviors of users in the internet products, and to enhance operational achievements of the enterprise.
[0003] Means for winning over and converting in the state of the art are mostly based on blind advertising promotion, but it has been found in practice that, since target users are indefinite, what the large input in advertisement brings about is only a limited number of users won over and converted, whereby there is an obvious contradiction between the advertisement cost as input and the conversion rate as acquired, thus exposing the deficiencies of high cost and low efficiency in the state of the art where users are won Date Recue/Date Received 2022-06-21 over and converted through the mode of blind advertising promotion.
SUMMARY OF THE INVENTION
[0004] An objective of the present invention it is to provide a data processing method and a data processing system based on a similarity model, employs data processing technical means based on a similarity model, and can effectively enhance the conversion rate of customers with decreased cost.
[0005] In order to achieve the above objective, according to one aspect, the present invention provides a data processing method based on a similarity model, the method comprises:
[0006] collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data;
[0007] extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data;
[0008] sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors;
[0009] employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation;
[0010] employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and
[0011] screening out any potential customer according to the final similarity distances.
[0012] Preferably, the step of extracting continuous label data from each piece of customer data, Date Recue/Date Received 2022-06-21 subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data includes:
[0013] performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data;
[0014] performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom; and
[0015] employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.
[0016] Preferably, the step of performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom includes:
[0017] cleaning and filtering invalid label features in the various groups of continuous label initial data sequentially in accordance with a missing rate filter condition, a tantile filter condition, and a proportion of categories filter condition of the label data, and correspondingly obtaining plural groups of continuous label data.
[0018] Preferably, the step of sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors includes:
[0019] employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;
[0020] calculating an IV value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV values;

Date Recue/Date Received 2022-06-21
[0021] employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees;
[0022] employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors; and
[0023] respectively invoking other groups of discrete label data to repeat the above calculating steps, and correspondingly obtaining plural groups of new discrete label data.
[0024] Preferably, the step of employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation includes:
[0025] selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data;
[0026] selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the gradient boosting decision tree algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; and
[0027] performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter performing summation to obtain weight results of plural groups of discrete factors.
[0028] Preferably, the step of employing a Manhattan distance algorithm to calculate a final Date Recue/Date Received 2022-06-21 similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors includes:
[0029] multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data; and
[0030] employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
[0031] Exemplarily, the step of screening out any potential customer according to the final similarity distances includes:
[0032] reversely arranging the final similarity distances according to value sizes, screening out top-ranking N pieces of customer data and marking the same as potential customers.
[0033] In comparison with prior-art technology, the marketing method based on a similarity model provided by the present invention achieves the following advantageous effects:
[0034] In the data processing method based on a similarity model provided by the present invention, plural pieces of customer data are collected to construct a dataset, the dataset contains positive-sample data of converted customers and negative-sample data of not converted customers, the label data of each piece of customer data in the dataset is thereafter correspondingly output to obtain plural groups of continuous label data, at this time, in order to verify each label feature in the continuous label data, namely the prominence of contribution of each discrete factor to the model, it is further needed to employ a binning transformation method to subject the various groups of continuous label data respectively to discrete processing, and to correspondingly obtain plural groups of discrete label data, in which one discrete factor in the discrete label data represents one label feature, and by performing similarity distance calculation on the discrete factors in Date Recue/Date Received 2022-06-21 each group of discrete label data, it is realized to score the various discrete factors, for instance, the smaller the value of the calculation result of a discrete factor is, this indicates the closer will be the discrete factor to the contribution degree of the positive-sample data, otherwise it indicates the farther will be the discrete factor to the contribution degree of the positive-sample data, until similarity distance calculations on the discrete factors in the various groups of discrete label data have been completed, obviously invalid discrete factors are eliminated from the various groups of discrete label data to form plural groups of prominently contributive discrete label data, the random forest algorithm and the gradient boosting decision tree algorithm are thereafter respectively employed to calculate importance indices of variables of various discrete factors in each group of discrete label data, weighted summation is performed on the calculation results of the two algorithms to thereafter obtain weight results of the discrete factors, and the Manhattan distance algorithm is finally employed to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors, to realize value estimation of each piece of customer data. As easily understandable, the smaller the final similarity distance is, the closer will be to the positive-sample data, and the higher will be the value of such customers, in other words, they are more likely to be converted to converted customers; to the contrary, the larger the final similarity distance is, the farther will be to the positive-sample data, and the lower will be the value of such customers, in other words, they are less likely to be converted to converted customers, till now, it is possible to screen out potential customers that meet the requirement according to the final similarity distance of each customer, and to hence carry out precision marketing on them.
[0035] Seen as such, the present invention brings about the following technical effects to winning over and converting of platform businesses:
[0036] Through the design of customer value degree appraising function, it is made possible to provide marketing activities of platforms with customer data support; relative to the blind Date Recue/Date Received 2022-06-21 advertising promotion in the state of the art, the present invention markedly reduces promotion cost of marketing activities at the same time of enhancing conversion rate of customers, and guarantees effects of the marketing activities.
[0037] Use of the similarity model can pertinently calculate the final similarity distance of each piece of customer data in accordance with label features in different customer data, hence appraise the value degree of each piece of customer data, and accurately screen out potential high-value customers.
[0038] According to another aspect, the present invention provides a data processing system based on a similarity model, the system is applied in the data processing method based on a similarity model as recited in the foregoing technical solution, and the system comprises:
[0039] an information collecting unit, for collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data;
[0040] a binning transforming unit, for extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data;
[0041] a label screening unit, for sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors;
[0042] a weight calculating unit, for employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation;
[0043] a similarity distance calculating unit, for employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and Date Recue/Date Received 2022-06-21
[0044] a marketing unit, for screening out any potential customer according to the final similarity distances.
[0045] Preferably, the binning transforming unit includes:
[0046] an initial data extracting module, for performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data;
[0047] a data cleaning module, for performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom; and
[0048] a binning processing module, for employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.
[0049] Preferably, the label screening unit includes:
[0050] an evidence weight algorithm module, for employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;
[0051] an information value calculating module, for calculating an IV value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV
values;
[0052] a Lasso regression algorithm module, for employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees; and
[0053] a ridge regression algorithm module, for employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors.

Date Recue/Date Received 2022-06-21
[0054] Preferably, the weight calculating unit includes:
[0055] a random forest algorithm module, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data;
[0056] a gradient boosting decision tree algorithm module, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a gradient boosting decision tree algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; and
[0057] a weighted assignment module, for performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter performing summation to obtain weight results of plural groups of discrete factors.
[0058] Preferably, the similarity distance calculating unit includes:
[0059] a label feature similarity distance module, for multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data; and
[0060] a customer data similarity distance module, for employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
[0061] In comparison with prior-art technology, the advantageous effects achieved by the data processing system based on a similarity model provided by the present invention are Date Recue/Date Received 2022-06-21 identical with the advantageous effects achievable by the data processing method based on a similarity modal provided by the foregoing technical solution, so these are not redundantly described in this context.
BRIEF DESCRIPTION OF THE DRAWINGS
[0062] The drawings described here are meant to provide further understanding of the present invention, and constitute part of the present invention. The exemplary embodiments of the present invention and the descriptions thereof are meant to explain the present invention, rather than to restrict the present invention. In the drawings:
[0063] Fig. 1 is a flowchart schematically illustrating a data processing method based on a similarity model in Embodiment 1 of the present invention;
[0064] Fig. 2 is an exemplary view illustrating customer data in Fig. 1; and
[0065] Fig. 3 is a block diagram illustrating the structure of a data processing system based on a similarity model in Embodiment 2 of the present invention.
[0066] Reference numerals:
[0067] 1 ¨ information collecting unit 2 ¨ binning transforming unit
[0068] 3 ¨ label screening unit 4 ¨ weight calculating unit
[0069] 5 ¨ similarity distance calculating unit 6 ¨ marketing unit
[0070] 21 ¨ initial data extracting module 22¨ data cleaning module
[0071] 23 ¨ binning processing module 31 ¨ evidence weight algorithm module
[0072] 32 ¨ information value calculating module 33 ¨ Lasso regression algorithm module
[0073] 34 ¨ ridge regression algorithm module 41 ¨ random forest algorithm module Date Recue/Date Received 2022-06-21
[0074] 42 ¨ gradient boosting decision tree 43 ¨ weighted assignment module algorithm module
[0075] 51 ¨ weighted assignment module 52 ¨ customer data similarity distance module DETAILED DESCRIPTION OF THE INVENTION
[0076] To make more lucid and clear the objectives, features and advantages of the present invention, the technical solutions in the embodiments of the present invention are clearly and comprehensively described below with reference to the accompanying drawings in the embodiments of the present invention. Apparently, the embodiments as described are merely partial, rather than the entire, embodiments of the present invention.
All other embodiments obtainable by persons ordinarily skilled in the art on the basis of the embodiments in the present invention without spending creative effort shall all fall within the protection scope of the present invention.
[0077] Embodiment 1
[0078] Fig. 1 is a flowchart schematically illustrating a data processing method based on a similarity model in Embodiment 1 of the present invention. Please refer to Fig. 1, the data processing method based on a similarity model provided by this embodiment comprises:
[0079] collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data; extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data; sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors; employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on Date Recue/Date Received 2022-06-21 discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation; employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and screening out any potential customer according to the final similarity distances.
[0080] In the data processing method based on a similarity model provided by this embodiment, plural pieces of customer data are collected to construct a dataset, the dataset contains positive-sample data of converted customers and negative-sample data of not converted customers, the label data of each piece of customer data in the dataset is thereafter correspondingly output to obtain plural groups of continuous label data, at this time, in order to verify each label feature in the continuous label data, namely the prominence of contribution of each discrete factor to the model, it is further needed to employ a binning transformation method to subject the various groups of continuous label data respectively to discrete processing, and to correspondingly obtain plural groups of discrete label data, in which one discrete factor in the discrete label data represents one label feature, and by performing similarity distance calculation on the discrete factors in each group of discrete label data, it is realized to score the various discrete factors, for instance, the smaller the value of the calculation result of a discrete factor is, this indicates the closer will be the discrete factor to the contribution degree of the positive-sample data, otherwise it indicates the farther will be the discrete factor to the contribution degree of the positive-sample data, until similarity distance calculations on the discrete factors in the various groups of discrete label data have been completed, obviously invalid discrete factors are eliminated from the various groups of discrete label data to form plural groups of prominently contributive discrete label data, the random forest algorithm and the gradient boosting decision tree algorithm are thereafter respectively employed to calculate importance indices of variables of various discrete factors in each group of discrete label data, weighted summation is performed on the calculation results of the two algorithms Date Recue/Date Received 2022-06-21 to thereafter obtain weight results of the discrete factors, and the Manhattan distance algorithm is finally employed to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors, to realize value estimation of each piece of customer data. As easily understandable, the smaller the final similarity distance is, the closer will be to the positive-sample data, and the higher will be the value of such customers, in other words, they are more likely to be converted to converted customers; to the contrary, the larger the final similarity distance is, the farther will be to the positive-sample data, and the lower will be the value of such customers, in other words, they are less likely to be converted to converted customers, till now, it is possible to screen out potential customers that meet the requirement according to the final similarity distance of each customer, and to hence carry out precision marketing on them.
[0081] Seen as such, this embodiment brings about the following technical effects to winning over and converting of platform businesses:
[0082] 1. Through the design of customer value degree appraising function, it is made possible to provide marketing activities of platforms with customer data support;
relative to the blind advertising promotion in the state of the art, this embodiment markedly reduces promotion cost of marketing activities at the same time of enhancing conversion rate of customers, and guarantees effects of the marketing activities.
[0083] 2. Use of the similarity model can pertinently calculate the final similarity distance of each piece of customer data in accordance with label features in different customer data, hence appraise the value degree of each piece of customer data, and accurately screen out potential high-value customers.
[0084] To facilitate comprehension, please refer to Fig. 2, financial platform financing is taken for example for description, customer data can be collected from a database of the financial platform, in which positive-sample data means data of quality customers who Date Recue/Date Received 2022-06-21 have bought financing products, while negative-sample data means data of common customers who have not bought any financing product; during the process of collecting positive-sample data and negative-sample data, a timeline point is firstly selected, and a period of time after the timeline point is then taken as a performance period, data of customers who have bought financing products within the performance period is defined as positive-sample data, and data of customers who have not bought any financing product within the performance period is defined as negative-sample data, more specifically, the positive-sample data and the negative-sample data both contain identification feature attribute discrete factors, such as account numbers of Yihubao, member genders, and member birth dates, etc., historical consumption behavior attribute discrete factors, such as latest shopping payment dates, latest water fee recharging dates, and latest electricity fee recharging dates, etc., member assets status attribute discrete factors, such as recent subscription amounts at Change Treasure, recent subscription amounts for funds, and subscription amounts for periodical financing, etc., and online behavior trajectory attribute discrete factors, such as numbers of in-depth financing pages accessed by members, numbers of in-depth crowd-funding pages accessed by members, and numbers of in-depth insurance pages accessed by members, etc.
[0085] The method of extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data in the foregoing embodiment includes:
[0086] performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data; performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom; employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.

Date Recue/Date Received 2022-06-21
[0087] Specifically, the method of performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom includes: cleaning and filtering invalid label features in the various groups of continuous label initial data sequentially in accordance with a missing rate filter condition, a tantile filter condition, and a proportion of categories filter condition of the label data, and correspondingly obtaining plural groups of continuous label data.
[0088] During the process of specific implementation, the entire label features in the various groups of continuous label initial data are firstly counted, and label features that do not satisfy the missing rate filter condition are then cleaned away from the label features, for instance, the missing rate filter condition can be so set that label features with missing rate exceeding 90% are to be cleaned away, afterwards label features that do not satisfy the tantile filter condition are then cleaned away from the remaining label features, for instance, the tantile filter condition can be so set that label features with tantile being smaller than or equal to 0.1 are to be cleaned away, thereafter label features that do not satisfy the missing rate filter condition are again cleaned away from the remaining label features, and continuous label data is finally output; the above steps are repeated to perform data cleaning with respect to the various groups of continuous label initial data respectively, and plural groups of continuous label data can be correspondingly obtained.
This embodiment makes it possible to remove invalid label features through the data cleaning steps, whereby it is avoided that noises as occur should reduce the precision of the model.
[0089] Moreover, the method of employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data includes the following:
Date Recue/Date Received 2022-06-21
[0090] The optimum binning strategy is employed with respect to the continuous label data, i.e., the attribute of the positive-sample data or the negative-sample data is used as a dependent variable, each continuous variable (label feature) serves as an independent variable, and a conditional inference tree algorithm is employed to discretize the continuous variables; it is firstly supposed that all independent variables and dependent variables are independent, chi-square independence test is subsequently carried out thereon, independent variables with P value being smaller than a set threshold are screened out, and split points are finally selected from each screened independent variable through displacement detection, whereby is achieved the objective of discretizing the continuous variables and finally forming discrete label data. As should be stressed, use of the optimum binning strategy to discretize continuous variables pertains to technical means frequently employed in this field of technology, so this is not redundantly described in this embodiment.
[0091] Specifically, the method of sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors in the foregoing embodiment includes:
[0092] employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;
calculating an IV
value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV
values; employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees;
employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors; and respectively invoking other groups of discrete label data to repeat the above calculating steps, and correspondingly obtaining plural groups of new discrete label data.

Date Recue/Date Received 2022-06-21
[0093] During specific implementation, the evidence weight algorithm in this embodiment indicates a WOE algorithm, the use of which can score the variables of various discrete factors in the discrete label data, the smaller the score of a variable of the discrete factor is, this indicates the higher will be its contribution to the positive sample, and it is further needed, after the variables of the discrete factors have been scored, to normalize them to form similarity distance WOE, in which i expresses the ith discrete factor (label feature), I expresses the jth variable (the variable here can also be understood as a classification) in the ith discrete factor, and the variable is a further definitive description of the discrete factor, for instance, when the discrete factor is member gender, it can be further defined to be classified into two types, the first type is male, and the second type is female;
alternatively, when the discrete factor is date, such as the latest shopping payment date, its further definition can be classified according to time lengths from a timeline point, the first type is within 10 days therefrom, the second type is within 30 days therefrom, and the third type is out of 30 days therefrom; when the discrete factor is numeral, such as the recent subscription amount at Change Treasure, its further definition can be classified according to a numerical gradient, for instance, the first type is within an amount of 5,000 Yuans RMB, the second type is within an amount of 50,000 Yuans RMB, and the third type is outside an amount of 50,000 Yuans RMB, the result of WOEy is in the range of [0,1] on completion of calculation. In practical operation, the number of classifications can be specifically set according to actual circumstances, and this embodiment makes no redundant description thereto. In addition, the evidence weight algorithm is an existing algorithm in this field of technology, however, in order to facilitate comprehension, specific formulae are given in this embodiment for explanation thereof:
#0ii #0,i (p01 #cliT = 41 WOEii = In _____________ ,, In ____ In __ p if # 1 i j #00,
[0094] iT\#l1.
[0095] where WOEy expresses the score of the jth variable in the ith discrete factor, pOu expresses the probability of the jth variable in the ith discrete factor being a negative sample, ply Date Recue/Date Received 2022-06-21 expresses the probability of the jth variable in the ith discrete factor being a positive sample, #0y expresses the number of the jth variable in the ith discrete factor being negative samples, #0,T expresses the total number of variables that are negative samples in the ith discrete factor, My expresses the number of the jth variable in the ith discrete factor being positive samples, and #LT expresses the total number of variables of the ith variable being positive samples.
[0096] After the similarity distance of each discrete factor has been calculated, it is needed to further calculate the IV (information value) value of each discrete factor, and the IV value calculation formula is as follows:
IV =1()0 u -ply) * = -WOE
[0097]
[0098] where n expresses the total number of variables in discrete factor i, and j expresses the jth variable in discrete factor i.
[0099] After the IV value of each discrete factor has been completed, the Lasso regression algorithm is then employed to calculate identification degrees of the various label features, and discrete factors with high identification degrees are screened out therefrom;
optionally, the screening condition of identification degrees is to screen out the minimum X, that satisfies the condition, and to retain discrete factors that satisfy the minimum X, to form a variable set. Subsequently the ridge regression algorithm is then employed to screen out discrete factors with prominent importance from the variable set, and the screening condition of identification degrees is to screen out discrete factors with P value <0.1; through the aforementioned three rounds of screening, prominently contributive discrete label data are finally retained, and the discrete factors remaining at this time can be generally classified into three large types, which are, respectively, customers' own attributes, customer accessing behaviors, and customer transaction behaviors.
As is understandable, X, is a Lagrange operator representing a coefficient of a first-order model Date Recue/Date Received 2022-06-21 norm penalty term in the Lasso regression algorithm.
[0100] As should be noted, both the Lasso regression algorithm and the ridge regression algorithm are regression algorithms frequently employed by persons skilled in this field of technology, and their specific formulae are not redundantly described in this context.
[0101] Preferably, the method of employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation in the foregoing embodiment includes:
[0102] selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the gradient boosting decision tree algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; and performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter summating the same data to obtain weight results of plural groups of discrete factors.
[0103] During specific implementation, the random forest algorithm is employed to classify discrete factors in each group of discrete label data, and to obtain importance indices (Wrfl, Wrf2, = .. Wrfn) to which various variables of each discrete factor correspond, the gradient boosting decision tree (GBDT) algorithm is further employed at the same time to classify discrete factors in each group of discrete label data, and to obtain importance indices Date Recue/Date Received 2022-06-21 (WGBDT1, WGBDT2, = = = WGBDTn) to which various variables of each discrete factor correspond, and weighted assignment is thereafter performed on the same discrete label data; preferably, 0.3 is weighted-assigned to the importance indices obtained by employing the random forest algorithm, 0.7 is weighted-assigned to the importance indices obtained by employing the gradient boosting decision tree algorithm, and weight results (W
, 1, W2,..., W) ¨0.3*(Wrfl, Wrf2,= = =, Wrfn)+0.7*(WGBDT1, WGBDT2, WGBDTn) of the various variables of the discrete factor can be obtained after summation. The random forest algorithm and the gradient boosting decision tree algorithm are both algorithmic formulae frequently employed by persons skilled in this field of technology, and are hence not redundantly described in this embodiment.
[0104] Moreover, the method of employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors in the foregoing embodiment includes:
[0105] multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data; and employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
[0106] During specific implementation, the final weights (Wi, W2, ... Wn) of the various variables in each discrete factor are multiplied with the score WOE,J of the WOE of the various variables in the discrete factor (Wi*W0Ey) to obtain the similarity distance between the customer and the positive sample on a single discrete factor. The Manhattan distance algorithm is thereafter employed to summate the similarity distances of all discrete factors in each piece of customer data, to obtain a final similarity distance between each piece of customer data and the positive-sample data. The Manhattan Date Recue/Date Received 2022-06-21 distance algorithm formula is as follows:
distance =
[0107] j =
[0108] The n expresses the number of discrete factors in the discrete label data, I, expresses the value of the jth classification of the corresponding ith discrete factor in the positive-sample data, in which I, represents an indicator matrix valuated as 0 or 1, for instance, when the ith discrete factor (such as gender) of a male member user is valuated as j (male), the corresponding Iy (Igender, male) is valuated as 1, and other variables (such as Igender, female) on the ith discrete factor are valuated as 0.
[0109] Specifically, the method of screening out any potential customer according to the final similarity distances in the foregoing embodiment includes: reversely arranging the final similarity distances according to value sizes, screening out top-ranking N
pieces of customer data and marking the same as potential customers. Preferably, N is valuated as 5,000, 5,000 customers with the least final similarity distances are then searched for and marked as "potential quality customers", and precision marketing is thereafter performed thereon, so as to entice them to purchase products of the platform.
[0110] Embodiment 2
[0111] Please refer to Fig. 1 and Fig. 3, this embodiment provides a data processing system based on a similarity model, the system comprises:
[0112] an information collecting unit 1, for collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data;
[0113] a binning transforming unit 2, for extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data;
[0114] a label screening unit 3, for sequentially performing similarity distance calculation on a Date Recue/Date Received 2022-06-21 discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors;
[0115] a weight calculating unit 4, for employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation;
[0116] a similarity distance calculating unit 5, for employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and
[0117] a marketing unit 6, for screening out any potential customer according to the final similarity distances.
[0118] Specifically, the binning transforming unit 2 includes:
[0119] an initial data extracting module 21, for performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data;
[0120] a data cleaning module 22, for performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom; and
[0121] a binning processing module 23, for employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.
[0122] Specifically, the label screening unit 3 includes:
[0123] an evidence weight algorithm module 31, for employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;

Date Recue/Date Received 2022-06-21
[0124] an information value calculating module 32, for calculating an IV value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV
values;
[0125] a Lasso regression algorithm module 33, for employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees; and
[0126] a ridge regression algorithm module 34, for employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors.
[0127] Specifically, the weight calculating unit 4 includes:
[0128] a random forest algorithm module 41, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data;
[0129] a gradient boosting decision tree algorithm module 42, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a gradient boosting decision tree algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; and
[0130] a weighted assignment module 43, for performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter performing summation to obtain weight results of plural groups of discrete factors.
[0131] Specifically, the similarity distance calculating unit 5 includes:

Date Recue/Date Received 2022-06-21
[0132] a label feature similarity distance module 51, for multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data; and
[0133] a customer data similarity distance module 52, for employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
[0134] In comparison with prior-art technology, the advantageous effects achieved by the data processing system based on a similarity model provided by this embodiment of the present invention are identical with the advantageous effects achievable by the data processing method based on a similarity modal provided by the foregoing Embodiment 1, so these are not redundantly described in this context.
[0135] As understandable to persons ordinarily skilled in the art, the entire or partial steps realizing the method of the present invention can be completed via a program that instructs relevant hardware, the program can be stored in a computer-readable storage medium, and subsumes the various steps of the method in the foregoing embodiment when it is executed, while the storage medium can be an ROM/RAM, a magnetic disk, an optical disk, or a memory card, etc.
[0136] What the above describes is merely directed to specific modes of execution of the present invention, but the protection scope of the present invention is not restricted thereby. Any change or replacement easily conceivable to persons skilled in the art within the technical range disclosed by the present invention shall be covered by the protection scope of the present invention. Accordingly, the protection scope of the present invention shall be based on the protection scope as claimed in the Claims.

Date Recue/Date Received 2022-06-21

Claims (12)

What is claimed is:
1. A data processing method based on a similarity model, characterized in comprising:
collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data;
extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data;
sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors;
employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation;
employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and screening out any potential customer according to the final similarity distances.
2. The method according to Claim 1, characterized in that the step of extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data includes:
performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data;
performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom; and employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.
3. The method according to Claim 2, characterized in that the step of performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid label feature therefrom includes:
cleaning and filtering invalid label features in the various groups of continuous label initial data sequentially in accordance with a missing rate filter condition, a tantile filter condition, and a proportion of categories filter condition of the label data, and correspondingly obtaining plural groups of continuous label data.
4. The method according to Claim 1, characterized in that the step of sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors includes:
employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;
calculating an IV value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV values;
employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees;
employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors; and respectively invoking other groups of discrete label data to repeat the above calculating steps, and correspondingly obtaining plural groups of new discrete label data.
5. The method according to Claim 1, characterized in that the step of employing a random forest Date Recue/Date Received 2022-06-21 algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation includes:
selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data;
selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing the gradient boosting decision tree algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data; and performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter performing summation to obtain weight results of plural groups of discrete factors.
6. The method according to Claim 1, characterized in that the step of employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors includes:
multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data; and employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
7. The method according to Claim 1, characterized in that the step of screening out any potential Date Recue/Date Received 2022-06-21 customer according to the final similarity distances includes:
reversely arranging the final similarity distances according to value sizes, screening out top-ranking N pieces of customer data and marking the same data as potential customers.
8. A data processing system based on a similarity model, characterized in comprising:
an information collecting unit, for collecting plural pieces of customer data, wherein the customer data are positive-sample data or negative-sample data;
a binning transforming unit, for extracting continuous label data from each piece of customer data, subjecting the same data to binning transformation to thereafter correspondingly obtain plural groups of discrete label data;
a label screening unit, for sequentially performing similarity distance calculation on a discrete factor in each group of discrete label data, and simultaneously screening out plural groups of new discrete label data consisting of prominently contributive discrete factors;
a weight calculating unit, for employing a random forest algorithm and a gradient boosting decision tree algorithm to respectively perform weight calculation on discrete factors in the new discrete label data, and obtaining weight results of plural groups of discrete factors after weighted summation;
a similarity distance calculating unit, for employing a Manhattan distance algorithm to calculate a final similarity distance between each piece of customer data and the positive-sample data on the basis of the weight results of the various groups of discrete factors and similarity distances of the various discrete factors; and a marketing unit, for screening out any potential customer according to the final similarity distances.
9. The system according to Claim 8, characterized in that the binning transforming unit includes:
an initial data extracting module, for performing label feature extraction on each piece of customer data, and obtaining plural groups of continuous label initial data;
a data cleaning module, for performing data cleaning with respect to the various groups of continuous label initial data, and retaining continuous label data after having removed any invalid Date Recue/Date Received 2022-06-21 label feature therefrom; and a binning processing module, for employing an optimum binning strategy to perform optimum binning processing on the various pieces of continuous label data respectively, and correspondingly obtaining plural groups of discrete label data, wherein each group of discrete label data includes plural label features discrete from one another.
10. The system according to Claim 8, characterized in that the label screening unit includes:
an evidence weight algorithm module, for employing an evidence weight algorithm to perform similarity distance calculation on variables of various discrete factors in one group of discrete label data;
an information value calculating module, for calculating an IV value to which each discrete factor corresponds through an information value formula, and screening out discrete factors with high value degrees on the basis of sizes of the IV values;
a Lasso regression algorithm module, for employing a Lasso regression algorithm to screen discrete factors with high identification degrees out of the discrete factors with high value degrees;
and a ridge regression algorithm module, for employing a ridge regression algorithm to further screen discrete factors with prominent importance out of the discrete factors with high identification degrees, and constituting plural groups of new discrete label data consisting of prominently contributive discrete factors.
11. The system according to Claim 8, characterized in that the weight calculating unit includes:
a random forest algorithm module, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a random forest algorithm to calculate importance indices of various variables of the discrete factors in the various pieces of discrete label data;
a gradient boosting decision tree algorithm module, for selecting data in a positive sample as a target variable, taking the discrete factor in each piece of discrete label data as a dependent variable, and employing a gradient boosting decision tree algorithm to calculate importance Date Recue/Date Received 2022-06-21 indices of various variables of the discrete factors in the various pieces of discrete label data; and a weighted assignment module, for performing weighted assignment on the importance indices of the various variables of the discrete factors obtained by employing the random forest algorithm and on the importance indices of the various variables of the discrete factors obtained by employing the gradient boosting decision tree algorithm in the same piece of discrete label data, and thereafter performing summation to obtain weight results of plural groups of discrete factors.
12. The system according to Claim 8, characterized in that the similarity distance calculating unit includes:
a label feature similarity distance module, for multiplying the weight results of the various groups of discrete factors with the similarity distances of the various discrete factors, and calculating a similarity distance between each discrete factor in the customer data and the positive-sample data;
and a customer data similarity distance module, for employing the Manhattan distance algorithm to summate the similarity distances of all discrete factors in each piece of customer data, and obtaining a final similarity distance between each piece of customer data and the positive-sample data.
Date Recue/Date Received 2022-06-21
CA3165582A 2018-12-21 2019-09-20 Data processing method and system based on similarity model Pending CA3165582A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201811570074.3A CN109636482B (en) 2018-12-21 2018-12-21 Data processing method and system based on similarity model
CN201811570074.3 2018-12-21
PCT/CN2019/106858 WO2020125106A1 (en) 2018-12-21 2019-09-20 Similarity model-based data processing method and system

Publications (1)

Publication Number Publication Date
CA3165582A1 true CA3165582A1 (en) 2020-06-25

Family

ID=66076419

Family Applications (1)

Application Number Title Priority Date Filing Date
CA3165582A Pending CA3165582A1 (en) 2018-12-21 2019-09-20 Data processing method and system based on similarity model

Country Status (3)

Country Link
CN (1) CN109636482B (en)
CA (1) CA3165582A1 (en)
WO (1) WO2020125106A1 (en)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109636482B (en) * 2018-12-21 2021-07-27 南京星云数字技术有限公司 Data processing method and system based on similarity model
CN111754253A (en) * 2019-06-20 2020-10-09 北京沃东天骏信息技术有限公司 User authentication method, device, computer equipment and storage medium
CN110990857B (en) * 2019-12-11 2021-04-06 支付宝(杭州)信息技术有限公司 Multi-party combined feature evaluation method and device for protecting privacy and safety
CN111597179B (en) * 2020-05-18 2023-12-05 北京思特奇信息技术股份有限公司 Method and device for automatically cleaning data, electronic equipment and storage medium
CN111564223B (en) * 2020-07-20 2021-01-12 医渡云(北京)技术有限公司 Infectious disease survival probability prediction method, and prediction model training method and device
US20230206294A1 (en) * 2021-12-29 2023-06-29 Rakuten Group, Inc. Information processing apparatus, information processing method, and recording medium

Family Cites Families (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140324523A1 (en) * 2013-04-30 2014-10-30 Wal-Mart Stores, Inc. Missing String Compensation In Capped Customer Linkage Model
CN104699717B (en) * 2013-12-10 2019-01-18 中国银联股份有限公司 Data digging method
CN105354210A (en) * 2015-09-23 2016-02-24 深圳市爱贝信息技术有限公司 Mobile game payment account behavior data processing method and apparatus
CN107273909A (en) * 2016-04-08 2017-10-20 上海市玻森数据科技有限公司 The sorting algorithm of high dimensional data
CN106355449B (en) * 2016-08-31 2021-09-07 腾讯科技(深圳)有限公司 User selection method and device
CN106503873A (en) * 2016-11-30 2017-03-15 腾云天宇科技(北京)有限公司 A kind of prediction user follows treaty method, device and the computing device of probability
CN107103050A (en) * 2017-03-31 2017-08-29 海通安恒(大连)大数据科技有限公司 A kind of big data Modeling Platform and method
CN108876076A (en) * 2017-05-09 2018-11-23 中国移动通信集团广东有限公司 The personal credit methods of marking and device of data based on instruction
CN108460087A (en) * 2018-01-22 2018-08-28 北京邮电大学 Heuristic high dimensional data visualization device and method
CN108876436A (en) * 2018-05-25 2018-11-23 广东工业大学 A kind of electric business discount coupon based on integrated model uses probability forecasting method
CN108876444A (en) * 2018-05-25 2018-11-23 平安科技(深圳)有限公司 Client's classification analysis method, device, computer equipment and storage medium
CN108960505A (en) * 2018-05-31 2018-12-07 试金石信用服务有限公司 Quantitative estimation method, device, system and the storage medium of personal finance credit
CN108776922A (en) * 2018-06-04 2018-11-09 北京至信普林科技有限公司 Finance product based on big data recommends method and device
CN109034658A (en) * 2018-08-22 2018-12-18 重庆邮电大学 A kind of promise breaking consumer's risk prediction technique based on big data finance
CN109636482B (en) * 2018-12-21 2021-07-27 南京星云数字技术有限公司 Data processing method and system based on similarity model

Also Published As

Publication number Publication date
CN109636482B (en) 2021-07-27
CN109636482A (en) 2019-04-16
WO2020125106A1 (en) 2020-06-25

Similar Documents

Publication Publication Date Title
CA3165582A1 (en) Data processing method and system based on similarity model
US10482079B2 (en) Data de-duplication systems and methods
US9916584B2 (en) Method and system for automatic assignment of sales opportunities to human agents
WO2007106786A2 (en) Methods and systems for multi-credit reporting agency data modeling
CN111667307B (en) Method and device for predicting financial product sales volume
Rahmaty et al. Customer churn modeling via the grey wolf optimizer and ensemble neural networks
CN117391313B (en) Intelligent decision method, system, equipment and medium based on AI
CN114997916A (en) Prediction method, system, electronic device and storage medium of potential user
CN111861759A (en) Matching method and system of product and customer group
CN115545886A (en) Overdue risk identification method, overdue risk identification device, overdue risk identification equipment and storage medium
Hu Predicting and improving invoice-to-cash collection through machine learning
CN113420909A (en) User response information prediction model establishing method and information prediction method
CN117114812A (en) Financial product recommendation method and device for enterprises
US10664742B1 (en) Systems and methods for training and executing a recurrent neural network to determine resolutions
Shih et al. Developing target marketing models for personal loans
Branch A case study of applying som in market segmentation of automobile insurance customers
JP6287280B2 (en) Information processing method, program, and information processing apparatus
CN113516511A (en) Financial product purchase prediction method and device and electronic equipment
CN113763032A (en) Commodity purchase intention identification method and device
JP2012238073A (en) Credit purchase assessment support system and credit purchase assessment support method
Deshmukh et al. Risky business: Predicting cancellations in imbalanced multi-classification settings
EP4428770A1 (en) Identifying recurring events using automated semi-supervised classifiers
Huang et al. A hybrid model for portfolio selection based on grey relational analysis and RS theories
CN112801563B (en) Risk assessment method and device
Konda et al. An In-Depth Evaluation of Machine Learning Techniques for Anticipating Effective Human Health Outcomes

Legal Events

Date Code Title Description
EEER Examination request

Effective date: 20220916

EEER Examination request

Effective date: 20220916

EEER Examination request

Effective date: 20220916

EEER Examination request

Effective date: 20220916

EEER Examination request

Effective date: 20220916