CN108108455A - Method for pushing, device, storage medium and the electronic equipment of destination - Google Patents
Method for pushing, device, storage medium and the electronic equipment of destination Download PDFInfo
- Publication number
- CN108108455A CN108108455A CN201711461519.XA CN201711461519A CN108108455A CN 108108455 A CN108108455 A CN 108108455A CN 201711461519 A CN201711461519 A CN 201711461519A CN 108108455 A CN108108455 A CN 108108455A
- Authority
- CN
- China
- Prior art keywords
- sample
- destination
- feature
- classification
- sample set
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9537—Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/29—Geographical information databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/243—Classification techniques relating to the number of classes
- G06F18/24323—Tree-organised classifiers
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/52—Network services specially adapted for the location of the user terminal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/55—Push-based network services
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Remote Sensing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The embodiment of the present application discloses a kind of method for pushing of destination, device, storage medium and electronic equipment, wherein, the method for pushing is included when detecting that user determines destination, and the corresponding multidimensional characteristic in acquisition destination builds the corresponding sample set in destination as sample;Sample classification is carried out to sample set for the information gain of sample classification according to feature, to construct the decision-tree model of destination, the output of decision-tree model is corresponding destination;When detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;Corresponding destination is predicted according to forecast sample and decision-tree model.The automatic push of destination are realized with this, improve the push accuracy rate of destination.
Description
Technical field
This application involves fields of communication technology, and in particular to a kind of method for pushing of destination, device, storage medium and electricity
Sub- equipment.
Background technology
At present, with the high speed development of terminal technology, among increasingly going deep into people’s lives such as smart mobile phone, user is past
It is largely applied toward that can be installed on smart mobile phone, such as chat application, game application, map application.
Wherein, user spreads out the map in application, generally require manually to search or search for destination, waste user when
Between, and operating process is comparatively laborious, although present map application has some simple destination push, the pin of push
It is not high to property, it solves the above problems therefore, it is necessary to provide a kind of method.
The content of the invention
In view of this, the embodiment of the present application provides a kind of method for pushing of destination, device, storage medium and electronics and sets
It is standby, the push accuracy rate of destination can be improved.
In a first aspect, a kind of method for pushing for the destination that the embodiment of the present application provides, including:
When detecting that user determines destination, the corresponding multidimensional characteristic in the destination is gathered as sample, and structure
Build the corresponding sample set in the destination;
Sample classification is carried out to the sample set for the information gain of sample classification according to the feature, to construct
The decision-tree model of destination is stated, the output of the decision-tree model is corresponding destination;
When detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;
Corresponding destination is predicted according to the forecast sample and the decision-tree model.
Second aspect, a kind of pusher for the destination that the embodiment of the present application provides, including:
First collecting unit, for when detecting that user determines destination, gathering the corresponding multidimensional in the destination
Feature builds the corresponding sample set in the destination as sample;
Construction unit, for carrying out sample point to the sample set for the information gain of sample classification according to the feature
Class, to construct the decision-tree model of the destination, the output of the decision-tree model is corresponding destination;
Second collecting unit detects that user spreads out the map in application, gathering current corresponding multidimensional characteristic work for working as
For forecast sample;
Predicting unit, for predicting corresponding destination according to the forecast sample and the decision-tree model.
The third aspect, storage medium provided by the embodiments of the present application, is stored thereon with computer program, when the computer
When program is run on computers so that the computer performs the push side of the destination provided such as the application any embodiment
Method.
Fourth aspect, electronic equipment provided by the embodiments of the present application, including processor and memory, the memory has meter
Calculation machine program, which is characterized in that the processor is by calling the computer program, for performing such as any implementation of the application
The method for pushing for the destination that example provides.
The embodiment of the present application is by the way that when detecting that user determines destination, the corresponding multidimensional characteristic in acquisition destination is made
For sample, and build the corresponding sample set in destination;Sample is carried out to sample set for the information gain of sample classification according to feature
This classification, to construct the decision-tree model of destination, the output of decision-tree model is corresponding destination;When detecting user
It spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;It is pre- according to forecast sample and decision-tree model
Measure corresponding destination.The automatic push of destination are realized with this, improve the push accuracy rate of destination.
Description of the drawings
To describe the technical solutions in the embodiments of the present invention more clearly, make required in being described below to embodiment
Attached drawing is briefly described, it should be apparent that, the accompanying drawings in the following description is only some embodiments of the present invention, for
For those skilled in the art, without creative efforts, it can also be obtained according to these attached drawings other attached
Figure.
Fig. 1 is the application scenarios schematic diagram of the method for pushing of destination provided by the embodiments of the present application.
Fig. 2 is a flow diagram of the method for pushing of destination provided by the embodiments of the present application.
Fig. 3 is a kind of schematic diagram of decision tree provided by the embodiments of the present application.
Fig. 4 is the schematic diagram of another decision tree provided by the embodiments of the present application.
Fig. 5 is another flow diagram of the method for pushing of destination provided by the embodiments of the present application.
Fig. 6 is a structure diagram of the pusher of destination provided by the embodiments of the present application.
Fig. 7 is another structure diagram of the pusher of destination provided by the embodiments of the present application.
Fig. 8 is a structure diagram of electronic equipment provided by the embodiments of the present application.
Fig. 9 is another structure diagram of electronic equipment provided by the embodiments of the present application.
Specific embodiment
Schema is refer to, wherein identical element numbers represent identical component, the principle of the application is to implement one
It is illustrated in appropriate computing environment.The following description be based on illustrated the application specific embodiment, should not be by
It is considered as limitation the application other specific embodiments not detailed herein.
In the following description, the specific embodiment of the application will be with reference to as the step performed by one or multi-section computer
And symbol illustrates, unless otherwise stating clearly.Therefore, these steps and operation will have to mention for several times is performed by computer, this paper institutes
The computer execution of finger includes by representing with the computer processing unit of the electronic signal of the data in a structuring pattern
Operation.This operation is converted at the data or the position being maintained in the memory system of the computer, reconfigurable
Or in addition change the running of the computer in a manner of known to the tester of this field.The data structure that the data are maintained
For the provider location of the memory, there is the specific feature as defined in the data format.But the application principle is with above-mentioned text
Word illustrates that be not represented as a kind of limitation, this field tester will appreciate that plurality of step as described below and behaviour
Also may be implemented among hardware.
Term as used herein " module " can regard the software object to be performed in the arithmetic system as.It is as described herein
Different components, module, engine and service can be regarded as the objective for implementation in the arithmetic system.And device as described herein and side
Method can be implemented in a manner of software, can also be implemented certainly on hardware, within the application protection domain.
Term " first ", " second " and " the 3rd " in the application etc. is for distinguishing different objects rather than for retouching
State particular order.In addition, term " comprising " and " having " and their any deformations, it is intended that cover non-exclusive include.
Such as contain the step of process, method, system, product or the equipment of series of steps or module is not limited to list or
Module, but some embodiments further include the step of not listing or module or some embodiments further include for these processes,
Method, product or equipment intrinsic other steps or module.
Referenced herein " embodiment " is it is meant that a particular feature, structure, or characteristic described can wrap in conjunction with the embodiments
It is contained at least one embodiment of the application.Each position in the description occur the phrase might not each mean it is identical
Embodiment, nor the independent or alternative embodiment with other embodiments mutual exclusion.Those skilled in the art explicitly and
Implicitly understand, embodiment described herein can be combined with other embodiments.
The embodiment of the present application provides a kind of method for pushing of destination, and the executive agent of the method for pushing of the destination can be with
It is the pusher of destination provided by the embodiments of the present application or is integrated with the electronic equipment of the pusher of the destination,
Hardware may be employed in the pusher of the wherein destination or the mode of software is realized.Wherein, electronic equipment can be intelligence
The equipment such as mobile phone, tablet computer, palm PC, laptop or desktop computer.
Referring to Fig. 1, Fig. 1 is the application scenarios schematic diagram of the method for pushing of destination provided by the embodiments of the present application, with
Exemplified by the pusher of destination integrates in the electronic device, when electronic equipment detects that user determines destination, acquisition
The corresponding multidimensional characteristic in destination builds the corresponding sample set in destination as sample;According to feature for sample classification
Information gain carries out sample classification to sample set, to construct the decision-tree model of destination;When detecting that user spreads out the map
In application, the current corresponding multidimensional characteristic of acquisition is as forecast sample;Correspondence is predicted according to forecast sample and decision-tree model
Destination.
It specifically, can be in historical time section, when electronic equipment detects that user determines purpose such as shown in Fig. 1
During ground, the corresponding multidimensional characteristic (weather characteristics, initially period feature, feature etc.) in acquisition destination is used as sample, and structure
Build the corresponding sample set in destination;According to feature (weather characteristics, initially period feature, feature etc.) for sample classification
Information gain carries out sample classification to sample set, to construct the decision-tree model of destination;When detecting that user spreads out the map
During using a, gather current corresponding multidimensional characteristic (weather characteristics, initially period feature, feature etc.) and be used as forecast sample,
Corresponding destination is predicted according to forecast sample and decision-tree model.
Referring to Fig. 2, Fig. 2 is the flow diagram of the method for pushing of destination provided by the embodiments of the present application.The application
The idiographic flow of the method for pushing for the destination that embodiment provides can be as follows:
201st, when detecting that user determines destination, the corresponding multidimensional characteristic in acquisition destination is as sample, and structure
Build the corresponding sample set in destination.
Destination mentioned by the present embodiment spreads out the map using the destination inputted for user, it is big to be specifically as follows XX
Tall building, XX stations, XX supermarkets etc., when electronic equipment detects that user has selected certain destination, when gathering the selected destination
Corresponding multidimensional characteristic is as sample.
Wherein, which has the dimension of certain length, the corresponding characterization destination of the parameter in each of which dimension
A kind of characteristic information, i.e., the multidimensional characteristic breath be made of multiple features.When the plurality of feature can be including selected destination
Relevant characteristic information, such as:Current state of weather;The current period;Initially information etc..
Wherein, the sample set of destination can include multiple samples, and each sample includes the corresponding multidimensional characteristic in destination.
In the sample set of destination, it can be included in historical time section, multiple samples of the destination of acquisition.Historical time section, example
It such as can be 7 days, 14 days in the past.It is understood that the multi-dimensional feature data of the destination once gathered forms a sample
This, multiple samples form sample set.
After sample set is formed, each sample in sample set can be marked, obtain the sample of each sample
Label, since this implementation will be accomplished that the destination of prediction map application, the sample label marked includes various mesh
Ground namely the different destination of sample class.
202nd, sample classification is carried out to sample set for the information gain of sample classification according to feature, to construct destination
Decision-tree model.
The embodiment of the present application can carry out sample classification for the information gain of sample classification with feature based to sample set, with
Build the decision-tree model of destination.For example, can decision-tree model be built based on ID3 algorithms.
Wherein, decision tree is a kind of a kind of tree relied on decision-making and set up.In machine learning, decision tree is a kind of
Prediction model, representative is a kind of a kind of mapping relations between object properties and object value, some is right for each node on behalf
As, each diverging paths in tree represent some possible property value, and each leaf node then correspond to from root node to
The value for the object represented by path that the leaf node is undergone.Decision tree only has single output, can be with if there is multiple outputs
Independent decision tree is established respectively to handle different output.
Wherein, ID3 (Iterative Dichotomiser 3,3 generation of iteration binary tree) algorithm is one kind of decision tree, it
It is based on "ockham's razor" principle, i.e., does more things with less thing with trying one's best.In information theory, it is expected that information is got over
It is small, then information gain is bigger, so as to which purity is higher.The core concept of ID3 algorithms is exactly to be belonged to information gain to measure
Property selection, the attribute of information gain maximum is into line splitting after selection division.The algorithm uses top-down greedy search time
Go through possible decision space.
Wherein, information gain exactly sees a feature t for feature one by one, and system has it and do not have
It when information content be respectively how many, the difference of the two is exactly the information content that this feature is brought to system, i.e. information gain.
The process classified based on information gain to sample set is described in detail below, for example, assorting process can wrap
Include following steps:
Corresponding root node is generated, and using sample set as the nodal information of root node;
The sample set of root node is determined as current target sample collection to be sorted;
Obtain the information gain that feature is classified for sample set in target sample collection;
Current division feature is chosen from feature according to information gain selection;
Sample set is divided according to division feature, obtains several subsample collection;
The division feature of sample in sub- sample set is removed, subsample collection after being removed
The child node of present node is generated, and using subsample collection after removal as the nodal information of child node;
Judge whether child node meets default classification end condition;
If it is not, target sample collection then is updated to subsample collection after removing, and returns to execution and obtain spy in target sample collection
Levy the information gain for sample set classification;
If so, using child node as leaf node, the classification for concentrating sample according to subsample after removal sets leaf section
The output of point, the classification of sample is corresponding destination.
Wherein, division is characterized as the feature chosen according to the information gain that each feature is classified for sample set from feature,
For classifying to sample set.Wherein, there are many modes that division feature is chosen according to information gain, such as in order to promote sample point
The accuracy of class, can choose maximum information gain it is corresponding be characterized as division feature.
Wherein, the classification of sample is corresponding multiple destination classifications.
When child node meets default classification end condition, can it stop to the son using child node as leaf node
The sample set classification of node, and can concentrate the classification of sample that the output of the leaf node is set based on subsample after removal.
Classification based on sample is set there are many modes of the output of leaf node.It for example, can be by sample number in sample set after removal
Measure output of most classifications as the leaf node.
Wherein, presetting classification end condition can set according to actual demand, and child node meets default classification and terminates item
During part, using current node as leaf node, stop carrying out participle classification to the corresponding sample set of child node;Child node is not
When meeting default classification end condition, continue to classify to the corresponding volume sample set of child node.For example, default classification terminates item
Part can include:The categorical measure of sample is " to judge with default quantity namely step in the set of subsample after the removal of child node
Whether child node meets default classification end condition " it can include:
Subsample concentrates whether the categorical measure of sample is default quantity after judging the corresponding removal of child node;
If so, determine that child node meets default classification end condition;
If not, it is determined that child node is discontented with default classified terminal end condition.
For example, default classification end condition can include:The classification of sample is concentrated in subsample after the corresponding removal of child node
Quantity be 1 namely the sample set of child node in only there are one classification sample.At this point, if child node meets the default classification
End condition, then, the classification of sample is concentrated into as the output of the leaf node in subsample.Subsample is concentrated only after such as removing
When having the sample that classification is " XX mansions ", then, it can be by the output of " XX mansions " as the leaf node.
In one embodiment, in order to promote the accuracy of determination of decision-tree model, a gain threshold can also be set;When
When maximum information gain is more than the threshold value, just choose the information gain for feature be division feature.That is, step " root
Current division feature is chosen from feature according to information gain selection " it can include:
Maximum target information gain is chosen from information gain;
Judge whether target information gain is more than predetermined threshold value;
If so, the corresponding feature of target information gain is chosen as current division feature.
It in one embodiment, can be using present node as leaf section when target information gain is not more than predetermined threshold value
Point, and choose output of the most sample class of sample size as the leaf node.Wherein, sample class is corresponding purpose
Ground.
Wherein, predetermined threshold value can be set according to actual demand, such as 0.9,0.8.
For example, when feature 1 for sample classification information gain 0.9 be maximum information gain when, predetermined threshold value 0.8
When, since maximum information gain is more than predetermined threshold value, at this point it is possible to using feature 1 as division feature.
In another example when predetermined threshold value is 1, then maximum information gain is less than predetermined threshold value, at this point it is possible to will work as prosthomere
Point understands sample set analysis classification is most for the sample size at " XX stations ", is " XX more than other classifications as leaf node
The sample size of mansion ", at this point it is possible to by the output of " XX stations " as the leaf node.
Wherein, sample is carried out there are many modes of classifying and dividing according to division feature, for example, division feature can be based on
Characteristic value sample set divided.Namely step " being divided according to division feature to sample set " can include:
Obtain the characteristic value that feature is divided in sample set;
Sample set is divided according to characteristic value.
It is concentrated for example, can will divide the identical sample of characteristic value in sample set and be divided into same subsample.For example, divide
The characteristic value of feature includes:0th, 1,2, then at this point it is possible to the sample that the characteristic value for dividing feature is 0 be classified as it is a kind of, by feature
The sample being worth for 1 is classified as sample that is a kind of, being 2 by characteristic value and is classified as one kind.
For example, for sample set A { sample 1, sample 2 ... sample i ... samples n }, wherein sample 1 includes feature 1, spy
Sign 2 ... feature m, sample i include feature 1, feature 2 ... feature m, sample n include feature 1, feature 2 ... feature m.
First, all samples in sample set are initialized, then, generate a root node a, and using sample set as
The nodal information of root node a, such as with reference to figure 3.
Calculate each feature information gain g1, g2 ... gm that for example feature 1, feature 2 ... feature m are classified for sample set;
Maximum information gain gmax is chosen, if gi is maximum information gain.
When maximum information gain gmax is less than predetermined threshold value ε, current node chooses sample number as leaf node
Measure output of most sample class as leaf node.
When maximum information gain gmax is more than predetermined threshold value ε, the corresponding feature i of information gain gmax can be chosen and made
To divide feature t, sample set A { sample 1, sample 2 ... sample i ... samples n } is divided according to feature i, such as by sample
This collection is divided into two sub- sample set A1 { sample 1, sample 2 ... sample k } and A2 { sample k+1 ... samples n }.
Feature t removal will be divided in subsample collection A1 and A2, at this point, in subsample collection A1 and A2 sample include feature 1,
Feature 2 ... feature i-1, feature i+1 ... features n }.Generate the child node a1 and a2 of root node a with reference to figure 3, and by increment
This collection A1 as the nodal information of child node a1, using subsample collection A2 as the nodal information of child node a2.
Then, for each child node, by taking child node a1 as an example, judge whether child node meets default classification and terminate item
Part, if so, using current child node a1 as leaf node, and according to the class of the corresponding subsample concentration samples of child node a1
The leaf node is not set to export.
When child node is unsatisfactory for default classification end condition, by the way of the above-mentioned classification based on information gain, continue
Classify to the corresponding subsample collection of child node, can such as be calculated by taking child node a2 as an example in A2 sample sets each feature compared with
The information gain g of sample classification chooses maximum information gain gmax, when maximum information gain gmax is more than predetermined threshold value ε
When, can choose information gain gmax it is corresponding be characterized as division feature t, based on division feature t A2 is divided into several sons
A2 can be such as divided into subsample collection A21, A22, A23 by sample set, then, by the division in subsample collection A21, A22, A23
Feature t is removed, and generates child node a21, a22, a23 of present node a2, will remove the sample set A21 after division feature t,
A22, A23 are respectively as the nodal information of child node a21, a22, a23.
And so on, decision tree as shown in Figure 4 is may be constructed out using the above-mentioned mode based on information gain classification,
The output of the leaf node of the decision tree includes " XX destinations ".
In the embodiment of the present application, empirical entropy that can be based on sample classification and feature are for the item of sample set classification results
Part entropy obtains the information gain that feature is classified for sample set.Namely " feature is for sample set in acquisition target sample collection for step
The information gain of classification " can include:
Obtain the empirical entropy of sample classification;
Obtain conditional entropy of the feature for sample set classification results;
According to conditional entropy and empirical entropy, the information gain that feature is classified for sample set is obtained.
Wherein it is possible to obtain the probability that each destination sample occurs in sample set, which is sample class
Not Wei corresponding destination, according to each probability obtain sample empirical entropy.
For example, for sample set Y { sample 1, sample 2 ... sample i ... samples n }, if sample class is XX mansions
The sample size of sample is j, and the sample size of XX stations sample is n-j;At this point, appearance of the XX mansions sample in sample set Y
Probability of occurrence p2=n-j/n of Probability p 1=j/n, the XX station in sample set Y.Then, the calculating based on following empirical entropy is public
Formula calculates the empirical entropy H (Y) of sample classification:
Wherein, pi is probability of occurrence of the sample in sample set Y.In decision tree classification problem, information gain is exactly certainly
The difference of plan tree information before Attributions selection division is carried out and after division.
In one embodiment, can sample set be divided by several subsample collection according to feature t, then, obtains each increment
The probability that the comentropy of this collection classification and each characteristic value of this feature t occur in sample set, according to the comentropy and is somebody's turn to do
Probability can be divided after comentropy, i.e., this feature t is for the conditional entropy of sample set classification results.
For example, for sample characteristics X, sample characteristics X can be by following for the conditional entropy of sample set Y classification results
Formula is calculated:
Wherein, n is characterized the value kind number of X, i.e. characteristic value number of types.At this point, it is i-th kind of value that pi, which is X characteristic values,
The probability that occurs in sample set Y of sample, xi is i-th kind of value of X.H (Y | X=xi) it is the experience that subsample collection Yi classifies
Entropy, the X characteristic values of sample are i-th kind of value in the collection i of the subsample.
For example, using the value kind number of feature X as 3, i.e., exemplified by x1, x2, x3, at this point it is possible to which feature X is by sample set Y { samples
1st, sample 2 ... sample i ... samples n } three sub- sample sets are divided into, characteristic value is Y1 { sample 1, sample 2 ... the sample of x1
This d }, the Y2 { sample d+1 ... samples e } that characteristic value is x2, the Y3 { sample e+1 ... samples n } that characteristic value is x3.D, e is equal
For positive integer, and less than n.
At this point, feature X is for the conditional entropy of sample set Y classification results:
H (Y | X)=p1H (Y | x1)+p2H (Y | x2)+p3H (Y | x3);
Wherein, p1=Y1/Y, p2=Y2/Y, p2=Y3/Y;
H (Y | x1) it is the comentropy that subsample collection Y1 classifies, i.e. empirical entropy, the calculation formula of above-mentioned empirical entropy can be passed through
It is calculated.
Obtaining the conditional entropy H (Y | X) of the empirical entropy H (Y) and feature X of sample classification for sample set Y classification results
Afterwards, the information gain that feature X classifies for sample set Y can be calculated, is such as calculated by the following formula:
G (Y, X)=H (Y)-H (Y | X)
Namely feature X is for the sample set Y information gains classified:Empirical entropy H (Y) and feature X classifies for sample set Y
As a result the difference of conditional entropy H (Y | X).
203rd, when detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample.
Wherein, when detect user spread out the map in application, illustrate user need carry out destination lookup.It is corresponding, it adopts
The current corresponding multidimensional characteristic of collection is as default sample.
It should be strongly noted that in the embodiment of the present application, the multidimensional characteristic gathered in step 201 and 203 is identical spy
Sign, such as:Current state of weather;The current period;Initially information etc..
204th, corresponding destination is predicted according to forecast sample and decision-tree model.
Specifically, corresponding output is obtained according to forecast sample and decision-tree model as a result, being determined pair according to output result
The destination answered.Wherein, exporting result includes each destination.
For example, can corresponding leaf node be determined according to the feature and decision-tree model of forecast sample, by the leaf section
The output of point is as prediction output result.Such as (feature is divided according to the branch condition of decision tree using the feature of forecast sample
Characteristic value) determine current leaf node, take the leaf node output as predict result.It is defeated due to leaf node
Go out including multiple destinations.
For example, after gathering current multidimensional characteristic, it can be with the branch condition according to decision tree in decision tree shown in Fig. 4
Corresponding leaf node is searched as an1, the output of leaf node an1 is destination 1, at this point, the destination for just determining push is
Destination 1.
From the foregoing, it will be observed that the embodiment of the present application is by the way that when detecting that user determines destination, acquisition destination is corresponding
Multidimensional characteristic builds the corresponding sample set in destination as sample;According to feature for sample classification information gain to sample
This collection carries out sample classification, and to construct the decision-tree model of destination, the output of decision-tree model is corresponding destination;When
Detect that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;According to forecast sample and certainly
Plan tree-model predicts corresponding destination.The automatic push of destination are realized with this, the push for improving destination is accurate
Rate.
Further, due in each sample of sample set, including reflection user the behavior of destination is selected to practise usually
Used multiple characteristic informations, therefore the embodiment of the present application can so that the push to destination is more personalized and intelligent.
Further, realize that the push of destination is predicted based on decision tree prediction model, destination push can be promoted
Accuracy, more be bonded user use habit.
On the basis of the method that will be described below in above-described embodiment, the sorting technique of the application is described further.Ginseng
Fig. 5 is examined, the method for pushing of the destination can include:
301st, when detecting that user determines destination, the corresponding multidimensional characteristic in acquisition destination is as sample, and structure
Build the corresponding sample set in destination.
Wherein, when detecting that user inputs destination by map application, when inputting " XX mansions " such as user, acquisition choosing
Corresponding multidimensional characteristic is as sample during the fixed destination.
The multidimensional characteristic information of application has the dimension of certain length, the corresponding characterization purpose of the parameter in each of which dimension
A kind of characteristic information on ground, i.e. the multidimensional characteristic information are made of multiple characteristic informations.The plurality of characteristic information can include choosing
Determine relevant characteristic information during destination, such as:Current state of weather;The current period;Initially information, date letter
Breath etc..
In the sample set of destination, multiple samples of acquisition in historical time section can be included in.Historical time section, such as
It can be 7 days, 14 days etc. in the past.It is understood that the multi-dimensional feature data for once gathering destination forms a sample, it is more
A sample forms sample set.
One specific sample can be as shown in table 1 below, and the characteristic information including multiple dimensions is, it is necessary to explanation, 1 institute of table
The characteristic information shown is only for example, and in practice, the quantity for the characteristic information that a sample is included can be more than than shown in table 1
The quantity of information can also be less than the quantity of information shown in table 1, and the specific features information taken can also be different from shown in table 1,
It is not especially limited herein.
Dimension | Characteristic information |
1 | Current state of weather |
2 | The current period |
3 | Initially information |
4 | Date information |
5 | Current wireless network state, such as wifi connection status |
Table 1
302nd, the sample in sample set is marked, obtains the sample label of each sample.
Since this implementation will be accomplished that prediction destination, the sample label marked includes each destination.It should
The sample label of sample characterizes the sample class of the sample.At this point, sample class can " XX mansions ", " XX stations " etc..
In one embodiment, the sample in sample set is marked in this, after obtaining the sample label of each sample,
It further includes:
(1) detect in sample set with the presence or absence of the sample that the corresponding characteristic value of multidimensional characteristic is identical but destination classification is different
This;
Wherein, when there are the corresponding characteristic value of multidimensional characteristic is completely the same, but sample label is inconsistent in sample set, i.e.,
The different sample of destination classification, such as state of weather current in sample 1, initially current period, information, date information
And currently the corresponding characteristic information of wireless network state and sample 2 are completely the same, but 1 corresponding destination classification of sample is
" XX mansions ", and 2 corresponding destination classification of sample is then judged to detecting that there are multidimensional characteristics in sample set for " XX stations "
The sample that corresponding characteristic value is identical but destination classification is different performs step (2).
(2) the most sample of destination classification sample size is retained.
Wherein, due to the corresponding characteristic value of multidimensional characteristic is identical but different samples that destination classification is different can shadow
Ring the decision tree structure in later stage, thus when detect the presence of that the corresponding characteristic value of multidimensional characteristic is identical but destination classification
During different different samples, the most sample of wherein destination classification sample size is retained, other samples are then deleted, such as
Destination classification is the sample that the sample size of " XX mansions " is more than that destination classification is " XX stations ", then retains destination classification
For all samples of " XX mansions ".
In one embodiment, if different destinations classification sample size is consistent, time when being stored according to sample,
Retain storage time and the immediate destination classification sample of current time.
303rd, the root node of decision-tree model is generated, and using sample set as the nodal information of root node.
For example, with reference to figure 3, for sample set A { sample 1, sample 2 ... sample i ... samples n }, can first generate certainly
The root node a of plan tree, and using sample set A as the nodal information of root node a.
304th, determine sample set for current target sample collection to be sorted.
Namely the sample set of definite root node is as current target sample collection to be sorted.
305th, the information gain that each feature is classified for sample set in target sample collection is obtained, and determines that maximum information increases
Benefit.
For example, for sample set A, each feature can be calculated as feature 1, feature 2 ... feature m classify for sample set
Information gain g1, g2 ... gm;Choose maximum information gain gmax.
Wherein, following manner acquisition may be employed in the information gain that feature is classified for sample set:
Obtain the empirical entropy of sample classification;Obtain conditional entropy of the feature for sample set classification results;According to conditional entropy and
Empirical entropy obtains the information gain that feature is classified for sample set.
For example, each destination classification can be obtained.
The probability that each destination sample occurs in sample set can be obtained, which is that sample class is pair
The destination answered obtains the empirical entropy of sample according to each probability.
For example, there was only " XX mansions " and " exemplified by XX stations " by the sample class of destination sample, for sample set Y
{ sample 1, sample 2 ... sample i ... samples n }, if the sample size that sample class is " XX mansions " is j, " XX stations "
Sample size be n-j;At this point, the probability of occurrence p1=j/n of " XX mansions " in sample set Y, " XX stations " is in sample set Y
Probability of occurrence p2=n-j/n.Then, the calculation formula based on following empirical entropy calculates the empirical entropy H (Y) of sample classification:
In decision tree classification problem, information gain is exactly decision tree information after carrying out Attributions selection and dividing preceding and division
Difference.
In one embodiment, can sample set be divided by several subsample collection according to feature t, then, obtains each increment
The probability that the comentropy of this collection classification and each characteristic value of this feature t occur in sample set, according to the comentropy and is somebody's turn to do
Probability can be divided after comentropy, i.e., this feature t is for the conditional entropy of sample set classification results.
For example, for sample characteristics X, sample characteristics X can be by following for the conditional entropy of sample set Y classification results
Formula is calculated:
Wherein, n is characterized the value kind number of X, i.e. characteristic value number of types.At this point, it is i-th kind of value that pi, which is X characteristic values,
The probability that occurs in sample set Y of sample, xi is i-th kind of value of X.H (Y | X=xi) it is the experience that subsample collection Yi classifies
Entropy, the X characteristic values of sample are i-th kind of value in the collection i of the subsample.
For example, using the value kind number of feature X as 3, i.e., exemplified by x1, x2, x3, at this point it is possible to which feature X is by sample set Y { samples
1st, sample 2 ... sample i ... samples n } three sub- sample sets are divided into, characteristic value is Y1 { sample 1, sample 2 ... the sample of x1
This d }, the Y2 { sample d+1 ... samples e } that characteristic value is x2, the Y3 { sample e+1 ... samples n } that characteristic value is x3.D, e is equal
For positive integer, and less than n.
At this point, feature X is for the conditional entropy of sample set Y classification results:
H (Y | X)=p1H (Y | x1)+p2H (Y | x2)+p3H (Y | x3);
Wherein, p1=Y1/Y, p2=Y2/Y, p3=Y3/Y;
H (Y | x1) it is the comentropy that subsample collection Y1 classifies, i.e. empirical entropy, the calculation formula of above-mentioned empirical entropy can be passed through
It is calculated.
Obtaining the conditional entropy H (Y | X) of the empirical entropy H (Y) and feature X of sample classification for sample set Y classification results
Afterwards, the information gain that feature X classifies for sample set Y can be calculated, is such as calculated by the following formula:
G (Y, X)=H (Y)-H (Y | X)
Namely feature X is for the sample set Y information gains classified:Empirical entropy H (Y) and feature X classifies for sample set Y
As a result the difference of conditional entropy H (Y | X).
306th, judge whether maximum information gain is more than predetermined threshold value, if so, step 307 is performed, if it is not, then performing
Step 313.
Such as, it can be determined that whether maximum information gain gmax is more than default threshold epsilon, which can be according to reality
Border demand setting.
307th, the corresponding feature of maximum information gain is chosen as division feature, and according to the characteristic value of the division feature
Sample set is divided, obtains several subsample collection.
For example, when the maximum corresponding features of information gain gmax are characterized i, it can be using selected characteristic i as division feature.
Specifically, can sample set be divided by several subsample collection, subsample according to the characteristic value kind number of division feature
The quantity of collection is identical with characteristic value kind number.For example, the identical sample of characteristic value can will be divided in sample set is divided into same son
In sample set.For example, dividing the characteristic value of feature includes:0th, 1,2, then at this point it is possible to the sample that the characteristic value for dividing feature is 0
Originally it is classified as sample that is a kind of, being 1 by characteristic value and is classified as sample that is a kind of, being 2 by characteristic value being classified as one kind.
308th, the division feature that subsample is concentrated sample removes, subsample collection after being removed.
For example, the value of division feature i there are two types of when, sample set A can be divided into A1 { sample 1, sample 2 ... sample
This k } and A2 { sample k+1 ... samples n }.It is then possible to the division feature i in subsample collection A1 and A2 is removed.
309th, the child node of present node is generated, and using subsample collection after removal as the nodal information of corresponding child node.
Wherein, a sub- sample set corresponds to a child node.For example, with reference to figure 3 generate root node a child node a1 and
A2, and using subsample collection A1 as the nodal information of child node a1, using subsample collection A2 as the nodal information of child node a2.
310th, judge whether the subsample collection of child node meets default classification end condition, if so, step 311 is performed,
If it is not, then perform step 312.
Wherein, presetting classification end condition can set according to actual demand, and child node meets default classification and terminates item
During part, using current node as leaf node, stop carrying out participle classification to the corresponding sample set of child node;Child node is not
When meeting default classification end condition, continue to classify to the corresponding volume sample set of child node.For example, default classification terminates item
Part can include:The categorical measure of sample is and default quantity in the set of subsample after the removal of child node.
For example, default classification end condition can include:The classification of sample is concentrated in subsample after the corresponding removal of child node
Quantity be 1 namely the sample set of child node in only there are one classification sample.
The 311st, target sample collection is updated to the subsample collection of child node, and return and perform step 305.
312nd, using the child node as leaf node, and concentrate sample class that the leaf is set according to the subsample of child node
The output of node.
For example, default classification end condition can include:The classification of sample is concentrated in subsample after the corresponding removal of child node
Quantity be 1 namely the sample set of child node in only there are one classification sample.
At this point, if child node meets the default classification end condition, then, by the destination class of subsample concentration sample
Output not as the leaf node.Such as remove after subsample concentrate only purposefully classification be " XX stations " sample when, that
, can be by the output of " XX stations " as the leaf node
313rd, using present node as leaf node, and the most sample class of sample size is chosen as the leaf node
Output.
Wherein, sample class includes each destination.
For example, in the subsample collection A1 classification of child node a1, if maximum information gain is small and predetermined threshold value, at this point,
It can be using the most sample class of sample size in the collection A1 of subsample as the output of the leaf node.Such as the sample of " XX mansions "
Quantity is most, then can be by the output of " XX mansions " as leaf node a1.
314th, after decision-tree model has been built, when detecting that user spreads out the map in application, gathering current corresponding more
Dimensional feature is as forecast sample.
Wherein, when detect user spread out the map in application, illustrate user need carry out destination lookup.It is corresponding, it adopts
The current corresponding multidimensional characteristic of collection is as default sample.
315th, corresponding destination is predicted according to forecast sample and decision-tree model.
For example, can corresponding leaf node be determined according to the feature and decision-tree model of forecast sample, by the leaf section
The output of point is as prediction output result.Such as (feature is divided according to the branch condition of decision tree using the feature of forecast sample
Characteristic value) determine current leaf node, take the leaf node output as predict result.It is defeated due to leaf node
Go out including each destination, therefore, can determine to need the destination pushed based on decision tree at this time.
For example, after gathering current multidimensional characteristic, it can be with the branch condition according to decision tree in decision tree shown in Fig. 4
Corresponding leaf node is searched as an2, the output of leaf node an2 is destination 2, at this point, the destination for just determining push is
Destination 2.
From the foregoing, it will be observed that the embodiment of the present application is by the way that when detecting that user determines destination, acquisition destination is corresponding
Multidimensional characteristic builds the corresponding sample set in destination as sample;According to feature for sample classification information gain to sample
This collection carries out sample classification, and to construct the decision-tree model of destination, the output of decision-tree model is corresponding destination;When
Detect that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;According to forecast sample and certainly
Plan tree-model predicts corresponding destination.The automatic push of destination are realized with this, the push for improving destination is accurate
Rate..
Further, due in each sample of sample set, including reflection user the behavior of destination is selected to practise usually
Used multiple characteristic informations, therefore the embodiment of the present application can so that the push to destination is more personalized and intelligent.
Further, realize that the push of destination is predicted based on decision tree prediction model, destination push can be promoted
Accuracy, more be bonded user use habit.
A kind of pusher of destination is additionally provided in one embodiment.Referring to Fig. 6, Fig. 6 is the embodiment of the present application
The structure diagram of the pusher of the destination of offer.Wherein the pusher of the destination is applied to electronic equipment, the mesh
Ground pusher include the first collecting unit 401, construction unit 402, the second collecting unit 403 and predicting unit 404,
It is as follows:
First collecting unit 401 is corresponding more for when detecting that user determines destination, gathering the destination
Dimensional feature builds the corresponding sample set in the destination as sample;
Construction unit 402, for carrying out sample to the sample set for the information gain of sample classification according to the feature
This classification, to construct the decision-tree model of the destination, the output of the decision-tree model is corresponding destination;
Second collecting unit 403 detects that user spreads out the map in application, gathering current corresponding multidimensional characteristic for working as
As forecast sample;
Predicting unit 404, for predicting corresponding destination according to the forecast sample and the decision-tree model.
In one embodiment, with reference to figure 7, construction unit 402 can include:
First node generates subelement 4021, for generating corresponding root node, and using the sample set as described
The nodal information of node;The sample set of the root node is determined as current target sample collection to be sorted;
Gain obtains subelement 4022, increases for obtaining the feature in target sample collection for the information that sample set is classified
Benefit;
Feature determination subelement 4023, for choosing current division from the feature according to described information gain selection
Feature;
Classification subelement 4024 for being divided according to the division feature to the sample set, obtains several increments
This collection;
Section point generates subelement 4025, and the division feature for concentrating sample to the subsample is gone
It removes, subsample collection after being removed;The child node of present node is generated, and using subsample collection after the removal as the sub- section
The nodal information of point;
Judgment sub-unit 4026, for judging whether child node meets default classification end condition, if it is not, then by the mesh
Mark sample set is updated to subsample collection after the removal, and triggers the gain and obtain in subelement execution acquisition target sample collection
The step of information gain that the feature is classified for sample set;If so, using the child node as leaf node, according to institute
Subsample concentrates the classification of sample to set the output of the leaf node after stating removal, and the classification of the sample is corresponding purpose
Ground.
Wherein, classification subelement 4024 can be used for obtaining in the sample set and divide the characteristic value of feature;
The sample set is divided according to the characteristic value.Identical sample is divided into identical subsample collection.
Wherein, feature determination subelement 4023, can be used for:
Maximum target information gain is chosen from described information gain;
Judge whether the target information gain is more than predetermined threshold value;
If so, the corresponding feature of the target information gain is chosen as current division feature.
In one embodiment, gain obtains subelement 4022, can be used for:
Obtain the empirical entropy of sample classification;
Obtain conditional entropy of the feature for sample set classification results;
According to the conditional entropy and the empirical entropy, the information gain that the feature is classified for the sample set is obtained.
For example, gain obtains subelement 4022, can be used for:Obtain each destination sample occur in sample set it is general
Rate, the destination sample are that sample class is corresponding destination, and the empirical entropy of sample is obtained according to each probability.
In one embodiment, judgment sub-unit 4025 can be used for judging subsample after the corresponding removal of the child node
Whether the categorical measure for concentrating sample is default quantity;
If so, determine that the child node meets default classification end condition.
In one embodiment, feature determination subelement 4023 can be also used for being not more than default threshold when target information gain
During value, using present node as leaf node, and the most sample class of sample size is chosen as the defeated of the leaf node
Go out.
In one embodiment, with reference to figure 7, the pusher of the destination further includes:
Detection unit 405, for detecting in the sample set with the presence or absence of the corresponding characteristic value of multidimensional characteristic is identical but mesh
The different sample of ground classification.
Stick unit 406, for when detecting in the sample set the identical but mesh there are the corresponding characteristic value of multidimensional characteristic
Ground classification different sample when, retain the most sample of destination classification sample size.
Wherein, the step of each unit performs in the pusher of destination may be referred to the side of above method embodiment description
Method step.The pusher of the destination can integrate in the electronic device, such as mobile phone, tablet computer.
When it is implemented, Yi Shang unit can be independent entity realization, can also be combined, as
Same or several entities realize, more than the specific implementation of each unit can be found in the embodiment of front, details are not described herein.
From the foregoing, it will be observed that the pusher of the present embodiment destination can be by detecting user when the first collecting unit 401
When destination is determined, the corresponding multidimensional characteristic in acquisition destination builds the corresponding sample set in destination as sample;Structure
Unit 402 carries out sample classification for the information gain of sample classification according to feature to sample set, to construct determining for destination
Plan tree-model, the output of decision-tree model is corresponding destination;It is answered when the second collecting unit 403 detects that user spreads out the map
Used time gathers current corresponding multidimensional characteristic as forecast sample;Predicting unit 404 is pre- according to forecast sample and decision-tree model
Measure corresponding destination.The automatic push of destination are realized with this, improve the push accuracy rate of destination.
The embodiment of the present application also provides a kind of electronic equipment.Referring to Fig. 8, electronic equipment 500 include processor 501 and
Memory 502.Wherein, processor 501 is electrically connected with memory 502.
The processor 500 is the control centre of electronic equipment 500, is set using various interfaces and the entire electronics of connection
Standby various pieces computer program in memory 502 and are called by operation or load store and are stored in memory
Data in 502 perform the various functions of electronic equipment 500 and handle data, so as to carry out whole prison to electronic equipment 500
Control.
The memory 502 can be used for storage software program and module, and processor 501 is stored in memory by operation
502 computer program and module, so as to perform various functions application and data processing.Memory 502 can mainly include
Storing program area and storage data field, wherein, storing program area can storage program area, the computer needed at least one function
Program (such as sound-playing function, image player function etc.) etc.;Storage data field can be stored uses institute according to electronic equipment
Data of establishment etc..In addition, memory 502 can include high-speed random access memory, non-volatile memories can also be included
Device, for example, at least a disk memory, flush memory device or other volatile solid-state parts.Correspondingly, memory
502 can also include Memory Controller, to provide access of the processor 501 to memory 502.
In the embodiment of the present application, the processor 501 in electronic equipment 500 can be according to the steps, by one or one
The corresponding instruction of process of a above computer program is loaded into memory 502, and is stored in by the operation of processor 501
Computer program in reservoir 502, it is as follows so as to fulfill various functions:
When detecting that user determines destination, the corresponding multidimensional characteristic in the destination is gathered as sample, and structure
Build the corresponding sample set in the destination;
Sample classification is carried out to the sample set for the information gain of sample classification according to the feature, to construct
The decision-tree model of destination is stated, the output of the decision-tree model is corresponding destination;
When detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;
Corresponding destination is predicted according to the forecast sample and the decision-tree model.
In some embodiments, the sample set is being carried out for the information gain of sample classification according to the feature
Sample classification, during constructing the decision-tree model of the application, processor 501 can specifically perform following steps:
Corresponding root node is generated, and using the sample set as the nodal information of the root node;
The sample set of the root node is determined as current target sample collection to be sorted;
Obtain the information gain that the feature is classified for sample set in target sample collection;
Current division feature is chosen from the feature according to described information gain;
The sample set is divided according to the division feature, obtains several subsample collection;
The division feature for concentrating sample to the subsample is removed, subsample collection after being removed;
The child node of present node is generated, and using subsample collection after the removal as the nodal information of the child node;
Judge whether child node meets default classification end condition;
If it is not, the target sample collection then is updated to subsample collection after the removal, and returns to execution and obtain target sample
The step of information gain that the feature is classified for sample set in this collection;
If so, using the child node as leaf node, the classification for concentrating sample according to subsample after the removal is set
The output of the leaf node is put, the classification of the sample is corresponding destination.
In some embodiments, according to gathering the corresponding multidimensional characteristic in the destination as sample, and institute is built
After stating the corresponding sample set in destination, processor 501 can also specifically perform following steps:
It detects in the sample set with the presence or absence of the sample that the corresponding characteristic value of multidimensional characteristic is identical but destination classification is different
This;
When detecting samples identical there are the corresponding characteristic value of multidimensional characteristic in the sample set but that destination classification is different
This when, retains the most sample of destination classification sample size.
In some embodiments, when choosing current division feature from the feature according to described information gain,
Processor 501 can specifically perform following steps:
Maximum target information gain is chosen from described information gain;
Judge whether the target information gain is more than predetermined threshold value;
If so, the corresponding feature of the target information gain is chosen as current division feature.
In some embodiments, processor 501 can also specifically perform following steps:
When target information gain is not more than predetermined threshold value, using present node as leaf node, and sample size is chosen
Output of most sample class as the leaf node.
In some embodiments, the information gain that the feature is classified for sample set in target sample collection is being obtained
When, processor 501 can specifically perform following steps:
Obtain the empirical entropy of sample classification;
Obtain conditional entropy of the feature for sample set classification results;
According to the conditional entropy and the empirical entropy, the information gain that the feature is classified for the sample set is obtained.
It can be seen from the above, the electronic equipment of the embodiment of the present application, by when detecting that user determines destination, gathering
The corresponding multidimensional characteristic in destination builds the corresponding sample set in destination as sample;According to feature for sample classification
Information gain carries out sample classification to sample set, and to construct the decision-tree model of destination, the output of decision-tree model is pair
The destination answered;When detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;According to
Forecast sample and decision-tree model predict corresponding destination.The automatic push of destination are realized with this, improve purpose
The push accuracy rate on ground.
Also referring to Fig. 9, in some embodiments, electronic equipment 500 can also include:Display 503, radio frequency electrical
Road 504, voicefrequency circuit 505 and power supply 506.Wherein, wherein, display 503, radio circuit 504, voicefrequency circuit 505 and
Power supply 506 is electrically connected respectively with processor 501.
The display 503 is displayed for by information input by user or is supplied to the information of user and various figures
Shape user interface, these graphical user interface can be made of figure, text, icon, video and its any combination.Display
503 can include display panel, in some embodiments, liquid crystal display (Liquid Crystal may be employed
Display, LCD) or the forms such as Organic Light Emitting Diode (Organic Light-Emitting Diode, OLED) match somebody with somebody
Put display panel.
The radio circuit 504 can be used for transceiving radio frequency signal, to pass through wireless communication and the network equipment or other electricity
Sub- equipment establishes wireless telecommunications, the receiving and transmitting signal between the network equipment or other electronic equipments.
The voicefrequency circuit 505 can be used for providing the audio between user and electronic equipment by loud speaker, microphone
Interface.
The power supply 506 is used to all parts power supply of electronic equipment 500.In some embodiments, power supply 506
Can be logically contiguous by power-supply management system and processor 501, so as to realize management charging by power-supply management system, put
The functions such as electricity and power managed.
Although not shown in Fig. 9, electronic equipment 500 can also include camera, bluetooth module etc., and details are not described herein.
The embodiment of the present application also provides a kind of storage medium, and the storage medium is stored with computer program, when the meter
When calculation machine program is run on computers so that the computer performs the push side of the destination in any of the above-described embodiment
Method, such as:When detecting that user determines destination, the corresponding multidimensional characteristic in the destination is gathered as sample, and structure
Build the corresponding sample set in the destination;Sample is carried out to the sample set for the information gain of sample classification according to the feature
This classification, to construct the decision-tree model of the destination, the output of the decision-tree model is corresponding destination;Work as inspection
User is measured to spread out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;According to the forecast sample and
The decision-tree model predicts corresponding destination.
In the embodiment of the present application, storage medium can be magnetic disc, CD, read-only memory (Read Only Memory,
ROM) or random access memory (Random Access Memory, RAM) etc..
In the above-described embodiments, all emphasize particularly on different fields to the description of each embodiment, there is no the portion being described in detail in some embodiment
Point, it may refer to the associated description of other embodiment.
It should be noted that for the method for pushing of the destination of the embodiment of the present application, this field common test personnel
It is appreciated that realizing all or part of flow of the method for pushing of the destination of the embodiment of the present application, being can be by computer journey
Sequence controls relevant hardware to complete, and the computer program can be stored in a computer read/write memory medium, such as deposited
Storage performs in the memory of electronic equipment, and by least one processor in the electronic equipment, can wrap in the process of implementation
Include the flow of the embodiment of the method for pushing such as destination.Wherein, the storage medium can be magnetic disc, CD, read-only storage
Device, random access memory etc..
For the pusher of the destination of the embodiment of the present application, each function module can be integrated in a processing core
In piece or modules are individually physically present, can also two or more modules be integrated in a module.On
The form realization that hardware had both may be employed in integrated module is stated, can also be realized in the form of software function module.The collection
If into module realized in the form of software function module and be independent production marketing or in use, can also be stored in
In one computer read/write memory medium, the storage medium is for example read-only memory, disk or CD etc..
Method for pushing, device, storage medium and the electronics of a kind of destination provided above the embodiment of the present application are set
Standby to be described in detail, the principle and implementation of this application are described for specific case used herein, more than
The explanation of embodiment is only intended to help to understand the present processes and its core concept;Meanwhile for those skilled in the art
Member, according to the thought of the application, there will be changes in specific embodiments and applications, in conclusion this explanation
Book content should not be construed as the limitation to the application.
Claims (16)
1. a kind of method for pushing of destination, which is characterized in that including:
When detecting that user determines destination, the corresponding multidimensional characteristic in the destination is gathered as sample, and builds institute
State the corresponding sample set in destination;
Sample classification is carried out to the sample set for the information gain of sample classification according to the feature, to construct the mesh
Ground decision-tree model, the output of the decision-tree model is corresponding destination;
When detecting that user spreads out the map in application, gathering current corresponding multidimensional characteristic as forecast sample;
Corresponding destination is predicted according to the forecast sample and the decision-tree model.
2. the method for pushing of destination as described in claim 1, which is characterized in that according to the feature for sample classification
Information gain carries out sample classification to the sample set, to construct the decision-tree model of the application, including:
Corresponding root node is generated, and using the sample set as the nodal information of the root node;
The sample set of the root node is determined as current target sample collection to be sorted;
Obtain the information gain that the feature is classified for sample set in target sample collection;
Current division feature is chosen from the feature according to described information gain;
The sample set is divided according to the division feature, obtains several subsample collection;
The division feature for concentrating sample to the subsample is removed, subsample collection after being removed;
The child node of present node is generated, and using subsample collection after the removal as the nodal information of the child node;
Judge whether child node meets default classification end condition;
If it is not, the target sample collection then is updated to subsample collection after the removal, and returns to execution and obtain target sample collection
The step of information gain that the interior feature is classified for sample set;
If so, using the child node as leaf node, the classification for concentrating sample according to subsample after the removal sets institute
The output of leaf node is stated, the classification of the sample is corresponding destination.
3. the method for pushing of destination as claimed in claim 2, which is characterized in that the destination is corresponding more according to gathering
Dimensional feature is further included as sample, and after building the corresponding sample set in the destination:
It detects in the sample set with the presence or absence of the sample that the corresponding characteristic value of multidimensional characteristic is identical but destination classification is different;
When detecting samples identical there are the corresponding characteristic value of multidimensional characteristic in the sample set but that destination classification is different,
Retain the most sample of destination classification sample size.
4. the method for pushing of destination as claimed in claim 2, which is characterized in that according to described information gain from the feature
It is middle to choose current division feature, including:
Maximum target information gain is chosen from described information gain;
Judge whether the target information gain is more than predetermined threshold value;
If so, the corresponding feature of the target information gain is chosen as current division feature.
5. the method for pushing of destination as claimed in claim 4, which is characterized in that the method for pushing of the destination also wraps
It includes:
When target information gain is not more than predetermined threshold value, using present node as leaf node, and it is most to choose sample size
Output of the sample class as the leaf node.
6. the method for pushing of destination as claimed in claim 2, which is characterized in that judge whether child node meets default classification
End condition, including:
Subsample concentrates whether the categorical measure of sample is default quantity after judging the corresponding removal of the child node;
If so, determine that the child node meets default classification end condition.
7. such as the method for pushing of claim 2-6 any one of them destination, which is characterized in that obtain institute in target sample collection
The information gain that feature is classified for sample set is stated, including:
Obtain the empirical entropy of sample classification;
Obtain conditional entropy of the feature for sample set classification results;
According to the conditional entropy and the empirical entropy, the information gain that the feature is classified for the sample set is obtained.
8. the method for pushing of destination as claimed in claim 7, which is characterized in that according to the conditional entropy and the experience
Entropy obtains the information gain that the feature is classified for the sample set, including:
G (Y, X)=H (Y)-H (Y | X)
Wherein, g (Y, X) is characterized the information gain that X classifies for sample set Y, the empirical entropy that H (Y) classifies for sample set Y, H (Y
| X) it is characterized conditional entropies of the X for sample set Y classification results.
9. the method for pushing of destination as claimed in claim 8, which is characterized in that the empirical entropy of sample classification is obtained, including:
The probability that each destination sample occurs in the sample set is obtained, the destination sample is that sample class is corresponding
Destination;
The empirical entropy of the sample is obtained according to the probability.
10. a kind of pusher of destination, which is characterized in that including:
First collecting unit, for when detecting that user determines destination, gathering the corresponding multidimensional characteristic in the destination
As sample, and build the corresponding sample set in the destination;
Construction unit, for carrying out sample classification to the sample set for the information gain of sample classification according to the feature,
To construct the decision-tree model of the destination, the output of the decision-tree model is corresponding destination;
Second collecting unit detects that user spreads out the map in application, gathering current corresponding multidimensional characteristic as pre- for working as
Test sample sheet;
Predicting unit, for predicting corresponding destination according to the forecast sample and the decision-tree model.
11. the pusher of destination as claimed in claim 10, which is characterized in that the construction unit includes:
First node generates subelement, for generating corresponding root node, and using the sample set as the section of the root node
Point information;The sample set of the root node is determined as current target sample collection to be sorted;
Gain obtains subelement, for obtaining the information gain that the feature is classified for sample set in target sample collection;
Feature determination subelement, for choosing current division feature from the feature according to described information gain selection;
Classification subelement for being divided according to the division feature to the sample set, obtains several subsample collection;
Section point generates subelement, and the division feature for concentrating sample to the subsample is removed, and is gone
Except rear subsample collection;The child node of present node is generated, and using subsample collection after the removal as the node of the child node
Information;
Judgment sub-unit, for judging whether child node meets default classification end condition, if it is not, then by the target sample collection
Subsample collection after the removal is updated to, and triggers the gain and obtains subelement and perform and obtain the feature in target sample collection
For sample set classification information gain the step of;If so, using the child node as leaf node, after the removal
Subsample concentrates the classification of sample to set the output of the leaf node, and the classification of the sample is corresponding destination.
12. the pusher of destination as claimed in claim 11, which is characterized in that described device further includes:
Detection unit, for detecting in the sample set with the presence or absence of the corresponding characteristic value of multidimensional characteristic is identical but destination classification
Different samples;
Stick unit, for when detecting in the sample set the identical but destination classification there are the corresponding characteristic value of multidimensional characteristic
During different sample, retain the most sample of destination classification sample size.
13. the pusher of destination as claimed in claim 11, which is characterized in that feature determination subelement is used for:
Maximum target information gain is chosen from described information gain;
Judge whether the target information gain is more than predetermined threshold value;
If so, the corresponding feature of the target information gain is chosen as current division feature.
14. the pusher of destination as claimed in claim 11, which is characterized in that the gain obtains subelement, is used for:
Obtain the empirical entropy of sample classification;
Obtain conditional entropy of the feature for sample set classification results;
According to the conditional entropy and the empirical entropy, the information gain that the feature is classified for the sample set is obtained.
15. a kind of storage medium, is stored thereon with computer program, which is characterized in that when the computer program is in computer
During upper operation so that the computer performs the method for pushing of destination as described in any one of claim 1 to 9.
16. a kind of electronic equipment, including processor and memory, the memory has computer program, which is characterized in that described
Processor is by calling the computer program, for performing the push side of destination as described in any one of claim 1 to 9
Method.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711461519.XA CN108108455B (en) | 2017-12-28 | 2017-12-28 | Destination pushing method and device, storage medium and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201711461519.XA CN108108455B (en) | 2017-12-28 | 2017-12-28 | Destination pushing method and device, storage medium and electronic equipment |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108108455A true CN108108455A (en) | 2018-06-01 |
CN108108455B CN108108455B (en) | 2020-06-16 |
Family
ID=62214189
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201711461519.XA Active CN108108455B (en) | 2017-12-28 | 2017-12-28 | Destination pushing method and device, storage medium and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108108455B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109635069A (en) * | 2018-12-21 | 2019-04-16 | 北京航天泰坦科技股份有限公司 | A kind of geographical spatial data self-organizing method based on comentropy |
CN110347760A (en) * | 2019-05-30 | 2019-10-18 | 中国地质大学(武汉) | A kind of data analysing method for the crowd's space-time positioning service that is missing |
CN110363304A (en) * | 2019-06-18 | 2019-10-22 | 深圳壹账通智能科技有限公司 | Location model construction method, device, computer equipment and storage medium |
CN110647929A (en) * | 2019-09-19 | 2020-01-03 | 京东城市(北京)数字科技有限公司 | Method for predicting travel destination and method for training classifier |
CN110689158A (en) * | 2018-07-05 | 2020-01-14 | 北京嘀嘀无限科技发展有限公司 | Method, device and storage medium for predicting destination |
WO2020042579A1 (en) * | 2018-08-27 | 2020-03-05 | 平安科技(深圳)有限公司 | Group classification method and device, electronic device, and storage medium |
CN110986985A (en) * | 2019-12-17 | 2020-04-10 | 广州小鹏汽车科技有限公司 | Vehicle travel pushing method and device, medium, control terminal and automobile |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102214213A (en) * | 2011-05-31 | 2011-10-12 | 中国科学院计算技术研究所 | Method and system for classifying data by adopting decision tree |
CN105868298A (en) * | 2016-03-23 | 2016-08-17 | 华南理工大学 | Mobile phone game recommendation method based on binary decision tree |
CN106407406A (en) * | 2016-09-22 | 2017-02-15 | 国信优易数据有限公司 | A text processing method and system |
CN106557846A (en) * | 2016-11-30 | 2017-04-05 | 成都寻道科技有限公司 | Based on university students school data graduation whereabouts Forecasting Methodology |
CN106934412A (en) * | 2015-12-31 | 2017-07-07 | 中国科学院深圳先进技术研究院 | A kind of user behavior sorting technique and system |
-
2017
- 2017-12-28 CN CN201711461519.XA patent/CN108108455B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102214213A (en) * | 2011-05-31 | 2011-10-12 | 中国科学院计算技术研究所 | Method and system for classifying data by adopting decision tree |
CN106934412A (en) * | 2015-12-31 | 2017-07-07 | 中国科学院深圳先进技术研究院 | A kind of user behavior sorting technique and system |
CN105868298A (en) * | 2016-03-23 | 2016-08-17 | 华南理工大学 | Mobile phone game recommendation method based on binary decision tree |
CN106407406A (en) * | 2016-09-22 | 2017-02-15 | 国信优易数据有限公司 | A text processing method and system |
CN106557846A (en) * | 2016-11-30 | 2017-04-05 | 成都寻道科技有限公司 | Based on university students school data graduation whereabouts Forecasting Methodology |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110689158A (en) * | 2018-07-05 | 2020-01-14 | 北京嘀嘀无限科技发展有限公司 | Method, device and storage medium for predicting destination |
CN110689158B (en) * | 2018-07-05 | 2022-05-31 | 北京嘀嘀无限科技发展有限公司 | Method, device and storage medium for predicting destination |
WO2020042579A1 (en) * | 2018-08-27 | 2020-03-05 | 平安科技(深圳)有限公司 | Group classification method and device, electronic device, and storage medium |
CN109635069A (en) * | 2018-12-21 | 2019-04-16 | 北京航天泰坦科技股份有限公司 | A kind of geographical spatial data self-organizing method based on comentropy |
CN109635069B (en) * | 2018-12-21 | 2021-08-10 | 北京航天泰坦科技股份有限公司 | Geographic space data self-organizing method based on information entropy |
CN110347760A (en) * | 2019-05-30 | 2019-10-18 | 中国地质大学(武汉) | A kind of data analysing method for the crowd's space-time positioning service that is missing |
CN110347760B (en) * | 2019-05-30 | 2021-07-09 | 中国地质大学(武汉) | Data analysis method for lost crowd space-time positioning service |
CN110363304A (en) * | 2019-06-18 | 2019-10-22 | 深圳壹账通智能科技有限公司 | Location model construction method, device, computer equipment and storage medium |
CN110647929A (en) * | 2019-09-19 | 2020-01-03 | 京东城市(北京)数字科技有限公司 | Method for predicting travel destination and method for training classifier |
CN110986985A (en) * | 2019-12-17 | 2020-04-10 | 广州小鹏汽车科技有限公司 | Vehicle travel pushing method and device, medium, control terminal and automobile |
CN110986985B (en) * | 2019-12-17 | 2022-07-12 | 广州小鹏汽车科技有限公司 | Vehicle travel pushing method and device, medium, control terminal and automobile |
Also Published As
Publication number | Publication date |
---|---|
CN108108455B (en) | 2020-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108108455A (en) | Method for pushing, device, storage medium and the electronic equipment of destination | |
CN107704070A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN107678845A (en) | Application program management-control method, device, storage medium and electronic equipment | |
CN107678799B (en) | Application program management-control method, device, storage medium and electronic equipment | |
CN105320957B (en) | Classifier training method and device | |
CN108280458B (en) | Group relation type identification method and device | |
CN109948641A (en) | Anomaly groups recognition methods and device | |
CN108197225A (en) | Sorting technique, device, storage medium and the electronic equipment of image | |
CN109961077A (en) | Gender prediction's method, apparatus, storage medium and electronic equipment | |
CN107894827A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN108337358A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN107678531A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN107704289A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN107678800A (en) | Background application method for cleaning, device, storage medium and electronic equipment | |
CN107077489A (en) | Automatic for multidimensional data is seen clearly | |
US11010220B2 (en) | System and methods for decomposing events from managed infrastructures that includes a feedback signalizer functor | |
CN109961075A (en) | User gender prediction method, apparatus, medium and electronic equipment | |
CN111797288A (en) | Data screening method and device, storage medium and electronic equipment | |
CN107807730B (en) | Using method for cleaning, device, storage medium and electronic equipment | |
CN111562541B (en) | Software platform for realizing electric energy meter detection data management by applying CART algorithm | |
CN109961163A (en) | Gender prediction's method, apparatus, storage medium and electronic equipment | |
CN107643925A (en) | Background application method for cleaning, device, storage medium and electronic equipment | |
CN107943537A (en) | Using method for cleaning, device, storage medium and electronic equipment | |
US20160330066A1 (en) | System and methods for decomposing events from managed infrastructures that includes a floating point unit | |
CN107704364B (en) | Background application management-control method, device, storage medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
CB02 | Change of applicant information | ||
CB02 | Change of applicant information |
Address after: Changan town in Guangdong province Dongguan 523860 usha Beach Road No. 18 Applicant after: GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., Ltd. Address before: Changan town in Guangdong province Dongguan 523860 usha Beach Road No. 18 Applicant before: GUANGDONG OPPO MOBILE TELECOMMUNICATIONS CORP., Ltd. |
|
GR01 | Patent grant | ||
GR01 | Patent grant |