CN107392318A - Complex machines learning model means of interpretation and device based on local linearization - Google Patents

Complex machines learning model means of interpretation and device based on local linearization Download PDF

Info

Publication number
CN107392318A
CN107392318A CN201710620391.0A CN201710620391A CN107392318A CN 107392318 A CN107392318 A CN 107392318A CN 201710620391 A CN201710620391 A CN 201710620391A CN 107392318 A CN107392318 A CN 107392318A
Authority
CN
China
Prior art keywords
mrow
msub
munder
msubsup
learning model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710620391.0A
Other languages
Chinese (zh)
Inventor
郑乐
胡伟
李勇
王春明
徐遐龄
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
STATE GRID CENTER CHINA GRID Co Ltd
Tsinghua University
State Grid Corp of China SGCC
Original Assignee
STATE GRID CENTER CHINA GRID Co Ltd
Tsinghua University
State Grid Corp of China SGCC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by STATE GRID CENTER CHINA GRID Co Ltd, Tsinghua University, State Grid Corp of China SGCC filed Critical STATE GRID CENTER CHINA GRID Co Ltd
Priority to CN201710620391.0A priority Critical patent/CN107392318A/en
Publication of CN107392318A publication Critical patent/CN107392318A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N7/00Computing arrangements based on specific mathematical models
    • G06N7/01Probabilistic graphical models, e.g. probabilistic networks

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Pure & Applied Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Algebra (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Other Investigation Or Analysis Of Materials By Electrical Means (AREA)

Abstract

The invention discloses a kind of complex machines learning model means of interpretation and device based on local linearization, wherein, method includes:Collecting sample concentrates any point as sample point, and stochastical sampling obtains multiple sampled points around sample point;In expression of space, Euler's distance between sample point and each sampled point is obtained, using the weight as each sampled point;Machine learning model to be explained is obtained according to the weight of each sampled point and linear model and explains the gap of the fitting result of function, to obtain optimization problem;The linear regression problem of regularization term penalty factor is used in Optimization Solution optimization problem, and obtains explanation results.This method can explain in each data neighborhood of a point to complex machines learning model, take into full account the local characteristicses of sample space, the dominant characteristics of sample space different zones not only can be effectively found, and it is more directly perceived, convenient, it is applicable to the explanation of a variety of machine learning models.

Description

Complex machines learning model means of interpretation and device based on local linearization
Technical field
The present invention relates to machine learning application and analysis technical field, more particularly to a kind of complexity based on local linearization Machine learning model means of interpretation and device.
Background technology
At the beginning of machine learning field is started, researchers begin to inquire into the explanation of machine learning algorithm (Interpretability/Comprehensibility) problem.Here so-called " explanation ", it is containing for machine learning field Justice, it is desirable to provide the quantitative relationship (qualitative understanding) between input variable and model output.Researcher It is believed that the precision of model, complexity and interpretation into opposite relation, i.e. the interpretation of naive model is strong, but It is that precision is relatively low;And complex model can obtain higher computational accuracy, it can be difficult to visual interpretation.
At present, researcher is more likely to first draw the higher model of precision using complex model, then utilizes naive model Obtained high-precision model is explained, i.e., goes to be fitted the output valve of complex model using naive model.Instructed in correlation technique Practice explanation of the decision-tree model as neural network model, tree-model complexity is used to indicate the interpretable energy of the interpretation model Power.(Validity-Interval Analysis) is analyzed to be carried out to neural network model by valid interval in correlation technique Explain, it is consistent with the general principle of decision-tree model.Correlation technique pilot scale uses first order logic expression (First-order Logic Formulate) and two kinds of algorithms such as Bayesian network (Bayesian Network) explain matrix decomposition algorithm.Always From the point of view of knot, the means of interpretation in correlation technique utilizes such as simple mould of linear model, decision-tree model in all input spaces The characteristics of type is to explain complex model, but can not consider input space part, which is that means of interpretation is maximum in correlation technique, asks Topic, has much room for improvement.
The content of the invention
It is contemplated that at least solves one of technical problem in correlation technique to a certain extent.
Therefore, it is an object of the present invention to propose a kind of complex machines learning model explanation based on local linearization Method, this method not only can effectively find the dominant characteristics of sample space different zones, and more directly perceived, convenient, can fit Explanation for a variety of machine learning models.
It is another object of the present invention to propose that a kind of complex machines learning model based on local linearization is explained to fill Put.
To reach above-mentioned purpose, one aspect of the present invention embodiment proposes a kind of complex machines based on local linearization Model explanation method is practised, including:Collecting sample concentrates any point as sample point, and the stochastical sampling around the sample point Multiple sampled points are obtained, and obtain according to former machine learning model the machine learning mould of each sampled point of the multiple sampled point Type prediction result;In expression of space, Euler's distance between the sample point and each sampled point is obtained, using as described The weight of each sampled point;According to the weight of each sampled point and linear model obtain machine learning model to be explained and The gap of the fitting result of function is explained, and obtains the complexity of the explanation function, to obtain optimization problem;Optimization Solution institute The linear regression problem that regularization term penalty factor is used in optimization problem is stated, and obtains explanation results.
The complex machines learning model means of interpretation based on local linearization of the embodiment of the present invention, can be in each data Complex machines learning model is explained in neighborhood of a point, takes into full account the local characteristicses of sample space, not only can be effective The dominant characteristics of sample space different zones are found, and it is more directly perceived, convenient, it is applicable to the solution of a variety of machine learning models Release.
In addition, the complex machines learning model means of interpretation according to the above embodiment of the present invention based on local linearization is also There can be following additional technical characteristic:
Further, in one embodiment of the invention, the weight of each sampled point is:
Wherein, xiFor the sample point,For the multiple sampled point, σ is the standard deviation of all distances.
Further, in one embodiment of the invention, the gap of the fitting result is:
Wherein, f is the machine learning model to be explained, and g is the explanation function, and Γ is the difference of the fitting result Away from;It is described explain function complexity be:
Wherein, Ω is the complexity.
Further, in one embodiment of the invention, the optimization problem is:
Alternatively, in one embodiment of the invention, by SGD (Saccharomyces Genome Database, Stochastic gradient descent) linear regression problem described in Algorithm for Solving.
To reach above-mentioned purpose, another aspect of the present invention embodiment proposes a kind of complex machines based on local linearization Learning model interpreting means, including:Acquisition module, any point is concentrated as sample point for collecting sample, and in the sample This point surrounding stochastical sampling obtains multiple sampled points, and obtains each of the multiple sampled point according to former machine learning model and adopt The machine learning model prediction result of sampling point;First acquisition module, in expression of space, obtain the sample point with it is described Euler's distance between each sampled point, using the weight as each sampled point;Second acquisition module, for according to described every The weight and linear model of individual sampled point obtain machine learning model to be explained and explain the gap of the fitting result of function, and The complexity of the explanation function is obtained, to obtain optimization problem;Explanation module, for making in optimization problem described in Optimization Solution With the linear regression problem of regularization term penalty factor, and obtain explanation results.
The complex machines learning model interpreting means based on local linearization of the embodiment of the present invention, can be in each data Complex machines learning model is explained in neighborhood of a point, takes into full account the local characteristicses of sample space, not only can be effective The dominant characteristics of sample space different zones are found, and it is more directly perceived, convenient, it is applicable to the solution of a variety of machine learning models Release.
In addition, the complex machines learning model interpreting means according to the above embodiment of the present invention based on local linearization are also There can be following additional technical characteristic:
Further, in one embodiment of the invention, the weight of each sampled point is:
Wherein, xiFor the sample point,For the multiple sampled point, σ is the standard deviation of all distances.
Further, in one embodiment of the invention, in addition to:The gap of the fitting result is:
Wherein, f is the machine learning model to be explained, and g is the explanation function, and Γ is the difference of the fitting result Away from;It is described explain function complexity be:
Wherein, Ω is the complexity.
Further, in one embodiment of the invention, the optimization problem is:
Alternatively, in one embodiment of the invention, the explanation module passes through linear regression described in SGD Algorithm for Solving Problem.
The additional aspect of the present invention and advantage will be set forth in part in the description, and will partly become from the following description Obtain substantially, or recognized by the practice of the present invention.
Brief description of the drawings
Of the invention above-mentioned and/or additional aspect and advantage will become from the following description of the accompanying drawings of embodiments Substantially and it is readily appreciated that, wherein:
Fig. 1 is the complex machines learning model means of interpretation based on local linearization according to one embodiment of the invention Flow chart;And
Fig. 2 is the complex machines learning model interpreting means based on local linearization according to one embodiment of the invention Structural representation.
Embodiment
Embodiments of the invention are described below in detail, the example of the embodiment is shown in the drawings, wherein from beginning to end Same or similar label represents same or similar element or the element with same or like function.Below with reference to attached The embodiment of figure description is exemplary, it is intended to for explaining the present invention, and is not considered as limiting the invention.
The study mould of the complex machines based on local linearization proposed according to embodiments of the present invention is described with reference to the accompanying drawings Type means of interpretation and device, the complexity based on local linearization proposed according to embodiments of the present invention is described with reference to the accompanying drawings first Machine learning model means of interpretation.
Fig. 1 is the flow of the complex machines learning model means of interpretation based on local linearization of one embodiment of the invention Figure.
Comprise the following steps as shown in figure 1, being somebody's turn to do the complex machines learning model means of interpretation based on local linearization:
In step S101, collecting sample concentrates any point as sample point, and stochastical sampling obtains around sample point To multiple sampled points, and obtain according to former machine learning model the machine learning model prediction of each sampled point of multiple sampled points As a result.
It is understood that in an embodiment of the present invention, any point of this concentration is sampled first as sample point xi, Stochastical sampling obtains the i.e. multiple sampled points of N number of point around the sample point, is designated as respectivelyAnd utilize former machine Device learning model, calculate the machine learning model prediction result of each sampled pointIt should be noted that the present invention is implemented The method of example does local explanation to machine learning model in each data vertex neighborhood, can take into full account the different input spaces The otherness of middle feature, realize that the local linearization of machine learning model is explained.
In step s 102, in expression of space, Euler's distance between sample point and each sampled point is obtained, using as every The weight of individual sampled point.
Wherein, in one embodiment of the invention, the weight of each sampled point can be:
Wherein, xiFor sample point,For multiple sampled points, σ is the standard deviation of all distances.
It is understood that in expression of space, Euler's distance between sample point and sampled point is calculated, is designated asWherein, σ is the standard deviation of all distances.It should be noted that sampled point is nearer with being explained point, distance is got over Small, weight is bigger, plays a part of in the study for explaining expression bigger.It should be noted that the embodiment of the present invention uses Europe Draw the purpose of distance:First, in order to be confined to explain analysis in neighborhood, second, to strengthening the robustness to sampling noiset.
In step s 103, according to the weight of each sampled point and linear model obtain machine learning model to be explained and The gap of the fitting result of function is explained, and obtains explaining the complexity of function, to obtain optimization problem.
Alternatively, in one embodiment of the invention, the gap of fitting result is:
Wherein, f is machine learning model to be explained, and for g to explain function, Γ is the gap of fitting result;Explain function Complexity be:
Wherein, Ω is complexity.
Alternatively, in one embodiment of the invention, optimization problem is:
It is understood that the method for the embodiment of the present invention in order in each data vertex neighborhood to machine learning model Local explanation is done, takes into full account the otherness of feature in the different input spaces, realizes the local linear neutralizing of machine learning model Release, it is f to assume initially that the machine learning model to be explained, simple to explain that function is g, is originally inputted as x, then to machine learning Solution to model, which is released, to be equivalent to solve g to meet formula 1:
g(x)≈f(x)。 (1)
In an embodiment of the present invention, the requirement to explaining function g has at 2 points, and one is accuracy (Fidelity), i.e., Explain that the model that function g learns wants faithful to master mould f;Another is interpretation, that is, explains that function g form will to the greatest extent can Can be simple.Requirement of the embodiment of the present invention to explaining function g can be converted into optimization problem as shown in Equation 2:
Min ξ (x)=Γ (f, g)+Ω (g), (2)
Wherein, Γ represents f and g fitting results gap, and it has reacted g accuracy, and the smaller then g of Γ values is closer to f;Ω G complexity is represented, is represented as decision-tree model with the number of decision tree leaf node, linear model typically uses non-zero The number of coefficient represents.
It should be noted that, although the parameter that can explain function is trained on all samples, but the table of simple function Danone power is general, can not meet the requirement of accuracy.For this defect, the method for the embodiment of the present invention proposes local explanation The thinking of model, i.e., for arbitrary sample point xi, train to obtain in the sample neighborhood of a point and explain function gi, present invention implementation Example will be converted into master mould f explanation solves a series of local explanation function g=[gi].The method of the embodiment of the present invention uses Linear model, which is used as, explains function, referred to as LLI (LocalLinear Interpretability, local linearization are explained), such as Shown in formula 3;
In step S104, the linear regression problem of regularization term penalty factor is used in Optimization Solution optimization problem, and Obtain explanation results.
It is understood that in an embodiment of the present invention, define Γ (f, g) and Ω (g) such as formula 4 and public affairs respectively first Shown in formula 5;
Further, formula 2 can be rewritten as:
As shown in Equation 6, left side item is square error sum term, and the right item is L1Regularization term, therefore the present invention is implemented Local linear can be explained and is converted into using L by example1The linear regression problem of penalty factor, i.e. Lasso problems.
Alternatively, in one embodiment of the invention, SGD Algorithm for Solving linear regression problems are passed through.
It is understood that being directed to above-mentioned Lasso problems, the method for the embodiment of the present invention can be entered using SGD algorithms Row Optimization Solution.Meanwhile L1Regularization has the effect for making model coefficient rarefaction so that the explanation mould that the embodiment of the present invention obtains The number of type nonzero coefficient is effective, effectively excludes the interference of useless feature, adds the solution to model property released.
The complex machines learning model means of interpretation based on local linearization proposed according to embodiments of the present invention, Ke Yi Complex machines learning model is explained in each data neighborhood of a point, takes into full account the local characteristicses of sample space, not only The dominant characteristics of sample space different zones can be effectively found, and it is more directly perceived, convenient, it is applicable to a variety of machine learning Solution to model is released.
The complex machines based on local linearization proposed according to embodiments of the present invention referring next to accompanying drawing description learn mould Type interpreting means.
Fig. 2 is the structure of the complex machines learning model interpreting means based on local linearization of one embodiment of the invention Schematic diagram.
As shown in Fig. 2 being somebody's turn to do the complex machines learning model interpreting means 10 based on local linearization includes:Acquisition module 100th, the first acquisition module 200, the second acquisition module 300 and explanation module 400.
Wherein, acquisition module 100 concentrates any point as sample point for collecting sample, and in alternatively sample point week Enclose stochastical sampling and obtain multiple sampled points, and each sampled point of alternatively multiple sampled points is obtained according to former machine learning model Machine learning model prediction result.First acquisition module 200 be used in expression of space, obtain alternatively sample point with it is optional Euler's distance between each sampled point in ground, with the weight of optionally each sampled point.Second acquisition module 300 is used for basis Alternatively the weight of each sampled point and linear model obtain machine learning model to be explained and explain the fitting result of function Gap, and the complexity of function is alternatively explained, to obtain optimization problem.Explanation module 400 can for Optimization Solution The linear regression problem of regularization term penalty factor is used in selection of land optimization problem, and obtains explanation results.The embodiment of the present invention Device 10 can explain in each data neighborhood of a point to complex machines learning model, take into full account the office of sample space Portion's characteristic, the dominant characteristics of sample space different zones not only can be effectively found, and it is more directly perceived, convenient, it is applicable to The explanation of a variety of machine learning models.
Alternatively, in one embodiment of the invention, the weight of each sampled point is:
Wherein, xiFor sample point,For multiple sampled points, σ is the standard deviation of all distances.
Alternatively, in one embodiment of the invention, the gap of fitting result is:
Wherein, f is machine learning model to be explained, and for g to explain function, Γ is the gap of fitting result;Explain function Complexity be:
Wherein, Ω is complexity.
Alternatively, in one embodiment of the invention, optimization problem is:
Alternatively, in one embodiment of the invention, explanation module passes through SGD Algorithm for Solving linear regression problems.
It should be noted that the foregoing solution to the complex machines learning model means of interpretation embodiment based on local linearization The complex machines learning model interpreting means based on local linearization that explanation is also applied for the embodiment are released, it is no longer superfluous herein State.
The complex machines learning model interpreting means based on local linearization proposed according to embodiments of the present invention, the device Complex machines learning model can be explained in each data neighborhood of a point, take into full account the local special of sample space Property, the device not only can effectively find the dominant characteristics of sample space different zones, and more directly perceived, convenient, applicable In the explanation of a variety of machine learning models.
In the description of the invention, it is to be understood that term " " center ", " longitudinal direction ", " transverse direction ", " length ", " width ", " thickness ", " on ", " under ", "front", "rear", "left", "right", " vertical ", " level ", " top ", " bottom " " interior ", " outer ", " up time The orientation or position relationship of the instruction such as pin ", " counterclockwise ", " axial direction ", " radial direction ", " circumference " be based on orientation shown in the drawings or Position relationship, it is for only for ease of and describes the present invention and simplify description, rather than indicates or imply that signified device or element must There must be specific orientation, with specific azimuth configuration and operation, therefore be not considered as limiting the invention.
In addition, term " first ", " second " are only used for describing purpose, and it is not intended that instruction or hint relative importance Or the implicit quantity for indicating indicated technical characteristic.Thus, define " first ", the feature of " second " can be expressed or Implicitly include at least one this feature.In the description of the invention, " multiple " are meant that at least two, such as two, three It is individual etc., unless otherwise specifically defined.
In the present invention, unless otherwise clearly defined and limited, term " installation ", " connected ", " connection ", " fixation " etc. Term should be interpreted broadly, for example, it may be fixedly connected or be detachably connected, or integrally;Can be that machinery connects Connect or electrically connect;Can be joined directly together, can also be indirectly connected by intermediary, can be in two elements The connection in portion or the interaction relationship of two elements, limited unless otherwise clear and definite.For one of ordinary skill in the art For, the concrete meaning of above-mentioned term in the present invention can be understood as the case may be.
In the present invention, unless otherwise clearly defined and limited, fisrt feature can be with "above" or "below" second feature It is that the first and second features directly contact, or the first and second features pass through intermediary mediate contact.Moreover, fisrt feature exists Second feature " on ", " top " and " above " but fisrt feature are directly over second feature or oblique upper, or be merely representative of Fisrt feature level height is higher than second feature.Fisrt feature second feature " under ", " lower section " and " below " can be One feature is immediately below second feature or obliquely downward, or is merely representative of fisrt feature level height and is less than second feature.
In the description of this specification, reference term " one embodiment ", " some embodiments ", " example ", " specifically show The description of example " or " some examples " etc. means specific features, structure, material or the spy for combining the embodiment or example description Point is contained at least one embodiment or example of the present invention.In this manual, to the schematic representation of above-mentioned term not Identical embodiment or example must be directed to.Moreover, specific features, structure, material or the feature of description can be with office Combined in an appropriate manner in one or more embodiments or example.In addition, in the case of not conflicting, the skill of this area Art personnel can be tied the different embodiments or example and the feature of different embodiments or example described in this specification Close and combine.
Although embodiments of the invention have been shown and described above, it is to be understood that above-described embodiment is example Property, it is impossible to limitation of the present invention is interpreted as, one of ordinary skill in the art within the scope of the invention can be to above-mentioned Embodiment is changed, changed, replacing and modification.

Claims (10)

1. a kind of complex machines learning model means of interpretation based on local linearization, it is characterised in that comprise the following steps:
Collecting sample concentrates any point as sample point, and stochastical sampling obtains multiple sampled points around the sample point, And the machine learning model prediction result of each sampled point of the multiple sampled point is obtained according to former machine learning model;
In expression of space, Euler's distance between the sample point and each sampled point is obtained, each to be adopted as described The weight of sampling point;
Machine learning model to be explained is obtained according to the weight of each sampled point and linear model and explains the plan of function The gap of result is closed, and obtains the complexity of the explanation function, to obtain optimization problem;And
The linear regression problem of regularization term penalty factor is used in optimization problem described in Optimization Solution, and obtains explanation results.
2. the complex machines learning model means of interpretation according to claim 1 based on local linearization, it is characterised in that The weight of each sampled point is:
<mrow> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>=</mo> <mi>exp</mi> <mrow> <mo>(</mo> <mo>-</mo> <mfrac> <mrow> <mi>d</mi> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <msup> <mi>&amp;sigma;</mi> <mn>2</mn> </msup> </mfrac> <mo>)</mo> </mrow> <mo>,</mo> </mrow>
Wherein, xiFor the sample point,For the multiple sampled point, σ is the standard deviation of all distances.
3. the complex machines learning model means of interpretation according to claim 2 based on local linearization, it is characterised in that The gap of the fitting result is:
<mrow> <mtable> <mtr> <mtd> <mrow> <mi>&amp;Gamma;</mi> <mrow> <mo>(</mo> <mi>f</mi> <mo>,</mo> <mi>g</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mi>&amp;Sigma;</mi> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <mi>g</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <munder> <mi>&amp;Sigma;</mi> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <munder> <mi>&amp;Sigma;</mi> <mi>j</mi> </munder> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <msubsup> <mi>x</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </mtd> </mtr> </mtable> <mo>,</mo> </mrow>
Wherein, f is the machine learning model to be explained, and g is the explanation function, and Γ is the gap of the fitting result;
It is described explain function complexity be:
<mrow> <mi>&amp;Omega;</mi> <mrow> <mo>(</mo> <mi>g</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <mo>|</mo> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <mo>|</mo> <mo>,</mo> </mrow>
Wherein, Ω is the complexity.
4. the complex machines learning model means of interpretation according to claim 3 based on local linearization, it is characterised in that The optimization problem is:
<mrow> <mtable> <mtr> <mtd> <mi>min</mi> </mtd> <mtd> <mrow> <munder> <mi>&amp;xi;</mi> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> </munder> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <msubsup> <mi>x</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <mi>&amp;lambda;</mi> <mo>&amp;CenterDot;</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <mo>|</mo> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <mo>|</mo> </mrow> </mtd> </mtr> </mtable> <mo>.</mo> </mrow>
5. the complex machines learning model means of interpretation based on local linearization according to claim any one of 1-4, lead to Cross linear regression problem described in stochastic gradient descent SGD Algorithm for Solving.
A kind of 6. complex machines learning model interpreting means based on local linearization, it is characterised in that including:
Acquisition module, any point is concentrated as sample point for collecting sample, and stochastical sampling obtains around the sample point To multiple sampled points, and obtain according to former machine learning model the machine learning model of each sampled point of the multiple sampled point Prediction result;
First acquisition module, in expression of space, obtaining Euler's distance between the sample point and each sampled point, Using the weight as each sampled point;
Second acquisition module, machine learning mould to be explained is obtained for the weight according to each sampled point and linear model The gap of the fitting result of type and explanation function, and the complexity of the explanation function is obtained, to obtain optimization problem;And
Explanation module, for using the linear regression problem of regularization term penalty factor in optimization problem described in Optimization Solution, and Obtain explanation results.
7. the complex machines learning model interpreting means according to claim 6 based on local linearization, it is characterised in that The weight of each sampled point is:
<mrow> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>=</mo> <mi>exp</mi> <mrow> <mo>(</mo> <mo>-</mo> <mfrac> <mrow> <mi>d</mi> <msup> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>,</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <msup> <mi>&amp;sigma;</mi> <mn>2</mn> </msup> </mfrac> <mo>)</mo> </mrow> <mo>,</mo> </mrow>
Wherein, xiFor the sample point,For the multiple sampled point, σ is the standard deviation of all distances.
8. the complex machines learning model interpreting means according to claim 7 based on local linearization, it is characterised in that The gap of the fitting result is:
<mrow> <mtable> <mtr> <mtd> <mrow> <mi>&amp;Gamma;</mi> <mrow> <mo>(</mo> <mi>f</mi> <mo>,</mo> <mi>g</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mi>&amp;Sigma;</mi> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <mi>g</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </mtd> </mtr> <mtr> <mtd> <mrow> <mo>=</mo> <munder> <mi>&amp;Sigma;</mi> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <munder> <mi>&amp;Sigma;</mi> <mi>j</mi> </munder> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <msubsup> <mi>x</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> </mtd> </mtr> </mtable> <mo>,</mo> </mrow>
Wherein, f is the machine learning model to be explained, and g is the explanation function, and Γ is the gap of the fitting result;
It is described explain function complexity be:
<mrow> <mi>&amp;Omega;</mi> <mrow> <mo>(</mo> <mi>g</mi> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <mo>|</mo> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <mo>|</mo> <mo>,</mo> </mrow>
Wherein, Ω is the complexity.
9. the complex machines learning model interpreting means according to claim 8 based on local linearization, it is characterised in that The optimization problem is:
<mrow> <mtable> <mtr> <mtd> <mi>min</mi> </mtd> <mtd> <mrow> <munder> <mi>&amp;xi;</mi> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> </munder> <mrow> <mo>(</mo> <msub> <mi>x</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>=</mo> <munder> <mo>&amp;Sigma;</mo> <mi>k</mi> </munder> <msub> <mi>&amp;Psi;</mi> <mi>k</mi> </msub> <mo>&amp;CenterDot;</mo> <msup> <mrow> <mo>(</mo> <mi>f</mi> <mo>(</mo> <msubsup> <mi>x</mi> <mi>i</mi> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> <mo>-</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <msubsup> <mi>x</mi> <mrow> <mi>i</mi> <mo>,</mo> <mi>j</mi> </mrow> <mrow> <mo>(</mo> <mi>k</mi> <mo>)</mo> </mrow> </msubsup> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <mi>&amp;lambda;</mi> <mo>&amp;CenterDot;</mo> <munder> <mo>&amp;Sigma;</mo> <mi>j</mi> </munder> <mo>|</mo> <msub> <mi>&amp;beta;</mi> <mi>j</mi> </msub> <mo>|</mo> </mrow> </mtd> </mtr> </mtable> <mo>.</mo> </mrow>
10. the complex machines learning model interpreting means based on local linearization according to claim any one of 6-9, institute State explanation module and pass through linear regression problem described in SGD Algorithm for Solving.
CN201710620391.0A 2017-07-26 2017-07-26 Complex machines learning model means of interpretation and device based on local linearization Pending CN107392318A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710620391.0A CN107392318A (en) 2017-07-26 2017-07-26 Complex machines learning model means of interpretation and device based on local linearization

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710620391.0A CN107392318A (en) 2017-07-26 2017-07-26 Complex machines learning model means of interpretation and device based on local linearization

Publications (1)

Publication Number Publication Date
CN107392318A true CN107392318A (en) 2017-11-24

Family

ID=60341057

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710620391.0A Pending CN107392318A (en) 2017-07-26 2017-07-26 Complex machines learning model means of interpretation and device based on local linearization

Country Status (1)

Country Link
CN (1) CN107392318A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902833A (en) * 2018-12-05 2019-06-18 阿里巴巴集团控股有限公司 Machine learning model means of interpretation and device
CN111262887A (en) * 2020-04-26 2020-06-09 腾讯科技(深圳)有限公司 Network risk detection method, device, equipment and medium based on object characteristics
CN111639688A (en) * 2020-05-19 2020-09-08 华中科技大学 Local interpretation method of Internet of things intelligent model based on linear kernel SVM
CN111754287A (en) * 2020-02-14 2020-10-09 北京沃东天骏信息技术有限公司 Article screening method, apparatus, device and storage medium
US11651276B2 (en) 2019-10-31 2023-05-16 International Business Machines Corporation Artificial intelligence transparency

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109902833A (en) * 2018-12-05 2019-06-18 阿里巴巴集团控股有限公司 Machine learning model means of interpretation and device
CN109902833B (en) * 2018-12-05 2023-06-27 创新先进技术有限公司 Machine learning model interpretation method and device
US11651276B2 (en) 2019-10-31 2023-05-16 International Business Machines Corporation Artificial intelligence transparency
CN111754287A (en) * 2020-02-14 2020-10-09 北京沃东天骏信息技术有限公司 Article screening method, apparatus, device and storage medium
CN111262887A (en) * 2020-04-26 2020-06-09 腾讯科技(深圳)有限公司 Network risk detection method, device, equipment and medium based on object characteristics
CN111262887B (en) * 2020-04-26 2020-08-28 腾讯科技(深圳)有限公司 Network risk detection method, device, equipment and medium based on object characteristics
CN111639688A (en) * 2020-05-19 2020-09-08 华中科技大学 Local interpretation method of Internet of things intelligent model based on linear kernel SVM

Similar Documents

Publication Publication Date Title
CN107392318A (en) Complex machines learning model means of interpretation and device based on local linearization
EP2419795B1 (en) Method for determining local accelerations, dynamic load distributions and aerodynamic data in an aircraft
CN106055827B (en) A kind of reservoir numerical simulation parameters sensitivity analysis device and method
DE102005058081B9 (en) Method for the reconstruction of gusts and structural loads in aircraft, in particular commercial aircraft
CN104020135B (en) Calibration model modeling method based near infrared spectrum
CN106886543A (en) The knowledge mapping of binding entity description represents learning method and system
van Gent et al. Knowledge architecture supporting collaborative MDO in the AGILE paradigm
DE112005002773T5 (en) Reynolds number correction function for a mass flow rate sensor
DE102011003314A1 (en) Method for producing a component
CN105931153A (en) Indirect questionnaire assessment method based on neural network prediction analysis model
Troy Baisden et al. An analytical approach to ecosystem biogeochemistry modeling
CN102736616B (en) Dulmage-Mendelsohn (DM)-decomposition-based measuring point optimal configuration method for closed loop system
DE112020004471T5 (en) Inference device, training device, inference method and training method
CN107481218A (en) Image aesthetic feeling appraisal procedure and device
Görtz et al. Collaborative multi-level MDO process development and application to long-range transport aircraft
Wold Chemometrics, why, what and where to next?
CN115577436A (en) Combined deep learning method for solving wind-induced vibration response of uncertain structure
EP1327959B1 (en) Neural network for modelling a physical system and method for building the neural network
Ordaz et al. Integration of off-track sonic boom analysis in conceptual design of supersonic aircraft
CN106485057A (en) A kind of coefficient of collapsibility of loess computational methods
Jategaonkar et al. ESTIMA–an integrated software tool for nonlinear parameter estimation
Jiang et al. Analytical target cascading using ensemble of surrogates for engineering design problems
CN106203618A (en) A kind of method of the neutral net building band border constraint
CN103411913A (en) Infrared spectroscopy PLS (Partial Least Square) modeling method based on genetic algorithm self-adaptive selection segment point
CN114741969B (en) Sulfur recovery soft measurement modeling method based on FIR-NMA model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20171124

RJ01 Rejection of invention patent application after publication