CN105426911A - Dirichlet process mixture model based TAC clustering method - Google Patents

Dirichlet process mixture model based TAC clustering method Download PDF

Info

Publication number
CN105426911A
CN105426911A CN201510779512.7A CN201510779512A CN105426911A CN 105426911 A CN105426911 A CN 105426911A CN 201510779512 A CN201510779512 A CN 201510779512A CN 105426911 A CN105426911 A CN 105426911A
Authority
CN
China
Prior art keywords
tac
class
mixture model
classification
relevant information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201510779512.7A
Other languages
Chinese (zh)
Other versions
CN105426911B (en
Inventor
刘华锋
王婷
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zhejiang University ZJU
Original Assignee
Zhejiang University ZJU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zhejiang University ZJU filed Critical Zhejiang University ZJU
Priority to CN201510779512.7A priority Critical patent/CN105426911B/en
Publication of CN105426911A publication Critical patent/CN105426911A/en
Application granted granted Critical
Publication of CN105426911B publication Critical patent/CN105426911B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)
  • Complex Calculations (AREA)

Abstract

The invention discloses a Dirichlet process mixture model based TAC clustering method. The method comprises: initializing a Dirichlet process mixture model; iteratively calculating a conditional probability; and performing sampling until an iterative stop condition is met. According to the method, the TAC clustering is performed by using the Dirichlet process mixture model, so that the problem in TAC clustering performed under the condition of an unknown class number is effectively solved; and compared with other clustering algorithms, the method has the advantage that the complexity of the Dirichlet process mixture model can be increased with an increase in volume of obtained data.

Description

A kind of TAC clustering method based on Di Li Cray process mixture model
Technical field
The invention belongs to clustering technique field, be specifically related to a kind of TAC clustering method based on Di Li Cray process mixture model.
Background technology
Positron emission tomography imaging (positronemissiontomography, PET) is a kind of medicine imaging technique.Dynamic pet imaging is gathered by continuous data, obtain multiframe physiological status space distribution, it can by rebuilding the distribution on the Time and place of the bio-matrix of radiopharmaceutical agent mark in biological tissue, provide about different biologies or physiology course quantification and the information of non-intrusion type.In practice, dynamic PET images is often divided into different area-of-interests (regionofinterest, ROI), then from each region, extracts time activity curve (timeactivitycurve, TAC).TAC analyzedly further can estimate physiologic parameters, and as volume of blood flow, metabolism and acceptor density, this depends on the characteristic of tracer agent.
Only have several ROI based on dynamic PET images and each ROI is that uniform this is true, we wish to carry out cluster to TAC, namely split dynamic PET images.More importantly, we wish to carry out cluster to TAC by feature based, make noise robust more.Di Li Cray process (DirichletProcess, DP) is the representational stochastic process of most in nonparametric bayes method.Different from other clustering methods, Di Li Cray process mixture model can carry out cluster when not knowing data class number to data, and along with the getable data grows of model many, it adaptive characteristic according to data self can carry out cluster to data and provides the information of data class number.
In the field of machine learning, probability model is used to the distribution of modeling based on observation data.Traditional parameter model uses fixing and a limited number of parameter, time this makes unmatched between the complexity (often weighing by the number of parameter) and available data volume of model, traditional parameter model is easy to suffer over-fitting or poor fitting.Therefore, the selection of model, has the selection of the model of correct complexity in other words, is very important problem in parameter model.Then, no matter we are that the selection of model is all very difficult using cross validation or marginal probability as the basis selected.Bayes's nonparametric technique is an option of parameter model and selection, it has the model of unlimited complexity by use one, the situation of poor fitting can alleviate, and meanwhile, uses bayes method calculating and the approximate complete posteriority based on parameter to alleviate the situation of over-fitting.Di Li Cray process mixture model is one of model most popular in Bayes's nonparametric model, and it is potential class models, is commonly used in clustering problem.
Summary of the invention
The invention provides a kind of TAC clustering method based on Di Li Cray process mixture model, the problems such as the poor fitting easily occurred when can solve Model Selection difficulty and existing clustering method modeling and over-fitting.
Based on a TAC clustering method for Di Li Cray process mixture model, comprise the steps:
(1) initializing set class number K, focuses parameters α, class degree of separation correlation parameter ss, s0, and the value of the degree of freedom v of the special covariance priori in inverse Vichy;
(2) based on the initialization information in step (1), initialization is carried out to the classification that every bar TAC belongs to, set up Di Li Cray process mixture model and calculate the relevant information q determining each classification in mixture model;
(3) for all TAC, successively every bar TAC shifted out from mixture model and calculate its conditional probability belonging to each classification and new classification;
(4) according to described conditional probability, this TAC is sampled selection classification again, then by its add-back mixture model and Renewal model parameter;
(5) travel through all TAC according to step (3) and (4) and take that iteration once as, iterating until class number K no longer changes, and preserve final result.
Carry out initialization to the classification that every bar TAC belongs in described step (2), calculate the relevant information q determining each classification in mixture model, detailed process is as follows:
2.1 represent the value of the class belonging to all TAC with z, z ifor the element of i-th in z, z ivalue obtained by generation random natural number, its span is 1≤z i≤ K, i are natural number and 1≤i≤N, N is the sum of TAC;
Relevant information q described in 2.2 comprises three components: qn, qo and qm; For arbitrary classification k, its relevant information q (k) of initialization, that is:
q·n(k)=q·o(k)=q·m(k)=0
Wherein: q (k) represents the relevant information of kth class, qn (k), qo (k) and qm (k) correspond to three components in relevant information q (k), and k is class number and is natural number, 1≤k≤K;
2.3 for arbitrary TAC, carries out renewal rewards theory by following relational expression to the relevant information q of each classification:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm be all elements in m and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
2.4 travel through every bar TAC according to step 2.3, and the result after traversal upgrades is finally as the relevant information qq of each classification in mixture model.
Calculate the conditional probability that every bar TAC belongs to each classification and new classification in described step (3), detailed process is as follows:
First current TAC shifts out by 3.1 from mixture model, and upgrades the relevant information q of each classification in mixture model by following relational expression:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal iclass relevant information q (z i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
After relational expression in 3.2 completing steps 3.1 calculates, if qn is (z i) ' value be 0, then current z is described ithe class corresponding to value in element number be 0, namely in such not containing any element, at this moment need the class of deleting this sky, if current z ivalue be k, need upgrade K, qn, qo, qm and z ivalue, concrete update mode is:
K′=K-1;
The qn of qn '=delete a kth element
The qo of qo '=delete a kth element
The qm of qm '=delete a kth element
Z ' i=z i-1 (if z i> k)
Wherein: K, qn, qo, qm, z iinformation before corresponding renewal, K ', qn ', qo ', qm ', z ' iinformation after corresponding renewal, for z irenewal need to travel through each TAC, namely i is natural number and its value is the sum of TAC from 1 to N, N; If after the relational expression in completing steps 3.1 calculates, if qn is (z i) ' value be not 0, then without the need to this single stepping;
Then 3.3 calculate this TAC by following relational expression, and to belong to the condition of each classification and new classification general
p1=log([q·n,α])
p 2 ( k ) = p 1 ( k ) = ( q . n ( k ) + 1 ) * d 2 * log ( π ) - d 2 * log ( ss 2 s 0 2 + 1 ) -
( v + 1 ) * Σ ( log ( d i a g ( c h o l 2 ( c h o l 1 ( c h o l ( ss 2 * I d * v ) , x ( i ) ) , x ( i ) ss 2 s 0 2 + 1 ) ) ) )
+ q . n ( k ) * d 2 * l o g ( π ) + d 2 * l o g ( ss 2 s 0 2 )
+ v * Σ ( log ( d i a g ( c h o l 2 ( c h o l ( ss 2 * I d * v ) , O d , 1 ) ) ) )
p3=exp(p2-max(p2))
p=p3/sum(p3)
Wherein: x (i) is i-th TAC and current TAC, d represents the dimension of TAC, p is the vector that i-th TAC belongs to the conditional probability composition of each class, namely element p (k) wherein represents that i-th TAC belongs to the conditional probability of kth class, wherein, last element representation i-th TAC in p belongs to the conditional probability of a new class, k is class number and is natural number, 1≤k≤k+1, α is focuses parameters, s0 and ss is class degree of separation correlation parameter, v is the degree of freedom of the special covariance priori in inverse Vichy, qn (k) is one of them component in kth class relevant information q (k), O d, 1be line number to be d columns be 1 null vector, chol (A) is Cholesky analytic function, if matrix A is symmetric positive definite, then Cholesky decomposes product matrix A being resolved into a lower triangular matrix and upper triangular matrix, if represent upper triangular matrix with R, then lower triangular matrix is its transposition, i.e. A=R ' R, if R=chol (A) is the Cholesky factor of A, so chol1 (R, X) result that function obtains is the Cholesky factor of (A-X*X '), chol2 (R, X) result that function obtains is the Cholesky factor of (A+X ' * X), chol1 (R, and chol2 (R X), X) function all only uses the diagonal angle of R and upper triangle element, the lower triangle element of R is all left in the basket, the effect of diag (X) function is that the element on the diagonal line of X is taken out the new vector of composition one, here summation symbol ∑ is sued for peace to each element in vector in bracket.
Again to sample selection classification to this TAC according to described conditional probability in described step (4), then by its add-back mixture model and Renewal model parameter, detailed process is as follows:
Each element in the p obtained in 4.1 pairs of steps 3.3 is sampled, and obtains sampling results, and the conditional probability drawing a kth element in p is p (k);
If 4.2 what draw is last element, i.e. (K+1) individual element in p, then need the class that structure one is new, concrete operations mode is as follows:
K′=K+1;
q·n′=[q·n,0]
q·o′=[q·o,0]
q·m′=[q·m,0]
z′ i=K′
Wherein, the information before K, qn, qo, qm correspondence upgrades, the information after K ', qn ', qo ', qm ' correspondence upgrade; If what draw is not last element in p, then without the need to this single stepping;
4.3 then by current TAC add-back mixture model, and at this moment need to revise three component qn of q, the value of qo, qm again, concrete modification mode is as follows:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC.
The span of described initialization class number K is 100 ~ 100000, the span of focuses parameters α is 0.001 ~ 10000, the span of class degree of separation correlation parameter ss and s0 is 0.01 ~ 1000, and the span of the degree of freedom v of the special covariance priori in inverse Vichy is 10 ~ 10000.
The present invention carries out cluster by using Di Li Cray process mixture model to TAC, effectively solve the problem of carrying out cluster when the unknown of class number, and the complexity of Di Li Cray mixture model can increase along with the increase of the data bulk obtained, this is the advantage not available for other clustering algorithms.
Accompanying drawing explanation
Fig. 1 is the steps flow chart schematic diagram of TAC clustering method of the present invention.
Fig. 2 be true value corresponding TAC figure; Wherein horizontal ordinate is frame number, and ordinate is the grey scale change of each pixel with frame number.
The cluster result schematic diagram that Fig. 3 (a) is true value; Wherein the pixel of the position of different gray scale represents different classes, and namely the TAC representated by pixel of same gray scale belongs to same class.
Fig. 3 (b) is for class number in cluster process is with the change schematic diagram of iterations.
Embodiment
In order to more specifically describe the present invention, below in conjunction with the drawings and the specific embodiments, TAC clustering method of the present invention is described in detail.
As shown in Figure 1, the present invention is based on the TAC clustering method of Di Li Cray process mixture model, comprise the steps:
S1. initialization parameters: need initialized parameter to have class number K, focuses parameters α, class degree of separation correlation parameter ss, s0, and the value of the degree of freedom v of the special covariance priori in inverse Vichy;
S2. initialization DP mixture model: initialization is carried out to the classification that every bar TAC belongs to, calculate the relevant information q determining each classification in mixture model, detailed process is as follows:
2.1 represent the value of the class belonging to all TAC with z, z ifor the element of i-th in z, z ivalue obtained by generation random natural number, its span is 1≤z i≤ K, i are natural number and 1≤i≤N, N is the sum of TAC;
Relevant information q described in 2.2 comprises three components: qn, qo and qm; For arbitrary classification k, its relevant information q (k) of initialization, namely
q·n(k)=q·o(k)=q·m(k)=0
Wherein: q (k) represents the relevant information of kth class, qn (k), qo (k) and qm (k) correspond to three components in relevant information q (k), and k is class number and is natural number, 1≤k≤K;
2.3 for arbitrary TAC, carries out renewal rewards theory by following relational expression to the relevant information q of each classification:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm be all elements in m and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
2.4 travel through every bar TAC according to step 2.3, and the result after traversal upgrades is finally as the relevant information qq of each classification in mixture model.
S3. design conditions probability: calculate the conditional probability that every bar TAC belongs to each classification and new classification, detailed process is as follows:
First current TAC shifts out by 3.1 from mixture model, and upgrades the relevant information q of each classification in mixture model by following relational expression:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal iclass relevant information q (z i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
After 3.2 relational expressions completed in 3.1 calculate, if qn is (z i) ' value be 0, then current z is described ithe class corresponding to value in element number be 0, namely in such not containing any element, at this moment need the class of deleting this sky, if current z ivalue be k, need upgrade K, qn, qo, qm and z ivalue, concrete update mode is:
K′=K-1;
The qn of qn '=delete a kth element
The qo of qo '=delete a kth element
The qm of qm '=delete a kth element
Z ' i=z i-1 (if z i> k)
Wherein: K, qn, qo, qm, z iinformation before corresponding renewal, K ', qn ', qo ', qm ', z ' iinformation after corresponding renewal, for z irenewal need to travel through each TAC, namely i is natural number and its value is the sum of TAC from 1 to N, N; If after the relational expression completed in 3.1 calculates, if qn is (z i) ' value be not 0, then without the need to this single stepping;
Then 3.3 calculate by following relational expression the conditional probability that this TAC belongs to each classification and new classification:
p1=log([q·n,α])
p 2 ( k ) = p 1 ( k ) = ( q . n ( k ) + 1 ) * d 2 * log ( π ) - d 2 * log ( ss 2 s 0 2 + 1 ) -
( v + 1 ) * Σ ( log ( d i a g ( c h o l 2 ( c h o l 1 ( c h o l ( ss 2 * I d * v ) , x ( i ) ) , x ( i ) ss 2 s 0 2 + 1 ) ) ) )
+ q . n ( k ) * d 2 * l o g ( π ) + d 2 * l o g ( ss 2 s 0 2 )
+ v * Σ ( log ( d i a g ( c h o l 2 ( c h o l ( ss 2 * I d * v ) , O d , 1 ) ) ) )
p3=exp(p2-max(p2))
p=p3/sum(p3)
Wherein: x (i) is i-th TAC and current TAC, d represents the dimension of TAC, p is the vector that i-th TAC belongs to the conditional probability composition of each class, namely element p (k) wherein represents that i-th TAC belongs to the conditional probability of kth class, wherein, last element representation i-th TAC in p belongs to the conditional probability of a new class, k is class number and is natural number, 1≤k≤K+1, α is focuses parameters, s0 and ss is class degree of separation correlation parameter, v is the degree of freedom of the special covariance priori in inverse Vichy, qn (k) is one of them component in kth class relevant information q (k), O d, 1be line number to be d columns be 1 null vector, chol (A) is Cholesky analytic function, if matrix A is symmetric positive definite, then Cholesky decomposes product matrix A being resolved into a lower triangular matrix and upper triangular matrix, if represent upper triangular matrix with R, then lower triangular matrix is its transposition, i.e. A=R ' R, if R=chol (A) is the Cholesky factor of A, so chol1 (R, X) result that function obtains is the Cholesky factor of (A-X*X '), chol2 (R, X) result that function obtains is the Cholesky factor of (A+X ' * X), chol1 (R, and chol2 (R X), X) function all only uses the diagonal angle of R and upper triangle element, the lower triangle element of R is all left in the basket, the effect of diag (X) function is that the element on the diagonal line of X is taken out the new vector of composition one, here summation symbol ∑ is sued for peace to each element in vector in bracket.
S4. sampling Renewal model: this TAC is sampled selection classification again according to the conditional probability calculated in S3, then by its add-back mixture model and Renewal model parameter, detailed process is as follows:
Each element in the p obtained in 4.1 couples 3.3 is sampled, and obtains sampling results, and the conditional probability drawing a kth element in p is p (k).
If 4.2 what draw is last element, i.e. (K+1) individual element in p, then need the class that structure one is new, concrete operations mode is as follows:
K′=K+1;
q·n′=[q·n,0]
q·o′=[q·o,0]
q·m′=[q·m,0]
z′ i=K′
Wherein, the information before K, qn, qo, qm correspondence upgrades, the information after K ', qn ', qo ', qm ' correspondence upgrade; If what draw is not last element in p, then without the need to this single stepping;
4.3 then by current TAC add-back mixture model, and at this moment need to revise three component qn of q, the value of qo, qm again, concrete modification mode is as follows:
m m = Σ k = 1 M m ( k )
q·n(z i)′=q·n(z i)+1
q·o(z i)′=q·o(z i)+x(i)
q·m(z i)′=q·m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, qn (z i), qo (z i) and qm (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, qn (z i) ', qo (z i) ' and qm (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC.
S5. repeat step S3 and S4, until the class number K of data no longer changes, now iteration stopping, preserves the value of q, K, z.
We verify practicality and the reliability of present embodiment by experiment true value (groundtruth) being carried out to cluster below.True value is made up of 18 two field pictures, and the size of every two field picture is 128 pixel * 128 pixels, and the gray-scale value of the same pixel on every two field picture can form ten octuple vectors, i.e. TAC, we can draw TAC curve, as shown in Figure 2, horizontal ordinate is frame number, and ordinate is gray-scale value.As shown in Figure 3, the pixel of different gray areas belongs to different classes to the operation result of program, and namely the pixel of same grayscale belongs to same class, the corresponding TAC of each pixel.The program of present embodiment can obtain correct cluster result with the probability (because have the composition of random sampling inside in program, therefore not can obtain correct cluster result, this is normal phenomenon) of more than 95% at every turn.

Claims (5)

1., based on a TAC clustering method for Di Li Cray process mixture model, comprise the steps:
(1) value of the degree of freedom v of initializing set class number K, focuses parameters α, the special covariance priori of class degree of separation correlation parameter ss and s0 and inverse Vichy;
(2) based on the initialization information in step (1), initialization is carried out to the classification that every bar TAC belongs to, set up Di Li Cray process mixture model and calculate the relevant information q determining each classification in mixture model;
(3) for all TAC, successively every bar TAC shifted out from mixture model and calculate its conditional probability belonging to each classification and new classification;
(4) according to described conditional probability, this TAC is sampled selection classification again, then by its add-back mixture model and Renewal model parameter;
(5) travel through all TAC according to step (3) and (4) and take that iteration once as, iterating until class number K no longer changes, and preserve final result.
2. the TAC clustering method based on Di Li Cray process mixture model according to claim 1, it is characterized in that: in described step (2), initialization is carried out to the classification that every bar TAC belongs to, calculate the relevant information q determining each classification in mixture model, detailed process is as follows:
2.1 represent the value of the class belonging to all TAC with z, z ifor the element of i-th in z, z ivalue obtained by generation random natural number, its span is 1≤z i≤ K, i are natural number and 1≤i≤N, N is the sum of TAC;
Relevant information q described in 2.2 comprises three components: q.n, q.o and q.m; For arbitrary classification k, its relevant information q (k) of initialization, that is:
q.n(k)=q.o(k)=q.m(k)=0
Wherein: q (k) represents the relevant information of kth class, q.n (k), q.o (k) and q.m (k) correspond to three components in relevant information q (k), and k is class number and is natural number, 1≤k≤K;
2.3 for arbitrary TAC, carries out renewal rewards theory by following relational expression to the relevant information q of each classification:
m m = Σ k = 1 M m ( k )
q.n(z i)′=q.n(z i)+1
q.o(z i)′=q.o(z i)+x(i)
q.m(z i)′=q.m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm be all elements in m and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, q.n (z i), q.o (z i) and q.m (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, q.n (z i) ', q.o (z i) ' and q.m (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
2.4 travel through every bar TAC according to step 2.3, and the result after traversal upgrades is finally as the relevant information qq of each classification in mixture model.
3. the TAC clustering method based on Di Li Cray process mixture model according to claim 1, is characterized in that: calculate the conditional probability that every bar TAC belongs to each classification and new classification in described step (3), detailed process is as follows:
First current TAC shifts out by 3.1 from mixture model, and upgrades the relevant information q of each classification in mixture model by following relational expression:
m m = Σ k = 1 M m ( k )
q.n(z i)′=q.n(z i)-1
q.o(z i)′=q.o(z i)-z(i)
q.m(z i)′=q.m(z i)-mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, q.n (z i), q.o (z i) and q.m (z i) correspond to the front z of renewal iclass relevant information q (z i) in three components, q.n (z i) ', q.o (z i) ' and q.m (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC;
After calculating in 3.2 completing steps 3.1, if q.n is (z i) ' value be 0, then current z is described ithe class corresponding to value in element number be 0, namely in such not containing any element, at this moment need the class of deleting this sky, if current z ivalue be k, need upgrade K, q.n, q.o, q.m and z ivalue, concrete update mode is:
K′=K-1;
The q.n of q.n '=delete a kth element
The q.o of q.o '=delete a kth element
The q.m of q.m '=delete a kth element
Z ' i=z i-1 (if z i> k)
Wherein: K, q.n, q.o, q.m, z iinformation before corresponding renewal, K ', q.n ', q.o ', q.m ', z ' iinformation after corresponding renewal, for z irenewal need to travel through each TAC, namely i is natural number and its value is the sum of TAC from 1 to N, N;
If after the calculating in completing steps 3.1, q.n (z i) ' value be not 0, then without the need to carrying out aforesaid operations;
Then 3.3 calculate by following relational expression the conditional probability that this TAC belongs to each classification and new classification:
p1=log([q.n,α])
p 2 ( k ) = p 1 ( k ) - ( q . n ( k ) + 1 ) * d 2 * log ( π ) - d 2 * log ( ss 2 s 0 2 + 1 ) - ( v + 1 ) * Σ ( log ( d i a g ( c h o l 2 ( c h o l 1 ( c h o l ( ss 2 * I d * v ) , x ( i ) ) , x ( i ) ss 2 s 0 2 + 1 ) ) ) ) + q . n ( k ) * d 2 * log ( π ) + d 2 * log ( ss 2 s 0 2 )
+ v * Σ ( l o g ( d i a g ( c h o l 2 ( c h o l ( ss 2 * I d * v ) , O d , 1 ) ) ) )
p3=exp(p2-max(p2))
p=p3/sum(p3)
Wherein: x (i) is i-th TAC and current TAC, d represents the dimension of TAC, p is the vector that i-th TAC belongs to the conditional probability composition of each class, namely element p (k) wherein represents that i-th TAC belongs to the conditional probability of kth class, wherein, last element representation i-th TAC in p belongs to the conditional probability of a new class, k is class number and is natural number, 1≤k≤K+1, α is focuses parameters, s0 and ss is class degree of separation correlation parameter, v is the degree of freedom of the special covariance priori in inverse Vichy, q.n (k) is one of them component in kth class relevant information q (k), O d, 1be line number to be d columns be 1 null vector, chol (A) is Cholesky analytic function, if matrix A is symmetric positive definite, then Cholesky decomposes product matrix A being resolved into a lower triangular matrix and upper triangular matrix, if represent upper triangular matrix with R, then lower triangular matrix is its transposition, i.e. A=R ' R, if R=chol (A) is the Cholesky factor of A, so chol1 (R, X) result that function obtains is the Cholesky factor of (A-X*X '), chol2 (R, X) result that function obtains is the Cholesky factor of (A+X ' * X), chol1 (R, and chol2 (R X), X) function all only uses the diagonal angle of R and upper triangle element, the lower triangle element of R is all left in the basket, the effect of diag (X) function is that the element on the diagonal line of X is taken out the new vector of composition one, summation symbol ∑ in above formula is sued for peace to each element in vector in bracket.
4. the TAC clustering method based on Di Li Cray mixture model according to claim 3, it is characterized in that: according to described conditional probability, this TAC is sampled selection classification again in described step (4), then by its add-back mixture model and Renewal model parameter, detailed process is as follows:
Each element in the p obtained in 4.1 pairs of steps 3.3 is sampled, and obtains sampling results, and the conditional probability drawing a kth element in p is p (k);
If 4.2 what draw is last element, i.e. (K+1) individual element in p, then need the class that structure one is new, concrete operations mode is as follows:
K′=K+1;
q.n′=[q.n,0]
q.o′=[q.o,0]
q.m′=[q.m,0]
z′ i=K′
Wherein, the information before K, q.n, q.o, q.m correspondence upgrades, the information after K ', q.n ', q.o ', q.m ' correspondence upgrade;
If what draw is not last element in p, then without the need to aforesaid operations;
4.3 then by current TAC add-back mixture model, and at this moment need to revise three component q.n of q, the value of q.o, q.m again, concrete modification mode is as follows:
m m = Σ k = 1 M m ( k )
q.n(z i)′=q.n(z i)+1
q.o(z i)′=q.o(z i)+x(i)
q.m(z i)′=q.m(z i)+mm
Wherein: x (i) is i-th TAC and current TAC, m is the vector of all elements composition not being 0 in i-th TAC, m (k) is the element of the kth in m, M is the dimension of this vector, namely be not the number of all elements of 0 in current TAC, mm and m be middle all elements and, z irepresent the value of the class belonging to i-th TAC, its span is 1≤z i≤ K, q.n (z i), q.o (z i) and q.m (z i) correspond to the front z of renewal irelevant information q (the z of class i) in three components, q.n (z i) ', q.o (z i) ' and q.m (z iz after) ' correspond to upgrades iclass relevant information q (z i) in three components, i is natural number and 1≤i≤N, N is the sum of TAC.
5. the TAC clustering method based on Di Li Cray process mixture model according to claim 1, it is characterized in that: in described step (1), the span of initialization class number K is 100 ~ 100000, the span of focuses parameters α is 0.001 ~ 10000, the span of class degree of separation correlation parameter ss and s0 is 0.01 ~ 1000, and the span of the degree of freedom v of the special covariance priori in inverse Vichy is 10 ~ 10000.
CN201510779512.7A 2015-11-13 2015-11-13 A kind of TAC clustering method based on Di Li Cray process mixed model Active CN105426911B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510779512.7A CN105426911B (en) 2015-11-13 2015-11-13 A kind of TAC clustering method based on Di Li Cray process mixed model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510779512.7A CN105426911B (en) 2015-11-13 2015-11-13 A kind of TAC clustering method based on Di Li Cray process mixed model

Publications (2)

Publication Number Publication Date
CN105426911A true CN105426911A (en) 2016-03-23
CN105426911B CN105426911B (en) 2018-12-25

Family

ID=55505109

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510779512.7A Active CN105426911B (en) 2015-11-13 2015-11-13 A kind of TAC clustering method based on Di Li Cray process mixed model

Country Status (1)

Country Link
CN (1) CN105426911B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446339A (en) * 2016-08-29 2017-02-22 北京化工大学 Rotating machine operation state anomaly detection method based on Dirichlet mixture model
CN109801073A (en) * 2018-12-13 2019-05-24 中国平安财产保险股份有限公司 Risk subscribers recognition methods, device, computer equipment and storage medium
CN113095542A (en) * 2021-03-01 2021-07-09 华中科技大学 Photovoltaic output power prediction error fitting method and system based on DPMM

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060083415A1 (en) * 2004-10-15 2006-04-20 Georges El Fakhri Factor analysis in medical imaging
CN1969295A (en) * 2004-05-10 2007-05-23 皇家飞利浦电子股份有限公司 Image data processing system for compartmental analysis
CN101582080A (en) * 2009-06-22 2009-11-18 浙江大学 Web image clustering method based on image and text relevant mining
CN104484346A (en) * 2014-11-28 2015-04-01 浙江大学 Hierarchical theme modeling method based on mixed distance and relying on Chinese restaurant process

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1969295A (en) * 2004-05-10 2007-05-23 皇家飞利浦电子股份有限公司 Image data processing system for compartmental analysis
US20060083415A1 (en) * 2004-10-15 2006-04-20 Georges El Fakhri Factor analysis in medical imaging
CN101582080A (en) * 2009-06-22 2009-11-18 浙江大学 Web image clustering method based on image and text relevant mining
CN104484346A (en) * 2014-11-28 2015-04-01 浙江大学 Hierarchical theme modeling method based on mixed distance and relying on Chinese restaurant process

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MIKA NAGANAWA .ETC: ""Extraction of a Plasma Time-Activity Curve From Dynamic Brain PET Images Based on Independent Component Analysis"", 《IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING》 *
ROSTOM MABROUK .ETC: ""Dynamic Cardiac PET Imaging: Extraction of Time-Activity Curves Using ICA and a Generalized Gaussian Distribution Model"", 《IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING》 *
S.D. WOLLENWEBER.ETC: ""A Simple On-Line Arterial Time-Activity Curve Detector for [O- 151 Water PET Studies"", 《IEEE TRANSACTIONS ON NUCLEAR SCIENCE》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446339A (en) * 2016-08-29 2017-02-22 北京化工大学 Rotating machine operation state anomaly detection method based on Dirichlet mixture model
CN106446339B (en) * 2016-08-29 2019-04-30 北京化工大学 Rotation Mechanical Running Condition method for detecting abnormality based on Di Li Cray mixed model
CN109801073A (en) * 2018-12-13 2019-05-24 中国平安财产保险股份有限公司 Risk subscribers recognition methods, device, computer equipment and storage medium
CN113095542A (en) * 2021-03-01 2021-07-09 华中科技大学 Photovoltaic output power prediction error fitting method and system based on DPMM
CN113095542B (en) * 2021-03-01 2023-11-14 华中科技大学 Fitting method and system for photovoltaic output power prediction error based on DPMM

Also Published As

Publication number Publication date
CN105426911B (en) 2018-12-25

Similar Documents

Publication Publication Date Title
Tompson et al. Accelerating eulerian fluid simulation with convolutional networks
Kleiner et al. The big data bootstrap
Nowozin et al. Decision tree fields
Li et al. Fast local image inpainting based on the Allen–Cahn model
US20210366168A1 (en) Pet image reconstruction method, computer storage medium, and computer device
US9665954B2 (en) Method for reconstructing PET image using GPU parallel computing
Klodt et al. A convex framework for image segmentation with moment constraints
CN104657950B (en) Dynamic PET (positron emission tomography) image reconstruction method based on Poisson TV
CN105938628A (en) Direct computation of image-derived biomarkers
Roy et al. A new method of brain tissues segmentation from MRI with accuracy estimation
CN109993808B (en) Dynamic double-tracing PET reconstruction method based on DSN
Liang et al. An EM approach to MAP solution of segmenting tissue mixtures: a numerical analysis
CN107346556A (en) A kind of PET image reconstruction method based on block dictionary learning and sparse expression
Allassonniere et al. Convergent Stochastic Expectation Maximization algorithm with efficient sampling in high dimension. Application to deformable template model estimation
Yu et al. Modeling spatial extremes via ensemble-of-trees of pairwise copulas
CN105426911A (en) Dirichlet process mixture model based TAC clustering method
CN107146263B (en) A kind of dynamic PET images method for reconstructing based on the constraint of tensor dictionary
Chen et al. A machine learning based solver for pressure Poisson equations
Li et al. MPMNet: A data-driven MPM framework for dynamic fluid-solid interaction
Xiang et al. Towards bi-directional skip connections in encoder-decoder architectures and beyond
Zhang et al. Learning to estimate and refine fluid motion with physical dynamics
Wu et al. Image segmentation via Fischer-Burmeister total variation and thresholding
Tobon-Gomez et al. Automatic construction of 3D-ASM intensity models by simulating image acquisition: Application to myocardial gated SPECT studies
Li et al. Feature pre-inpainting enhanced transformer for video inpainting
CN104077798B (en) High-reality-sense animation synthesis method for deformable object

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant