US20120177128A1 - System and method for designing of dictionaries for sparse representation - Google Patents

System and method for designing of dictionaries for sparse representation Download PDF

Info

Publication number
US20120177128A1
US20120177128A1 US13/425,142 US201213425142A US2012177128A1 US 20120177128 A1 US20120177128 A1 US 20120177128A1 US 201213425142 A US201213425142 A US 201213425142A US 2012177128 A1 US2012177128 A1 US 2012177128A1
Authority
US
United States
Prior art keywords
dictionary
signal
processing system
signal processing
atom
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/425,142
Inventor
Michal Aharon
Michael Elad
Alfred Bruckstein
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Technion Research and Development Foundation Ltd
Original Assignee
Technion Research and Development Foundation Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Technion Research and Development Foundation Ltd filed Critical Technion Research and Development Foundation Ltd
Priority to US13/425,142 priority Critical patent/US20120177128A1/en
Publication of US20120177128A1 publication Critical patent/US20120177128A1/en
Priority to US13/958,644 priority patent/US20140037199A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/90Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
    • H04N19/97Matching pursuit coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/213Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
    • G06F18/2136Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on sparsity criteria, e.g. with an overcomplete basis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/7715Feature extraction, e.g. by transforming the feature space, e.g. multi-dimensional scaling [MDS]; Mappings, e.g. subspace methods

Definitions

  • the present invention relates to a system and method for the sparse representation of signals.
  • the invention is particularly relevant for applications such as compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • a signal y ⁇ n can be represented as a sparse linear combination of these atoms.
  • the vector x ⁇ K displays the representation coefficients of the signal y.
  • ⁇ 0 is the l 0 norm, counting the non zero entries of a vector.
  • a second well known pursuit approach is the Basis Pursuit (BP). It suggests a convexisation of the problems posed in (1) and (2), by replacing the l 0 -norm with an ⁇ 1 -norm.
  • the Focal Under-determined System Solver (FOCUSS) is very similar, using the l p -norm with p ⁇ 1, as a replacement to the l 0 -norm.
  • FOCUSS Focal Under-determined System Solver
  • Lagrange multipliers are used to convert the constraint into a penalty term, and an iterative method is derived based on the idea of iterated reweighed least-squares that handles the l p -norm as an l 2 weighted one.
  • MAP Maximum A Posteriori
  • the MAP can be used to estimate the coefficients as random variables by maximizing the posterior P(x
  • the prior distribution on the coefficient vector x is assumed to be a super-Gaussian Independent Identically-Distributed (iid) distribution that favors sparsity. For the Laplace distribution this approach is equivalent to BP.
  • An overcomplete dictionary D that leads to sparse representations can either be chosen as a pre-specified set of functions, or designed by adapting its content to fit a given set of signal examples.
  • Sparse coding is the process of computing the representation coefficients, x, based on the given signal y and the dictionary D. This process, commonly referred to as “atom decomposition”, requires solving (1) or (2), and this is typically done by a “pursuit algorithm” that finds an approximate solution.
  • Three popular pursuit algorithms are the Orthogonal Matching Pursuit (OMP), Basis Pursuit (BP) and the Focal Under-determined System Solver (FOCUSS).
  • Orthogonal Matching Pursuit is a greedy step-wise regression algorithm. At each stage this method selects the dictionary element having the maximal projection onto the residual signal. After each selection, the representation coefficients with regarding to the so far chosen atoms are found via least-squares.
  • the algorithm can be stopped after a predetermined number of steps, hence after having selected a fixed number of atoms.
  • the stopping rule can be based on norm of the residual, or on the maximal inner product computed in the next atom selection stage.
  • OMP is an appealing and very simple to implement algorithm. Unlike other methods, it can be easily programmed to supply a representation with an a priori fixed number of non-zero entries ⁇ a desired outcome in the training of dictionaries.
  • Focal Under-determined System Solver is an approximating algorithm for finding the solutions of either (1) or (2), by replacing the l 0 -norm with an l p one for p ⁇ 1.
  • a regularization can, and should, be introduced to avoid near-zero entries in the weight matrix ⁇ (x).
  • K-Means algorithm also known as generalized Lloyd algorithm—GLA
  • GLA generalized Lloyd algorithm
  • the approaches to dictionary design that have been tried so far are very much in line with the two-step process described above.
  • the first step finds the coefficients given the dictionary—a step we shall refer to as “sparse coding”.
  • the dictionary is updated assuming known and fixed coefficients.
  • the differences between the various algorithms that have been proposed are in the method used for the calculation of coefficients, and in the procedure used for modifying the dictionary.
  • the second assumption is critical and refers to the “hidden variable” x.
  • the ingredients of the likelihood function are computed using the relation
  • the prior distribution of the representation vector x is assumed to be such that the entries of x are zero-mean iid, with Cauchy or Laplace distributions. Assuming for example a Laplace distribution we get
  • An iterative method was suggested for solving (16). It includes two main steps in each iteration: (i) calculate the coefficients x i using a simple gradient descent procedure; and then (ii) update the dictionary using
  • the Method of Optimal Directions follows more closely the K-Means outline, with a sparse coding stage that uses either the OMP or FOCUSS, followed by an update of the dictionary.
  • the main contribution of the MOD method is its simple way of updating the dictionary.
  • the first two terms are the same ones as in (17).
  • the last term compensates for deviations from the constraint. This case allows different columns in D to have different norm values. As a consequence, columns with small norm values tend to be under-used, as the coefficients they need are larger and as such more penalized.
  • d i (n+1) d i (n) + ⁇ ( I ⁇ d i (n) d i (n) T ) E ⁇ x i T . (22)
  • the coefficients of the sparse representations X can be decomposed to L pieces, each referring to a different ortho-basis.
  • X [X 1 , X 2 , . . . , X L ] T ,
  • X i is the matrix containing the coefficients of the orthonormal dictionary Di.
  • the proposed algorithm updates each orthonormal basis D j sequentially.
  • the update of D j is done by first computing the residual matrix
  • Grinbonval suggested a slightly different method. Apart from the fact that here the dictionary is structured, handling a union of orthonormal bases, it updates each orthonormal bases sequentially, and thus reminds the sequential updates done in the K-means. Experimental results show weak performance compared to previous methods. This could partly be explained by the fact that the update of D j depends strongly on the coefficients X j .
  • the K-Means can be extended to suggest a fuzzy assignment and a covariance matrix per each cluster, so that the data is modeled as a mixture of Gaussians.
  • the dictionary of VQ codewords is typically trained using the K-Means algorithm.
  • C the codebook matrix
  • c K the codewords being the columns
  • each signal is represented as its closest codeword (under l 2 -norm distance).
  • y i Cx i
  • the index j is selected such that
  • the VQ training problem is to find a codebook C that minimizes the error E, subject to the limited structure of X, whose columns must be taken from the trivial basis,
  • the K-Means algorithm is an iterative method used for designing the optimal codebook for VQ. In each iteration there are two stages—one for sparse coding that essentially evaluates X, and one for updating the codebook.
  • the sparse coding stage assumes a known codebook C (J ⁇ 1) , and computes a feasible X that minimizes the value of (25).
  • the dictionary update stage fixes X as known, and seeks an update of C so as to minimize (25).
  • the minimization step is optimal under the assumptions. As the MSE is bounded from below by zero, and the algorithm ensures a monotonic decrease of the MSE, convergence to at least a local minimum solution is guaranteed. Stopping rules for the above-described algorithm can vary a lot but are quite easy to handle.
  • These dictionaries have the potential to outperform commonly used pre-determined dictionaries.
  • the invention thus relate to a novel system and algorithm for adapting dictionaries so as to represent signals sparsely.
  • Given a set of training signals ⁇ y i ⁇ i 1 N , we seek the dictionary D that leads to the best possible representations for each member in this set with strict sparsity constraints.
  • the invention introduces the K-SVD algorithm that addresses the above task, generalizing the K-Means algorithm.
  • the K-SVD is an iterative method that alternates between sparse coding of the examples based on the current dictionary, and an update process for the dictionary atoms so as to better fit the data.
  • the update of the dictionary columns is done jointly with an update of the sparse representation coefficients related to it, resulting in accelerated convergence.
  • the K-SVD algorithm is flexible and can work with any pursuit method, thereby tailoring the dictionary to the application in mind.
  • the sparse representation problem can be viewed as a generalization of the VQ objective (25), in which we allow each input signal to be represented by a linear combination of codewords, which we now call dictionary elements. Therefore the coefficients vector is now allowed more than one nonzero entry, and these can have arbitrary values.
  • the minimization corresponding to Equation (25) is that of searching the best possible dictionary for the sparse representation of the example set Y,
  • K-Means generalization methods freeze X while finding a better D.
  • the approach of the invention is different, as we change the columns of D sequentially, and allow changing the relevant coefficients. In a sense, this approach is a more direct generalization of the K-Means algorithm, because it updates each column separately, as done in K-Means.
  • the invention is useful for a variety of applications in signal processing including but not limited to: compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • all the training signals involved are from the same family and thus have common traits.
  • the signals can all be pictures, music, speech etc.
  • FIG. 1 is a description of the K-SVD algorithm of the invention.
  • FIG. 2 is graph of synthetic results comparing K-SVD against two known algorithms, MOD and MAP-based algorithms. For each of the tested algorithms and for each noise level, 50 trials were conducted and their results sorted. The graph labels represent the mean number of detected atoms (out of 50) over the ordered tests in groups of 10 experiments.
  • FIG. 3 is a collection of 500 random block patches of size 8 ⁇ 8 pixels, taken from a database of face images, which were used for training with the K-SVD algorithm, sorted by their variance.
  • FIG. 4A depicts the learned dictionary (a K-SVD trained dictionary of size 64 ⁇ 441). Its elements are sorted in an ascending order of their variance, and stretched to maximal range for display purposes.
  • FIGS. 4B and 4C depict the overcomplete separable Haar dictionary and the overcomplete DCT dictionary, respectively, of the same size (shown for comparison).
  • FIG. 5 is the RMSE for 594 new blocks with missing pixels using the learned dictionary og FIG. 4A , the overcomplete Haar dictionary and the overcomplete DCT dictionary.
  • FIGS. 6A-6H compare two corrupted images and their reconstruction, with the missing pixels marked as points ( 6 A, 50% of missing pixels; 6 E, 70% of missing pixels), and the reconstructed results by the learned dictionary ( 6 B, 6 F), the overcomplete Haar dictionary ( 6 C, 6 G), and the overcomplete DCT dictionary ( 6 D, 6 H), respectively.
  • FIG. 7 depicts Rate-Distortion graphs for the compression results for each dictionary.
  • FIGS. 8A-8C show sample compression results for the K-SVD, overcomplete DCT and complete DCT dictionaries, respectively.
  • the present invention relates to a signal processing method adapted for sparse representation of signals and a system for implementing said method, said system comprising:
  • the training signals are typically from the same family and thus all training signals share common traits and have common behavior patterns.
  • all training signals can be pictures, including pictures of human faces, or the training signals can be sound files including music files, speeches, and the like.
  • the purpose of the dictionary of the present invention is to discover the common building blocks with which all the training signals can be represented. All the training signals can be represented by linear combinations of the dictionary atoms (building blocks).
  • the term “atom” as referred to herein means dictionary atom or signal-atom.
  • the building blocks or some of the building blocks of the training signals are known or can be approximated intuitively, while in other cases the invention helps to discover them.
  • the dictionary is updated one atom at a time. It is possible however to also update the dictionary a group of atoms at a time, for example two or three atoms at a time, or defining the group of atoms to be updated containing any number of atoms.
  • the dictionary is an overcomplete dictionary.
  • An overcomplete dictionary contains more atoms (building blocks, functions) than strictly necessary to represent the signal space.
  • An overcomplete dictionary thus allows a suitable representation of a signal with fewer encoded atoms. This is important for applications in which a low bit rate is required.
  • each training signal can be represented in many forms.
  • the representation of each training signal is a coefficient matrix.
  • the representation of the training signals may take any other form such as a vector.
  • the generation of the coefficients matrix is achieved by a pursuit algorithm.
  • the pursuit algorithm can include: Orthogonal Matching Pursuit, Matching Pursuit, Basis Pursuit, FOCUSS or any combination or variation thereof.
  • Updating the dictionary can be performed sequentially or in any other order.
  • the dictionary is updated in a predefined order of the signal-atoms. Depending on the application used and the nature of the training signals, updating the dictionary in a predefined order of signal-atoms will yield different results and thus can be exploited by the application.
  • only selected signal-atoms of said dictionary are updated. Again, depending on the nature of the application in mind, one may decide to leave certain signal-atoms (building blocks) fixed, and consequently only update the remaining signal-atoms.
  • a dictionary is built wherein two signal-atoms are very similar to each other but not equal to each other.
  • the similarity for the purpose of the application used, may be too big, and thus the differentiation between the two atoms may be considered negligible.
  • the application will thus wish to modify one of the similar atoms.
  • a signal-atom is modified when the difference between said signal-atom to another signal atom is below a predefined value.
  • a signal-atom may be defined by the system as a building block for representing the training signals, but the actual signal-atom may never be used to construct any of the given training signals. One may thus wish to modify this atom.
  • a signal-atom is modified when it is not used in any representation.
  • a signal-atom may be found to be used only rarely to construct training signals. It may be thus preferred not to work with such a building block, and modify this atom to one used more frequently in training signals representation.
  • a signal-atom is modified when its usage frequency in the representation of signal-atoms is below a predefined value.
  • MSE Mean Square Error
  • the dictionary can be shift-invariant.
  • a system is shift-invariant if f(x ⁇ ,y ⁇ ) ⁇ g(x ⁇ ,y ⁇ ) for arbitrary ⁇ and ⁇ .
  • Another embodiment of the invention may design a dictionary with non-negative dictionary values, wherein each atom contains only non-negative entries. Another option is to force zeros in predetermined places in the dictionary. It is possible to design the dictionary with any matrix structure.
  • Multiscale dictionaries or zeros in predefined places are two examples of a structure, but any structure can be used depending on the nature of the training signals and application in mind A person skilled in the art will easily design other properties in the dictionary according the training signals and the nature of the application. Such custom properties are all considered to be with the scope of the present invention.
  • multiscale dictionaries are built.
  • An image for example, can be defined using multiscale dictionaries, wherein each dictionary represents the image in a different size. Obviously, a smaller image will show fewer details than a bigger image.
  • the invention can be used for a variety of applications, including but not limited to: for compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • ⁇ k as a matrix of size N ⁇
  • x k R x k T ⁇ k
  • E R k E k ⁇ k , implying a selection of error columns that correspond to examples that use the atom d k .
  • K-SVD This algorithm has been herein named “K-SVD” to parallel the name K-Means. While K-Means applies K computations of means to update the codebook, the K-SVD obtains the updated dictionary by K SVD computations, each determining one column. A full description of the algorithm is given in FIG. 1 .
  • Step 1 generation of the data to train on:
  • White Gaussian noise with varying Signal to Noise Ration (SNR) was added to the resulting data signals.
  • SNR Signal to Noise Ration
  • Step 2 applying the K-SVD: The dictionary was initialized with data signals. The coefficients were found using OMP with fixed number of 3 coefficients. The maximum number of iterations was set to 80.
  • Step 3 compare to other reported works: we implemented the MOD algorithm, and applied it on the same data, using OMP with fixed number of 3 coefficients, and initializing in the same way.
  • Results the computed dictionary was compared against the known generating dictionary. This comparison was done by sweeping through the columns of the generating dictionary, and finding the closest column (in l 2 distance) in the computed dictionary, measuring the distance via
  • Training Data The training data was constructed as a set of 11,000 examples of block patches of size 8 ⁇ 8 pixels, taken from a database of face images (in various locations). A random collection of 500 such blocks, sorted by their variance, is presented in FIG. 3 .
  • Running the K-SVD We applied the K-SVD, training a dictionary of size 64 ⁇ 441.
  • the coefficients were computed using the OMP with fixed number of coefficients, where the maximal number of coefficients is 10. A better performance can be obtained by switching to FOCUSS.
  • the test was conducted using OMP because of its simplicity and fast execution.
  • the trained dictionary is presented in FIG. 4A .
  • the trained dictionary was compared with the overcomplete Haar dictionary which includes separable basis functions, having steps of various sizes and in all locations (total of 441 elements).
  • we built an overcomplete separable version of the DCT dictionary by sampling the cosine wave in different frequencies to result a total of 441 elements.
  • the overcomplete Haar dictionary is presented in FIG. 4B and the overcomplete DCT dictionary is presented in FIG. 4C .
  • the K-SVD results were used, denoted here as the learned dictionary, for two different applications on images. All tests were performed on one face image which was not included in the training set. The first application is filling-in missing pixels: random pixels in the image were deleted, and their values were filled using the various dictionaries decomposition. Then the compression potential of the learned dictionary decomposition was tested, and a rate-distortion graph was presented.
  • FIG. 6A shows a face with 50% missing pixels
  • FIGS. 6B , 6 C and 6 D show a learned reconstruction, a Haar reconstruction and a complete DCT reconstruction respectively
  • FIG. 6E shows a face with 70% missing pixels
  • FIGS. 6F , 6 G and 6 H show a learned reconstruction, a Haar reconstruction and a complete DCT reconstruction respectively.
  • PSNR 10 ⁇ log 10 ⁇ ( 1 e 2 ) . ( 35 )
  • R a ⁇ # ⁇ Blocks + # ⁇ coefs ⁇ ( b + Q ) # ⁇ pixels , ( 36 )
  • FIG. 7 presents the best obtained R-D curves for each dictionary.
  • the K-SVD dictionary outperforms all other dictionaries, and achieves up to 1-2 dB better for bit rates less than 1.5 bits-per-pixel (where the sparsity model holds true).
  • Samples results are presented in FIGS. 8A-8C .
  • FIG. 8A shows the result using the K-SVD dictionary
  • FIGS. 8B and 8C show the results using the overcomplete DCT dictionary and the complete DCT dictionary, respectively.

Abstract

A signal processing system adapted for sparse representation of signals is provided, comprising: (i)one or more training signals; (ii) a dictionary containing signal-atoms; (iii) a representation of each training signal using a linear combination of said dictionary's signal-atoms; (iv) means for updating the representation of the training signal; (v) means for updating the dictionary one group of atoms at a time, wherein each atom update may include all representations referring to said updated atom; and (vi) means for iterating (iv) and (v) until a stopping rule is fulfilled. The system uses the K-SVD algorithm for designing dictionaries for sparse representation of signals.

Description

    FIELD OF THE INVENTION
  • The present invention relates to a system and method for the sparse representation of signals. The invention is particularly relevant for applications such as compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • BACKGROUND OF THE INVENTION Sparse Representation of Signals
  • In recent years there has been a growing interest in the study of sparse representation of signals. Using an overcomplete dictionary that contains prototype signal-atoms, signals are described as sparse linear combinations of these atoms. Applications that use sparse representation are many and include compression, regularization in inverse problems, feature extraction, and more. Recent activity in this field concentrated mainly on the study of pursuit algorithms that decompose signals with respect to a given dictionary. Designing dictionaries to better fit the above model can be done by either selecting one from a pre-specified set of linear transforms, or by adapting the dictionary to a set of training signals. Both these techniques have been considered in recent years, but this topic is largely still open.
  • Using an overcomplete dictionary matrix D∈
    Figure US20120177128A1-20120712-P00001
    n×K that contains K prototype signal-atoms for columns, {dj}j=1 K, it is assumed that a signal y∈
    Figure US20120177128A1-20120712-P00001
    n can be represented as a sparse linear combination of these atoms. The representation of y may either be exact y=Dx, or approximate, y≈Dx, satisfying ∥y−Dx∥p≦ε. The vector x∈
    Figure US20120177128A1-20120712-P00001
    K displays the representation coefficients of the signal y. In approximation methods, typical norms used for measuring the deviation are the lp-norms for p=1, 2 and ∞.
  • If n<K and D is a full-rank matrix, an infinite number of solutions are available for the representation problem, hence constraints on the solution must be set. The solution with the fewest number of nonzero coefficients is certainly an appealing representation. This, sparsest representation, is the solution of either
  • ( P 0 ) min x x 0 subject to y = Dx . or ( 1 ) ( P 0 , ε ) min x x 0 subject to y - Dx 2 ε , ( 2 )
  • where ∥·∥0 is the l0 norm, counting the non zero entries of a vector.
  • Applications that can benefit from the sparsity and overcompleteness concepts (together or separately) include compression, regularization in inverse problems, feature extraction, and more. Indeed, the success of the JPEG2000 coding standard can be attributed to the sparsity of the wavelet coefficients of natural images. In denoising (removal of noise from noisy data so to obtain the unknown or original signal), wavelet methods and shift-invariant variations that exploit overcomplete representation, are among the most effective known algorithms for this task. Sparsity and overcompleteness have been successfully used for dynamic range compression in images, separation of texture and cartoon content in images, inpainting (changing an image so that the change is not noticeable by an observer) and restoration, and more.
  • In order to use overcomplete and sparse representations in applications, one needs to fix a dictionary D, and then find efficient ways to solve (1) or (2). Recent activity in this field has been concentrated mostly on the study of so called pursuit algorithms that represent signals with respect to a known dictionary, and approximate the solutions of (1) and (2). Exact determination of sparsest representations proves to be an NP-hard problem. Thus, approximate solutions are considered instead, and several efficient pursuit algorithms have been proposed in the last decade. The simplest ones are the Matching Pursuit (MP) and the Orthogonal Matching Pursuit (OMP) algorithms. Both are greedy algorithms that select the dictionary atoms sequentially. These methods are very simple, involving the computation of inner products between the signal and dictionary columns' and possibly deploying some least squares solvers. Both (1) and (2) are easily addressed by changing the stopping rule of the algorithm.
  • A second well known pursuit approach is the Basis Pursuit (BP). It suggests a convexisation of the problems posed in (1) and (2), by replacing the l0-norm with an λ1-norm. The Focal Under-determined System Solver (FOCUSS) is very similar, using the lp-norm with p≦1, as a replacement to the l0-norm. Here, for p<1 the similarity to the true sparsity measure is better, but the overall problem becomes non-convex, giving rise to local minima that may divert the optimization. Lagrange multipliers are used to convert the constraint into a penalty term, and an iterative method is derived based on the idea of iterated reweighed least-squares that handles the lp-norm as an l2 weighted one.
  • Both the BP and FOCUSS can be motivated based on Maximum A Posteriori (MAP) estimation and indeed several works used this reasoning directly. The MAP can be used to estimate the coefficients as random variables by maximizing the posterior P(x|y,D)αP(y|D,x)P(x). The prior distribution on the coefficient vector x is assumed to be a super-Gaussian Independent Identically-Distributed (iid) distribution that favors sparsity. For the Laplace distribution this approach is equivalent to BP.
  • Extensive study of these algorithms in recent years has established that if the sought solution, x, is sparse enough, these techniques recover it well in the exact case. Further work considered the approximated versions and has shown stability in recovery of x. The recent front of activity revisits those questions within a probabilistic setting, obtaining more realistic assessments on pursuit algorithms performance and success. The properties of the dictionary D set the limits that may be assumed on the sparsity that consequently ensure successful approximation. Interestingly, in all the works mentioned so far, there is a preliminary assumption that the dictionary is known and fixed. There is a great need to address the issue of designing the proper dictionary in order to better fit the sparsity model imposed.
  • The Choice of the Dictionary
  • An overcomplete dictionary D that leads to sparse representations can either be chosen as a pre-specified set of functions, or designed by adapting its content to fit a given set of signal examples.
  • Choosing a pre-specified transform matrix is appealing because it is simpler. Also, in many cases it leads to simple and fast algorithms for the evaluation of the sparse representation. This is indeed the case for overcomplete wavelets, curvelets, contourlets, steerable wavelet filters, short-time-Fourier transforms, and more. Preference is typically given to tight frames that can easily be pseudo-inverted. The success of such dictionaries in applications depends on how suitable they are to sparsely describe the signals in question. Multiscale analysis with oriented basis functions and a shift-invariant property are guidelines in such constructions.
  • There is need to develop a different route for designing dictionaries D based on learning, and find the dictionary D that yields sparse representations for the training signals. Such dictionaries have the potential to outperform commonly used pre-determined dictionaries. With ever-growing computational capabilities, computational cost may become secondary in importance to the improved performance achievable by methods which adapt dictionaries for special classes of signals.
  • Sparse coding is the process of computing the representation coefficients, x, based on the given signal y and the dictionary D. This process, commonly referred to as “atom decomposition”, requires solving (1) or (2), and this is typically done by a “pursuit algorithm” that finds an approximate solution. Three popular pursuit algorithms are the Orthogonal Matching Pursuit (OMP), Basis Pursuit (BP) and the Focal Under-determined System Solver (FOCUSS).
  • Orthogonal Matching Pursuit is a greedy step-wise regression algorithm. At each stage this method selects the dictionary element having the maximal projection onto the residual signal. After each selection, the representation coefficients with regarding to the so far chosen atoms are found via least-squares. Formally, given a signal y ∈
    Figure US20120177128A1-20120712-P00001
    n, and a dictionary D with K l2-normalized columns {dk}k=1 K, one starts by setting r0=y, k=1, and performing the following steps:
    • 1) Select the index of the next dictionary element ik=argmaxw|(rk−1, dw)|;
    • 2) Update the current approximation yk=argminy k ∥y−yk2 2 such that yk ∈ span {di 1 , di 2 , . . . di k }; and
    • 3) Update the residual rk=y−yk.
  • The algorithm can be stopped after a predetermined number of steps, hence after having selected a fixed number of atoms. Alternatively, the stopping rule can be based on norm of the residual, or on the maximal inner product computed in the next atom selection stage.
  • OMP is an appealing and very simple to implement algorithm. Unlike other methods, it can be easily programmed to supply a representation with an a priori fixed number of non-zero entries−a desired outcome in the training of dictionaries. There are several variants of the OMP that suggest (i) skipping the least-squares and using inner product itself as a coefficient, or (ii) applying least-squares per every candidate atom, rather than just using inner-products at the selection stage, or (iii) doing faster and less precise search, where instead of searching for the maximal inner product, a nearly maximal one is selected, thereby speeding up the search.
  • Theoretic study has shown that the OMP solution is also the sparsest available one (solving (1)) if some conditions on the dictionary and on the exact solution prevail. More recent work has shown that the above is also true for the approximation version (2). These results and some later ones that apply to the basis pursuit and FOCUSS involve a key feature of the dictionary D called the mutual incoherence and defined as:
  • μ = max i j d i T d j . ( 3 )
  • This measure quantifies how similar two columns of the dictionary can be. Given μ, the sparse representation to be found has fewer than O(1/μ) non-zeros, the OMP and its variants are guaranteed to succeed in recovering it.
  • Basis Pursuit (BP) algorithm proposes the replacement of the l0-norm in (1) and (2) with an l1-norm. Hence solutions of:
  • ( P 1 ) min x x 1 subject to y = Dx , ( 4 )
  • in the exact representation case, and
  • ( P 1 , ε ) min x x 1 subject to y - Dx ε , ( 5 )
  • in the approximate one, lead to the BP representations. Solution of (4) amounts to linear programming, and thus there exists efficient solvers for such problems.
  • Recent research addressed the connection between the (P0) and (P1). The essential claims are quite similar to the ones of OMP, namely, if the signal representation to be found has fewer than O(1/μ) non-zeros, the BP is guaranteed to succeed in recovering it. Similar results exist for the approximated case, proving that recovered representations are very close to the original sparse one in case of high sparsity.
  • Focal Under-determined System Solver (FOCUSS) is an approximating algorithm for finding the solutions of either (1) or (2), by replacing the l0-norm with an lp one for p≦1.
  • For the exact case problem, (P0), this method requires solving
  • ( P p ) min x x p subject to y = Dx . ( 6 )
  • The use of a Lagrange multiplier vector λ∈
    Figure US20120177128A1-20120712-P00001
    n here yields the Lagrangian function

  • ζ(x, λ)=∥x∥ pT(y−Dx).   (7)
  • Hence necessary conditions for a pair x, λ to be a solution of 6 are

  • xζ(x, λ)=pΠ(x)x−D Tλ=0 and ∇λζ(x, λ)=Dx−y=0,   (8)
  • where Π(x) is defined to be a diagonal matrix with |xi|p−2 as its (i, i)th entry. The split of the lp-norm derivative into a linear term multiplied by a weight matrix is the core of the FOCUSS method, and this follows the well-known idea of iterated reweighed least-squares. Several simple steps of algebra leads to the solution:

  • x=Π(x)−1 D T((x)−1 D T)−1 y.   (9)
  • While it is impossible to get a closed form solution for x from the above result, an iterative replacement procedure can be proposed, where the right hand side is computed based on the currently known xk−1, and this leads to the updating process,

  • x k=Π(x k−1)−1 D T((x k−1)−1 D T)−1 y.   (10)
  • A regularization can, and should, be introduced to avoid near-zero entries in the weight matrix Π(x).
  • For the treatment of (P0,ε) via the (Pp,ε) parallel expressions can be derived quite similarly, although in this case the determination of the Lagrange multiplier is more difficult and must be searched within the algorithm.
  • Recent work analyzed the (Pp) problem and showed its equivalence to the (P0), under conditions similar in flavor to the sparsity conditions mentioned above. Hence, this approach too enjoys the support of some theoretical justification, like BP and OMP. However, the analysis does not say anything about local minima traps and prospects in hitting those in the FOCUSS-algorithm.
  • Design of Dictionaries: Prior Art
  • There has been some work in the field regarding the training of dictionaries based on a set of examples. Given such set Y={yd}d=1 N, we assume that there exists a dictionary D that gave rise to the given signal examples via sparse combinations, i.e., we assume that there exists D, so that solving (P0) for each example yk gives a sparse representation xk. It is in this setting that the question is raised what the proper dictionary D is.
  • A. Generalizing the K-Means
  • There is an intriguing relation between sparse representation and clustering (i.e., vector quantization). In clustering, a set of descriptive vectors {dk}k=1 K is learned, and each sample is represented by one of those vectors (the one closest to it, usually in the l2 distance measure). One can think of this as an extreme sparse representation, where only one atom is allowed in the signal decomposition, and furthermore, the coefficient multiplying it must be 1. There is a variant of the vector quantization (VQ) coding method, called Gain-Shape VQ, where this coefficient is allowed to vary. In contrast, in sparse representations relevant to the invention, each example is represented as a linear combination of several vectors {dk}k=1 K. Thus, sparse representations can be referred to as a generalization of the clustering problem.
  • Since the K-Means algorithm (also known as generalized Lloyd algorithm—GLA) is the most commonly used procedure for training in the vector quantization setting, it is natural to consider generalizations of this algorithm when turning to the problem of dictionary training The K-Means process applies two steps per each iteration: (i) given {dk}k=1 K, assign the training examples to their nearest neighbor; and (ii) given that assignment, update {dk}k=1 K to better fit the examples.
  • The approaches to dictionary design that have been tried so far are very much in line with the two-step process described above. The first step finds the coefficients given the dictionary—a step we shall refer to as “sparse coding”. Then, the dictionary is updated assuming known and fixed coefficients. The differences between the various algorithms that have been proposed are in the method used for the calculation of coefficients, and in the procedure used for modifying the dictionary.
  • B. Maximum Likelihood Methods
  • Maximum likelihood methods use probabilistic reasoning in the construction of D. The proposed model suggests that for every example y the relation

  • y=Dx+v,   (11)
  • holds true with a sparse representation x and Gaussian white residual vector v with variance σhu 2. Given the examples Y={yi}i=1 N these works consider the likelihood function P (Y|D) and seek the dictionary that maximizes it. Two assumptions are required in order to proceed—the first is that the measurements are drawn independently, readily providing
  • P ( Y D ) = i = 1 N P ( y i D ) . ( 12 )
  • The second assumption is critical and refers to the “hidden variable” x. The ingredients of the likelihood function are computed using the relation

  • P(y i |D)=∫P(y i , x|D)dx=∫P(y i |x, DP(x)dx   (13)
  • Returning to the initial assumption in (11), we have
  • P ( y i x , D ) = Const · exp { 1 2 σ 2 Dx - y i 2 } . ( 14 )
  • The prior distribution of the representation vector x is assumed to be such that the entries of x are zero-mean iid, with Cauchy or Laplace distributions. Assuming for example a Laplace distribution we get
  • P ( y i D ) = P ( y i x , D ) · P ( x ) x = Const . exp { 1 2 σ 2 Dx - y i 2 } · exp { λ x 1 } x ( 15 )
  • This integration over x is difficult to evaluate, and indeed, it has been handled by replacing it with the extremal value of P(yi, x|D). The overall problem turns into
  • D = arg max D i = 1 N max x i { P ( y i , x i D ) } = arg min D i = 1 N min x i { Dx i - y i 2 + λ x i 1 } ( 16 )
  • This problem does not penalize the entries of D as it does for the ones of xi. Thus, the solution will tend to increase the dictionary entries' values, in order to allow the coefficients to become closer to zero. This difficulty has been handled by constraining the l2-norm of each basis element, so that the output variance of the coefficients is kept at an appropriate level.
  • An iterative method was suggested for solving (16). It includes two main steps in each iteration: (i) calculate the coefficients xi using a simple gradient descent procedure; and then (ii) update the dictionary using
  • D ( n + 1 ) = D ( n ) - η i = 1 N ( D ( n ) x i - y i ) x i T ( 17 )
  • This idea of iterative refinement, mentioned before as a generalization of the K-Means algorithm, was later used again by other researchers, with some variations.
  • A different approach to handle the integration in (15) has been suggested. It consisted in approximating the posterior as a Gaussian, enabling an analytic solution of the integration. This allows an objective comparison of different image models (basis or priors). It also removes the need for the additional re-scaling that enforces the norm constraint. However, this model may be too limited in describing the true behaviors expected. This technique and closely related ones have been referred to as approximated ML techniques.
  • There is an interesting relation between the maximum likelihood method and the Independent Component Analysis (ICA) algorithm. The latter handles the case of a complete dictionary (the number of elements equals the dimensionality) without assuming additive noise. The maximum likelihood method is then similar to ICA in that the algorithm can be interpreted as trying to maximize the mutual information between the inputs (samples) and the outputs (the coefficients).
  • C. The Method of Optimal Directions
  • The Method of Optimal Directions (MOD), a dictionary-training algorithm, follows more closely the K-Means outline, with a sparse coding stage that uses either the OMP or FOCUSS, followed by an update of the dictionary. The main contribution of the MOD method is its simple way of updating the dictionary.
  • Assuming that the sparse coding for each example is known, we define the errors ei=yi−Dxi. The overall representation mean square error is given by

  • ∥E∥ F 2 =∥[e 1 , e 2 , . . . , e N]∥F 2 =∥Y−DX∥ F 2.   (18)
  • Here we have concatenated all the examples y, as columns of the matrix Y, and similarly gathered the representations coefficient vectors x, to build the matrix X. The notation ∥A∥F stands for the Frobenius Norm, defined as ∥A∥F=√{square root over (ΣUAU 2)}.
  • Assuming that X is fixed, we can seek an update to D such that the above error is minimized. Taking the derivative of (10) with respect to D we obtain the relation (Y−DX)XT=0, leading to

  • D (N+1) =YX (n) T ·(X (n) X (n) T )−1   (19)
  • In updating the dictionary, the update relation given in (19) is the best that can be achieved for fixed X. The iterative steepest descent update in (17) is far slower. Interestingly, in both stages of the algorithm, the difference is in deploying a second order (Newtonian) update instead of a first-order one. Looking closely at the update relation in (17), it could be written as
  • D ( n + 1 ) = D ( n ) + η EX ( n ) T = D ( n ) + η ( Y - D ( n ) X ( n ) ) X ( n ) T = D ( n ) ( 1 - η X ( n ) X ( n ) T ) + η YX ( n ) T . ( 20 )
  • Using infinitely many iterations of this sort, and using small enough η, this leads to a steady state outcome, which is exactly the MOD update matrix (19). Thus, while the MOD method assumes known coefficients at each iteration, and derives the best possible dictionary, the ML method by Olshausen and Field only gets closer to this best current solution, and then turns to calculate the coefficients. Note, however, that in both methods a normalization of the dictionary columns is required and done.
  • D. Maximum A-Posteriori Probability Approach
  • The same researchers that conceived the MOD method also suggested a maximum a-posteriori probability (MAP) setting for the training of dictionaries, attempting to merge the efficiency of the MOD with a natural way to take into account preferences in the recovered dictionary. This probabilistic point of view is very similar to the ML methods discussed above. However, rather than working with the likelihood function P(Y|D), the posterior P(D|Y) is used. Using Bayes rule, we have P(D|Y)αP(Y|D)P(D), and thus we can use the likelihood expression as before, and add a prior on the dictionary as a new ingredient.
  • Research currents considered several priors P(D) and per each proposed an update formula for the dictionary. The efficiency of the MOD in these methods is manifested in the efficient sparse coding, which is carried out with FOCUSS. The proposed algorithms in this family deliberately avoid a direct minimization with respect to D as in MOD, due to the prohibitive n×n matrix inversion required. Instead, iterative gradient descent is used.
  • When no prior is chosen, the update formula is the very one used in (17). A prior that constrains D to have a unit Frobenius norm leads to the update formula

  • D (n+1) =D (n) +ηEX T +η·tr(XE T D (n))D (n).   (21)
  • As can be seen, the first two terms are the same ones as in (17). The last term compensates for deviations from the constraint. This case allows different columns in D to have different norm values. As a consequence, columns with small norm values tend to be under-used, as the coefficients they need are larger and as such more penalized.
  • This led to the second prior choice, constraining the columns of D to have a unit l2-norm. The new update equation formed is given by

  • d i (n+1) =d i (n)+η(I−d i (n) d i (n) T )E·x i T.   (22)
  • where xT i is the i-th column in the matrix XT.
  • Compared to the MOD, this line of work provides slower training algorithms.
  • E. Unions of Orthogonal Bases
  • Recent work considered a dictionary composed as a union of orthonormal bases

  • D=[D1; D2. . . , DL],
  • where Dj
    Figure US20120177128A1-20120712-P00001
    n×n, j=1, 2, . . . , L are orthonormal matrices. Such a dictionary structure is quite restrictive, but its updating may potentially be made more efficient.
  • The coefficients of the sparse representations X can be decomposed to L pieces, each referring to a different ortho-basis. Thus,

  • X=[X 1 , X 2 , . . . , X L]T,
  • where Xi is the matrix containing the coefficients of the orthonormal dictionary Di.
  • One of the major advantages of the union of ortho-bases is the relative simplicity of the pursuit algorithm needed for the sparse coding stage. The coefficients are found using the Block Coordinate Relaxation (BCR) algorithm. This is an appealing way to solve (P1,ε) as a sequence of simple shrinkage steps, such that at each stage Xi is computed, while keeping all the other pieces of X fixed. Thus, this evaluation amounts to a simple shrinkage.
  • Assuming known coefficients, the proposed algorithm updates each orthonormal basis Dj sequentially. The update of Dj is done by first computing the residual matrix
  • E j = [ e 1 , e 2 , , e N ] = Y - i j D i X i ,
  • Then, by computing the singular value decomposition of the matrix EjXT j=UΛVT, the update of the j-th ortho-basis is done by Dj=UVT. This update rule is obtained by solving a constrained least squares problem with ∥Ej−DjXjF 2 as the penalty term, assuming fixed coefficients Xj and error Ej. The constraint is over the feasible matrices Dj, which are forced to be orthonormal.
  • This way the proposed algorithm improves each matrix Dj separately, by replacing the role of the data matrix Y in the residual matrix Ej, as the latter should be represented by this updated basis.
  • Grinbonval suggested a slightly different method. Apart from the fact that here the dictionary is structured, handling a union of orthonormal bases, it updates each orthonormal bases sequentially, and thus reminds the sequential updates done in the K-means. Experimental results show weak performance compared to previous methods. This could partly be explained by the fact that the update of Dj depends strongly on the coefficients Xj.
  • K-Means Algorithm for Vector Quantization (VQ)
  • In VQ, a codebook C that includes K codewords (representatives) is used to represent a wide family of vectors (signals) Y={yi}l=1 N(N>K) by a nearest neighbor assignment. This leads to an efficient compression or description of those signals, as clusters in
    Figure US20120177128A1-20120712-P00001
    n surrounding the chosen codewords. Based on the expectation maximization procedure, the K-Means can be extended to suggest a fuzzy assignment and a covariance matrix per each cluster, so that the data is modeled as a mixture of Gaussians.
  • The dictionary of VQ codewords is typically trained using the K-Means algorithm. We denote the codebook matrix by C=[c1, c2, . . . , cK], the codewords being the columns When C is given, each signal is represented as its closest codeword (under l2-norm distance). We can write yi=Cxi, where xi=ej is a vector from the trivial basis, with all zero entries except a one in the j-th position. The index j is selected such that

  • k≠j ∥y i −Ce j2 2 ≦∥y i −Ce k2 2.
  • This is considered as an extreme case of sparse coding in the sense that only one atom is allowed to participate in the construction of yi and the coefficient is forced to be 1. The representation MSE per yi is defined as

  • e x 2 =∥y i −Cx i2 2.   (23)
  • and the overall MSE is
  • E = i = 1 K ɛ i 2 = Y - CX F 2 . ( 24 )
  • The VQ training problem is to find a codebook C that minimizes the error E, subject to the limited structure of X, whose columns must be taken from the trivial basis,
  • min C , X { Y - CX F 2 } subject to i , x j = e k for some k . ( 25 )
  • The K-Means algorithm is an iterative method used for designing the optimal codebook for VQ. In each iteration there are two stages—one for sparse coding that essentially evaluates X, and one for updating the codebook.
  • The sparse coding stage assumes a known codebook C(J−1), and computes a feasible X that minimizes the value of (25). Similarly, the dictionary update stage fixes X as known, and seeks an update of C so as to minimize (25). Clearly, at each iteration either a reduction or no change in the MSE is ensured. Furthermore, at each such stage, the minimization step is optimal under the assumptions. As the MSE is bounded from below by zero, and the algorithm ensures a monotonic decrease of the MSE, convergence to at least a local minimum solution is guaranteed. Stopping rules for the above-described algorithm can vary a lot but are quite easy to handle.
  • SUMMARY OF THE PRIOR ART
  • Almost all previous methods can essentially be interpreted as generalizations of the K-Means algorithm, and yet, there are marked differences between these procedures. In the quest for a successful dictionary training algorithm, there are several desirable properties:
  • (i) Flexibility: the algorithm should be able to run with any pursuit algorithm, and this way enable choosing the one adequate for the run-time constraints, or the one planned for future usage in conjunction with the obtained dictionary. Methods that decouple the sparse-coding stage from the dictionary update readily have such a property. Such is the case with the MOD and the MAP based methods.
  • (ii) Simplicity: much of the appeal of a proposed dictionary training method has to do with how simple it is, and more specifically, how similar it is to K-Means. It is desirable to have an algorithm that may be regarded as a natural generalization of the K-Means. The algorithm should emulate the ease with which the K-Means is explainable and implementable. Again, the MOD seems to have made a substantial progress in this direction, although there is still room for improvement.
  • (iii) Efficiency: the proposed algorithm should be numerically efficient and exhibit fast convergence. The above described methods are all quite slow. The MOD, which has a second-order update formula, is nearly impractical in reasonable dimensions, because of the matrix inversion step involved. Also, in all the above formulations, the dictionary columns are updated before turning to re-evaluate the coefficients. This approach inflicts a severe limitation on the training speed.
  • (iv) Well Defined Objective: for a method to succeed, it should have a well defined objective function that measures the quality of the solution obtained. This almost trivial fact was overlooked in some of the preceding work in this field. Hence, even though an algorithm can be designed to greedily improve the representation Mean Square Error (MSE) and the sparsity, it may happen that the algorithm leads to aimless oscillations in terms of a global objective measure of quality.
  • SUMMARY OF THE INVENTION
  • It is an object of the present invention to design a dictionary based on learning from training signals, wherein the dictionary yields sparse representations for a set of training signals. These dictionaries have the potential to outperform commonly used pre-determined dictionaries.
  • The invention thus relate to a novel system and algorithm for adapting dictionaries so as to represent signals sparsely. Given a set of training signals {yi}i=1 N, we seek the dictionary D that leads to the best possible representations for each member in this set with strict sparsity constraints. The invention introduces the K-SVD algorithm that addresses the above task, generalizing the K-Means algorithm. The K-SVD is an iterative method that alternates between sparse coding of the examples based on the current dictionary, and an update process for the dictionary atoms so as to better fit the data. The update of the dictionary columns is done jointly with an update of the sparse representation coefficients related to it, resulting in accelerated convergence. The K-SVD algorithm is flexible and can work with any pursuit method, thereby tailoring the dictionary to the application in mind.
  • The sparse representation problem can be viewed as a generalization of the VQ objective (25), in which we allow each input signal to be represented by a linear combination of codewords, which we now call dictionary elements. Therefore the coefficients vector is now allowed more than one nonzero entry, and these can have arbitrary values. For this case, the minimization corresponding to Equation (25) is that of searching the best possible dictionary for the sparse representation of the example set Y,
  • min D , X { Y - DX F 2 } subject to i , x i 0 T 0 . ( 26 )
  • A similar objective could alternatively be met by considering
  • min D , X i x i 0 subject to Y - DX F 2 ε , ( 27 )
  • for a fixed value ε. By disclosing the treatment for the first problem (26), any person skilled in the art would immediately realize that the treatment is very similar.
  • In the algorithm of the invention, we minimize the expression in (26) iteratively. First, we fix D and aim to find the best coefficient matrix X that can be found. As finding the truly optimal X is impossible, we use an approximation pursuit method. Any such algorithm can be used for the calculation of the coefficients, as long as it can supply a solution with a fixed and predetermined number of nonzero entries, T0.
  • Once the sparse coding task is done, a second stage is performed to search for a better dictionary. This process updates one column at a time, fixing all columns in D except one, dk, and finding a new column dk and new values for its coefficients that best reduce the MSE. This is markedly different from all the K-Means generalizations. K-Means generalization methods freeze X while finding a better D. The approach of the invention is different, as we change the columns of D sequentially, and allow changing the relevant coefficients. In a sense, this approach is a more direct generalization of the K-Means algorithm, because it updates each column separately, as done in K-Means.
  • The process of updating only one column of D at a time is a problem having a straightforward solution based on the singular value decomposition (SVD). Furthermore, allowing a change in the coefficients' values while updating the dictionary columns accelerates convergence, since the subsequent columns updates will be based on more relevant coefficients. The overall effect is very much in line with the leap from gradient descent to Gauss-Seidel methods in optimization.
  • A hypothetical alternative would be to skip the step of sparse coding, and use only updates of columns in D, along with their coefficients, applied in a cyclic fashion, again and again. This however will not work well, as the support of the representations will never be changed, and such an algorithm will necessarily fall into a local minimum trap.
  • The invention is useful for a variety of applications in signal processing including but not limited to: compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • Typically, all the training signals involved are from the same family and thus have common traits. For examples, the signals can all be pictures, music, speech etc.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a description of the K-SVD algorithm of the invention.
  • FIG. 2 is graph of synthetic results comparing K-SVD against two known algorithms, MOD and MAP-based algorithms. For each of the tested algorithms and for each noise level, 50 trials were conducted and their results sorted. The graph labels represent the mean number of detected atoms (out of 50) over the ordered tests in groups of 10 experiments.
  • FIG. 3 is a collection of 500 random block patches of size 8×8 pixels, taken from a database of face images, which were used for training with the K-SVD algorithm, sorted by their variance.
  • FIG. 4A depicts the learned dictionary (a K-SVD trained dictionary of size 64×441). Its elements are sorted in an ascending order of their variance, and stretched to maximal range for display purposes. FIGS. 4B and 4C depict the overcomplete separable Haar dictionary and the overcomplete DCT dictionary, respectively, of the same size (shown for comparison).
  • FIG. 5 is the RMSE for 594 new blocks with missing pixels using the learned dictionary og FIG. 4A, the overcomplete Haar dictionary and the overcomplete DCT dictionary.
  • FIGS. 6A-6H compare two corrupted images and their reconstruction, with the missing pixels marked as points (6A, 50% of missing pixels; 6E, 70% of missing pixels), and the reconstructed results by the learned dictionary (6B, 6F), the overcomplete Haar dictionary (6C, 6G), and the overcomplete DCT dictionary (6D, 6H), respectively.
  • FIG. 7 depicts Rate-Distortion graphs for the compression results for each dictionary.
  • FIGS. 8A-8C show sample compression results for the K-SVD, overcomplete DCT and complete DCT dictionaries, respectively.
  • DETAILED DESCRIPTION OF THE INVENTION
  • In the following detailed description of various embodiments, reference is made to the accompanying drawings that form a part thereof, and in which are shown by way of illustration specific embodiments in which the invention may be practiced. It is understood that other embodiments may be utilized and structural changes may be made without departing from the scope of the present invention.
  • In the present invention, we address the problem of designing dictionaries, and introduce the K-SVD algorithm for this task. We show how this algorithm can be interpreted as a generalization of the K-Means clustering process, and demonstrate its behavior in both synthetic tests and in applications on real data.
  • The present invention relates to a signal processing method adapted for sparse representation of signals and a system for implementing said method, said system comprising:
      • (i) one or more training signals;
      • (ii) a dictionary containing signal-atoms;
      • (iii) a representation of each training signal using a linear combination of said dictionary's signal-atoms;
      • (iv) means for updating the representation of the training signal;
      • (v) means for updating the dictionary one group of atoms at a time, wherein each atom update may include all representations referring to said updated atom; and
      • (vi) means for iterating (iv) and (v) until a stopping rule is fulfilled.
  • The training signals are typically from the same family and thus all training signals share common traits and have common behavior patterns. For example, all training signals can be pictures, including pictures of human faces, or the training signals can be sound files including music files, speeches, and the like.
  • The purpose of the dictionary of the present invention is to discover the common building blocks with which all the training signals can be represented. All the training signals can be represented by linear combinations of the dictionary atoms (building blocks). The term “atom” as referred to herein means dictionary atom or signal-atom.
  • In some cases, the building blocks or some of the building blocks of the training signals are known or can be approximated intuitively, while in other cases the invention helps to discover them.
  • According to a preferred embodiment, the dictionary is updated one atom at a time. It is possible however to also update the dictionary a group of atoms at a time, for example two or three atoms at a time, or defining the group of atoms to be updated containing any number of atoms.
  • In one embodiment of the present invention, the dictionary is an overcomplete dictionary. An overcomplete dictionary contains more atoms (building blocks, functions) than strictly necessary to represent the signal space. An overcomplete dictionary thus allows a suitable representation of a signal with fewer encoded atoms. This is important for applications in which a low bit rate is required.
  • Signals can be represented in many forms. In one embodiment of the present invention, the representation of each training signal is a coefficient matrix. The representation of the training signals may take any other form such as a vector.
  • There are many ways to generate a coefficient matrix representing the training signals. In one embodiment of the present invention, the generation of the coefficients matrix is achieved by a pursuit algorithm. The pursuit algorithm can include: Orthogonal Matching Pursuit, Matching Pursuit, Basis Pursuit, FOCUSS or any combination or variation thereof.
  • Updating the dictionary can be performed sequentially or in any other order. In yet another embodiment of the present invention, the dictionary is updated in a predefined order of the signal-atoms. Depending on the application used and the nature of the training signals, updating the dictionary in a predefined order of signal-atoms will yield different results and thus can be exploited by the application.
  • In another embodiment of the present invention, only selected signal-atoms of said dictionary are updated. Again, depending on the nature of the application in mind, one may decide to leave certain signal-atoms (building blocks) fixed, and consequently only update the remaining signal-atoms.
  • In some cases, it may happen that a dictionary is built wherein two signal-atoms are very similar to each other but not equal to each other. The similarity, for the purpose of the application used, may be too big, and thus the differentiation between the two atoms may be considered negligible. The application will thus wish to modify one of the similar atoms. In yet another embodiment of the present invention, a signal-atom is modified when the difference between said signal-atom to another signal atom is below a predefined value.
  • A signal-atom may be defined by the system as a building block for representing the training signals, but the actual signal-atom may never be used to construct any of the given training signals. One may thus wish to modify this atom. In a further embodiment of the present invention, a signal-atom is modified when it is not used in any representation.
  • A signal-atom may be found to be used only rarely to construct training signals. It may be thus preferred not to work with such a building block, and modify this atom to one used more frequently in training signals representation. In one embodiment of the present invention, a signal-atom is modified when its usage frequency in the representation of signal-atoms is below a predefined value.
  • When updating the dictionary, either a single atom or a group of atoms at a time, there are many possibilities to define the best results for the atom values. In yet another embodiment of the present invention, updating the group of atoms and their coefficients best reduces the Mean Square Error (MSE).
  • In some cases, again depending on the nature of the application used and of the training signals, it may be desired to design dictionaries with one or more custom properties. For example, the dictionary can be shift-invariant. A system is shift-invariant if f(x−α,y−β)→g(x−α,y−β) for arbitrary α and β. Another embodiment of the invention may design a dictionary with non-negative dictionary values, wherein each atom contains only non-negative entries. Another option is to force zeros in predetermined places in the dictionary. It is possible to design the dictionary with any matrix structure. Multiscale dictionaries or zeros in predefined places are two examples of a structure, but any structure can be used depending on the nature of the training signals and application in mind A person skilled in the art will easily design other properties in the dictionary according the training signals and the nature of the application. Such custom properties are all considered to be with the scope of the present invention.
  • In yet another embodiment of the present invention, multiscale dictionaries are built. An image, for example, can be defined using multiscale dictionaries, wherein each dictionary represents the image in a different size. Obviously, a smaller image will show fewer details than a bigger image.
  • The invention can be used for a variety of applications, including but not limited to: for compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
  • The K-SVD—Detailed Description
  • As mentioned previously, the objective function of the K-SVD is
  • min D , X { Y - DX F 2 } subject to i , x i 0 T 0 . ( 28 )
  • Let us first consider the sparse coding stage, where we assume that D is fixed, and consider the optimization problem as a search for sparse representations with coefficients summarized in the matrix X. The penalty term can be rewritten as
  • Y - DX F 2 = i = 1 N y i - Dx i 2 2 .
  • Therefore the problem posed in (28) can be decoupled to N distinct problems of the form
  • i = 1 , 2 , , N , min x i { y i - Dx 2 2 } subject to x i 0 T 0 . ( 29 )
  • This problem is adequately addressed by the pursuit algorithms mentioned before, and we have seen that if T0 is small enough, their solution is a good approximation to the ideal one that is numerically infeasible to compute.
  • We now turn to the second, and slightly more involved process of updating the dictionary together with the nonzero coefficients. Assume that both X and D are fixed, and we put in question only one column in the dictionary, dk, and the coefficients that correspond to it, the i-th row in X, denoted as xi T (this is not the vector xi which is the i-th column in X). Returning to the objective function (28), the penalty term can be rewritten as
  • Y - DX F 2 = Y - j = 1 K d j x T j F 2 = ( Y - j = 1 K d j x T j ) - d k x T k F 2 = E k - d k x T k F 2 . ( 30 )
  • We have decomposed the multiplication DX to the sum of K rank-1 matrices. Among those, K-1 terms are assumed fixed, and one—the k-th—remains in question. The matrix Ek stands for the error for all the N examples when the k-th atom is removed.
  • Here, it would be tempting to suggest the use of the SVD (Singular Value Decomposition) to find alternative dk and xk T. The SVD finds the closest rank-1 matrix (in Frobenius norm) that approximates Ek, and this will effectively minimize the error as defined in (30). However, such a step will be a mistake, because the new vector xk T is very likely to be filled, since in such an update of dk we do not enforce the sparsity constraint.
  • A remedy to the above problem, however, is simple and also quite intuitive. Define wi as the group of indices pointing to examples {yi} that use the atom dk, i.e., those where xk T (i) is nonzero. Thus,

  • w k ={i|1≦i≦K, x T k(i)≠0}.   (31)
  • Define Ωk as a matrix of size N×|wi|, with ones on the (wk(i), i)-th entries, and zeros elsewhere. When multiplying xk R=xk TΩk, this shrinks the row vector xk T by discarding of the zero entries, resulting with the row vector xk R of length |wk|. Similarly, the multiplication YR k=YΩk creates a matrix of size n×|wk| that includes a subset of the examples that are currently using the dk atom. The same effect happens with ER k=EkΩk, implying a selection of error columns that correspond to examples that use the atom dk.
  • With this notation, we can now return to (30) and suggest minimization with respect to both dk and xk T, but this time force the solution of xk T to have the same support as the original xk T. This is equivalent to the minimization of

  • ∥E kΩk −d k x T kΩkF 2 =∥E k R −d k x R kF 2.   (32)
  • and this time it can be done directly via SVD. Taking the restricted matrix ER k, SVD decomposes it to ER k=UΔVT. We define the solution for dk as the first column of U, and the coefficient vector xk R as the first column of V multiplied by Δ(1, 1). In this solution, we necessarily have that (i) the columns of D remain normalized; and (ii) the support of all representations either stays the same or gets smaller by possible nulling of terms.
  • This algorithm has been herein named “K-SVD” to parallel the name K-Means. While K-Means applies K computations of means to update the codebook, the K-SVD obtains the updated dictionary by K SVD computations, each determining one column. A full description of the algorithm is given in FIG. 1.
  • In the K-SVD algorithm we sweep through the columns and use always the most updated coefficients as they emerge from preceding SVD steps. Parallel versions of this algorithm can also be considered, where all updates of the previous dictionary are done based on the same X. Experiments show that while this version also converges, it yields an inferior solution, and typically requires more than 4 times the number of iterations. These parallel versions and variation are all encompassed by the present invention.
  • An important question that arises is: Will the K-SVD algorithm converge? Let us first assume we can perform the sparse coding stage perfectly, retrieving the best approximation to the signal y, that contains no more than T0 nonzero entries. In this case, and assuming a fixed dictionary D, each sparse coding step decreases the total representation error ∥Y−DX∥2 F, posed in (28). Moreover, at the update step for dk, an additional reduction or no change in the MSE is guaranteed, while not violating the sparsity constraint. Executing a series of such steps ensures a monotonic MSE reduction, and therefore, convergence to a local minimum is guaranteed.
  • Unfortunately, the above statement depends on the success of pursuit algorithms to robustly approximate the solution to (29), and thus convergence is not always guaranteed. However, when T0 is small enough relative to n, the OMP, FOCUSS, and BP approximating methods are known to perform very well. While OMP can be naturally used to get a fixed and pre-determined number of non-zeros (T0), both BP and FOCUSS require some slight modifications. For example, in using FOCUSS to lead to T0 non-zeros, the regularization parameter should be adapted while iterating. In those circumstances the convergence is guaranteed. We can ensure convergence by external interference—by comparing the best solution using the already given support to the one proposed by the new run of the pursuit algorithm, and adopting the better one. This way we shall always get an improvement. Practically, we saw in all our experiments that a convergence is reached, and there was no need for such external interference.
  • From K-SVD Back to K-Means
  • When the model order T0=1, this case corresponds to the gain-shape VQ, and as such it is important, as the K-SVD becomes a method for its codebook training. When T0=1, the coefficient matrix X has only one nonzero entry per column. Thus, computing the error ER k in (32), yields
  • E k R = E k Ω k = ( Y - j k d j x T j ) Ω k = Y Ω k = Y k R . ( 33 )
  • This is because the restriction Ωk takes only those columns in Ek that use the dk atom, and thus, necessarily, they use no other atoms, implying that for all j, xT jΩk=0.
  • The implication of the above outcome is that the SVD in the T0=1 case is done directly on the group of examples in wk. Also, the K updates of the columns of D become independent of each other, implying that a sequential process as before, or a parallel one, both lead to the same algorithm.
  • We could further constraint our representation stage and beyond the choice T0=1, limit the nonzero entries of X to be 1. This brings us back to the classical clustering problem as described earlier. In this case we have that xk R is filled with ones, thus xk R=1T. The K-SVD then needs to approximate the restricted error matrix ER k=YR k by a rank-1 matrix dk·1T. The solution is the mean of the columns of YR k, exactly as K-Means suggests.
  • K-SVD—Implementation Details
  • Just like the K-Means, the K-SVD algorithm is susceptible to local minimum traps. Our experiments show that improved results can be reached if the following variations are applied:
  • (i) When using approximation methods with a fixed number of coefficients, we found out that FOCUSS proves to be the best in terms of getting the best out of each iteration. However, from a run-time point of view, OMP was found to lead to far more efficient overall algorithm.
  • (ii) When a dictionary element is not being used “enough” (relative to the number of dictionary elements and to the number of samples) it could be replaced with the least represented data element, after being normalized (the representation is measured without the dictionary element that is going to be replaced). Since the number of data elements is much larger than the number of dictionary elements, and since our model assumption suggests that the dictionary atoms are of equal importance, such replacement is very effective in avoiding local minima and over-fitting.
  • (iii) Similar to the idea of removal of unpopular elements from the dictionary, we found that it is very effective to prune the dictionary from having too-close elements. If indeed such a pair of atoms is found (based on their absolute inner product exceeding some threshold), one of those elements should be removed and replaced with the least-represented signal.
  • Similarly to the K-Means, we can propose a variety of techniques to further improve the K-SVD algorithm. Appealing examples on this list are multi-scale approaches and tree-based training where the number of columns K is allowed to increase during the algorithm. All these variations, adaptations and improvements are encompassed by the present invention.
  • Synthetic Experiments
  • We have first tried the K-SVD algorithm on synthetic signals, to test whether this algorithm recovers the original dictionary that generated the data, and to compare its results with other reported algorithms.
  • Step 1—generation of the data to train on: A random matrix D (referred to later-on as the generating dictionary) of size 20×50 was generated with iid uniformly distributed entries. Each column was normalized to a unit l2-norm. Then, 1500 data signals {yi}i=1 1500 of dimension 20 were produced, each created by a linear combination of 3 different generating dictionary atoms, with uniformly distributed iid coefficients in random and independent locations. White Gaussian noise with varying Signal to Noise Ration (SNR) was added to the resulting data signals.
  • Step 2—applying the K-SVD: The dictionary was initialized with data signals. The coefficients were found using OMP with fixed number of 3 coefficients. The maximum number of iterations was set to 80.
  • Step 3—comparison to other reported works: we implemented the MOD algorithm, and applied it on the same data, using OMP with fixed number of 3 coefficients, and initializing in the same way. We executed the MOD algorithm for a total number of 80 iterations. We also executed the MAP-based algorithm of Rao and Kreutz-Delgado (Kreutz-Delgado et al., Dictionary learning algorithms for sparse representation. Neural Computation. 15(2):349-396, 2003). This algorithm was executed as is, therefore using FOCUSS as its decomposition method. Here, again, a maximum of 80 iterations were allowed.
  • Results: the computed dictionary was compared against the known generating dictionary. This comparison was done by sweeping through the columns of the generating dictionary, and finding the closest column (in l2 distance) in the computed dictionary, measuring the distance via

  • 1−|di T{tilde over (d)}i|,   (34)
  • where di is a generating dictionary atom, and {tilde over (d)}i is its corresponding element in the recovered dictionary. A distance less than 0.01 was considered a success. All trials were repeated 50 times, and the number of successes in each trial was computed. The results for the three algorithms and for noise levels of 10 dB, 20 dB, 30 dB and ∞ dB (no noise) are displayed in FIG. 2.
  • We should note that for different dictionary size (e.g., 20×30) and with more executed iterations, the MAP-based algorithm improves and gets closer to the
  • K-SVD detection rates.
  • Applications to Image Processing
  • Several experiments have been conducted on natural image data, trying to show the practicality of the algorithm of the invention and the general sparse coding theme. These preliminary tests prove the concept of using such dictionaries with sparse representations.
  • Training Data: The training data was constructed as a set of 11,000 examples of block patches of size 8×8 pixels, taken from a database of face images (in various locations). A random collection of 500 such blocks, sorted by their variance, is presented in FIG. 3.
  • Removal of the DC: Working with real images data we preferred that all dictionary elements except one has a zero mean. For this purpose, the first dictionary element, denoted as the DC, was set to include a constant value in all its entries, and was not changed afterwards. The DC takes part in all representations, and as a result, all other dictionary elements remain with zero mean during all iterations.
  • Running the K-SVD: We applied the K-SVD, training a dictionary of size 64×441. The choice K=441 came from our attempt to compare the outcome to the overcomplete Haar dictionary of the same size. The coefficients were computed using the OMP with fixed number of coefficients, where the maximal number of coefficients is 10. A better performance can be obtained by switching to FOCUSS. The test was conducted using OMP because of its simplicity and fast execution. The trained dictionary is presented in FIG. 4A.
  • Comparison Dictionaries: The trained dictionary was compared with the overcomplete Haar dictionary which includes separable basis functions, having steps of various sizes and in all locations (total of 441 elements). In addition, we built an overcomplete separable version of the DCT dictionary by sampling the cosine wave in different frequencies to result a total of 441 elements. The overcomplete Haar dictionary is presented in FIG. 4B and the overcomplete DCT dictionary is presented in FIG. 4C.
  • Applications. The K-SVD results were used, denoted here as the learned dictionary, for two different applications on images. All tests were performed on one face image which was not included in the training set. The first application is filling-in missing pixels: random pixels in the image were deleted, and their values were filled using the various dictionaries decomposition. Then the compression potential of the learned dictionary decomposition was tested, and a rate-distortion graph was presented. These experiments will be described in more detail hereafter.
  • A. Filling-In Missing Pixels
  • One random full face image was chosen, which consists of 594 non-overlapping blocks (none of which were used for training). For each block, the following procedure was conducted for r in the range {0.2, 0.9}:
  • (i) A fraction r of the pixels in each block, in random locations, were deleted (set to zero).
  • (ii) The coefficients of the corrupted block under the learned dictionary, the overcomplete Haar dictionary, and the overcomplete DCT dictionary were found using OMP with an error bound of ∥0.02·1∥2, where 1∈Rn is a vector of all ones (the input image is scald to the dynamic range [0, 1]), (allowing an error of ±5 gray-values in 8-bit images). All projections in the OMP algorithm included only the non-corrupted pixels, and for this purpose, the dictionary elements were normalized so that the non-corrupted indices in each dictionary element have a unit norm. The resulting coefficient vector of the block B is denoted xB.
  • (iii) The reconstructed block {tilde over (B)} was chosen as {tilde over (B)}=D·xB.
  • (iv) The reconstruction error was set to: √{square root over (∥B−{tilde over (B)}∥F 2/64)} (64 is the number of pixels in each block). The mean reconstruction errors (for all blocks and all corruption rates) were computed, and are displayed in FIG. 5. Two corrupted images and their reconstructions can be seen in FIGS. 6A-6H. FIG. 6A shows a face with 50% missing pixels, where FIGS. 6B, 6C and 6D show a learned reconstruction, a Haar reconstruction and a complete DCT reconstruction respectively. FIG. 6E shows a face with 70% missing pixels, where FIGS. 6F, 6G and 6H show a learned reconstruction, a Haar reconstruction and a complete DCT reconstruction respectively. As can be seen, higher quality recovery is obtained using the learned dictionary.
  • B. Compression
  • A compression comparison was conducted between the overcomplete learned dictionary, the overcomplete Haar dictionary, and the overcomplete DCT dictionary (as described before), all of size 64×441. In addition, a comparison was made to the regular (unitary) DCT dictionary (used by the JPEG algorithm). The resulted rate-distortion graph is presented in FIG. 7. In this compression test, the face image was partitioned (again) into 594 disjoint 8×8 blocks. All blocks were coded in various rates (bits-per-pixel values), and the Peak Signal-to-Noise Ratio (PSNR) was measured. Let I be the original image and Ĩ be the coded image, combined by all the coded blocks. We denote ε2 as the mean squared error between I and Ĩ, and calculate
  • PSNR = 10 · log 10 ( 1 e 2 ) . ( 35 )
  • In each test we set an error goal ε, and fixed the number of bits-per-coefficient Q. For each such pair of parameters, all blocks were coded in order to achieve the desired error goal, and the coefficients were quantized to the desired number of bits (uniform quantization, using upper and lower bounds for each coefficient in each dictionary based on the training set coefficients). For the overcomplete dictionaries, the OMP coding method was used. The rate value was defined as
  • R = a · # Blocks + # coefs · ( b + Q ) # pixels , ( 36 )
  • where
      • a holds the required number of bits to code the number of coefficients for each block.
      • b holds the required number of bits to code the index of the representing atom. Both a and b values were calculated using an entropy coder.
      • #·Blocks is the number of blocks in the image (594).
      • #·coefs is the total number of coefficients required to represent the whole image.
      • #·pixels is the number of pixels in the image (=64×Blocks).
        In the unitary DCT dictionary the coefficients were picked in a zigzag order, as done by JPEG, until the error bound ε is reached. Therefore, the index of each atom should not be coded, and the rate was defined by,
  • R = a · # Blocks + coefs · Q # pixels , ( 37 )
  • with the same notation as before.
  • By sweeping through various values of ε and Q we get per each dictionary several curves in the R-D plane. FIG. 7 presents the best obtained R-D curves for each dictionary. As can be seen, the K-SVD dictionary outperforms all other dictionaries, and achieves up to 1-2 dB better for bit rates less than 1.5 bits-per-pixel (where the sparsity model holds true). Samples results are presented in FIGS. 8A-8C. FIG. 8A shows the result using the K-SVD dictionary, while FIGS. 8B and 8C show the results using the overcomplete DCT dictionary and the complete DCT dictionary, respectively.
  • Although the invention has been described in detail, nevertheless changes and modifications, which do not depart from the teachings of the present invention, will be evident to those skilled in the art. Such changes and modifications are deemed to come within the purview of the present invention and the appended claims.

Claims (17)

1. A computerized signal processing system adapted for sparse representation of signals, said system comprising:
(i) one or more training signals;
(ii) a dictionary containing signal-atoms;
(iii) a representation of each training signal using a linear combination of said dictionary's signal-atoms;
(iv) a module for updating the representation of the training signal;
(v) a module for updating the dictionary a single atom or one group of atoms at a time, wherein each atom update includes an update of all representations referring to said updated atom; and
(vi) a module for iterating (iv) and (v) until a stopping rule is fulfilled.
2. The computerized signal processing system according to claim 1, wherein said group of atoms contains one atom only.
3. The computerized signal processing system according to claim 1, wherein all training signals belong to the same family and have one or more common traits.
4. The computerized signal processing system according to claim 1, wherein said dictionary is an overcomplete dictionary.
5. The computerized signal processing system according to claim 1, wherein the representation of all training signals is a coefficient matrix.
6. The computerized signal processing system according to claim 1, wherein the generation of each coefficients matrix is achieved by a pursuit algorithm.
7. The computerized signal processing system according to claim 6, wherein said pursuit algorithm includes Orthogonal Matching Pursuit, Matching Pursuit, Basis Pursuit, FOCUSS or any combination or variation thereof.
8. The computerized signal processing system according to claim 1, wherein said dictionary is updated in a predefined order of the signal-atoms.
9. The computerized signal processing system according to claim 1, wherein only selected signal-atoms of said dictionary are updated.
10. The computerized signal processing system according to claim 1, wherein a signal-atom is modified when the difference between said signal-atom to another signal atom is below a predefined value.
11. The computerized signal processing system according to claim 1, wherein a signal-atom is modified when it is not used in any representation.
12. The computerized signal processing system according to claim 1, wherein a signal-atom is modified when its usage frequency in the representation of signal-atoms is below a predefined value.
13. The computerized signal processing system according to claim 1, wherein updating said group of atoms and their coefficients best reduces the Mean Square Error.
14. The computerized signal processing system according to claim 1, wherein said dictionary is designed with custom properties.
15. The signal processing system according to claim 14, wherein said custom properties include:
(i) shift invariant;
(ii) non-negative dictionary values so that each atom contains only non-negative entries;
(iii) zeros in predetermined places; and
(iv) any structure of a matrix.
16. The computerized signal processing system according to claim 1, comprising multiscale dictionaries.
17. The computerized signal processing system according to claim 1, for compression, regularization in inverse problems, feature extraction, denoising, separation of texture and cartoon content in images, signal analysis, signal synthesis, inpainting and restoration.
US13/425,142 2005-04-04 2012-03-20 System and method for designing of dictionaries for sparse representation Abandoned US20120177128A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US13/425,142 US20120177128A1 (en) 2005-04-04 2012-03-20 System and method for designing of dictionaries for sparse representation
US13/958,644 US20140037199A1 (en) 2005-04-04 2013-08-05 System and method for designing of dictionaries for sparse representation

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US66827705P 2005-04-04 2005-04-04
PCT/IL2006/000423 WO2006106508A2 (en) 2005-04-04 2006-04-04 System and method for designing of dictionaries for sparse representation
US91056807A 2007-10-03 2007-10-03
US13/425,142 US20120177128A1 (en) 2005-04-04 2012-03-20 System and method for designing of dictionaries for sparse representation

Related Parent Applications (3)

Application Number Title Priority Date Filing Date
US11/910,568 Continuation US8165215B2 (en) 2005-04-04 2006-04-04 System and method for designing of dictionaries for sparse representation
PCT/IL2006/000423 Continuation WO2006106508A2 (en) 2005-04-04 2006-04-04 System and method for designing of dictionaries for sparse representation
US91056807A Continuation 2005-04-04 2007-10-03

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/958,644 Continuation-In-Part US20140037199A1 (en) 2005-04-04 2013-08-05 System and method for designing of dictionaries for sparse representation

Publications (1)

Publication Number Publication Date
US20120177128A1 true US20120177128A1 (en) 2012-07-12

Family

ID=37073848

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/910,568 Active 2029-07-23 US8165215B2 (en) 2005-04-04 2006-04-04 System and method for designing of dictionaries for sparse representation
US13/425,142 Abandoned US20120177128A1 (en) 2005-04-04 2012-03-20 System and method for designing of dictionaries for sparse representation

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/910,568 Active 2029-07-23 US8165215B2 (en) 2005-04-04 2006-04-04 System and method for designing of dictionaries for sparse representation

Country Status (2)

Country Link
US (2) US8165215B2 (en)
WO (1) WO2006106508A2 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2568929C1 (en) * 2014-04-30 2015-11-20 Самсунг Электроникс Ко., Лтд. Method and system for fast mri-images reconstruction from sub-sampled data
RU2661537C2 (en) * 2016-08-30 2018-07-17 Самсунг Электроникс Ко., Лтд. Method and system of superresolution by combined sparse approximation
US11614508B1 (en) 2021-10-25 2023-03-28 Q Bio, Inc. Sparse representation of measurements
US11614509B2 (en) 2019-09-27 2023-03-28 Q Bio, Inc. Maxwell parallel imaging
US11748642B2 (en) 2019-02-15 2023-09-05 Q Bio, Inc. Model parameter determination using a predictive model

Families Citing this family (69)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8407028B2 (en) * 2003-12-31 2013-03-26 Jeffrey M. Sieracki Indirect monitoring of device usage and activities
WO2006106508A2 (en) * 2005-04-04 2006-10-12 Technion Research & Development Foundation Ltd. System and method for designing of dictionaries for sparse representation
US20080162475A1 (en) * 2007-01-03 2008-07-03 Meggs Anthony F Click-fraud detection method
US20090083012A1 (en) * 2007-09-20 2009-03-26 Harris Corporation Geospatial modeling system providing wavelet decomposition and inpainting features and related methods
US8290251B2 (en) * 2008-08-21 2012-10-16 Adobe Systems Incorporated Image stylization using sparse representation
FR2935822B1 (en) * 2008-09-05 2010-09-17 Commissariat Energie Atomique METHOD FOR RECOGNIZING FORMS AND SYSTEM IMPLEMENTING THE METHOD
US20110052023A1 (en) * 2009-08-28 2011-03-03 International Business Machines Corporation Reconstruction of Images Using Sparse Representation
CN101882314B (en) * 2010-07-20 2012-06-20 上海交通大学 Infrared small target detection method based on overcomplete sparse representation
CN101958000B (en) * 2010-09-24 2012-08-15 西安电子科技大学 Face image-picture generating method based on sparse representation
CN101957993A (en) * 2010-10-11 2011-01-26 上海交通大学 Adaptive infrared small object detection method
WO2012051094A2 (en) * 2010-10-14 2012-04-19 Technicolor Usa, Inc Methods and apparatus for video encoding and decoding using motion matrix
CN102024152B (en) * 2010-12-14 2013-01-30 浙江大学 Method for recognizing traffic sings based on sparse expression and dictionary study
CN102073999B (en) * 2011-01-20 2012-08-29 西安电子科技大学 Natural image noise removal method based on dual redundant dictionary learning
ES2559040T3 (en) 2011-03-10 2016-02-10 Telefonaktiebolaget Lm Ericsson (Publ) Filling of subcodes not encoded in audio signals encoded by transform
CN102156975B (en) * 2011-04-22 2013-01-23 西安电子科技大学 Natural image de-noising method based on support value transform and multi-scale redundant dictionary learning
US9558762B1 (en) * 2011-07-03 2017-01-31 Reality Analytics, Inc. System and method for distinguishing source from unconstrained acoustic signals emitted thereby in context agnostic manner
US9886945B1 (en) * 2011-07-03 2018-02-06 Reality Analytics, Inc. System and method for taxonomically distinguishing sample data captured from biota sources
US9691395B1 (en) 2011-12-31 2017-06-27 Reality Analytics, Inc. System and method for taxonomically distinguishing unconstrained signal data segments
CN102496153B (en) * 2011-11-04 2013-08-14 西安电子科技大学 SAR image speckle suppression method based on dictionary learning in wavelet domain
CN102510438B (en) * 2011-11-21 2014-05-14 四川虹微技术有限公司 Acquisition method of sparse coefficient vector for recovering and enhancing video image
US8494305B2 (en) * 2011-12-20 2013-07-23 Mitsubishi Electric Research Laboratories, Inc. Image filtering by sparse reconstruction on affinity net
CN102592267B (en) * 2012-01-06 2014-09-03 复旦大学 Medical ultrasonic image filtering method based on sparse representation
US9167274B1 (en) 2012-03-21 2015-10-20 Google Inc. Generating synchronized dictionaries for sparse coding
US9122932B2 (en) * 2012-04-30 2015-09-01 Xerox Corporation Method and system for automatically detecting multi-object anomalies utilizing joint sparse reconstruction model
US9288484B1 (en) * 2012-08-30 2016-03-15 Google Inc. Sparse coding dictionary priming
US9092692B2 (en) 2012-09-13 2015-07-28 Los Alamos National Security, Llc Object detection approach using generative sparse, hierarchical networks with top-down and lateral connections for combining texture/color detection and shape/contour detection
US9152888B2 (en) 2012-09-13 2015-10-06 Los Alamos National Security, Llc System and method for automated object detection in an image
US9152881B2 (en) * 2012-09-13 2015-10-06 Los Alamos National Security, Llc Image fusion using sparse overcomplete feature dictionaries
JP2014063359A (en) * 2012-09-21 2014-04-10 Sony Corp Signal processing apparatus, signal processing method, output apparatus, output method, and program
US9137528B1 (en) * 2012-09-28 2015-09-15 Google Inc. Synchronizing sparse coding dictionaries for use in communication sessions
US8861872B2 (en) * 2012-11-05 2014-10-14 Raytheon Bbn Technologies Corp. Image analysis using coefficient distributions with selective basis feature representation
US9092890B2 (en) 2012-12-20 2015-07-28 Ricoh Company, Ltd. Occlusion-aware reconstruction of three-dimensional scenes from light field images
CN103093433B (en) * 2013-01-25 2015-04-08 西安电子科技大学 Natural image denoising method based on regionalism and dictionary learning
US9098749B2 (en) * 2013-03-14 2015-08-04 Xerox Corporation Dictionary design for computationally efficient video anomaly detection via sparse reconstruction techniques
US9300906B2 (en) 2013-03-29 2016-03-29 Google Inc. Pull frame interpolation
EP2989606B1 (en) 2013-04-26 2020-06-03 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Upsampling and signal enhancement
CN103268494B (en) * 2013-05-15 2016-06-15 江苏大学 Parasite egg recognition methods based on rarefaction representation
US20160021425A1 (en) * 2013-06-26 2016-01-21 Thomson Licensing System and method for predicting audience responses to content from electro-dermal activity signals
CN103440502A (en) * 2013-09-06 2013-12-11 重庆大学 Infrared small-target detection method based on mixing Gauss and sparse representation
US9317780B2 (en) * 2013-10-17 2016-04-19 Xerox Corporation Detecting multi-object anomalies utilizing a low rank sparsity model
CN103561276B (en) * 2013-11-07 2017-01-04 北京大学 A kind of image/video decoding method
TWI513291B (en) * 2013-12-12 2015-12-11 Ind Tech Res Inst Method and apparatus for image processing and computer readable medium
CN103854262B (en) * 2014-03-20 2016-06-29 西安电子科技大学 Medical image denoising method based on documents structured Cluster with sparse dictionary study
CN103886557A (en) * 2014-03-28 2014-06-25 北京工业大学 Denoising method of depth image
EP3055831B1 (en) 2014-06-10 2019-11-20 Ramot at Tel-Aviv University Ltd. Method and system for processing an image
CN104050644B (en) * 2014-06-23 2017-01-04 西北工业大学 A kind of SAR image denoising method based on non-local constraint rarefaction representation
US9286653B2 (en) 2014-08-06 2016-03-15 Google Inc. System and method for increasing the bit depth of images
CN104867119B (en) * 2015-05-21 2018-08-24 天津大学 The structural missing image fill method rebuild based on low-rank matrix
US10755395B2 (en) * 2015-11-27 2020-08-25 Canon Medical Systems Corporation Dynamic image denoising using a sparse representation
CN105738109B (en) * 2016-02-22 2017-11-21 重庆大学 Bearing fault classification diagnosis method based on rarefaction representation and integrated study
US10223358B2 (en) * 2016-03-07 2019-03-05 Gracenote, Inc. Selecting balanced clusters of descriptive vectors
US10136116B2 (en) 2016-03-07 2018-11-20 Ricoh Company, Ltd. Object segmentation from light field data
US10713761B2 (en) 2016-03-15 2020-07-14 The Trustees Of Columbia University In The City Of New York Method and apparatus to perform local de-noising of a scanning imager image
US10867142B2 (en) 2016-06-29 2020-12-15 Intel Corporation Multiplication-free approximation for neural networks and sparse coding
CN106685546A (en) * 2016-12-29 2017-05-17 深圳天珑无线科技有限公司 Wireless human body sensing method and server
US11915722B2 (en) 2017-03-30 2024-02-27 Gracenote, Inc. Generating a video presentation to accompany audio
CN107507139B (en) * 2017-07-28 2019-11-22 北京航空航天大学 The dual sparse image repair method of sample based on Facet directional derivative feature
CN107895063A (en) * 2017-10-16 2018-04-10 杭州电子科技大学 One kind compression EO-1 hyperion mask optimization method
CN108388692B (en) * 2018-01-17 2020-12-29 西安交通大学 Rolling bearing fault feature extraction method based on layered sparse coding
CN108828670B (en) * 2018-08-20 2019-11-29 成都理工大学 A kind of seismic data noise-reduction method
CN109470905B (en) * 2018-09-05 2022-03-04 中国电力科学研究院有限公司 Method and system for extracting corona current signal of extra-high voltage direct current transmission line
CN109671019B (en) * 2018-12-14 2022-11-01 武汉大学 Remote sensing image sub-pixel mapping method based on multi-objective optimization algorithm and sparse expression
CN111383652B (en) * 2019-10-25 2023-09-12 南京邮电大学 Single-channel voice enhancement method based on double-layer dictionary learning
CN111665050B (en) * 2020-06-04 2021-07-27 燕山大学 Rolling bearing fault diagnosis method based on clustering K-SVD algorithm
EP3940696A1 (en) * 2020-07-15 2022-01-19 Österreichische Akademie der Wissenschaften Inpainting based signal gap reconstruction with sparse representation
CN112304419A (en) * 2020-10-25 2021-02-02 广东石油化工学院 Vibration and sound detection signal reconstruction method and system by using generalized sparse coding
CN112713907B (en) * 2020-12-23 2022-03-15 中南大学 Marine CSEM noise reduction method and system based on dictionary learning
CN113836732B (en) * 2021-09-29 2024-04-02 华东理工大学 Vibration signal sparse reconstruction method guided by impact characteristic indexes and application
CN113922823B (en) * 2021-10-29 2023-04-21 电子科技大学 Social media information propagation graph data compression method based on constraint sparse representation

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6151414A (en) * 1998-01-30 2000-11-21 Lucent Technologies Inc. Method for signal encoding and feature extraction
US20030058943A1 (en) * 2001-07-18 2003-03-27 Tru Video Corporation Dictionary generation method for video and image compression
US7079986B2 (en) * 2003-12-31 2006-07-18 Sieracki Jeffrey M Greedy adaptive signature discrimination system and method
US7164494B1 (en) * 2000-02-07 2007-01-16 Adobe Systems Incorporated Color separation of pattern color spaces and form XObjects
US20090103602A1 (en) * 2003-03-28 2009-04-23 Digital Accelerator Corporation Overcomplete basis transform-based motion residual frame coding method and apparatus for video compression
US8165215B2 (en) * 2005-04-04 2012-04-24 Technion Research And Development Foundation Ltd. System and method for designing of dictionaries for sparse representation

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6151414A (en) * 1998-01-30 2000-11-21 Lucent Technologies Inc. Method for signal encoding and feature extraction
US7164494B1 (en) * 2000-02-07 2007-01-16 Adobe Systems Incorporated Color separation of pattern color spaces and form XObjects
US20030058943A1 (en) * 2001-07-18 2003-03-27 Tru Video Corporation Dictionary generation method for video and image compression
US20090103602A1 (en) * 2003-03-28 2009-04-23 Digital Accelerator Corporation Overcomplete basis transform-based motion residual frame coding method and apparatus for video compression
US7079986B2 (en) * 2003-12-31 2006-07-18 Sieracki Jeffrey M Greedy adaptive signature discrimination system and method
US8165215B2 (en) * 2005-04-04 2012-04-24 Technion Research And Development Foundation Ltd. System and method for designing of dictionaries for sparse representation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Gribonval, R., "Fast matching pursuit with a multiscale dictionary of Gaussian chirps", May 2001, IEEE Transsactions on Signal Processing, Volumn 49 Issue 5, pg. 994-1001 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2568929C1 (en) * 2014-04-30 2015-11-20 Самсунг Электроникс Ко., Лтд. Method and system for fast mri-images reconstruction from sub-sampled data
US11080847B2 (en) 2014-04-30 2021-08-03 Samsung Electronics Co., Ltd. Magnetic resonance imaging device and method for generating magnetic resonance image
RU2661537C2 (en) * 2016-08-30 2018-07-17 Самсунг Электроникс Ко., Лтд. Method and system of superresolution by combined sparse approximation
US11748642B2 (en) 2019-02-15 2023-09-05 Q Bio, Inc. Model parameter determination using a predictive model
US11614509B2 (en) 2019-09-27 2023-03-28 Q Bio, Inc. Maxwell parallel imaging
US11614508B1 (en) 2021-10-25 2023-03-28 Q Bio, Inc. Sparse representation of measurements
WO2023075781A1 (en) * 2021-10-25 2023-05-04 Q Bio, Inc. Sparse representation of measurements

Also Published As

Publication number Publication date
US20080170623A1 (en) 2008-07-17
WO2006106508A3 (en) 2007-05-31
WO2006106508A2 (en) 2006-10-12
US8165215B2 (en) 2012-04-24

Similar Documents

Publication Publication Date Title
US8165215B2 (en) System and method for designing of dictionaries for sparse representation
US20140037199A1 (en) System and method for designing of dictionaries for sparse representation
Aharon et al. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation
Wu et al. Deep compressed sensing
Aharon et al. K-SVD and its non-negative variant for dictionary design
Bruckstein et al. From sparse solutions of systems of equations to sparse modeling of signals and images
Yu et al. On compressing deep models by low rank and sparse decomposition
Aharon et al. On the uniqueness of overcomplete dictionaries, and a practical way to retrieve them
Ilin et al. Practical approaches to principal component analysis in the presence of missing values
Sadeghi et al. Learning overcomplete dictionaries based on atom-by-atom updating
Rubinstein et al. Dictionary learning for analysis-synthesis thresholding
Zayyani et al. An iterative Bayesian algorithm for sparse component analysis in presence of noise
Lu et al. Sparse coding from a Bayesian perspective
Bethge Factorial coding of natural images: how effective are linear models in removing higher-order dependencies?
US8942467B2 (en) Method for reducing blocking artifacts in images
Ichir et al. Hidden Markov models for wavelet-based blind source separation
Yang et al. An iterative reweighted method for tucker decomposition of incomplete tensors
Labusch et al. Robust and fast learning of sparse codes with stochastic gradient descent
Tirer et al. Generalizing CoSaMP to signals from a union of low dimensional linear subspaces
Borsos et al. Data summarization via bilevel optimization
Iqbal et al. An approach for sequential dictionary learning in nonuniform noise
Popat Conjoint probabilistic subband modeling
Van Luong et al. Interpretable Deep Recurrent Neural Networks via Unfolding Reweighted $\ell_1 $-$\ell_1 $ Minimization: Architecture Design and Generalization Analysis
Feng et al. Differentiable sparse unmixing based on Bregman divergence for hyperspectral remote sensing imagery
Yang et al. Learning overcomplete dictionaries with application to image Denoising

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION