US7783459B2 - Analog system for computing sparse codes - Google Patents
Analog system for computing sparse codes Download PDFInfo
- Publication number
- US7783459B2 US7783459B2 US12/035,424 US3542408A US7783459B2 US 7783459 B2 US7783459 B2 US 7783459B2 US 3542408 A US3542408 A US 3542408A US 7783459 B2 US7783459 B2 US 7783459B2
- Authority
- US
- United States
- Prior art keywords
- nodes
- node
- signal
- coefficients
- sparsely
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06G—ANALOGUE COMPUTERS
- G06G7/00—Devices in which the computing operation is performed by varying electric or magnetic quantities
- G06G7/12—Arrangements for performing computing operations, e.g. operational amplifiers
- G06G7/26—Arbitrary function generators
Definitions
- the present invention was made with government support under the following government grants or contracts: Office of Naval Research Grant Nos. N00014-06-1-0769, N00014-06-1-0829 and N00014-02-1-0353, U.S. Department of Energy Grant No. DE-FC02-01ER25462, and National Science Foundation Grant Nos. ANI-0099148, ANI-0099148 and IIS-0625717. The government has certain rights in the invention.
- the present invention relates to a system for computing sparse representations of data, i.e., where the data can be fully represented in terms of a small number of non-zero code elements, and for reconstructing compressively sensed images.
- Natural signals can be well-approximated by a small subset of elements from an over complete dictionary.
- the process of choosing a good subset of dictionary elements along with the corresponding coefficients to represent a signal is known as sparse approximation.
- Sparse approximation is a difficult non-convex optimization problem that is at the center of much research in mathematics and signal processing.
- the first general approach substitutes an alternate sparsity measure to convexify the l 0 norm.
- One well-known example is Basis Pursuit (BP) (Chen et al., 2001), which replaces the l 0 norm with the l 1 norm
- BPDN Basis Pursuit De-Noising
- BPDN provides the l 1 -sparsest approximation for a given reconstruction quality.
- the second general approach employed by signal processing researchers uses iterative greedy algorithms to constructively build up a signal representation (Tropp, 2004).
- the canonical example of a greedy algorithm is known in the signal processing community as Matching Pursuit (MP) (Mallat and Zhang, 1993).
- MP Matching Pursuit
- V1 population responses to natural stimuli may be the result of a sparse approximation. For example, it has been shown that V1 receptive fields are consistent with optimizing the coefficient sparsity when encoding natural images. Additionally, V1 recordings in response to natural scene stimuli show activity levels (corresponding to the coefficients ⁇ a m ⁇ ) becoming sparser as neighboring units are also stimulated. These populations are typically very overcomplete, allowing great flexibility in the representation of a stimulus. Using this flexibility to pursue sparse codes might offer many advantages to sensory neural systems, including enhancing the performance of subsequent processing stages, increasing the storage capacity in associative memories, and increasing the energy efficiency of the system.
- this implementation requires persistent (two-way) signaling between all units with overlapping receptive fields (e.g., even a node with a nearly zero value would have to continue sending inhibition signals to all similar nodes).
- spiking neural circuits can be constructed to implement MP.
- this type of circuit implementation relies on a temporal code that requires tightly coupled and precise elements to both encode and decode.
- a time-varying input signal s(t) is represented with a set of time-varying coefficients ⁇ a m (t) ⁇ . While temporal coefficient changes are necessary to encode stimulus changes, the most useful encoding would use coefficient changes that reflect the character of the stimulus. In particular, sparse coefficients should have smooth temporal variations in response to smooth changes in the image.
- most sparse approximation schemes have a single goal: select the smallest number of coefficients to represent a fixed signal. This single-minded approach can produce coefficient sequences for time-varying stimuli that are erratic, with drastic changes not only in the values of the coefficients but also in the selection of which coefficients are used. These erratic temporal codes are inefficient because they introduce uncertainty about which coefficients are coding the most significant stimulus changes, thereby complicating the process of understanding the changing stimulus content.
- the present invention is a parallel dynamical system based on the principles of thresholding and local competition that solves a family of sparse approximation problems corresponding to various sparsity metrics.
- LCAs Locally Competitive Algorithms
- nodes in a population continually compete with neighboring units using (usually one-way) lateral inhibition to calculate coefficients representing an input in an overcomplete dictionary.
- ODEs nonlinear ordinary differential equations
- Each LCA corresponds to an optimal sparse approximation problem that minimizes an energy function combining reconstruction mean-squared error (MSE) and a sparsity-inducing cost function.
- MSE mean-squared error
- the present invention is a neural architecture for locally competitive algorithms (“LCAs”) that correspond to a broad class of sparse approximation problems and possess three properties critical for a neurally plausible sparse coding system.
- LCAs locally competitive algorithms
- the LCA dynamical system is stable to guarantee that a physical implementation is well-behaved.
- the LCAs perform their primary task well, finding codes for fixed images that are have sparsity comparable to the most popular centralized algorithms.
- the LCAs display inertia, coding video sequences with a coefficient time series that is significantly smoother in time than the coefficients produced by other algorithms. This increased coefficient regularity better reflects the smooth nature of natural input signals, making the coefficients much more predictable and making it easier for higher-level structures to identify and understand the changing content in the time-varying stimulus.
- the present invention is an analog system for sparsely approximating a signal.
- the system comprises a matching system for calculating and outputting matching signals representative of how well-matched said signal is to a plurality of dictionary elements and a plurality of nodes.
- Each node receives one of said matching signals from said matching system.
- Each node comprises a source of an internal state signal and a thresholding element.
- the internal state signal in each node is calculated as a function of said matching signal received at said node and weighted outputs of all other nodes.
- the source of an internal state signal may comprise a low pass averaging system.
- the matching system may comprise a projection system for projecting a signal vector onto the plurality of dictionary elements.
- Each node may further comprise a plurality of weighting elements, each weighting element receiving an output from another one of the plurality of nodes and providing the weighted outputs to the source of an internal state signal.
- the internal state signal may be derived from the matching signal less a sum of weighted outputs from the other nodes.
- each node may further comprise a plurality of weighting elements for receiving an output of the thresholding element and providing a plurality of weighted outputs.
- the inputted signal may be a video signal or other type of signal.
- the source of an internal state signal may be a voltage source, current source or other source of electrical energy.
- the low pass averaging system may comprise a low pass averaging circuit such as a resistor and capacitor or any other type of low pass averaging circuit.
- the present invention is a parallel dynamical system for computing sparse representations of data.
- the system comprises a projection system for projecting the data onto projection vectors and a plurality of nodes.
- Each node receives one of the projection vectors from the projection system.
- Each node comprises a source of electrical energy, a low pass averaging circuit and a thresholding element.
- the source of electrical energy in each node comprises a projection vector received at the node less weighted outputs of all other nodes.
- Each node further comprises a plurality of weighting elements, each weighting element receiving an output from another one of the plurality of nodes and providing the weighted output to the source of electrical energy. Other arrangements of the weighting elements may be used with the present invention and with this embodiment.
- the present invention is a parallel dynamical system for computing sparse representations of data.
- the system comprises a plurality of nodes, each node being active or inactive.
- Each node comprises a leaky integrator element, wherein inputs to the leaky integrator element cause an activation potential to charge up, and a thresholding element for receiving the activation potential and for producing an output coefficient.
- the output coefficient is the result of an activation function applied to the activation potential and parameterized by a system threshold.
- Active nodes inhibit other nodes with inhibition signals proportional to both level of activity of the active nodes and a similarity of receptive fields of the active nodes.
- FIG. 1( a ) illustrates LCA nodes in accordance with a preferred embodiment of the present invention behaving as leaky integrators, charging with a speed that depends on how well the input matches the associated dictionary element and the inhibition received from other nodes.
- FIG. 1( b ) is a diagram of a system in accordance with a preferred embodiment of the present invention showing the inhibition signals being sent between nodes.
- node 2 is shown as being active (i.e., having a coefficient above threshold) and inhibiting its neighbors. Since the neighbors are inactive then the inhibition is one-way.
- FIG. 1 ( c ) illustrates a source of electrical energy in a node in accordance with a preferred embodiment of the present invention.
- FIGS. 2( a )-( f ) illustrate the relationship between the threshold function T ( ⁇ , ⁇ , ⁇ ) (•) and the sparsity cost function C(•). Only the positive half of the symmetric threshold and cost functions are plotted.
- FIG. 2( a ) illustrates a sigmoidal threshold function.
- FIG. 2( d ) illustrates the corresponding cost function.
- the dashed line shows the limit, but coefficients produced by the ideal thresholding function cannot take values in this range.
- FIG. 2( f ) illustrates the corresponding cost function.
- FIGS. 3( a ) and ( b ) respectively illustrate the top 200 coefficients from a BPDN solver sorted by magnitude and the same coefficients, sorted according to the magnitude ordering of the SLCA coefficients. While there is a gross decreasing trend noticeable, the largest SLCA coefficients are not in the same locations as the largest BPDN coefficients. While the solutions have equivalent energy functions, the two sets of coefficients differ significantly.
- FIG. 4( a ) illustrates an example of a dictionary having one “extra” element that comprises decaying combinations of all other dictionary elements.
- FIG. 4( b ) illustrates an input vector having a sparse representation in just a few dictionary elements.
- FIG. 4( c ) illustrates an MP initially choosing an “extra” dictionary element, preventing it from finding the optimally sparse representation (coefficients shown after 100 iterations).
- FIG. 4( d ) illustrate that, in contrast, the HLCA system finds the optimally sparse coefficients.
- FIG. 4( e ) illustrates how the time-dynamics of the HLCA system illustrate its advantage.
- the “extra” dictionary element is the first node to activate, followed shortly by the nodes corresponding to the optimal coefficients.
- the collective inhibition of the optimal nodes causes the “extra” node to die away.
- FIG. 5 illustrates SLCA and BPDN coefficients for a series of standard test images. Each line on the plot indicates the tradeoff between MSE and l 1 coefficient norm as ⁇ is varied. The results for SLCA and BPDN overlap exactly, illustrating that the systems are finding equivalent minima of the energy function.
- FIG. 6( a ) shows the MSE decay and
- FIG. 6( b ) shows the l 0 sparsity for HLCA.
- FIG. 6( c ) illustrates the MSE decay and
- FIG. 6( d ) illustrates the l 0 sparsity for SLCA.
- the error converges within 1-2 time constants and the sparsity often approximately converges within 3-4 time constants. In some cases sparsity is reduced with a longer running time.
- FIG. 7 illustrates the mean tradeoff between MSE and l 0 -sparsity for normalized (32 ⁇ 32) patches from a standard set of test images. For a given MSE range, the mean (a) and standard deviation (b) of the l 0 sparsity are plotted.
- FIGS. 8( a )-( d ) shows the HLCA and SLCA systems simulated on 200 frames of the “foreman” test video sequence. For comparison, MP coefficients and thresholded BPDN coefficients are also shown. Average values for each system are notated in the legend.
- FIG. 8( a ) shows Per-frame MSE for each coding scheme, designed to be approximately equal.
- FIG. 8( b ) shows the number of active coefficients in each frame.
- FIG. 8( c ) shows the number of changing coefficient locations for each frame, including the number of inactive nodes becoming active and the number of active nodes becoming inactive.
- FIG. 8( d ) shows the ratio of changing coefficients to active coefficients.
- a ratio near 2 (such as with MP) means that almost 100% of the coefficient locations are new at each frame.
- a ratio near 0.5 (such as with HLCA) means that approximately 25% of the coefficients are new at each frame.
- FIG. 9( a ) shows the marginal probabilities denoting the fraction of the time coefficients spent in the three states: negative, zero and positive ( ⁇ , 0, and +).
- FIG. 9( b ) shows the transition probabilities denoting the probability of a node in one state transitioning to another state on the next frame.
- +) is the probability that a node with an active positive coefficient will be inactive (i.e., zero) in the next frame.
- FIG. 10( a ) illustrates an example time-series coefficient for the HLCA and MP (top and bottom, respectively) encodings for the test video sequence.
- HLCA clusters non-zero entries together into longer runs while MP switches more often between states.
- FIG. 10( b ) illustrates the empirical conditional entropy of the coefficient states ( ⁇ ,0,+) during the test video sequence.
- FIG. 10( c ) illustrates the conditional entropy is calculated analytically while varying P (+
- +) the tendency of a system to group non-zero states together is the most important factor in determining the entropy.
- FIG. 11 is a diagram illustrating an LCA network for compressive sensing reconstruction and the nonlinear transformation applied to the state variable in each node in accordance with a preferred embodiment of the present invention.
- the LCA network found the best sparse approximation for the data vector x.
- the network input m equaled the data x directly and the interconnection strengths were given by ⁇ i , ⁇ l .
- the LCA structure was modified as indicated so that it could solve the compressive sensing reconstruction problem.
- the present invention is an analog system that compresses data before digitization, thereby saving time and energy that would have been wasted. More specifically, the present invention is a parallel dynamical system for computing sparse representations of data, i.e., where the data can be fully represented in terms of a small number of non-zero code elements. Such a system could be envisioned to perform data compression before digitization, reversing the resource wasting common in digital systems.
- a technique referred to as compressive sensing permits a signal to be captured directly in a compressed form rather than recording raw samples in the classical sense. With compressive sensing, only about 5-10% of the original number of measurements need to be made from the original analog image to retain a reasonable quality image. In compressive sensing, however, reconstruction involves solving an optimal sparse approximation problem which requires enormous calculations and memory.
- the present invention employs a locally competitive algorithm (“LCA”) that stylizes interacting neuron-like nodes to solve the sparse approximation problem.
- LCDA locally competitive algorithm
- the present invention uses thresholding functions to induce local (possibly one-way) inhibitory competitions among units, thus constituting a locally competitive algorithm (LCA).
- the LCA can be implemented as a circuit and can be shown to minimize weighted combination of mean-squared-error in describing the data and a cost function on neural activity. It demonstrates sparsity levels comparable to existing sparse coding algorithms, but in contrast to greedy algorithms that iteratively select the single best element, the circuit allows continual interaction among many units, allowing the system to reach more optimal solutions.
- the LCA coefficients for video sequences demonstrate inertial properties that are both qualitatively and quantitatively more regular, i.e., smoother and more predictable, than the coefficients produced by greedy algorithms.
- the LCAs associate each node with an element of a dictionary ⁇ m ⁇ D.
- the collection of nodes evolve according to fixed dynamics (described below) and settle on a collective output ⁇ a m (t) ⁇ , corresponding to the short-term average firing rate of the nodes.
- positive and negative coefficients are allowed, but rectified systems could use two physical units to implement one LCA node.
- Each node's sub-threshold value is represented by a time-varying internal state u m (t).
- the nodes best matching the stimulus will have internal state variables that charge at the fastest rates and become active soonest.
- active nodes inhibit other nodes with an inhibition signal proportional to both their activity level and the similarity of the nodes' receptive fields.
- LCA node dynamics are expressed by the non-linear ordinary differential equation (ODE)
- u . m ⁇ ( t ) 1 ⁇ [ b m ⁇ ( t ) - u m ⁇ ( t ) - ⁇ n ⁇ m ⁇ G m , n ⁇ a n ⁇ ( t ) ] . ( 4 )
- FIG. 1( a ) shows an LCA node circuit schematic and FIG. 1( b ) is a system diagram illustrating the lateral inhibition.
- the node 100 has a source of electrical energy 110 , a low pass averaging circuit 120 comprised of a resistor and a capacitor, and a thresholding element 130 .
- the source of electrical energy 110 is shown in FIG. 1( a ) as a voltage source, other arrangements such as a current source may be used in the present invention and such alternatives will readily apparent to those of ordinary skill in the art.
- the low pass averaging circuit 120 is shown as a simple resistor and capacitor arrangement in FIG. 1( a ), other arrangements may be used with the present invention and will be readily apparently to those of ordinary skill in the art.
- FIG. 1( c ) An embodiment of the source of electrical energy 110 is shown in greater detail in FIG. 1( c ).
- the source 110 is not a “source” in the sense that it generates electrical energy, but rather, it uses received signals to produce or “compute” the output provided to the low pass averaging circuit 120 and the thresholding element 130 . More specifically, the source 110 receives a projection vector ⁇ n ,s(t) from a projection system 200 (shown in FIG. 1( b )) and an output a n (t) from each other node 100 .
- the source 110 in each node 100 has a weighting element 112 corresponding to the output received from each other node for weighting that output.
- the source 110 outputs the difference
- V m ⁇ ( t ) ⁇ ⁇ m , s ⁇ ( t ) ⁇ - ⁇ n ⁇ m ⁇ T ⁇ ⁇ ( u n ⁇ ( t ) ) ⁇ ⁇ ⁇ m , ⁇ n ⁇ .
- a preferred embodiment of the analog system for compressing signals of the present invention has a projection system 200 for projecting a received signal s(t) onto a plurality of projection vectors ⁇ n , s(t) that are then provided to a plurality of nodes 100 .
- each node 100 receives a projection vector ⁇ n , s(t) and also the output a n (t) of each other node.
- FIGS. 1( a )-( c ) show the output of each node being pass directly back to each other node and the weighting of such outputs being performed inside the receiving node
- each node could calculate the weighting of its own output a n (t) and could then pass its own weighted output a n (t) ⁇ n , ⁇ m to the other nodes.
- LCA architecture described by equation (5) solves a family of sparse approximation problems corresponding to different sparsity measures. Specifically, LCAs descend an energy function combining the reconstruction MSE and a sparsity-inducing cost penalty C(•),
- the specific form of the cost function C(•) is determined by the form of the thresholding activation function T ⁇ (•). For a given threshold function, the cost function is specified (up to a constant) by
- Thresholding functions limit the lateral inhibition by allowing only “strong” units to suppress other units and forcing most coefficients to be identically zero.
- thresholding functions T ⁇ (•) have two distinct behaviors over their range: they are essentially linear with unit slope above threshold ⁇ , and essentially zero below threshold.
- T ( ⁇ , ⁇ ) ⁇ ( u m ) u m - ⁇ 1 + e - ⁇ ⁇ ( u m - ⁇ ) , ( 8 )
- ⁇ is a parameter controlling the speed of the threshold transition
- ⁇ [0,1] indicates what fraction of an additive adjustment is made for values above threshold.
- An example sigmoidal thresholding function is shown in FIG. 2 a .
- a LCA can optimize a variety of different sparsity measures depending on the choice of thresholding function.
- the SLCA is simply another solution method for the general BPDN problem described above.
- SLCA and BPDN solvers will find different sets of coefficients, as illustrated in FIG. 3 .
- the connection between soft-thresholding and BPDN is well-known in the case of orthonormal dictionaries (Chen et al., 2001), and recent results have given some justification for using soft-thresholding in over complete dictionaries.
- the SLCA provides another formal connection between the soft-thresholding function and the l 1 cost function.
- BPDN uses the l 1 -norm as its sparsity penalty, we often expect many of the resulting coefficients to be identically zero (especially when M>N). However, most numerical methods (including direct gradient descent and interior point solvers) will drive coefficients toward zero but will never make them identically zero. While an ad hoc threshold could be applied to the results of a BPDN solver, the SLCA has the advantage of incorporating a natural thresholding function that keeps coefficients identically zero during the computation unless they become active. In other words, while BPDN solvers often start with many non-zero coefficients and try to force coefficients down, the SLCA starts with all coefficients equal to zero and only lets a few grow up. This advantage is especially important for systems that must expend energy for non-zero values throughout the entire computation.
- the HLCA also has connections to known sparse approximation principles. If node m is fully charged, the inhibition signal it sends to other nodes would be exactly the same as the update step when the m th node is chosen in the MP algorithm. However, due to the continuous competition between nodes before they are fully charged, the HLCA is not equivalent to MP in general.
- e m is the canonical basis element (i.e., it contains a single 1 in the m th location) and k is a constant to make the vectors have unit norm.
- the input signal is sparsely represented in the first K dictionary elements,
- the first MP iteration chooses ⁇ M , introducing a residual with decaying terms. Even though s has an exact representation in K elements, MP iterates forever trying to atone for this bad initial choice.
- the HLCA initially activates the M th node but uses the collective inhibition from nodes 1, . . . , K to suppress this node and calculate the optimal set of coefficients. While this pathological example is unlikely to exactly occur in natural signals, it is often used as a criticism of greedy methods to demonstrate their shortsightedness. We mention it here to demonstrate the flexibility of LCAs and their differences from pure greedy algorithms.
- LCAs must exhibit several critical properties: the dynamic systems must remain stable under normal operating conditions, the system must produce sparse coefficients that represent the stimulus with low error, and coefficient sequences must exhibit regularity in response to time-varying inputs. We show that LCAs exhibit good characteristics in each of these three areas. We focus our analysis on the HLCA both because it yields the most interesting results and because it is notationally the cleanest to discuss. In general, the analysis principles we use will apply to all LCAs through straightforward (through perhaps laborious) extensions.
- M u(t) ⁇ m:
- the stability criteria are likely to be met under normal operating conditions for two reasons. First, small subsets of dictionary elements are unlikely to be linearly dependent unless the dictionary is designed with this property. Second, sparse coding systems are actively trying to select dictionary subsets so that they can use many fewer coefficients then the dimension of the signal space,
- the most powerful algorithm produces the lowest reconstruction MSE for a given sparsity.
- the sparsity measure is the l 1 norm
- the problem is convex and the SLCA produces solutions with equivalent l 1 -sparsity to interior point BPDN solvers (demonstrated in FIG. 5 ).
- the HLCA is appealing because of its l 0 -like sparsity penalty, but this objective function is not convex and the HLCA may find a local minimum.
- ⁇ ⁇ m , s ⁇ ( t ) - s ⁇ ⁇ ( t ) ⁇ ⁇ u m ⁇ ( t ) if ⁇ ⁇ ⁇ u m ⁇ ⁇ ⁇ 0 if ⁇ ⁇ ⁇ u m ⁇ > ⁇ .
- ⁇ min is the minimum eigenvalue of ⁇ t .
- the HLCA may use the m th node or a collection of other nodes to represent s, but it cannot use a combination of both. This result extends intuitively beyond one-sparse signals: each component in an optimal decomposition is represented by either the optimal node or another collection of nodes, but not both. While not necessarily finding the optimal representation, the system does not needlessly employ both the optimal and extraneous nodes.
- the LCAs achieve a combination of error and sparsity comparable with known methods.
- a dictionary consisting of the bandpass band of a steerable pyramid with one level and four orientation bands (i.e., the dictionary is approximately four times overcomplete).
- Image patches 32 ⁇ 32 were selected at random from a standard set of test images. The selected image patches were decomposed using the steerable pyramid and reconstructed using just the bandpass band. The bandpass image patches were also normalized to have unit energy.
- We simulated the system using a simple Euler's method approach i.e., first order finite difference approximation
- FIG. 6 shows the time evolution of the reconstruction MSE and l 0 sparsity for SLCA and HLCA responding to an individual image
- FIG. 7 shows the mean steady-state tradeoff between l 0 sparsity and MSE.
- BPDNthr a standard BPDN interior point solver followed by thresholding to enforce l 0 sparsity
- SLCAthr SLCA with the same threshold applied
- u . m ⁇ ( t ) 1 ⁇ ⁇ ⁇ ⁇ ⁇ m , ( s ⁇ ( t ) - s ⁇ ⁇ ( t ) ) ⁇ - u m ⁇ ( t ) when ⁇ ⁇ ⁇ u m ⁇ ( t ) ⁇ ⁇ ⁇ ⁇ ⁇ m , ( s ⁇ ( t ) - s ⁇ ⁇ ( t ) ) ⁇ - ⁇ when ⁇ ⁇ ⁇ u m ⁇ ( t ) ⁇ ⁇ ⁇ .
- a new residual signal drives the coefficient higher but suffers an additive penalty.
- This property induces a “king of the hill” effect: when a new residual appears, active nodes move virtually unimpeded to represent it while inactive nodes are penalized until they reach threshold. This inertia encourages inactive nodes to remain inactive unless the active nodes cannot adequately represent the new input.
- the HLCA uses approximately the same number of active coefficients as MP but chooses coefficients that more efficiently represent the video sequence.
- the HLCA is significantly more likely to re-use active coefficient locations from the previous frame without making significant sacrifices in the sparsity of the solution. This difference is highlighted when looking at the ratio of the number of changing coefficients to the number of active coefficients,
- MP has a ratio of 1.7, meaning that MP is finding almost an entirely new set of active coefficient locations for each frame.
- the HLCA has a ratio of 0.5, meaning that it is changing approximately 25% of its coefficient locations at each frame.
- SLCA and BPDNthr have approximately the same performance, with regularity falling between HLCA and MP. Though the two systems can calculate different coefficients, the convexity of the energy function appears to be limiting the coefficient choices enough so that SLCA cannot smooth the coefficient time series substantially more than BPDNthr.
- the simulation results indicate that the HLCA is producing time series coefficients that are much more regular than MP.
- This regularity is visualized in FIG. 10 by looking at the time-series of example HLCA and MP coefficients. Note that though the two coding schemes produce roughly the same number of non-zero entries, the HLCA does much better than MP at clustering the values into consecutive runs of positive or negative values. This type of smoothness better reflects the regularity in the natural video sequence input. We can quantify this increased regularity by examining the Markov state transitions. Specifically, each coefficient time-series is Markov chain with three possible states at frame n:
- ⁇ m ⁇ ( n ) ⁇ - if ⁇ ⁇ u m ⁇ ( n ) ⁇ - ⁇ 0 if ⁇ - ⁇ ⁇ u m ⁇ ( n ) ⁇ ⁇ + if ⁇ ⁇ u m ⁇ ( n ) > ⁇
- FIG. 9 shows the marginal probabilities P(•) of the states and the conditional probabilities P(•
- the HLCA and MP are equally likely to have non-zero states, but the HLCA is over five times more likely than MP to have a positive coefficient stay positive (P (+
- conditional entropy indicates how much uncertainty there is about the status of the current coefficients given the coefficients from the previous frame.
- the conditional entropy for MP is almost double the entropy for the HLCA, while SLCA is again similar to BPDNthr.
- the principle contributing factor to the conditional entropy appears to be the probability a non-zero node remains in the same state (i.e., P(+
- FIG. 10 shows the change in conditional entropy is almost linear with varying P(+
- ⁇ ) P(+
- the substantial decrease in the conditional entropy for the HLCA compared to MP quantifies the increased regularity in time-series coefficients due to the inertial properties of the LCAs.
- the HLCA in particular encourages coefficients to maintain their present state (i.e., active or inactive) if it is possible
- Sparse approximation is an important paradigm in modern sensing and signal processing, though mechanisms to calculate these codes using parallel analog computational elements instead of digital computers have remained unknown.
- a locally competitive algorithm that solves a series of sparse approximation problems (including BPDN as a special case).
- These LCAs can be implemented using a parallel network of simple elements that match well with parallel analog computational architectures, including analog circuits and sensory cortical areas such as V1. Though these LCA systems are non-linear, we have shown that they are well-behaved under nominal operating conditions.
- LCA systems are not generally guaranteed to find a globally optimal solution to their energy function, we have proven that the systems will be efficient in a meaningful sense.
- the SLCA system produces coefficients with sparsity levels comparable to BPDN solvers, but uses a natural physical implementation that is more energy efficient (i.e., it uses fewer non-zero inhibition signals between nodes).
- the HLCA produces coefficients with almost identical sparsity as MP. This is significant because greedy methods such as MP are widely used in signal processing practice because of their efficiency, but HLCA offers a much more natural parallel implementation.
- LCAs are particularly appropriate for time-varying data such as video sequences.
- the LCA ODE not only encourages sparsity but also introduces an inertia into the coefficient time-series that we have quantified using both raw counts of changing coefficient location and through the conditional entropy of the coefficient states.
- the LCAs produce smoother coefficient sequences that better reflect the structure of the time-varying stimulus. This property could prove valuable for higher levels of analysis that are trying to interpret the sensory scene from a set of sparse coefficients.
- LCAs By using simple computational primitives, LCAs also have the benefit of being implementable in analog hardware.
- An imaging system using VLSI to implement LCAs as a data collection front end has the potential to be extremely fast and energy efficient. Instead of digitizing all of the sensed data and using digital hardware to run a compression algorithm, analog processing would compress the data into sparse coefficients before digitization. In this system, time and energy resources would only be spent digitizing coefficients that are a critical component in the signal representation.
- the LCA network Since the LCA network represents an analog way of finding sparse representations, it can be modified it for the compressive sensing reconstruction problem.
- the compressed input signal is received by a projection system 300 which passes projection vectors to a plurality of nodes 400 .
- the compressive sensing reconstruction problem amounts to a constrained optimization problem that can be recast with Lagrange multipliers into an unconstrained optimization problem.
- the network's inputs must now equal ⁇ t y and the inner products ⁇ i , ⁇ l determine the connection strengths between nodes i and l.
- the diagonal values of ⁇ t ⁇ are not all equal to 1, which is required to find the optimal solution without error. As a result, the following equation accommodates this problem.
- D is a diagonal matrix whose entries equal those of the diagonal of ⁇ t ⁇ .
- Test ⁇ matrices were derived from the product of a randomly generated ⁇ matrix (Gaussian independent and identically distributed random variables with zero mean) and a ⁇ basis matrix (2D Haar or Daubechies 4 wavelets).
Landscapes
- Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Computer Hardware Design (AREA)
- General Physics & Mathematics (AREA)
- Image Analysis (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
Abstract
Description
In optimal sparse approximation, we seek the coefficients having the fewest number of non-zero entries by solving the minimization problem
where the l0 “norm” denotes the number of non-zero elements of a=[a1, a2, . . . , aM]. While this clearly is not a norm in the mathematical sense, the term here will be used as it is prevalent in the literature. Unfortunately, this combinatorial optimization problem is NP-hard.
In practice, the presence of signal noise often leads to using a modified approach called Basis Pursuit De-Noising (BPDN) (Chen et al., 2001) that makes a tradeoff between reconstruction mean-squared error (MSE) and sparsity in an unconstrained optimization problem:
where λ is a tradeoff parameter. BPDN provides the l1-sparsest approximation for a given reconstruction quality. There are many algorithms that can be used to solve the BPDN optimization problem, with interior point-type methods being the most common choice.
Though they may not be optimal in general, greedy algorithms often efficiently find good sparse signal representations in practice.
This correspondence between the thresholding function and the cost function can be seen by computing the derivative of E with respect to the active coefficients, {am}. If equation (7) holds, then letting the internal states {um} evolve according to
yields the equation for the internal state dynamics above in equation (4). Note that although the dynamics are specified through a gradient approach, the system is not performing direct gradient descent
As long as am and um are related by a monotonically increasing function, the {am} will also descend the energy function E.
where γ is a parameter controlling the speed of the threshold transition and αε[0,1] indicates what fraction of an additive adjustment is made for values above threshold. An example sigmoidal thresholding function is shown in
where I(•) is the indicator function evaluating to 1 if the argument is true and 0 if the argument is false. As with the SLCA, the HLCA also has connections to known sparse approximation principles. If node m is fully charged, the inhibition signal it sends to other nodes would be exactly the same as the update step when the mth node is chosen in the MP algorithm. However, due to the continuous competition between nodes before they are fully charged, the HLCA is not equivalent to MP in general.
where em is the canonical basis element (i.e., it contains a single 1 in the mth location) and k is a constant to make the vectors have unit norm. In words, the dictionary includes the canonical basis along with one “extra” element that is a decaying combination of all other elements (illustrated in
The first MP iteration chooses φM, introducing a residual with decaying terms. Even though s has an exact representation in K elements, MP iterates forever trying to atone for this bad initial choice. In contrast, the HLCA initially activates the Mth node but uses the collective inhibition from
-
- has a finite number of equilibrium points;
- has equilibrium points that are almost certainly isolated (no two equilibrium points are arbitrarily close together); and
- is almost certainly locally asymptotically stable for every equilibrium point.
for fixed inputs. While this is encouraging, it does not guarantee input-output stability. To appreciate this effect, note that the HLCA cost function is constant for nodes above threshold—nothing explicitly keeps a node from growing without bound once it is active.
where ηmin is the minimum eigenvalue of ΦΦt.
A new residual signal drives the coefficient higher but suffers an additive penalty. Inactive coefficients suffer an increasing penalty as they get closer to threshold while active coefficients only suffer a constant penalty αλ that can be very small (e.g., the HLCA has αλ=0). This property induces a “king of the hill” effect: when a new residual appears, active nodes move virtually unimpeded to represent it while inactive nodes are penalized until they reach threshold. This inertia encourages inactive nodes to remain inactive unless the active nodes cannot adequately represent the new input.
plotted in
ŝ=s−λD(ΘtΘ)−11 (13)
Thus, λD(ΘtΘ)−1 represents error. Note that this bias could be removed because it is data-independent. However, calculating it would require inverting a large matrix. Our goal is to reduce and control it as much as possible by other means.
ŝ=s−λ(ΘTΘ)−11 (14)
Claims (14)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/035,424 US7783459B2 (en) | 2007-02-21 | 2008-02-21 | Analog system for computing sparse codes |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US90267307P | 2007-02-21 | 2007-02-21 | |
US12/035,424 US7783459B2 (en) | 2007-02-21 | 2008-02-21 | Analog system for computing sparse codes |
Publications (2)
Publication Number | Publication Date |
---|---|
US20080270055A1 US20080270055A1 (en) | 2008-10-30 |
US7783459B2 true US7783459B2 (en) | 2010-08-24 |
Family
ID=39888011
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/035,424 Active 2028-12-11 US7783459B2 (en) | 2007-02-21 | 2008-02-21 | Analog system for computing sparse codes |
Country Status (1)
Country | Link |
---|---|
US (1) | US7783459B2 (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100077016A1 (en) * | 2008-09-24 | 2010-03-25 | Eduardo Perez | Estimating a Signal Based on Samples Derived from Random Projections |
EP2424117A2 (en) | 2010-08-24 | 2012-02-29 | Her Majesty the Queen in Right of Canada, as represented by the Minister of Industry, through The Communications Research Centre Canada | Sparse data compression |
US20120259590A1 (en) * | 2011-04-11 | 2012-10-11 | Jong Chul Ye | Method and apparatus for compressed sensing with joint sparsity |
US20130151924A1 (en) * | 2011-12-08 | 2013-06-13 | Harris Corporation, Corporation Of The State Of Delaware | Data system for interfacing with a remote data storage facility using compressive sensing and associated methods |
US8538329B2 (en) | 2011-12-08 | 2013-09-17 | Harris Corporation | System with sub-nyquist signal acquisition and transmission and associated methods |
US8570406B2 (en) | 2010-08-11 | 2013-10-29 | Inview Technology Corporation | Low-pass filtering of compressive imaging measurements to infer light level variation |
US8605786B2 (en) * | 2007-09-04 | 2013-12-10 | The Regents Of The University Of California | Hierarchical motion vector processing method, software and devices |
US8760542B2 (en) * | 2010-08-11 | 2014-06-24 | Inview Technology Corporation | Compensation of compressive imaging measurements based on measurements from power meter |
US9137528B1 (en) | 2012-09-28 | 2015-09-15 | Google Inc. | Synchronizing sparse coding dictionaries for use in communication sessions |
US9167274B1 (en) | 2012-03-21 | 2015-10-20 | Google Inc. | Generating synchronized dictionaries for sparse coding |
US9633306B2 (en) * | 2015-05-07 | 2017-04-25 | Siemens Healthcare Gmbh | Method and system for approximating deep neural networks for anatomical object detection |
US9823350B2 (en) | 2014-07-31 | 2017-11-21 | Raytheon Company | Linear mode computational sensing LADAR |
CN107667381A (en) * | 2015-06-04 | 2018-02-06 | 西门子保健有限责任公司 | The medical model represented using non-linear and non-negative sparse is classified |
US10049683B2 (en) | 2013-10-21 | 2018-08-14 | Dolby International Ab | Audio encoder and decoder |
CN108924148A (en) * | 2018-07-18 | 2018-11-30 | 中南大学 | A kind of source signal collaborative compression perception data restoration methods |
CN106993240B (en) * | 2017-03-14 | 2020-10-16 | 天津大学 | Multi-video abstraction method based on sparse coding |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8437564B2 (en) * | 2007-08-07 | 2013-05-07 | Ntt Docomo, Inc. | Image and video compression using sparse orthonormal transforms |
US8787186B2 (en) | 2008-12-12 | 2014-07-22 | Blackberry Limited | Mobility in a distributed antenna system |
US8693597B2 (en) * | 2008-12-12 | 2014-04-08 | Blackberry Limited | Sensor-based wireless communication systems using compressive sampling |
US20100290395A1 (en) * | 2009-04-15 | 2010-11-18 | Research In Motion Ltd | Sensor-based wireless communication systems using compressive sampling |
US8644244B2 (en) * | 2008-12-12 | 2014-02-04 | Research In Motion Limited | Sensor-based wireless communication systems using compressive sampling |
CA2751717C (en) * | 2008-12-22 | 2015-10-06 | Schlumberger Canada Limited | Automatic dispersion extraction of multiple time overlapped acoustic signals |
US8768095B2 (en) * | 2009-08-14 | 2014-07-01 | General Electric Company | System and method for processing data signals |
US8484024B2 (en) | 2011-02-24 | 2013-07-09 | Nuance Communications, Inc. | Phonetic features for speech recognition |
GB201114255D0 (en) * | 2011-08-18 | 2011-10-05 | Univ Antwerp | Smart sampling and sparse reconstruction |
CN105338357B (en) * | 2015-09-29 | 2018-08-24 | 湖北工业大学 | A kind of distributed video compressed sensing decoding method |
WO2019074744A1 (en) * | 2017-10-09 | 2019-04-18 | Fovia Inc. | Bit prediction method and system using a statistical model |
CN110766139A (en) * | 2019-10-21 | 2020-02-07 | 哈尔滨工业大学 | Composite neural network model and modeling method thereof |
Citations (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5303058A (en) * | 1990-10-22 | 1994-04-12 | Fujitsu Limited | Data processing apparatus for compressing and reconstructing image data |
US6298166B1 (en) * | 1998-03-30 | 2001-10-02 | Seiko Epson Corporation | Image transformations in the compressed domain |
US20010028683A1 (en) * | 1999-12-28 | 2001-10-11 | Vincent Bottreau | Video encoding method based on the matching pursuit algorithm |
US20030058943A1 (en) * | 2001-07-18 | 2003-03-27 | Tru Video Corporation | Dictionary generation method for video and image compression |
US20040268334A1 (en) * | 2003-06-30 | 2004-12-30 | Kalyan Muthukumar | System and method for software-pipelining of loops with sparse matrix routines |
US20060029279A1 (en) * | 2004-08-09 | 2006-02-09 | Donoho David L | Method and apparatus for compressed sensing |
US20070019723A1 (en) * | 2003-08-12 | 2007-01-25 | Koninklijke Philips Electronics N.V. | Video encoding and decoding methods and corresponding devices |
US20070027656A1 (en) * | 2005-05-10 | 2007-02-01 | Baraniuk Richard G | Method and Apparatus for Distributed Compressed Sensing |
US7345603B1 (en) * | 2006-11-07 | 2008-03-18 | L3 Communications Integrated Systems, L.P. | Method and apparatus for compressed sensing using analog projection |
US20080168324A1 (en) * | 2005-01-23 | 2008-07-10 | Zte Corporation | Basic Matrix Based on Irregular Ldpc, Codec and Generation Method Thereof |
US7450032B1 (en) * | 2005-12-27 | 2008-11-11 | At & T Intellectual Property Ii, L.P. | System and method for encoding a signal using compressed sensor measurements |
US7584396B1 (en) * | 2005-12-27 | 2009-09-01 | At&T Corp. | System and method for decoding a signal using compressed sensor measurements |
-
2008
- 2008-02-21 US US12/035,424 patent/US7783459B2/en active Active
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5303058A (en) * | 1990-10-22 | 1994-04-12 | Fujitsu Limited | Data processing apparatus for compressing and reconstructing image data |
US6298166B1 (en) * | 1998-03-30 | 2001-10-02 | Seiko Epson Corporation | Image transformations in the compressed domain |
US20010028683A1 (en) * | 1999-12-28 | 2001-10-11 | Vincent Bottreau | Video encoding method based on the matching pursuit algorithm |
US20030058943A1 (en) * | 2001-07-18 | 2003-03-27 | Tru Video Corporation | Dictionary generation method for video and image compression |
US7003039B2 (en) * | 2001-07-18 | 2006-02-21 | Avideh Zakhor | Dictionary generation method for video and image compression |
US20040268334A1 (en) * | 2003-06-30 | 2004-12-30 | Kalyan Muthukumar | System and method for software-pipelining of loops with sparse matrix routines |
US20070019723A1 (en) * | 2003-08-12 | 2007-01-25 | Koninklijke Philips Electronics N.V. | Video encoding and decoding methods and corresponding devices |
US20060029279A1 (en) * | 2004-08-09 | 2006-02-09 | Donoho David L | Method and apparatus for compressed sensing |
US7646924B2 (en) * | 2004-08-09 | 2010-01-12 | David Leigh Donoho | Method and apparatus for compressed sensing |
US20080168324A1 (en) * | 2005-01-23 | 2008-07-10 | Zte Corporation | Basic Matrix Based on Irregular Ldpc, Codec and Generation Method Thereof |
US20080129560A1 (en) * | 2005-05-10 | 2008-06-05 | Baraniuk Richard G | Method and Apparatus for Distributed Compressed Sensing |
US7271747B2 (en) * | 2005-05-10 | 2007-09-18 | Rice University | Method and apparatus for distributed compressed sensing |
US20070027656A1 (en) * | 2005-05-10 | 2007-02-01 | Baraniuk Richard G | Method and Apparatus for Distributed Compressed Sensing |
US7450032B1 (en) * | 2005-12-27 | 2008-11-11 | At & T Intellectual Property Ii, L.P. | System and method for encoding a signal using compressed sensor measurements |
US7584396B1 (en) * | 2005-12-27 | 2009-09-01 | At&T Corp. | System and method for decoding a signal using compressed sensor measurements |
US7345603B1 (en) * | 2006-11-07 | 2008-03-18 | L3 Communications Integrated Systems, L.P. | Method and apparatus for compressed sensing using analog projection |
Cited By (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8605786B2 (en) * | 2007-09-04 | 2013-12-10 | The Regents Of The University Of California | Hierarchical motion vector processing method, software and devices |
US8239436B2 (en) | 2008-09-24 | 2012-08-07 | National Instruments Corporation | Estimating a signal based on samples derived from dot products and random projections |
US20100077016A1 (en) * | 2008-09-24 | 2010-03-25 | Eduardo Perez | Estimating a Signal Based on Samples Derived from Random Projections |
US8885073B2 (en) | 2010-08-11 | 2014-11-11 | Inview Technology Corporation | Dedicated power meter to measure background light level in compressive imaging system |
US8760542B2 (en) * | 2010-08-11 | 2014-06-24 | Inview Technology Corporation | Compensation of compressive imaging measurements based on measurements from power meter |
US8570406B2 (en) | 2010-08-11 | 2013-10-29 | Inview Technology Corporation | Low-pass filtering of compressive imaging measurements to infer light level variation |
US8711015B2 (en) | 2010-08-24 | 2014-04-29 | Her Majesty The Queen In Right Of Canada As Represented By The Minister Of Industry, Through The Communications Research Centre Canada | Sparse data compression |
EP2424117A2 (en) | 2010-08-24 | 2012-02-29 | Her Majesty the Queen in Right of Canada, as represented by the Minister of Industry, through The Communications Research Centre Canada | Sparse data compression |
US20120259590A1 (en) * | 2011-04-11 | 2012-10-11 | Jong Chul Ye | Method and apparatus for compressed sensing with joint sparsity |
US8516340B2 (en) * | 2011-12-08 | 2013-08-20 | Harris Corporation | Data system for interfacing with a remote data storage facility using compressive sensing and associated methods |
US8538329B2 (en) | 2011-12-08 | 2013-09-17 | Harris Corporation | System with sub-nyquist signal acquisition and transmission and associated methods |
US20130151924A1 (en) * | 2011-12-08 | 2013-06-13 | Harris Corporation, Corporation Of The State Of Delaware | Data system for interfacing with a remote data storage facility using compressive sensing and associated methods |
US9167274B1 (en) | 2012-03-21 | 2015-10-20 | Google Inc. | Generating synchronized dictionaries for sparse coding |
US9137528B1 (en) | 2012-09-28 | 2015-09-15 | Google Inc. | Synchronizing sparse coding dictionaries for use in communication sessions |
US10049683B2 (en) | 2013-10-21 | 2018-08-14 | Dolby International Ab | Audio encoder and decoder |
US9823350B2 (en) | 2014-07-31 | 2017-11-21 | Raytheon Company | Linear mode computational sensing LADAR |
US9633306B2 (en) * | 2015-05-07 | 2017-04-25 | Siemens Healthcare Gmbh | Method and system for approximating deep neural networks for anatomical object detection |
CN107667381A (en) * | 2015-06-04 | 2018-02-06 | 西门子保健有限责任公司 | The medical model represented using non-linear and non-negative sparse is classified |
CN107667381B (en) * | 2015-06-04 | 2022-02-11 | 西门子保健有限责任公司 | Medical mode classification using non-linear and non-negative sparse representations |
CN106993240B (en) * | 2017-03-14 | 2020-10-16 | 天津大学 | Multi-video abstraction method based on sparse coding |
CN108924148A (en) * | 2018-07-18 | 2018-11-30 | 中南大学 | A kind of source signal collaborative compression perception data restoration methods |
CN108924148B (en) * | 2018-07-18 | 2020-07-24 | 中南大学 | Multi-source signal collaborative compressed sensing data recovery method |
Also Published As
Publication number | Publication date |
---|---|
US20080270055A1 (en) | 2008-10-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7783459B2 (en) | Analog system for computing sparse codes | |
Rozell et al. | Sparse coding via thresholding and local competition in neural circuits | |
Ballé et al. | Nonlinear transform coding | |
Ballé et al. | Variational image compression with a scale hyperprior | |
Iliadis et al. | Deepbinarymask: Learning a binary mask for video compressive sensing | |
Ballé et al. | Integer networks for data compression with latent-variable models | |
Skatchkovsky et al. | End-to-end learning of neuromorphic wireless systems for low-power edge artificial intelligence | |
Murray et al. | Learning sparse overcomplete codes for images | |
WO2023207836A1 (en) | Image encoding method and apparatus, and image decompression method and apparatus | |
Li et al. | Fixed-sign binary neural network: An efficient design of neural network for Internet-of-Things devices | |
CN117501696A (en) | Parallel context modeling using information shared between partitions | |
CN116309135A (en) | Diffusion model processing method and device and picture processing method and device | |
CN115361559A (en) | Image encoding method, image decoding method, image encoding device, image decoding device, and storage medium | |
Golts et al. | Image compression optimized for 3D reconstruction by utilizing deep neural networks | |
Tartaglione et al. | HEMP: high-order entropy minimization for neural network compression | |
EP4449313A1 (en) | Operation of a neural network with clipped input data | |
WO2023177318A1 (en) | Neural network with approximated activation function | |
Mahmoud et al. | Learning algorithm effect on multilayer feed forward artificial neural network performance in image coding | |
Murray et al. | Learning sparse overcomplete codes for images | |
Chang et al. | Informative sensing | |
Rozell et al. | Neurally plausible sparse coding via thresholding and local competition | |
US20240348783A1 (en) | Methods and apparatus for approximating a cumulative distribution function for use in entropy coding or decoding data | |
Jalali | Toward Theoretically Founded Learning-Based Compressed Sensing | |
US20230262237A1 (en) | System and methods for video analysis | |
Xu | Learning in Compressed Domains |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: THE REGENTS OF THE UNIVERSITY OF CALIFORNIA, CALIF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OLSHAUSEN, BRUNO ADOLPHUS;REEL/FRAME:021626/0548 Effective date: 20080401 Owner name: WILLIAM MARSH RICE UNIVERSITY, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ROZELL, CHRISTOPHER JOHN;JOHNSON, DON HERRICK;BARANIUK, RICHARD GORDON;AND OTHERS;REEL/FRAME:021626/0687;SIGNING DATES FROM 20080401 TO 20080404 Owner name: WILLIAM MARSH RICE UNIVERSITY, TEXAS Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ROZELL, CHRISTOPHER JOHN;JOHNSON, DON HERRICK;BARANIUK, RICHARD GORDON;AND OTHERS;SIGNING DATES FROM 20080401 TO 20080404;REEL/FRAME:021626/0687 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
FEPP | Fee payment procedure |
Free format text: PAT HOLDER NO LONGER CLAIMS SMALL ENTITY STATUS, ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: STOL); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
AS | Assignment |
Owner name: NAVY, SECRETARY OF THE UNITED STATES OF AMERICA, V Free format text: CONFIRMATORY LICENSE;ASSIGNOR:RICE UNIVERSITY;REEL/FRAME:042221/0020 Effective date: 20120326 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552) Year of fee payment: 8 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
FEPP | Fee payment procedure |
Free format text: 11.5 YR SURCHARGE- LATE PMT W/IN 6 MO, LARGE ENTITY (ORIGINAL EVENT CODE: M1556); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 12 |