US6531971B2  Method for monitoring information density and compressing digitized signals  Google Patents
Method for monitoring information density and compressing digitized signals Download PDFInfo
 Publication number
 US6531971B2 US6531971B2 US09854860 US85486001A US6531971B2 US 6531971 B2 US6531971 B2 US 6531971B2 US 09854860 US09854860 US 09854860 US 85486001 A US85486001 A US 85486001A US 6531971 B2 US6531971 B2 US 6531971B2
 Authority
 US
 Grant status
 Grant
 Patent type
 Prior art keywords
 sampling
 signal
 rate
 data
 time
 Prior art date
 Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
 Expired  Fee Related
Links
Images
Classifications

 G—PHYSICS
 G10—MUSICAL INSTRUMENTS; ACOUSTICS
 G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
 G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00G10L21/00
 G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00G10L21/00 specially adapted for particular use

 H—ELECTRICITY
 H03—BASIC ELECTRONIC CIRCUITRY
 H03M—CODING; DECODING; CODE CONVERSION IN GENERAL
 H03M7/00—Conversion of a code where information is represented by a given sequence or number of digits to a code where the same information or similar information or a subset of information is represented by a different sequence or number of digits
 H03M7/30—Compression; Expansion; Suppression of unnecessary data, e.g. redundancy reduction
Abstract
Description
This Application is entitled to the benefit of Provisional Patent Application Ser. No. 60/204,174, filed May 15, 2000.
Not Applicable.
Not Applicable.
This invention relates to digital data processing systems which compress and decompress digitized analogue signals, such as signals from microphones or other analogue measurement devises. The invention also relates to data processing systems which analyze and monitor digitized analogue signals for diagnostic and display purposes.
Historic Background Of fundamental importance for the digital processing of analogue signals is the socalled Shannon sampling theorem. It was introduced into information theory by C.E. Shannon in the 1940's. The theorem had been known already to Borel in 1898, according to R. J. Marks II, Introduction to Shannon Sampling and Interpolation Theory, Springer, New York, 1991.
The sampling theorem states that in order to record an analogue signal (such as a signal from a microphone) it is in fact not necessary to record the signal's amplitude continuously. Namely, if the amplitude of the signal is recorded only at sufficiently tightly spaced discrete points in time then from these data the signal's amplitude can be reconstructed at all points in time. To this end the spacing of the sampling points is sufficiently tight if it is smaller than half the period of the highest frequency component which possesses a substantial presence in the signal. It is important to note that for Shannon sampling the spacing of the sampling times must be equidistant.
To be precise (see e.g. the text by Marks mentioned above), the reconstruction of the signal from its discrete samples works as follows: Let us denote the maximal frequency in the signal by ω_{max}. Let us further denote the amplitude of the signal at time t by ƒ(t). Assume that a machine measured and recorded the amplitudes ƒ(t_{i}) of the signal at equidistantly spaced points in time, t_{i}, whose spacing Δt=t_{i+1}−t_{i }is sufficiently small,
Then, the amplitude of the analogue signal ƒ(t) at any arbitrary time t can be calculated from the measured values ƒ(t_{n}) in this way:
Here, G(t, t_{n}) is the socalled “cardinal series reconstruction kernel”, or “sampling kernel”:
This method of reconstructing an analogue signal's amplitude at arbitrary times from only its discretely taken samples can easily be implemented on computers—and it is of course in ubiquitous use.
Shannon Sampling Is Not Optimally Efficient
While this method, “Shannon sampling”, has been of enormous practical importance, it is clearly not efficient:
When using the Shannon sampling method the highest frequency that is present in the signal determines the rate at which all samples must be taken. Namely, the larger the highest frequency in the signal the more samples must be taken per unit time. This means, in particular, that even if high frequencies occur in a signal only for short durations one must nevertheless sample the entire signal at a high sampling rate.
In practise, it is clear that the “frequency content”, or “bandwidth”, or “information density” of common analogue signals is not constant in time and that high frequencies are present often only for short durations. Therefore, it should normally be possible to suitably lower the sampling rate whenever a signal's information density is low and to take samples at a high rate only whenever the signal's information density is high. The Shannon sampling method, however, does not allow one to adjust the sampling rate: Shannon sampling is wasteful in that it requires one to first determine the highest overall frequency component in the signal and then, second, to maintain a correspondingly high constant sampling rate throughout the recording of the signal.
This shortcoming of Shannon sampling is important because the sampling rate of digitized analogue signals is usually the major limiting factor for the availability of network transmission speed and for computer memory capacity. Therefore, in order to use data memory and data transmission resources most efficiently, it is highly desirable to find ways to continuously adapt the sampling rate to the varying information density of the signal.
For this purpose, one needs, 1) methods and systems for measuring how a signal's information density varies in time so that one can adjust the sampling rate accordingly and, 2) methods and systems for reconstructing the signal from its sotaken samples.
Any method that allows one to sample and reconstruct analogue signals at continuously adjusted rates that are lower than the constant Shannon sampling rate amounts to a data compression method.
It would be desirable to be able to implement such a compression method purely digitally: An analogue signal that has been sampled conventionally, i.e. equidistantly (and therefore wastefully), is digitally analyized for its timevarying information density, then digitally resampled at a correspondingly timevarying sampling rate (using the cardinal series sampling formula of above), and is later decompressed by resampling it at a constant high sampling rate using a new sampling kernel that replaces the cardinal series sampling kernel and is appropriate for the case of the timevarying sampling rate.
It is clear that for such a data compression method to be most useful, the quality of the subsequent reconstruction of the signal should be controllable.
It is also clear that means or method for reliably measuring the timevarying information density of analogue signals can also be used for monitoring and displaying the information density of analogue signals. The ability to measure a timevarying characteristic of an analogue signal, such as here the signal's timevarying information density, can be of great practical value, e.g. for monitoring or diagnostic purposes, as will be explained further below.
The present invention provides corresponding methods and means.
Prior Art Techniques for Adaptive Sampling Rates
Much prior art has strived to achieve methods of sampling and reconstruction which use adaptively lower sampling rates:
Kitamura et al.
The system described by Kitamura, in U.S. Pat. No. 4,370,643 samples signals from analogue at a variable rate. The sampling rate is adjusted according to the momentary amount of change in the signal's amplitude. The reconstruction quality is not controlled. The system described by Kitamura et al., in U.S. Pat. No. 4,568,912 improves on this by reconstructing the signal as joined line segments. The inventors aim is data compression by adaptive rate sampling and also elimination of quantization noise. However, neither aim is satisfactorily achieved: Large amplitude changes of low bandwidth signals lead to inefficient oversampling rather than to the desired compression. Also, the quantization noise is not effectively eliminated since it tends to reappear in the form of jitter.
Kitamura et al., in U.S. Pat. No. 4,626,827, recognize deficiencies in their prior system. In their new system they determine the variable sampling rate by optionally either zerocrossing counting or by Fourier transforming the signal in blocks. The sampling rates are submultiples of the basic rate.
However, zerocrossing counting is a very unreliable indicator of a signal's minimum sampling rate: a signal can be very wiggly (and thus information rich) over long intervals without crossing zero at all.
The alternatively described method of establishing the minimum sampling rate by Fourier analysis of a block (or “interval”, or “slice”, or “period”) of the signal is also unreliable. There are two main reasons:
First, there is the wellknown timefrequency uncertainty relation. Second, it is known that even low bandwidth signals can be arbitrarily quickly varying in arbitrarily long intervals, and vice versa. Therefore, any method that determines a variable sampling rate by Fourier analysis of blocks of the interval is necessarily prone to uncontrolled instances of overor undersampling.
Kitamura et al. recognize that there is a problem and try to repair these effects by sending the analogue signal before sampling through a lowpass filter which cuts off at the chosen rate, and there is a similar filter for the output. Also this is not fully satisfactory: The system still amounts to trying to do Shannon sampling for variable rates. However. since Shannon sampling requires constant sampling rates throughout there necessarily arise reconstruction errors wherever the sampling rate changes.
Podolak et al.
The system then described by Podolak et al. in U.S. Pat. No. 4,763,207 works with variable sampling rates, the rates being determined from a set of cascaded lowpass filters. The system which Podolak et al. later describe in U.S. Pat. No. 4,816,829 and U.S. Pat. No. 4,899,146 is similar but does the filtering digitally. In effect, also these systems try to use Shannon sampling for variable rates—even though it is wellknown that Shannon sampling is only correct for strictly constant rates. Indeed, the authors recognize that in their system, in order to reduce uncontrollable errors, the rate must always be held constant for some rather significant length before allowing it to change again.
Similar to Podolak et al. above, Page in U.S. Pat. No. 4,755,795 describes a system in which variable sampling rates are determined by dynamic shorttime bandwidth analysis. While Podolak's method keeps the sampling rate constant for significant stretches of time, Page's method leads to continuously varying sampling rates. As just discussed above, Podolak keeps the sampling rate constant in stretches because this allows Podolak to provide an approximate decompression method by using the Shannon sampling theorem, whose validity is restricted to constant rates. Page, due to his method's continuously varying sampling rates cannot use the Shannon sampling theorem for decompression. Indeed, Page is silent on how to decompress, i.e. on how to reconstruct the original signal with controllable error from samples taken at a varying rate.
Johnson et al.
Johnson et al., in U.S. Pat. No. 5,302,950 then describe a system for automatic detection of a signal's minimum constant sampling rate. This invention works with constant sampling rates and the scope of this invention is merely to provide users with information about memory versus quality options in recording sessions. The determination of the minimum sampling rate is optionally by zerocrossing counting or by blockwise Fourier transform.
Johnson et al., in U.S. Pat. No. 5,543,792 then build on this method. They describe a system which does data compression again by effectively doing Shannon sampling at blockwise constant rates. This system therefore too suffers from the above mentioned deficiencies. In particular, for the reconstruction purposes this system establishes a common time base of the various rates used. This is to fill in, i.e. to restore the samples previously dropped in the compression process. However, it is known that the reconstruction of lost samples requires some significant oversampling—because the reconstruction of lost samples is numerically highly unstable under small perturbations.
In addition, their methods of zerocrossing counting or Fourier transform in blocks for the purpose of determining the sampling rate suffers again of the above mentioned deficiencies. The system also does not provide control over the quality of the reconstruction.
In Johnson et al., U.S. Pat. Nos. 5,610,825 and 5,893,899 the authors add a system for informing users of potential information loss through undersampling. However, the method is not measuring the actual reconstruction quality. Instead, it merely compares the actual sampling rate to a theoretical rate. The theoretical rate is determined again by methods such as Fourier transform in blocks, the reliability of which is limited due to the above mentioned reasons.
Systems Using Psychoacoustic and Psychovisual Effects
Systems of the type described e.g. by Anderson in U.S. Pat. No. 5,388,181 utilize peculiarities of human perception, i.e. certain psychoacoustic and psychovisual effects. These methods compress data by dropping such information the loss of which humans are normally not likely to perceive. Such systems are limited to analogue signals which are audio or video signals, and among these only to those signals for which high quality reconstruction is inessential. Due to their loss of information such compression systems are unsuitable, e.g., for the compression of most medical and scientific analogue signals.
Prior Art Which Measures Signals' Variable Information Density for Monitoring and Diagnostic Purposes
Methods for measuring a signal's varying information density are useful not only for data compression purposes. Methods for measuring a signal's varying information density can also be a valuable tool for the technical and medical monitoring of the source of the analogue signal:
For example, it has been found that often characteristics of sounds emitted by machines change shortly before the machine fails. Similarly, much effort is put into the analysis and monitoring of seismic signals for changes which might indicate imminent quakes. The monitoring of signal characteristics has also proven useful in medical applications, for example in efforts to predict epileptic seizures from patients' EEG signals. Current methods and applications are described, e.g., in L. Cohen, Timefrequency analysis, Prentice Hall, 1995.
In prior art, methods which monitor signals by measuring the signals' varying information density use the same approaches as discussed before in the context of data compression, such as zero crossing counting and various windowed Fourier transforms. These methods therefore possess the same deficiencies as discussed above in the context of data compression.
Limitations of Prior Art
To summarize, prior art methods and systems do not reliably measure the analogue signals' timevarying information density. Therefore, prior art does not provide means or methods for optimally adjusting the sampling rate to the varying information density of signals. Therefore, prior art adaptive rate sampling does not yield efficient data compression. Prior art also does not provide efficient means or methods for decompressing data sampled at varying rate because prior art does not allow the user to satisfactorily control the amount and type of compression loss, namely, prior art does not allow the user to satisfactorily control the amount of possible deviation of the decompressed signal (i.e. of the reconstructed signal) from the original signal.
The present invention comprises a digital method for measuring a digitized analogue signal's varying information density. This part of the invention can stand alone, i.e. it alone is already useful. Namely, it provides a new method for monitoring how an important characteristic of analogue signals, namely their information density, varies over time. This monitoring ability can be useful both for technical and medical diagnostic purposes.
Based on this method for measuring the variable information density of signals, the present invention further comprises a digital method for compressing and decompressing digitized analogue signals. Namley, according to the present invention, 1) a signal's samples taken conventionally at a constant high rate are input, 2) the signal's timevariable information density is measured digitally, 3) the signal is numerically resampled from these samples taken at the constant high sampling rate to samples at a continuously variable lower sampling rate which corresponds to the measured timevarying information density and 4) those lower rate samples (and the rate information) are output as the compressed data. 5) The invention also provides a method for reconstructing the original signal from this output by using a new sampling kernel that replaces the usual cardinal series sampling kernel.
It is an object of the invention to thereby provide a method for approaching the theoretical limit of compression. It is a further object of the invention that it provides the user with full control over the amount and type of lossyness of the compression, i.e. over the quality of the reconstructed signal.
According to the present invention, the timevarying information density of analogue signals is measured by applying a line search algorithm. This new method has the important advantage that the sampling rate is adjusted truly continuously, thus adjusting optimally to the signal and avoiding the block boundary artifacts of prior art. The line that the algorithm searches for is the maximum sample spacing as a function of time. In the line search algorithm, trial timevarying sampling rates are systematically tried out and labeled acceptable if, for the signal at hand, an acceptance criterion calculated from the signal and the trial sampling rate is met. A preferred embodiment of a line search algorithm is explicitly given.
Also, several explicit embodiments of acceptance criteria are given. The preferred embodiment of an acceptance criterion is the criterion whether the reconstruction with a preferred choice of reconstruction kernel yields a reconstructed signal of predetermined maximum deviation from the original signal. This embodiment has the important advantage that it provides the user with full control over the amount of lossiness of the compression. The invention allows the use of arbitrary choices of functions G(t,t_{n}) as the reconstruction kernel for timevarying sampling rates. A preferred choice of G(t,t_{n}) is presented, which is optimal according to the present inventor's analysis. The line search outputs the lowest timevarying sampling rate that meets the acceptance criterion. The compression can be iterated, e.g. with successive levels of required reconstruction quality applied to the difference between the original and the reconstructed signals of the previous compression step.
For the underlying idea, please see the section “Theory of Operation”, below.
FIG. 1 shows the hardware configuration of the computing means for compression, decompression and information density measurement.
FIG. 2 shows the compression algorithm.
FIG. 3 gives the example reconstruction of a simple but important type of signal: the step function. The invention and conventional Shannon sampling are compared, demonstrating the great efficiency gain through the invention.
FIG 4 is a program listing with preferred embodiments of essential features of the invention. The program listing carries its own documentation.
The Compression
According to the present invention, computing means for compression comprise a CPU, memory for data, memory for programs—which may comprise both RAM and ROM—interfaces to import and export data, and optionally an interface for user control, such as a display and keybord.
The computing means for compression receive an analogue signal in conventionally digitized form, i.e. in the form of equidistantly sampled amplitude data. The computing means for compression process the signal by implementing the following algorithm as a computer program in a convenient programming language:
A pointer variable denotes the region of the signal which is currently being compressed, and this pointer's value successively increases from initially denoting the beginning to finally denoting the end of the signal. At any given time the computing means for compression need to retrieve and store only a sufficient number of signal data up and downstream from the pointer value. Sufficient data are as many as are needed to allow the convergence of the reconstruction kernel summations described below to converge up to a predetermined precision. For onthefly operation there is therefore a short lookahead.
As a subroutine, a line search routine is implemented. The line searched is the optimal sample spacing as it varies along the signal.
The line search routine tests variable trial sampling rates in a region of either predetermined or of ratedependent length downstream from the point currently designated by the pointer variable. As mentioned above, this stretch is to be chosen small enough as compared to the strech of original signal data currently loaded. This stretch is small enough if in the line search region the values of the signal are computable up to desired precision via the cardinal series reconstruction kernel from the original signal data currently loaded. The line search routine outputs the lowest variable sampling rate which passes an acceptance test.
When a line search ends, data representing the sofound optimal sampling rate are stored and the pointer is moved forward. It is moved forward by an amount such that the subsequent line search then applies to a stretch of the signal which is further downstream but which still possesses significant overlap with the stretch of the preceding line search. Thus, upstream of the pointer value the sampling rate is then fixed. Samples taken upstream, at this optimally low variable rate, together with data describing the samples' variable spacing can be output as soon as produced, i.e. whenever the pointer variable moves, or a posteriori. Those are the compressed data. The compression as a whole, may also be iterated. For example, successive compressions can be applied to the difference between the original and the reconstructed signal of the prior compression, thereby enforcing increasing reconstruction quality.
The invention offers users to choose any desired compromise between the processing speed and the level of resolution of the information density: The larger the overlap of successively line searched stretches the finer and therefore better the varying information density is being resolved. On the other hand, choosing a smaller overlap of successive line searches allows faster processing of the signal.
This user choice also affects the achievable compression ratio: The larger the overlap is chosen and therefore the finer the information density is resolved, the fewer samples need to be taken. However, this also means that more data are needed to specify the varying sampling rate. In the preferred embodiment an intermediate size overlap is implemented and the number of lines searched in the line search is restricted such that a small number of bits suffice to describe the variations of the sampling spacing.
In principle any arbitrary line search algorithm (many are known) can be used. The preferred embodiment is explicitly given in FIG. 4. In the preferred embodiment, the line search simply tries to increase the previously adopted spacing for the interval immediately prior to the pointer for the next spacing by a predetermined number N of positive and negative trial percentages, the trial increase being sustained for a predetermined number R of samples, followed by relaxing the spacing from there to the original sample spacing within a predetermined number, e.g. 20, of successive samples. In FIG. 4, the prefered embodiment is presented for R=1 in the form of the enclosed program listing. The finding of the largest acceptable trial variable sampling rate then terminates the line search and the result is stored by storing which of the N trial sampling rate changes has been accepted for the M samples. Consequently, the cost for storing the sampling spacing information is merely log_{2}N bits per M new samples.
However, this particular line search subroutine is only one embodiment and of course any one of the many known textbook line search routines or its variations may be used as the line search subroutine. The test upon which the line search subroutine is to decide the acceptance of a trial sampling rate is one of the following: In one embodiment, the acceptance test used by the line search routine is the criterion whether the error of reconstruction of the signal from the lower rate is below a predetermined threshold. To this end, the signal is computationally reconstructed by numerically resampling it from samples spaced at the trial rate. Recall here that Shannon sampling, which resamples using the cardinal series reconstruction kernel, i.e. which uses the sincfunction kernel, is not suitable for this purpose because Shannon sampling applies only to equidistantly spaced samples.
According to the present invention, any function G(a, b) of two variables can be used as the kernel in the reconstruction computation: ƒ(t)=Σ_{n}G(t,t_{n})ƒ(t_{n}). However, the quality of the achievable compression strongly depends on the choice of reconstruction kernel for the reconstruction from samples which are nonequidistantly spaced.
The preferred choice of reconstruction kernel is:
Here, the t_{n }are the (now nonequidistant) sampling points. The function s is s(t_{n})=t_{n}+1−t_{n }in the preferred embodiment. Alternatively, s can be any measure of the spacing around the point t_{n}, as described in more detail below. Equivalent expressions for this kernel will be given below. According to the author's analysis, the reconstruction kernel above allows one to approach optimal compression. (Note that in the special case of equidistant sampling spacing this preferred kernel reduces indeed to the conventional cardinal series kernel).
The evaluation of the acceptance test then consists in computing any one of the usual measures for the difference between two signals. in this case between the original and the reconstructed signal. The test results in acceptance of the trial line if the error is smaller than some predetermined value, i.e. if the numerical reconstruction meets a predetermined level of precision. In the preferred embodiment the difference between original and reconstructed signal is measured by the number of original sampling points within the current stretch, at which the original signal's and the reconstruction's amplitude differ by more than a predetermined amount. In the preferred embodiment this test is set to yield acceptance only if this number is zero. As a consequence, the preferred embodiment has the advantage that its operation guarantees that the reconstructed, i.e., the decompressed signal at no sampling point differs more than a predetermined amount from the original.
In a second embodiment, the acceptance test used by the line search sub routine is the criterion whether an alternating sum of samples with weighted coefficients, or “weighted alternating sum”, when taken at the trial rate, is consistently close to zero. Here, the sum is consistently close to zero if it is close to zero in all of a predetermined number computations of the weighted alternating sum for samples taken at the rate given by the trial line, but with different sampling points. The quality of the compression depends on the choice of the precise form of this weighted alternating sum. The preferred choice of formula for the weighted alternating sum improves upon the Fourier series by making it suitable for use with nonequidistantly spaced samples; it is the following formula:
Here, t′_{n }is any measure of the spacing around the point t_{n}. In the preferred embodiment it is simply t′_{n}=t_{n+1}−t_{n}.
In alternative embodiments of acceptance tests a combination of both of the above acceptance tests can be chosen: The former test has the advantage of providing absolute control over the reconstruction quality. The latter test has the advantage of being computationally fast. Both advantages can be utilized for example by implementing the fast test for initial rough stages of the line search and the fully controlled test for the finetuning in final stages of the line search.
The Decompression
According to the present invention, computing means for decompression comprise a CPU, memory for data, memory for programs—which may comprise both RAM and ROM—interfaces to import and export data, and optionally an interface for user control, such as a display and keybord.
The computing means for decompression operate by implementing the following algorithm as a computer program is a convenient programming language: The algorithm consists in numerically resampling the signal at a high constant rate from the compressed data. To this end again any function G(t,t_{n}) as a generalized reconstruction kernel for nonequidistantly spaced samples can be used. For anyone skilled in the art it is of course very easy to implement the numerical resampling of the signal at a constant high rate by numerically implementing ƒ(t)=Σ_{n}G(t,t_{n})ƒ(t_{n}) as a computer program, for whatever choice of G(t, t_{n}).
In the preferred embodiment G(t,t_{n}) is the same kernel as was used for the compression. If the data have been compressed without using a reconstruction kernel for nonequally spaced samples, e.g. by using the second above mentioned test for the line search, of if the nonequally spaced samples have been obtained by different methods, e.g. by methods of prior art, then the present invention allows users to decompress these data as if they were obtained by using the method which uses a reconstruction kernel for nonequally spaced samples, as disclosed above. In this case, the preferred reconstruction kernel for the decompression is the one disclosed in the above equation.
The Information Density Measurement
According to the present invention, computing means for the measurement of signals' information density comprise a CPU, memory for data, memory for programs—which may comprise both RAM and ROM—interfaces to import and export data, and optionally an interface for user control, such as a display and keybord.
The method of measuring the variable information density of signals for monitoring and display purposes consists of identical steps as the above described method of data compression. Only, in this case, the output consists merely of the data which describe the optimally low, variable, sampling rate.
Advantages
The invention has the advantage that it provides a much improved method for measuring the timevarying information density of digitized analogue signals. It therefore provides improved tools for monitoring sources of analogue signals. The invention can be particularly useful for example for medical and technical diagnostic purposes, such as e.g. for the prediction of machine failure through machine sound monitoring.
The invention has the further advantage that, based on the new method for measuring the timevarying information density of signals, it provides an improved method for the compression and decompression of digitized analogue signals. The compression method has the advantage to provide means for approaching the theoretical limit of compression. The invention therefore allows the user to more efficiently use the resources of computer memory and network capacity.
A further advantage of the invention is that it provides the user with full control over the lossiness of the compression, i.e. over the quality of the reconstructed signal: Indeed, the invention has the advantage that the user can choose seamlessly from lossless to lossy compression. The user can even choose both the amount and the type of tolerated compression loss a feature which is particularly useful when the quality of the decompressed signal is essential, as e.g. in medical applications, where uncontrolled reconstruction artifacts could lead to misdiagnoses.
FIG. 1 shows the hardware configuration of the computing means for compression, decompression and information density measurement. These are essentially the same for all three:
The computing means comprise a CPU, fast RAM memory for data, disk and fast RAM or ROM memory for programs. These receive and output data through an interface or port device which are able to communicate with external data storage, external computer networks or external realtime data acquisition (such as from A/D samplers) or realtime data postprocessing (such as to D/A converters or to displays). The computing means comprise, optionally, also an interface for human intervention, e.g. through display and keyboard. E.g. a human operator may wish to influence the settings for the tolerated lossyness of the compression. On the other hand, the invention can also be embodied in an entirely hardwired form, with all options preset. In this case, the program which implements the compression, or the decompression or the informations density measurement, may be stored in ROM memory. As such, the various ramifications of the invention can be incorporated into recording devices and into replaying devices and into monitoring devices, where they can operate onthefly, i.e. essentially realtime, up to a short delay due to the required short lookahead and processing time.
FIG. 2 shows the compression algorithm: In box 20, a pointer variable is set to zero. The pointer variable will always point to the position up to which the signal has been compressed. In Box 22, sample data, i.e. the equidistantly taken samples of the to be compressed signal are loaded for a finite length downstream. The length of this lookahead can be predetermined fixed or ratedependent, as will become clear after we return to box 22 from the algorithm. Box 24 contains a line search algorithm which finds—among the lines which this particular line search considers—the line which describes the largest generally variable sampling spacing in a stretch downstream from the pointer, the length of the stretch being predetermined or ratedependent. This optimal rate is stored. In box 26, the pointer is moved downstream by a predetermined or ratedependent amount. The sampling rate upstream from the pointer is now fixed up to the pointer. In Box 27 those upstream data are stored. For onthefly operation, samples taken upstream at the now established lower sampling rate, together with data that describe those samples' spacing can already be output. These are the compressed data. In box 28 the pointer is checked for having reached the end of the to be compressed signal's data file. If not, the algorithm proceeds to box 22, otherwise the algorithm reaches box 29. If the compressed data have not already been output on the fly, in box 17, then they are to be output now with the end of the algorithm in box 29. As mentionen above, the compression may be iterated. In this algorithm its speed and its capability of resolving details of the varying sampling rate clearly depend on how far downstream the line searched stretch is chosen, and on how far the pointer is moved downstream after each line search. Processing speed can obviously be increased by making these lengths about proportional to the sampling rate in the stretch considered. It is important to note that in all embodiments the lookahead is chosen sufficiently larger than the line searched stretch. This is because the line search obviously requires the computation of sample values through the cardinal series, and the lookahead must be long enough so that the series can be summed up to the required predetermined numerical precision.
Any one of the numerous textbook line search routines or their variations may be used in this compression algorithm. A simple but concrete embodiment is here disclosed as an accompanying program listing. Those skilled in the art will readily be able to implement the above compression algorithm with efficient line search routines in programming languages that are adapted to the fast numerical processing of large amounts of data. The line search finds for the considered downstream stretch the largest variable sampling spacing line which passes an acceptance test. I here disclose two basic types of test, with combinations being possible. An embodiment of the test is implemented in the enclosed program listing, FIG. 4.
Description of the Method for Measuring and Monitoring Signals' Varying Information Density
It is clear that the same figure, FIG. 2, can describe also the working of the disclosed method for measuring the information density of signals for the purpose of monitoring the sources of analogue signals. The operation is identical, up to the single point that the data to be output are merely the data which describe the variable sampling rate. This is because the varying sampling rate is by itself already the desired measure for the information density of the signal: Where the sampling rate is large, the information density is low and vice versa. Clearly, the varying sampling rate output can be visualized on displays or serve as the input of monitoring postprocessing.
Description of the Decompression Method
The decompression of the compressed data, i.e. of a file which contains sample amplitudes together with data which indicate the spacings at which these samples where taken, works as follows: A file is created of conventional, i.e. of equidistant samples of the signal. This is achieved by computing the reconstructed signal at a high equidistant sampling rate from the nonequidistantly spaced samples by using a reconstruction kernel G(t, t_{n}). If the signal was compressed using in the line search the first of the two tests, then the reconstruction kernel used in this test is the preferred kernel also for the decompression. If the data were not compressed in this way (i.e. either by using the second test, or even by using some other compression method that yields nonequally spaced samples and their spacing as its output) then the preferred reconstruction kernel is the one which has been explicitly disclosed above. The implementation of the numerical resampling of the signal from the unequally spaced data using the generalized sampling kernel is similar to the familiar Shannon sampling and those skilled in the art will therefore not require explanation in an extra flow chart.
FIG. 3 shows four graphs. The four graphs of FIG. 3 illustrate how the new sampling method handles the worst case scenario, namely the step function (which is of course not bandlimited at all). FIG. 3 demonstrates the large efficiency gain through the invention. Each graph of FIG. 3 shows in red the original signal, i.e. the step function. The ticks on the time axis indicate a choice of points at which the original signal is being sampled. The reconstruction of the signal is then made by using the preferred reconstruction kernel. The reconstructed functions are shown in blue.
1. The first graph, 3 a, shows the result at a constant low sampling rate. Since the sampling rate is constant this case is ordinary Shannon sampling. Clearly, since the sampling rate is low, only the low frequency components of the step are properly captured and this leads to a poor reconstruction of the signal. One recognizes the Gibbs phenomenon.
2. In second graph, 3 b, the flexibility provided by the new sampling method is used to choose a three times higher sampling rate close to the step while away from the step the sampling rate decreases down to the same sampling rate as in the first graph. We see that the increased sampling rate close to the step already improves the reproduction of the signal.
3. In the third graph, 3c, the sampling rate close to the step is chosen 10 times higher than the rate in the first graph. Again, away from the step, the sampling rate is chosen to linearly decrease down to the rate of the first graph. The reproduction is further improved.
4. In the fourth graph, 3 d, the 10fold sampling rate is chosen throughout the time axis. Since the sampling rate is now constant this is again ordinary Shannon sampling, with the expected result.
We can now compare the third with the fourth graph: The sampling rate used in the third graph is everywhere on the time axis only a tenth of the sampling rate used in the fourth graph, apart from a small region close to the step. Thus much fewer samples are needed in the third graph. Nevertheless, one sees that the reproduction of the original signal is in both graphs of comparable quality. It is clear that the new method as applied in the thrid graph does indeed increase sampling efficiency, namely by allowing comparable quality reconstruction from a highly reduced number of samples.
FIG. 4 is the program listing with a preferred embodiment. The listing carries its own documentation.
The present invention, through new techniques for determining signal's timevarying information density overcomes the limitations of prior art such as the above described limitations of zerocrossing counting or such as blockwise, or “windowed” Fourier transforming. The invention thereby yields a powerful new method for measuring a crucial timevarying characteristic of signals, namely a signal's timevarying information density.
The output, display and possible further processing of the timevarying information density of a signal can be useful in numerous contexts. For example, it can serve as an indicator of changes in the process that produces the analogue signal, e.g. for the purposes of monitoring machine noises to predict machine failure, or also for the monitoring of human vital functions for diagnosis or predictions, such as predictions of seizures from EEG's.
The present invention also provides new techniques for the compression and decompression of digitized analogue signals by adaptively lowering the sampling rate. The invention provides for a technique that utilizes truly continuously varying sampling rates, thereby allowing optimal adaptation of the sampling rate to the signal, namely by utilizing not the cardinal series sampling kernel but instead sampling kernels that are adapted to varying sampling rates.
This invention is not to be construed as limited to the particular preferred embodiments, since these are regarded as illustrative rather than restrictive. Moreover, variations and changes may be made by those skilled in the art without departing from the spirit of the invention.
Theory of Operation
Prior art methods to compressing analogue data such as speech or music by reducing the sampling rate have focussed on the idea that the “bandwidth” of signals can vary in time, and that therefore it should be possible to choose a variable sampling rate accordingly. In practice it is clear what is meant by a variable bandwidth, and it is clear that some form of data compression along these lines should be possible. However, progress in this direction has been hampered by a simple but important fact: The concept of bandwidth requires Fourier analysis. Fourier analysis, however, requires equidistant samples.
Much prior art has therefore been devoted to inventing methods that use blockwise constant sampling rates so as to be able to use conventional Shannon sampling within each block. However, since Shannon sampling strictly requires constant sampling rates, there then necessarily arise significant errors due to the finite size of these blocks.
The new invention solves the original problem. Namely, it provides a method to continuously adjust the sampling rate to the information density of the signal. To this purpose, the invention does not use Fourier methods. Instead, the starting point is to directly address the main goal: the measurement of the information density and the compression of data:
I propose that the true reason why typical signals are compressible when equidistantly sampled is not to be found with any Fourier based concept of variable bandwidth. Instead, the reason why most signals are compressible is the following:
In some regions of typical signals, neighboring samples will be strongly correlated while in other regions neighboring samples will be less correlated. For example: if a machine records only one sample per second from a music recording, then neighboring of those samples will be essentially uncorrelated. If, however, the machine takes 106^{6 }samples per second then the music signal changes little from one sample to the next, and therefore neighboring of those samples will be strongly correlated.
As is clear, however, correlated data can be compressed. Indeed, maximally compressed data are maximum entropy data. This is because if a signal is not maximum entropy and therefore possesses residual correlations among its data then these correlations could be exploited for further compression.
Therefore, in regions of a signal where equidistantly sampled neighboring samples are strongly correlated, compression is possible: to this end one can simply take fewer samples. This is because, as mentioned, wider spaced samples are less correlated.
The basis of the present data compression method is therefore the idea to increase the sample spacing to a timevarying degree throughout the signal, up to just before the samples start to become uncorrelated. The soobtained variable spacing is then inversely proportional to the information density of the signal.
In order to measure the correlation of the samples taken at a trial sampling rate, a machine can compute the above mentioned weighted alternating sum of samples: By the central limit theorem, in any fixed interval the weighted alternating sum has the same normal distribution independently of how the samples are spaced—if these samples are uncorrelated. For strongly correlated samples the alternating sum is essentially zero. Therefore, by computing the alternating sum for a trial sampling rate, a machine can test whether the signal is still compressible, namely by testing whether the samples are correlated: In this case, the weighted alternating sum is consistently essentially zero. Consistently essentially zero means that the sum will be essentially zero, i.e. only of the order of the average amplitude, no matter which set of sampling points with the trial spacing is considered. To exclude the case of a mistaken acceptance due to the sum for uncorrelated samples being accidentally zero, it is therefore preferred to run this test several times for each trial sampling rate line, each time choosing slightly different sampling points which obey the spacing indicated by the trial line. Since the calculation of the weighted alternating sum is computationally fast, even repeating the calculation a few times is still computationally fast.
After establishing a lower variable sampling rate in this way, the question is which reconstruction kernel applies. A mathematical result obtained by the inventor shows that the preferred reconstruction kernel, namely
reconstructs signals so that the alternating weighted sum indeed essentially zero. Here the t_{n }are the generally nonequidistantly spaced sampling points. The definition of the t′_{n }is as follows: As with Shannon sampling, also here, there is not only one set of sampling points which have the optimal spacing: There is a whole family of sampling grids, or onedimensional sampling lattices, which all have the optimum variable spacing. These grids differ essentially only by translation from another. One can parametrize the shifting from one optimum spacing sampling grid to the next by a continuous parameter, say α, so that the sampling points are now parametrized: t_{n }(α). Then, t′_{nl =dt} _{n}(α)/dα_{α=0}. The details of the parametrization are irrelevant because any overall factor for the t′_{n }drops out of both, the calculation of the weighted alternating sum and the calculation of the reconstruction kernel.
While this is the theoretical and strictly correct expression, in practice, a simple approximation can often be used: It is clear that t′_{n }is essentially the sampling spacing at t_{n}. Thus, as long as one does not attempt to resolve the information density to a precision on the order of a few sampling spacings, it is numerically satisfactory to simply set t′_{n }equal to the lattice spacing pre or post t_{n}, or any type of average of the two. However, when the information density is to be resolved to high precision, e.g. for monitoring purposes, or in order to achieve highest compression, then the line search subroutine is to compute the t′_{n }as the derivative of the line that is searched up to the precision to which this derivate makes sense, i.e. up to the precision to which the line is being resolved.
Crucially, the knowledge of the reconstruction kernel opens up a powerful new way of measuring the information density: The measurement of the signal's information density through the calculation of the weighted alternating sum is based on statistics. Namely, the algorithm distinguishes a zero sum from a sum that is gaussian distributed. Given the reconstruction kernel, it becomes possible to entirely avoid the uncertainties of statistics and arrive at full control over the measurement of the information density: Namely, the machine can directly compare the signal reconstructed from the samples taken at the trial rate with the original signal.
If one were to restrict the new compression method by allowing the use only of constant sampling rates, then one would recover Shannon sampling: The preferred weighted alternating sum then simplifies to become a Fourier series. Also, the preferred reconstruction kernel then simplifies and becomes the Shannon theorem's cardinal series sincfunction reconstruction kernel.
Claims (11)
Priority Applications (2)
Application Number  Priority Date  Filing Date  Title 

US20417400 true  20000515  20000515  
US09854860 US6531971B2 (en)  20000515  20010514  Method for monitoring information density and compressing digitized signals 
Applications Claiming Priority (1)
Application Number  Priority Date  Filing Date  Title 

US09854860 US6531971B2 (en)  20000515  20010514  Method for monitoring information density and compressing digitized signals 
Publications (2)
Publication Number  Publication Date 

US20010050953A1 true US20010050953A1 (en)  20011213 
US6531971B2 true US6531971B2 (en)  20030311 
Family
ID=26899256
Family Applications (1)
Application Number  Title  Priority Date  Filing Date 

US09854860 Expired  Fee Related US6531971B2 (en)  20000515  20010514  Method for monitoring information density and compressing digitized signals 
Country Status (1)
Country  Link 

US (1)  US6531971B2 (en) 
Cited By (8)
Publication number  Priority date  Publication date  Assignee  Title 

US20040178933A1 (en) *  20030311  20040916  Canon Kabushiki Kaisha  Encoding method and encoding apparatus, and computer program and computer readable stroage medium 
US20050261878A1 (en) *  20040521  20051124  Sandeep Shrivastava  Diagnostic image 
US20050261879A1 (en) *  20040521  20051124  Sandeep Shrivastava  Diagnostic context 
US20050273667A1 (en) *  20040521  20051208  Sandeep Shrivastava  Diagnostic instrumentation 
US20050273490A1 (en) *  20040521  20051208  Sandeep Shrivastava  Hierarchical debug 
US20060088123A1 (en) *  20041021  20060427  Jensen Henrik T  Method and system for Gaussian filter modification for improved modulation characteristics in Bluetooth RF transmitters 
US20090234645A1 (en) *  20060913  20090917  Stefan Bruhn  Methods and arrangements for a speech/audio sender and receiver 
WO2010025546A1 (en) *  20080903  20100311  4473574 Canada Inc.  Apparatus, method, and system for digital content and access protection 
Families Citing this family (3)
Publication number  Priority date  Publication date  Assignee  Title 

US7376534B2 (en) *  20040521  20080520  Bea Systems, Inc.  Watches and notifications 
JP4182448B2 (en) *  20060727  20081119  ソニー株式会社  Receiving apparatus, a receiving method, a program, and a recording medium 
JP4304632B2 (en) *  20061012  20090729  ソニー株式会社  Receiving apparatus, a receiving method, a program, and a recording medium 
Citations (12)
Publication number  Priority date  Publication date  Assignee  Title 

US4370643A (en)  19800506  19830125  Victor Company Of Japan, Limited  Apparatus and method for compressively approximating an analog signal 
US4568912A (en)  19820318  19860204  Victor Company Of Japan, Limited  Method and system for translating digital signal sampled at variable frequency 
US4626827A (en)  19820316  19861202  Victor Company Of Japan, Limited  Method and system for data compression by variable frequency sampling 
US4755795A (en) *  19861031  19880705  HewlettPackard Company  Adaptive sample rate based on input signal bandwidth 
US4763207A (en)  19851015  19880809  R. R. Donnelley & Sons Company  Digital method and system for reproducing analog data 
US4816829A (en)  19870430  19890328  R. R. Donnelley & Sons Company  Method of and apparatus for converting digital data between data formats 
US4899146A (en)  19870430  19900206  R. R. Donnelley & Sons Company  Method of and apparatus for converting digital data between data formats 
US5225787A (en) *  19910510  19930706  U.S. Philips Corporation  Sampling frequency converter including a sigmadelta modulator 
US5302950A (en)  19920717  19940412  International Business Machines Corp.  Method of and apparatus for providing automatic determination of information sampling rate 
US5388181A (en)  19900529  19950207  Anderson; David J.  Digital audio compression system 
US5512895A (en) *  19940425  19960430  Teradyne, Inc.  Sample rate converter 
US5610825A (en)  19941108  19970311  Johnson; William J.  Method and apparatus for the display of digitized analog signal loss 
Patent Citations (13)
Publication number  Priority date  Publication date  Assignee  Title 

US4370643A (en)  19800506  19830125  Victor Company Of Japan, Limited  Apparatus and method for compressively approximating an analog signal 
US4626827A (en)  19820316  19861202  Victor Company Of Japan, Limited  Method and system for data compression by variable frequency sampling 
US4568912A (en)  19820318  19860204  Victor Company Of Japan, Limited  Method and system for translating digital signal sampled at variable frequency 
US4763207A (en)  19851015  19880809  R. R. Donnelley & Sons Company  Digital method and system for reproducing analog data 
US4755795A (en) *  19861031  19880705  HewlettPackard Company  Adaptive sample rate based on input signal bandwidth 
US4816829A (en)  19870430  19890328  R. R. Donnelley & Sons Company  Method of and apparatus for converting digital data between data formats 
US4899146A (en)  19870430  19900206  R. R. Donnelley & Sons Company  Method of and apparatus for converting digital data between data formats 
US5388181A (en)  19900529  19950207  Anderson; David J.  Digital audio compression system 
US5225787A (en) *  19910510  19930706  U.S. Philips Corporation  Sampling frequency converter including a sigmadelta modulator 
US5302950A (en)  19920717  19940412  International Business Machines Corp.  Method of and apparatus for providing automatic determination of information sampling rate 
US5512895A (en) *  19940425  19960430  Teradyne, Inc.  Sample rate converter 
US5610825A (en)  19941108  19970311  Johnson; William J.  Method and apparatus for the display of digitized analog signal loss 
US5893899A (en)  19941108  19990413  Johnson; William J.  Method and apparatus for the display of digitized analog signal loss 
Cited By (17)
Publication number  Priority date  Publication date  Assignee  Title 

US20040178933A1 (en) *  20030311  20040916  Canon Kabushiki Kaisha  Encoding method and encoding apparatus, and computer program and computer readable stroage medium 
US6958714B2 (en) *  20030311  20051025  Canon Kabushiki Kaisha  Encoding method and encoding apparatus, and computer program and computer readable storage medium 
US20050261878A1 (en) *  20040521  20051124  Sandeep Shrivastava  Diagnostic image 
US20050261879A1 (en) *  20040521  20051124  Sandeep Shrivastava  Diagnostic context 
US20050273667A1 (en) *  20040521  20051208  Sandeep Shrivastava  Diagnostic instrumentation 
US20050273490A1 (en) *  20040521  20051208  Sandeep Shrivastava  Hierarchical debug 
US8490064B2 (en)  20040521  20130716  Oracle International Corporation  Hierarchical debug 
US7379849B2 (en) *  20040521  20080527  Bea Systems, Inc.  Diagnostic image 
US7395458B2 (en)  20040521  20080701  Bea Systems, Inc.  Diagnostic instrumentation 
US20060088123A1 (en) *  20041021  20060427  Jensen Henrik T  Method and system for Gaussian filter modification for improved modulation characteristics in Bluetooth RF transmitters 
US20090234645A1 (en) *  20060913  20090917  Stefan Bruhn  Methods and arrangements for a speech/audio sender and receiver 
US8214202B2 (en) *  20060913  20120703  Telefonaktiebolaget L M Ericsson (Publ)  Methods and arrangements for a speech/audio sender and receiver 
WO2010025546A1 (en) *  20080903  20100311  4473574 Canada Inc.  Apparatus, method, and system for digital content and access protection 
EP2321926A1 (en) *  20080903  20110518  4473574 Canada Inc.  Apparatus, method, and system for digital content and access protection 
EP2321926A4 (en) *  20080903  20130213  4473574 Canada Inc  Apparatus, method, and system for digital content and access protection 
CN102197623B (en)  20080903  20140129  4473574加拿大公司  Apparatus, method, and system for digital content and access protection 
US8964972B2 (en)  20080903  20150224  Colin Gavrilenco  Apparatus, method, and system for digital content and access protection 
Also Published As
Publication number  Publication date  Type 

US20010050953A1 (en)  20011213  application 
Similar Documents
Publication  Publication Date  Title 

Makhoul  Spectral linear prediction: Properties and applications  
US20040028244A1 (en)  Audio signal decoding device and audio signal encoding device  
Robinson  SHORTEN: Simple lossless and nearlossless waveform compression  
US5493516A (en)  Dynamical system analyzer  
US20030115052A1 (en)  Adaptive windowsize selection in transform coding  
US7433824B2 (en)  Entropy coding by adapting coding between level and runlength/level modes  
US20030115051A1 (en)  Quantization matrices for digital audio  
US20060155399A1 (en)  Method and system for generating acoustic fingerprints  
US20080243518A1 (en)  System And Method For Compressing And Reconstructing Audio Files  
US20070016404A1 (en)  Method and apparatus to extract important spectral component from audio signal and low bitrate audio signal coding and/or decoding method and apparatus using the same  
US6879955B2 (en)  Signal modification based on continuous time warping for low bit rate CELP coding  
US3976863A (en)  Optimal decoder for nonstationary signals  
Virtanen  Separation of sound sources by convolutive sparse coding  
US5067158A (en)  Linear predictive residual representation via noniterative spectral reconstruction  
US5400371A (en)  System and method for filtering random noise using data compression  
US20040172411A1 (en)  Method and device for producing a fingerprint and method and method and device for identifying an audio signal  
Hans et al.  Lossless compression of digital audio  
US20030115041A1 (en)  Quality improvement techniques in an audio encoder  
US6681204B2 (en)  Apparatus and method for encoding a signal as well as apparatus and method for decoding a signal  
US7516074B2 (en)  Extraction and matching of characteristic fingerprints from audio signals  
US7146313B2 (en)  Techniques for measurement of perceptual audio quality  
US5583784A (en)  Frequency analysis method  
US20070016405A1 (en)  Coding with improved time resolution for selected segments via adaptive block transformation of a group of samples from a subband decomposition  
US7333930B2 (en)  Tonal analysis for perceptual audio coding using a compressed spectral representation  
EP2743922A1 (en)  Method and apparatus for compressing and decompressing a higher order ambisonics representation for a sound field 
Legal Events
Date  Code  Title  Description 

REMI  Maintenance fee reminder mailed  
LAPS  Lapse for failure to pay maintenance fees  
FP  Expired due to failure to pay maintenance fee 
Effective date: 20070311 