CA2206426C - Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor - Google Patents

Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor Download PDF

Info

Publication number
CA2206426C
CA2206426C CA 2206426 CA2206426A CA2206426C CA 2206426 C CA2206426 C CA 2206426C CA 2206426 CA2206426 CA 2206426 CA 2206426 A CA2206426 A CA 2206426A CA 2206426 C CA2206426 C CA 2206426C
Authority
CA
Canada
Prior art keywords
image
pixel
size
pixels
black
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Lifetime
Application number
CA 2206426
Other languages
French (fr)
Other versions
CA2206426A1 (en
Inventor
Roderick J. Ragland
Spiros S. Dimolitsas
Franklin L. Corcoran
Neil H. Tender
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Comsat Corp
Original Assignee
Comsat Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US08/352,836 external-priority patent/US6181825B1/en
Application filed by Comsat Corp filed Critical Comsat Corp
Publication of CA2206426A1 publication Critical patent/CA2206426A1/en
Application granted granted Critical
Publication of CA2206426C publication Critical patent/CA2206426C/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Landscapes

  • Image Processing (AREA)
  • Compression Of Band Width Or Redundancy In Fax (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

Compression and decompression methods for transmitting facsimile images in which the facsimile image is compressed in both vertical and horizontal directions are implemented using two dimensional maximum differences algorithms. Asymmetrical and fractional image compression are provided by controlling the number of lines, i.e., rows or columns, being compressed and by controlling the bit segment length used during each differences determination. The compressed facsimile image is decompressed by selecting a reference bit in the decompressed facsimile image and expanding the bit in both vertical and horizontal directions in accordance with a binary value of bits near the selected reference bit. A transmission channel adapted to serially perform the encoding and decoding functions is also described.

Description

. CA 02206426 1997-0~-29 ~ f W O 96/17469 PCTnUS95/14945 DECODING DURING REAL-TIME FACSIMILE IMAGE COMPRESSION AND
APPARATUS

The present invention relates generally to rea~ time facsimiie image compres-sion. More specifically, the present invention relates to a plurality of efficient methods for the real-time encoding of bi-level graphics, e.g., facsimile images. The inventive methods are particularly suited for mobile communication applications 10 using a 2-dimensional bit-reduction, e.g., decimation, scheme. The methods developed are particularly applicable to the coding of handwritten and typed-text and are suitable for incorporation in the middle, rather than the terminating-points, of a communications network. An apparatus specifically adapted to perform the enumerated methods is also disclosed.

The emerging use of low-rate coding and digital transmission in mobile satellite communications and the increasing use of facsimile services have identified the desirability for visual service transparency over the same very narrow-band circuits. The economic viability of such a service, however, rests on the capability 20 of introducing further image message compression, either in real-time or off-line during a so-called store-and-forward operation. In addition to the economic side, however, long document transmission times negatively impact the customer perceived quality of service in that, over narrow-band circuits, several rninutes might be required for the transmission of documents that would normally need only CA 02206426 1997-0~-29 W O96/17469 PCTrUS9S/14945 seconds when transmitted through the public switch telephone network. In order to achieve these objectives, some form of image compression over and above what is provided by traditional Group 3 facsimile is needed.
Of general interest are U.S. Patent No. 4,779,266 to Chung et al., which 5 discloses methods and corresponding circuitry for encoding and decoding informa-tion with code patterns which engender the two-state equivalent of electronic orthogonal coding. These special code patterns, referred to as optimizing orthogo-nal codes, are useful in system having only two signal propagation states, e.g., optical processors and U.S. Patent No. 4,695,971 to Reimann, which discloses 10 circuitry for rapidly determining the greatest difference among three binary numeri-cal values which undertakes a classification and coding of the maximum differenGe within a number of different numerical ranges.
References of limited interest include U.S. Patent No. 3,971,88 to Ching et al., which discloses a synchronization system for variable length encoded signals, U.S. Patent No. 3,938,085 to Battail, which discloses a transmission system including a transmitting station and a receiving station for operating with a system-atic recurrent code, and U.S. Patent No. 3,927,372 to Zschunke, which discloses techniques for improving the reproduction of amplitude jumps in a differential pulse code modulation (DPCM) system based on the use of maximum difference vaiue zo code words.
Currently there are several methods in use for encoding bi-level graphics.
These include:
a. One-dimensional Huffman Coding, which is used to encode colored (black or white) strips of picture elements (pixels) when an image is raster-scanned and ,~ r W O 96117469 PCTnUS95/1494S

digitized. A well known algorithm belonging to this class of techniques is the 1-Dimensional Run-Length Coding IRLC) which has become an international standard.
~ç CCll~r Recommendation T-4-, "Standardization of Group 3 Facsimile Apparatus for Document Transmissions', Melbourne 1988, Fascicle Vl1.3 Volume Vll, Pages 5 21-47. The 1-Dimensional RLC consists of a Huffrrlan Code that has been suitably modified to increase its robustnèss in the presence of telephone network-type of impairments. The 1-Dimensional RLC is a powerful technique that permits the lossless coding of bi-level images and is able to achieve a bit-rate requirement reduction on the order of 10:1, depending on the statistical content of the irnage 0 encoded. However, since group 3 facsimile messages are already encoded using this technique there can be little, if any, benefit derived by further compressin~ such images in the network using the same, or a modified version of this approach.
b. To increase the compression achievable, a two dimensional versiori of the 1-Dimensional RLC technique has also been developed. See CCITT Recommenda-tion T.4. In the 2-Dimensional RLC method, only the first scan-line of image information is encoded in accordance to the 1 -Dimensional RLC. Subsequently, the differences between adjacent lines, rather than the actual scan-lines, are encoded using a technique that is essentially the same as 1-Dimensional RLC. The 2-Dimensional RLC is also a loss-less coding method. However, because of its 20 increased image redundancy removal, it is more susceptible to telephone network-type of transmission impairments. As a result, 1-Dimensional RLC of the actual scan-lines is used every-few lines to assure that re-synchronization can be estab-lished even when some of the encoded information has been corrupted. As a result, its performance is somewhat limited. Furthermore, since, as stated earlier, group CA 02206426 1997-0~-29 ~, , f ~
W O96/17469 PCTnUS9S/14945 3 facsimile messages are already encoded using this standardized technique, there can be little additional benefit realized, e.g., an additional 20%, by further com-pressing such images in the network using a 2-dimensional Huffman coding ap-proach, or a variant thereof.
C. If the communications channel can be assumed to be error free, e.g., when automatic repeat request procedures are employed, a slightly more efficientmethod of 2-D RLC can be derived. In this method, all image scan-lines are codedon an adjacent line differential basis and 1-Dimensional RLC is not repeated for re-synchronization. This method is somewhat more efficient in that a compression ratios of the order of 20:1 can be achieved. However, the underlying error-free channel assumption is noted. This method has also been standardized, and is covered by, for example, CCITT Recommendation T.6, entitled "~acsimile Coding Schemes and Coding Control Functions for Group 4 Facsimile Apparatus" (Mel-bourne 1988, Fascicle Vl1.3, Volume Vll).
d. Other techniques for the coding of facsimile images have also been used.
See K. Knowlton, "Progressive Transmission of Gray-Scale and Binary Images by Simple, Efficient and Lossless Coding Scheme", Proceedings of IEEE, pp. 885-896 (July 1980), and N.S Jayant et al., "Digital Coding of Waveforms", Prentice Hall(1984). Some of these employ transform domain techniques and promise to be effective, but primarily only when dealing with graphical information such as gray level, or highly detailed images. As a result, such methods have not demonstrated their optimal abilities when coding bi-level handwritten graphics for mobile commu-nication applications.

CA 02206426 1997-0~-29 ~_ f W O96117469 PCTnUS9S/1494S

e. Another technique which has proven to be powerful is based on the segmentation of an image into many sub-images and in subsequently matching the content of these sub-images by elements drawn from a code-book of elementary images. The process is completed by transmitting overthe communications channel a code-word representing the identity of the elementary image most closely resem-bling the image's sub-image. Such techniques combine pattern recognition princi-ples and vector quantization and have demonstrated that significant compression ratios can be achieved, e.g., more that 100:1, if the code-book used is well suited to the image content(on a microscope scale). See 0. Johnsen et al., "Coding of 0 Two-Level Pictures by Pattern Matching and Substitution", The Bell System Techni-cal Journal, Vol. 62, No. 8, pp. 2513-2545 (October 1983) as well as Super Fax Compression, COMSAT Laboratories Final Report under Contract MCS-10 ~Decem-ber 1991). Despite the impressive compression ability of these techniques, some limitations should be noted when the image structure is not well matched to the codebook contents. First, the compression ratios realized are significantly lower than 100:1 . Second, these techniques are information lossy, i.e., the reconstructed image quality is degraded when compared to the ori~inal image. In particular, if the image structure is not well matched to the codebook's contents, the reconstructed image quality can be significantly degraded. As a result, the performance of these techniques is essentially limited to compression ratios in the range of 40:1 to 100:1 and their use can be primarily confined to typed text, where the outline of the characters to be coded is well defined and uniform throughout a document.
f. A technique called Vector Encoding with a Distortion Criterion (VEDC), was proposed but not disclosed by COMSAT Corporation employees which models CA 02206426 1997-0~-29 f r W O96/17469 PCTnUS9~/1494S

each character within an image as a short sequence of interlocked 2-dimensional vectors whose lengths and directions are coded using variable length Huffman coding. The coding process is completed when a location is provided for the origin of the first vector associated with each character- This technique has the ability s to efficiently encode line graphics, as well as handwritten or typed text, while maintaining a constant level of distortion across the encoded page. This method is primarily geared towards the so-called store-and-forward operation, rather than real-time operations, or else the full compression efficiency benefits can not be realized .
9. Finally, a set of techniques known as "Maximum Differences" and "Analysis-by-Synthesis" have been proposed. See commonly-assigned ILI.S. Patent No. 5,293,251, which is incorporated herein by reference for all purposes. See also the articles by Spiros Dimolitsas and Franklin L. Corcoran, co-inventors of U.S.
Patent No. 5,293,251, entitled "Facsimile Compression forTransmission over ~00 bit/s to 2400 bit/s Digital Mobile Circuitsn, Conference Proceedings, IEEE Military Communications Conference, MILCOM '90, Monterey, CA, and "Compression of Facsimile Graphics for Transmission over Digital Mobile Satellite Channels", Conference Proceedings, IEEEMi/itary Communications Conference, MILCOM '91, McLean, VA, pp. 644-647 (November 4-7, 1991).
The present invention was motivated by a desire to overcome the perceived problems in the techniques enumerated immediately above. For example, while the techniques described in U.S. Patént No. 5,293,251 and the two MILCOM articles were suitable for the real-time network compression of facsimile images, it was not until the present invention was made that these techniques became usable, since CA 02206426 1997-0~-29 '. , f W O96/17469 PCTnUS9S11494S

these techniques not considered suitable for encoding either typed text or bi-level graphics which incorporated a significant amount of detail.

SUMMARY OF THE INVENTION
The present invention relates to method for reducing transmission require-5 ments related to facsimile images while maintaining high intelligibility in mobilecommunications environments.
According to one aspect of the invention, a low-cost facsimile interface unit (FIU) of the type disclosed in U.S. patent applications Serial Nos. 07/720,990 and 07/754,785, which is suitable for facsimile communication between low-power 10 mobile earth stations and fixed earth stations for both point-to-point and point-to-multipoint transmissions can be enhanced by improved and optimized coding and decoding method treating the intermediate form of the image being transmitted.
These method operates on small numbers of scan lines, thus, lending themselves to use in real-time operations over satellite channel rates ranging from 2400 bit/s 5 to 800 bit/s.
The method according to the invention are designed to emphasize the intelligibility retention of both typed-text and hand-written images. These methods offer hig~ reconstruction intelligibility and only little additional quality degradation when compared with standard resolution facsimile coding, which are limited to a compression ratio of approximately 12 to 1.
The basic idea behind the compression methods of the invention is the selective removal of pixel information from an intercepted document that is being transmitted from an intercepted document that is being transmitted using the CClrr CA 02206426 1997-0~-29 ~ r (~
W O96117469 PCT~US9S114945 Recommendation T.4 run-length coding (RLC)- Subsequently, an interm~diate sized image is generated using a selected two dimensional maximum differences algorithm and the resulting image is then encoded using optimized RLC prior to transmission over the digital channel.
These and other objects, features and advantages of the invention are disclosed in or will be apparent from the following description of preferred embodi-ments.

BRIEF DESCRIPTION OF THE DRAWINGS
The preferred embodiments are described with reference to the drawings in which like elements are denoted by like or similar numbers and in which:
Fig. 1 is an illustration displaying a portion of a facsimile image scan line which is useful in understanding run-length encodings;
Fig. 2 is a flow diagram of facslmile image data, which advantageously can be used according to the present invention, through each stage of the selected algorithm;
Fi~. 3 illustrates an example of a 2:1 vertical reduction using the so-called original maximum differences algorithm, wherein the shaded region forms an encoded scan line;
Fig. 4 illuSLraleS an example of two-dimensional Modified Maximum Differenc-es ~2-D MMD) algorithm (m = 2, 1 = 8) processing according to the present inven-tion, wherein shaded regions indicate selected segments of the original image;

CA 02206426 l997-0~-29 t ~ f wo 96/17469 Pcr/usss/1494s Fig. 5 illustrates an example of one-dimensional vertical 2:1 compression which advantageously employed in a first stage of a two-dimensional Asymmetrical Maximum Differences (2-D AMD) codec according to the present invention;
Fig. 6 illustrates an example of 3:2 vertical reduction which advantageousiy 5 can be employed as a first stage of the two-dirnensional Fractional Maximum Differences (2-D FMD) algorithm according to the present invention;
Fig. 7 illustrates the interpolation rules used in the original 4x4 decoder wherein Fig. 7(a) illustrates horizontal expansion, Fig. 7(b) illustrates vertical expansion and Fig. 7(c) illustrates exemplary interpolation rules under which 10 expansion is performed;
Fig. 8 illustrates two examples of pixel to 4x4 block expansions performed by the M-S decoder according to a preferred embodiment of the present invention;
Fig. 9 illustrates examples of 3x3 to 4x4 pixel block mappings in the T-D
decoder according to another embodiment of the present invention; and Fig. 10 illustrates encoder statistics generated during vertical compression using a two-dimensional fractional maximum differences encoder according to a preferred embodiment of the present invention.

DESCRIPTION OF THE PREFERRED EMBODIMENTS
Using the techniques disclosed here, the ability to achieve moderate coding efficiency (e.g., 2 to 4 times better than Group 3 facsimile, 1 -Dimensional RLC) and acceptable image quality retention has been demonstrated for both handwritten and typed text while at the same time it is suitable for real-time network operation. The sdvantages and distinctions between conventional compression and //or decompres-W O 96/17469 PCTAUS9S~14945 sion techniqueS and the preferred embodiments of the inventive method will best be understood by briefly describing the conventional techniques prior to presentdetailed descriptions of the preferred embodiments.
Most of the more popular facsimile machines and personal computer (PC) plug-in cards support the CClrr Recommendation T.4 standard for document transmission over Group 3 facsimile systems. The T.4 standard employs a one-dimensional lossless image compression scheme to reduce transmission time by a factor of 6 to 12, depending upon the contents of the document. Because of this compression, the time needed to transmit a single page over a 9600 bit/s line is0 about 15-30 seconds rather than 3 minutes.
The compression algorithm used by T.4 is a modified form of Huffman run-length coding. It is based on the observation that for any given scan line in a typical 1728x1100 (horizontal by vertical) standard mode facsimile image, there tend to be long "runs" of black and white pixels. These strings of pixels occur because the fax documents are made up of lines with an almost uniform thickness.i.e, strokes of the pen or characters in a font, that are made up of black runs and which are separated by white runs. In run-length coding, a scan Une is encoded as a series of numbers representing the lengths of these runs rather than as individ-ual pixels, thus resulting in significant compression gains. Fig. 1 contains an example run-length coding on a portion of a facsimile image scan line. The encoded run-lengths alternate between white and black runs across the scan line.
Another important characteristic of facsimile images is that some run-lengths occur more frequently than others. For example, one would expect a significantlylarger number of short black run-lengths than long black run-lengths because of the CA 02206426 1997-0~-29 , ~ r W O 96/17469 PCT~US95/14945 many thin lines, i.e., pen strokes or font, on the page. By comparison, not manyIong strings of black pixels occur in the ordinary handwritten or typed documents that tend to be transmitted by a fax machine. A sim~ilar effect can be observed for white run-lengths, which occur most frequently in-between black pen strokes or in association with blank image scan-lines.
Because of this property, a codin9 scheme that gives an equal number of bits to each run-length would have some statistical redundancy and would therefore besub-optimal. To improve the compression gain further, the T.4 standard use Huffman variable-length codes for run-lengths. Runs that have a high probabilityof occurring are assigned shorter code words, and runs that have a low probability of occurring are given longer code words. The result is a much more efficient coding scheme that achieves compression gains on the order of 6 to 12, dependingupon the entropy of the particular image.
The Maximum Differences facsimile image encoding algorithm, as disclosed in U.S. Patent No. 5,293,251, can compress a fax document by a significantly greater factor than the T.4 standard can, albeit at the expense of a distortionless transmission. At the encoding end, the 1 728x1 100 standard mode image is first pixel-decimated by a fixed amount both horizontally and vertically using the Maxi-mum Differences algorithm. Thus, the image is shrunk down in a manner so as to preserve as much of the detail as possible. It will be appreciated that this is where loss is introduced into the image in a controlled manner. Next, the bit-reduced image is passed through a lossless Huffman run-length encoder very similar to that used by the CCITT Rec. T.4 standard. The image decimation process has an effect on the run-length probability density functions (pdfs) that degrades the performance CA 02206426 1997-0~-29 W O 96/17469 PCTnUS9~114945 of the run-length coding, but the total compression, i-e-, decimation and run-length coding, is still significantly greater than the achieved by T.4 coding alone.
At the decoding end, an inverse Huffman run-length coding routine recovers the decimated image bitmap that was generated by the encoder by detecting the individual code words and converting them back to run-lengths. This lossless stage of the expansion process introduces no new distortion. Finally, the condensed image is expanded by a fixed amount both horizontally and vertically back in to a 1728x1100 standard mode facsimile image. The expansion routine attempts to interpolate the missing pixel information by taking into account the properties of 10 typical handwritten or typed facsimile pages. The reconstructed image is not identical to the ori~qinal, but maintains most of the essential details and is perfectly intelligible for the types of images transmiKed through facsimile equipment.
In order to realize the benefits obtained by employing image compression in facsimile transmissions, a device that allows end-user terminal equipment to transmit ~acsimile images as demodulated bits over low-rate digital links is required.
This device, a so-called Facsimile Interface Unit (FIU), must permit end-userterminal equipment to transmit facsimile images as demodulated bits over low-rate digital links.
With an FIU at each end of the link, two fax machines can communicate as if they were connected through an ordinary telephone line, even though the actual transfer of data takes place through the lower rate connection. With facsimile image compression employed, the FlUs can use even lower bit-rate links to transmit the same facsimile pages. It will be noted that this change is invisible to the CA 02206426 1997-0~-29 W 096/17469 ~iU~9~1494S

facsimile terminal equipment, while the actual amount of transmitted bits is reduced significantly.
Fig. 2 contains a block diagram of the different stages of facsimile compres-sion as implemented on an FIU platform which allows fax machines communicating at 4800 bit/s to transfer data over 2400 bit/s channels. In the example shown inFig. 2, it is assumed that the image is cornpressed by a factor of 2 using a encod-er/decoder pair which are parametrically designated as n 3/2x2" . At the transmitting FIU, 4800 bit/s of T.4-encoded data from the transmitting fax equipment is firstdecoded on a scan-line basis. It will be appreciated that when the entire irnage is processed, this corresponds to a full size image of 1728x1100 pixels. Then, a Maximum Differences type of image compression algorithm is applied which generates a smaller 864x732 image, once the entire image has been fully decode and decompressed. Next, the pixels are Huffman run-length encoded with an optimized set of code words, where the selected code words are based on the new reduced ima~e statistics. These compression steps yie~d a gain of 2:1 over the T.4-encoded data, allowing the data to be transmitted over a 2400 bit/s channel. It will be noted that at the receiving FIU, the Huffman run-length encoded data is converted back into the reduced 864x732 bitmap, an exact copy of the intermedi-ate image generated by the encoder. This intermediate image is then expanded back to its original size by an al~orithm that attempts to interpolate all of the needed pixel information, thereby resulting in the reconstructed image. Finally, the ~ull size 1 728x1 100 reconstructed bitmap is encoded with the T.4 compression scheme and transmitted at 4800 bit/s to the receiving fax equipment.

CA 02206426 1997-0~-29 c ~ r f W O 96117469 PCTnUS9~1494S
All of the above steps are performed in real-time to minimize the propagation delay of the transmission of scan lines.
During its development, the Maximum Differences class of encoding algo-rithms went through several distinct stages. To understand the motivations behind a number of aspects of its design requires a familiarity with the progress of the algorithm through each of the stages.
An early Maximum Differences coder disclosed by Spiros Dimolitsas and Franklin L. Corcoran to the other co-inventors of the present invention, and de-scribed in an unpublished article entitled nFacsimile Compression in Mobile Commu-~o nications," was geared more towards higher compression gains and less towardsretaining the reconstructed image's quality. This provided fixed compression factors of 4:1 along each axis for fine resolution facsimile documents (1728 x 2200 pixels).
This resulted in an effective compression ratio of approximately 3:1 compared to a standard resolution facsimile, but the image quality degradation resulting from this process rendered the coder suitable only for the compression of large-type handwrit-ten text documents.
To simplify the development of stimulation software, the Maximurn Differenc-es compression was only applied in the vertical direction for decimation by a factor of 2. Horizontal reduction by a factor of 2 was achieved by a simpler dithering pattern pixel-removing procedure. Reduction by another factor of 2 along each axis was achieved by an inclusive ORing process. Table 1 lists the four stages of this procedure.

~. CA 02206426 1997-0~-29 ~

W O 96/17469 PCTrUS9~/1494S

Stage 1 Vertical reduction by factor of 2 using Maximum Differences Stage 2 Horizontal reduction by factor of 2 using dither-ing pattern Stage 3 Vertical reduction by additional factor of 2 using inclusive OR
Stage 4 Horizontal reduction by additional factor of 2 using inclusive OR

Stage 1 uses the original Maximum Differences algorithm to reduce every pair of source scan lines to a single encoded scan line, achieving a vertical reduction of 2 to 1. Fig. 3 illustrates how this is accomplished. Starting from the top of the page, two scan lines and a reference line are divided into 8-pixel segments, i.e., 10 octets. For each of these segments, the pixels from one of the two scan lines are selected and placed on the reduced output line. The particular scan line that is chosen for each segment is the one with the greater number of bit-wise differences between the segment itself and the corresponding portion of the reference line.
If the number of pixels that differ between the reference line and each of the scan 15 lines is the same, then the scan line whose segment has the greater number of internal transitions is chosen. When the entire scan line has been processed, the reduced output line becomes the reference line that is used for the compression of the next two scan lines and the process repeats. Thus, each line of the reduced image contains pixels from a pair of source scan lines. In Fig. 3, the shaded segments are the ones that are to be piaced on the output line (i.e., octet 1-line, octet 2-line 2-line 2, octet 3-line 2).

CA 02206426 1997-0~-29 : - -W O 96/17469 PCTrUS95/1494~

It should be noted that the reference line always contains the output scan line from the reduction of the previous pair of lines- For the compression of the first two lines, the reference line can take on an initial value of all white pixels (since the page has a white background) or it can simply be a copy of the first scan line.
Stage 2 of the reduction process uses a dithering pattern to compress the image by a factor of 2 horizontally. In essence, the odd-numbered pixels are removed from the even-numbered scan lines, while the even-numbered pixels are removed from the odd-numbered scan lines.
During Stage 3 of the reduction process, a black pixel-preserving OR algo-rithm achieves another factor of 2 in vertical compression. For every pair of source scan lines, a single reduced output scan line is computed as the logical OR of the two source lines, where a white pixel is represented by a O and a black pixel isrepresented by a 1. In other words, if either the two source pixels in a given horizontal pixel position is black, the output scan line pixel at that position is also 15 black. Otherwise, the output pixel is white.
The fourth and final stage of compression is a lo~ical OR in the horizontal direction. For every two columns of pixels, a single reduced output column is computed as the logical OR of the two source columns. The output pixel for a column on a given line is black only if either of the pixels on the source columns of this line is black.
With the above background firmly in mind, the preferred embodiments accordin~ to the present will now be described. One such preferred embodiment is an improved version of the Maximum Differences algorithm, which is outlined in Table 2.

CA 02206426 1997-0~-29 .. ---W O96117469 PCTnUS9~1494S

Stage 1 Simultaneous horizontal and vertical reduction - by a factor of m using the Maximum Differences algorithm with pixel segments of size /
Stage 2 Vertical reduction by additional factor of n using inclusive OR
Stage 3 Horizontal reduction by additional factor of n usin~ inclusive OR

This modified version of the Maximum Differences encoder, hereinafter referred to as a 2-Dimensional Modified Maximum Differences (2-D MMD), imple-ments a new approach for coding images. It will be readily appreciated by compar-ing Tables 1 and 2 that the 2-D MMD algorithm has basic differences from the original that make it more flexible and allow it to be more selsctive about the o removal of pixels from the source image.
First of all, the compression factors are variable in each of the three stages employed by the 2-D MMD algorithm. Stage 1, the so-called Maximum Differences reduction stage, advantageously can decrease the size of an image by an integral factor of m along each axis. It uses the same algorithm as before, but compares m (rather than 2) pixel segments to the reference line in order to select one line segment. Preferably, the inclusive OR reduction stages. i.e., stages 2 and 3, logically combine every n rows and n columns into one.
Second, the 2-D MMD algorithm preferablyallows variable size pixel compari-son segments during performance of the Maximum Differences stage, i.e., stage 1. Advantageously, the segments do not necessary have to be 8 bits wide; any integral size / can be used. Finally, in the 2-D MMD algorithm, both horizontal and CA 02206426 l997-0~-29 t 1 r r wo 96/17469 ~cr/usssll4s4s vertical compression advantageously can be performed simultaneously during the Maximum Differences stage.
Fig. 4 is an exemplary illustration which is useful in understanding the two-dimensional image reduction method according to the present invention, which G reduces an image by a factor of 2 along each axis with a comparison segment size of 8. Both a reference line and a reference column are maintained, and the horizon-tal and vertical pixel segments that have the most differences from their respective reference segments are selected. These se~ments are indicated by the shaded regions in Fig. 4, each consisting of 8 contiguous pixels. The reduced image output o is constructed from the pixels at the intersections of the horizontal and vertical pixel segments. Thus, intersections of the shaded regions in Fig. 4 form the 4x4 reduced image shown in the bottom right corner.
As with the one-dimensional encoder described above, each of these seg-ments becomes the reference segment for the next comparison. It will be appreciat-5 ed that this ~eneralized Maximum Differences algorithm degenerates to the originalone-dimensionalverticallycOmPressin9technique when the simultaneous horizontal compression is skipped.
Experirnents with different comparison segment sizes indicate that the 2-D
MMD algorithm introduces less visual distortion when short segments are used.
20 It should be noted that the decision to choose one of two pixels should be based on the nearby or neighboring pixels; a pixel that is 8 positions away is not likely to have any useful information about what can be safely discarded.
From the above observations, it will be appreciated that it is desirable to use a small value for /, the comparison segment size in the 2-D MMD algorithm.

CA 02206426 1997-0~-29 r ,~
W O96/17469 PCTnUS9~1494~

Unfortunately, there is a restriction imposed on / such that I be greater than or equal to m, the compression factor, i-e-, 1 2 m- If / were less than m, a comparison segment could not fill a complete row or column of an m x m block and an intersec-tion between selected horizontal and vertical segments could not be guaranteed.
Thus, if a compression factor of 2 is selected, the smallest comparison size possible is Z.
- In order to achieve a smaller comparison size (/=1), the~constraint that the horizontal and vertical image reduction steps be performed simultaneous!y advanta-geously can be relaxed, resulting in an asymmetrical encoder. Table 3 lists the two 0 steps that are performed by this 2-Dimensional Asymmetrical Maximum Differences (2-D AMD) encoder. It should be noted that the 2-D AMD represents a new approach for coding images and is considered to be another preferred embodiment according to the present invention. First, a vertical compression by a factor of my is achieved by using the original one-dimensional encoder with a comparison size of 1. With such a low value for /, each pixel segment contains just one pixel, and the criteria for selecting a pixel segment degenerates into just one, i.e., if any of the pixels differs from the reference pixel in color, then that pixel is used in the reduced image; otherwise, the reference pixel is used. The original rule concerning bit transitions no longer applies since there is only one pixel now. As before, the selected pixels become the new reference line for the next comparison. Finally, a horizontal reduction by a factor of my is achieved in an identical manner, but along the other axis, for a total reduction of m" x my. Fig. 5 illustrates the one-dimension-al 2:1 vertical compression technique, the first stage of the 2-D AMD encoder.

CA 02206426 1997-0~-29 W O 96/17469 PCT~US9S11494S

Stage 1 Vertical reduction by a factor of my using one-dimensional Maximum Differences algorithm with a comparison pixel segment size of 1 Stage 2 Horizontal reduction by a factor of mx using one-dimensional Maximum Differences algorithm with a comparison pixel segment size of 1 The 2-D AMD algorithm compromises symmetry for a small comparison pixelsegment size, but experiments have shown that the lack of symmetry during the reduction process does not introduce any noticeable distortion into the reconstruct-ed image. It should also be noted that the horizontal compression is intentionally performed last because typical handwritten and typed documents tend to have moreredundancy in the vertical direction than in the horizontal direction. In short, there 10 is a significant amount of vertical strokes in most letters, so the horizontal reduc-tion stage can be more sensitive to pixel-removal decisions when the compressionfactor used is high enough to compensate for its extra resolution in standard mode.
It will be appreciated that there are two other minor differences between the 2-D AMD algorithm and the 2-D MMD algorithms. First, according to the 2-D AMD
algorithm, the reduction factors for the horizontal and vertical stages are no longer constrained to be the same. This extra parameter advantageously was added to - accommodate standard mode facsimile images, which do not have equal horizontaland vertical resolutions. The other difference is that the essentially "brute-force"
inclusive OR stages in the 2-D MMD algorithm have been omitted from the 2-D
20 AMD algorithrn. The demand for higher quality reconstructed images led to a CA 02206426 1997-0~-29 ~ f' ' W O96/17469 PCTnUS95114945 decrease in the amount of image reduction possible, making only one reduction level necessary.
The 2-D AMD encoder maintains important details when compressing facsimite documents. However, it is only capable of scaling down images by 6 integral compression factors. To permit non-integer compression ratios to be realized, the 2-D AMD algorithm was further modified to allow the use of fractional compression factors, thus giving rise to the 2-Dimensional Fractional Maximum Differences (2-D FMD) algorithm, which represents a third preferred embodiments according to the present invention. The 2-D FMD algorithm is also asymmetrical o and performs the two compression stages in the same order, vertical compression followed by horizontal compression. It will be readily appreciated that the 2-D FMD
can be considered as the convolution of 1-D FMD and its transpose. The only high-level difference is that the new "fractional" reduction algorithm compresses every mvn lines down to mvd and every mxn columns down to mxd, rather than my lines down to 1 and mx columns down to 1, as illustrated in Table 4. This advanta-geously allows fractional reductions, such as 3/2:1, 4/3:1 and 5/3:1, along each axis.
TABLE ~

Stage 1 Vertical reduction by a factor of mvd/mvd using one-dimensional fractional Maximum Differences algorithm with a comparison pixel segment size of 1 Stage 2 Horizontal reduction by a factor of mxnlmxd using one-dimensional fractional Maximum Differences algorithm with a comparison pixel segment size of 1 C _ ,_ W O 96/17469 PCTnDS9S/1494S

The new 2-D FMD algorithm reduces an image by detectin9 contiguous pixels and removing them individually when possible- The rules that dictate which pixels are to be eliminated a~vantageously can be developed heuristically and are based on the rules employed by the 2-D AMD algorithm discussed in greater detail above.
As before, the horizontal reduction rules are identical to the vertical rules, but their respective compression factors advantageously can differ. A reference line or column must still be maintained, and is always taken from the bottommost line or right-most column of the previous line/column reduction's output. During vertical compression, every column of an myn line portion of the page forms an myn pixel o segment that must be replaced with a shorter myd pixel segment. Similarly, for horizontal compression, every mxn pixel horizontal segment in an mxn column region must be replaced with a shorter mxd pixel segment.
When reducing mn pixels down tO md (along either axis), there are two types of redundancies that can be exploited to eliminate a pixel. First, there is the likely possibility that two adjacent pixels in the segment are of the same color. When this situation is detected, the encoder throws away one of the two pixels. Second, the color of the first pixel in the segment (upper or leftmost) is compared with the color of the corresponding reference pixel. If they are the same, then the first pixel in the segment is eliminated. In either of these reduction strategies, when a pixel iS removed, another pixel of the same color always remains in the adjacent position.
Because of this, it is unlikely that a significant amount of visual distortion will be introduced, since the human eye cannot easily detect the displacement of anything by just one pixel position.

CA 02206426 l997-0~-29 WO 96/17469 ~ r~,llU;,~5/1494S

If no redundancy is detected in the pixel segment, the encoder is forced to adopt a third rule and remove pixels from the end of the segment until it is down to the correct size. Unfortunately, this action is likely to introduce noticeable visual distortion. However, this situation only arises when documents with tiny fonts or small details are encoded- It will be appreciated that these images are not suitable for reliable facsimile transmission anyway.
Fig. 6 is an exemplary case illustrating the operation of one of the compo-nents of the 2-D FMD algorithm, i.e., a one-dimensional 3:2 vertical reduction (1-D
FMD) stage. The diagram is useful in understanding how 3 scan lines can be 0 replaced by 2 scan lines by removing unnecessary redundancy. The first five co!umns are examples of the first pixel-elimlrlatiorl rule, discussed immediately above. The sixth and seventh coiumns demonstrate the second rule, also discussedimmediately above. Finally, the last column is an example where neither rule could be applied and the se~ment had to be truncated.
It will be appreciated that one feature of the 2-D FMD algorithm is that it is a superset of the 2-D AMD encoder. That is, if the reduction ratio happens to reduce to an inteyer (has a 1 in the denominator), the 2-D FMD encoder gives theexact same results as would be obtained using the 2-D AMD encoder. By eliminat-ing adjacent pixels of the same color, the single pixel that is left behind is always zo the one that is "most different" from the reference pixel.
For typed documents with small fonts, a noticeable form of distortion has appeared on the encoded images along the upper and lower edges of many charac-ters. The "jaggedness" appears to be the result of the poor vertical resolution of the standard resolution Group 3 facsimile image. The Maximum Differences family CA 02206426 1997-0~-29 W O 96117469 PCTnUS9S/14945 of algorithms, i.e., 1-D MD, 2-D MMD, 2-D AMD and 2-D FMD, were designed working under the assumption that the displacement of a pixel by one position is negli~ible, i.e., unnoticed by the human eye- However, for standard mode images, this does not seem to be the case with respect to the vertical direction. Often, a single black pixel on the edge of a typed charactèr is moved up or down by one pixel with respect to its neighbors to preserve details inside the character. Unfortu-nately, this pixel then sticks out enou~h to make the letter appear unpleasantly jagged .
The problem appears to be related to the priorities used for the selection of 10 a pixel-removal strategy. Previously, two forms of redundancy were described that can be eliminated when reducing a pixel segment of size m~, to one of size m~,d.
The compression algorithm gave the highest priority to the removal of the first type of redundancy, adjacent pixels of the same color. If that condition was not present, the second pixel-removing strategy was attempted. The undesirabie distortion described occurs frequently in places where different pixel-removal rules are applied to adjacent columns.
To help alleviate this problem, the pixel-removal priority system for vertical compression advantageously can be modified. When reducing a segment, the encoder attempts to use the same scheme that was used for the compression of the segment immediately to its left. If this is not possible, then it attempts the other pixel-removal strate~y. With this new set of priorities, the chances of a particular column being bit-reduced by a different scheme than its neighbors are greatly reduced . This technique is only employed during vertical reduction; standard CA 02206426 1997-0~-29 W O 96/17469 PCT~US9S/1494S

mode facsimile images have a high enough horizontal resolution that it is unneces-sary during horizontal compression.
At ~he decodiny end of the communications channel, the bit-reduced image is expanded back to the original size. The expansion (or decoding) algorithm must interpolate all of the missing pixel information so that the reconstructed image has smooth, well-formed typed characters and pen strokes. Several differenttechniques according to addition preferred embodiments according to the present invention advanta~eouslY can be employed during decoding. These techniques are described in greater detail immediately below.
,0 All of the decoding algorithms according to the present invention work on the same basic principle. To uniformly expand the facsimile page, each pixel (or ~roup of pixels) must be replaced by a larger block of pixels. The size of this block depends upon the particular compression factors that were used in the encoder side of the transmission channel. For example, a 2x2 compressed image can be expand-ed by replacing each pixel with a square consisting of four pixels. For the decoder to produce a reconstructed image of ~ood quality, the contents of this larger block must be coherent with respect to the surrounding pixels. That is, the decoder must attempt to interpolate and fill in this block with pixel patterns that "connect" well with those in the surrounding blocks. In orderto do this, the expansion algorithm(s) advantageously use information not only from the pixel or group of pixels being expanded, but also from adjacent, i.e., neighboring, pixels in the reduced image.
The expansion algorithm, which was previously disclosed in U.S. Patent No.
5,293,251, was designed to decode images that were reduced by a factor of 4 along each axis by the original Maximum Differences encoder, i.e., algorithm, ~ CA 02206426 1997-0~-29 ' ' f W O 96/17469 P~ sll494s described earlier. For example, the 1-D MD decoder uses a simple set of interpola-tion rules to convert a 432X550 reduced image back into a 1 728x2200 fine mode facsimile document.
The routine employed by the 1-D MD decoder consists of an asymmetrical algorithm that performs the horizontal and vertical expansions in two separate stages, as listed in Table 5. The image is first augmented in the horizontal direction by a procedure that replaces each pixel with a straight segment of four pixels in a row. The decoder determines the contents of this segment by examining the pixel being expanded and six of its neighboring pixels and by then employing a set of interpolation rules designed to smoothly connect them. Subsequently, the vertical expansion is performed using the exact same procedure, but transposed. That is, each pixel in the horizontally expanded image is replaced by a column of four pixels whose contents are determined by the same set of interpolation rules. The sur-rounding pixels that are examined when interpolating are also the same as those ~5 from the horizontal expansion, but rotated by 90 degrees.

Stage 1 Horizontal expansion by a factor of 4 using one-dimensional decoder based on interpolation rules Stage 2 Vertical expansion by a factor of 4 using one-dimensional decoder based on interpolation rules The interpolation rules, which were disclosed in U.S. Patent No. 5,293,251, are illustrated in detail in Fig. 7. In Fig. 7(a), the symbols aO through a~ represent the pixel ~o be expanded and its nei~hbors in the encoded image, and b 1 through b4 represent the expanded segment replacing it. If the source pixel aO is white, CA 02206426 1997-0~-29 ( W O96/17469 PCTrUS9S11494S

then it is always expanded to a string of white pixels. e.g., b1 = b2 = b3 = b4 = white. If aO is black but is surrounded by white or black pixels on either side, a2 = a5 = white OR a2 = a5 = black, then aO is expanded to a string of black pixels, i.e., b 1 = b2 = b3 = b4 = black. Finally, if aO is black but only one of its non-diagonal neighbors is black, e.g., a2 and a~ are of a different color, then aO is expanded to a string of black pixels only if one of the opposite corners contains a black pixel, i.e., EITHER a2 = black and a4 or a6 = black OR a5 =
~ black and a1 or a3 = black). Otherwise, the decoded segment replacing aO
contains two contiguous black pixels on the side with the black neighbor and two10 contiguous white pixels on the side with the white neighbor. These instructions are followed during both stages of decompression: the aO through a6 and bO
through b4 used for horizontal decoding are simply transposed for vertical decoding .
It should be noted that because this algorithm is asymmetrical, the order in which each stage of expansion is performed has a significant effect on the results obtained. For example, as a consequence of the fact that vertical expansion is performed last, strings of pixels occasionally form vertical lines one pixel wide on the insides of letters with small openings (such as a lowercase 'e'). Had the order of expansion been reversed, this same effect would have been observed with horizontal strings of pixels.
In orderto resolve any difficulties that arise when expansions along each axis do not impact the reconstructed image in the same manner, a new symmetrical version of the 4x4 decoder, i.e., decoding algorithm, advantageously can be provided as yet another preferred embodiment according to the present invention.The modified-symmetrical (M-S) decoder performs both horizontal and vertical r W O96/17469 PCTrUS95/1494S

expansions simultaneouslY by replacing each pixel in the source image by a 4x4 pixel block, resulting in a factor of 4 expansion along each axis. The particular rules that would be applied during the horizontal and vertical expansion stages of the asymmetrical decoder are combined to produce this decoded block. The examples in Fig. 8 illustrate exactly how this is done. The 4~pixel segments prodluced from applying the interpolation rules alon~ each axis are replicated 4 times and arranyed into two 4x4 blocks. Then, these two blocks are combined into one block by a logical AND operation which advantageously preserves white pixels, resulting in the decoded 4x4 block. It will be appreciated that this technique is capable of interpo-~o lating for missing pixels during image expansion in exactly the same manner alongeach axis.
The development of the interpolation rules shown in Fig. 7 was achieved on a heuristic basis regardin9 what pixels patterns would best fill in the missing space in an encoded document in each of the possible surroundin~ pixel scenarios. They were then refined following a number of experiments using typical handwritten documents. Thus, althou~h the performance of the 4x4 decoder with interpolation rules has demonstrated their soundness, they may not be the optimal set of rules.
One limitation of these decoding algorithms is that they are incapable of constructing expanded images with pen strokes that are as smooth as in the original. This is because the interpolation rules employed never allow any entities smaller than a two-pixel segment in the asymmetrical decoder or a 2x2 pixel block in symmetrical decoder to appear in the reconstructed image. In effect, this lowers the resolution of the expanded document by a factor of 2 in one dimension for the asymmetrical decoder and in both dimensions for the sy~ nel~ical decoder. As a CA 02206426 l997-0~-29~
,_ WO 96/17469 PCr/US9S/14945 result, decoded images may appear coarse and exhibit jagged lines. This distortion does not affect the intelli~ibility of the document, but is noticeable.
In order to improve the quality of the reconstructed documents produced by the decoder in the communications channel, it is necessary to introduce a ~reater number of possible pixel sequences. This would ensure enou~h diversity in the image to allow finer detail and thus smoother interpolation. Rather than generating a longer and more sophisticated set of interpolation rules to accomplish this, a more general approach has been developed. The symmetrical 4x4 decoder was modified to use a lookup table rather than a set of rules to generate the pixel segments in .0 the decoded image. That way, every pixel combination in the encoded image can map to its own reconstructed sequence, resulting in the best possible use of theavailable information.
The Table-Driven (T-D) decoder achieves a factor of 4 expansion along each axis by replacing every pixel in the encoded image by a 4x4 block of pixels. Thecontents of the source pixel and its eight adjacent neighboring pixels form a 3x3 block, which is the index of the lookup table. For each index, the table contains a 4x4 block that is to be placed in the reconstructed document. Thus, the lookuptable is a basically 3x3 to 4x4 pixel block mapping. It contains 28 = 256 entries of 16 bits each for a total size of 512 bytes.
The T-D decoding approach is general because any set of interpolation rules can be implemented with it simply by modifying the table, including those of thesymmetrical 4x4 decoder. However, because of the much larger number of possible 4x4 blooks capable of being generated, there is an opportunity for much more diversity and detail in the reconstructed image than with the interpolation rule-based ,_ ,~

W O96117469 PCTnUS~S114945 methods. The mappin~s that are actually used by this decoder are a set of ad-hoc associations derived heuristically- They are designed to connect lines and curve smoothly so that the expanded ima~e is of good quality. Fig. 9 contains some examples of the mappin9s in this table. Although there are 256 entries, only 104 actually needed to be constructed from scralcl~: the remainin~ were generated by taking symmetry into account in the 3x3 and 4x4 blocks and applyin9 each of the 104 mappings to 1, 2, 4, or 8 entries in the table when possible.
The quality of the reconstructed images produced by the 4x4 lookup table decoder is clearly better than that of the interpolation rule-based decoders. Howev-10 er, because the table was generated on an ad hoc basis rather than by training thereis scope for further adaptive optimization of the T-D decoder based upon the statistics of the documents being encoded.
Results from the interpolatiQn-based rule and table-driven decoders have shown that when expandin~ ima~es, the contents of the pixels that are adjacent to the pixel being expanded are very useful in deciding what sequence to expand it to. However, neither of the 4x4 decoders make the best possible use of this information; both employ a "best guess" strategy based upon intuition and common sense rather that the actual statistical properties of the encoded documents.
For the 2-D FMD encoder described above, which reduces a sequence of mn ' pixels down to md pixels, it was observed that any given sequence of md pixels could only have been arrived at by a limited number of mn pixel sequence combina-tions. In addition, the particular combination that let to it seemed to be heavily related to the contents of the neighboring pixels to that segment. These observa-CA 02206426 l997-0~-29 r WO 96/17469 PCTIUS9~/1494S

tions were the motivation for the development of a new ~lali~lical decoding technique.
Fig. 10 is a compilation of contains some statistics on the 3:2 vertical compression of a set of typical handwritten and typed documents by the 2-D FMD
encoder discussed in greater detail above. Fi~. 10 lists the number of times each s 2-pixel output se~ment possibility occurred for each 3-pixel input cornbination. To determine the statistical significance of this mapping on the encoded pixel sequen-ce's neighboring pixels, each output segment is split into four parts which are tallied separately, each representing a possible combination of the two neighboring pixels on either end of the output segment. The two output pixels and their neighbors 10 thus form a straight line in the image; the 2-pixel output segment is in parentheses in the table.
Fig. 10 shows a clear correlation between the particular encoding decision that was made and the 2 neighboring pixels in the encoded image. A notable example is the output pixel sequence BB (two adjacent black pixels). When it is surrounded by black pixels on each side, there is a 100% chance that the corre-sponding pixel sequence in the original document was BBB. When it is surrounded by two white pixels, on the other hand, here is roughly a 50% chance that the original pixel sequence was BBB and a 50% chance that it was WBB. Thus, the two outer surrounding pixels in the encoded image contain valuable information regarding how the segment was encoded.
A slali~lical decoder according to yet another preferred embodiment of the present invention advantageously makes optimal use of this information and at the same time permits simpler implementations to be realized. The asymmetrical algorithm expands an image by fractional amounts first horizontally and tlien CA 02206426 1997-0~-29 WO 96117469 PCI~/US9S11494S

vertically, essentially "undoing" the steps performed by the 2-D FMD encoder. That is, each m"d pixel horizontal segment is replaced by a larger m,~n-pixel segment, and then each m~pixel vertical segment is replaced by a larger my~pixel segment, In order to do this, two tables are first generated from the horizontal and vertical s encoder statistics which contain the optimal m~pixel sequence; based upon its two outer neighbors in the encoded image- The particular mn-pixel segments selected for the table are those that would result in the fewest number of pixel errors in the reconstructed document, based upon the number of occurrences of each encoder mapping. These tables are then used by the decoder when expanding each pixel o segment.
One property of the human visual system that can be exploited when generat-ing a decoder table is the fact that the eye is far more sensitive to pixel errors where a pixel that was black in the original image is incorrectly decoded to a white pixel (a white pixel error) than to errors where a white pixel is decoded to a black pixel (a black pixel error). This makes intuitive sense because white pixel errors leave gaps in the lines that construct handwritten letters or typed characters and make them harder to recognize. Black pixel errors, on the other hand, merely make the lines thicker and distort their shape somewhat, but do little to~degrade an image's readability.
To ~ake this property into account, the decoder table generator was generat-ed by computiny the perceived error of a reconstructed document as a weighted sum of the number of black pixel errors and the number of white pixel errors. The coefficients that were chosen preferably are based upon the assumption that a white pixel error introduces twice as much distortion to the reconstructed image CA 02206426 1997-0~-29 - -..
W O 96117469 PCTnUS9S11494S

as a black pixel error does, resulting in a slight bias towards black pixels errors.
It should be noted that other weighing factors are possible and, indeed, may be more suitable. The algorithm then generates a decoder table that would minimize this weighted sum in the expanded ima~e. Table 6 contains example mappings for five training documents.

B(B~)B BBB B(BW)B BWW
B(B~)W BBB B(BW)W BBW
W(BB)B BBB W(BW)B BBW
o W(BB)W BBB W(BW)W BBW
B(WB)B WWB B(WW)B WWW
B(WB)W WBB B(WW)W WWW
W(WB)B WWB W(WW)B WWW
W(WB)W WBB W(WW)W WWW

Once an image has been bit-reduced in the manner defined in greater detail above, the resulting horizontal run-lengths are encoded using a variable length Huffman code. This codebook used for this process needs to be optimized to exploit the bit-reduced image run-length statistics (i.e., the probability density function?. It should be noted that the re-optimized Huffman code does not achieve the level of compression that Huffman coding realizes over uncompressed images.
However, the combination of bit-reduction and re-optimized Huffman coding achieves a total compression ratio that exceeds that offered by standard T.4 encoding. The methods for generating, encoding and decoding Huffman-based codes are described adequately in the publication of Richard W. Hamming, entitled CA 02206426 1997-0~-29 - - - -W O 96/17469 PCTnUS9S/1494S

"Coding of Information Theory", Prentice-Hall, Englewood Cliffs, NJ 07632, (1980), which is incorporated by reference herein for all purposes.
It will be appreciated that the present invention is not limited to the above- -described encoding and decoding methods. As shown in Fig. 2, a communications channel includes an inverse coder 10, a 2-D maximum differences encoder 20, an optimizer Huffman run-length coder 30, a transmission link 35, an inverse-optimized Huffman run-length decoder 40, a 2-D maximum differences decoder 50 and a conventional T4 Huffman run-length coder 60. The encoder 20 can be selected from the 2-D MMD, the 2-D AMD and the 2-D FMD encoders. The decoder 50 10 advantageously may be selected from the M-S decoder or the T-D decoder, which are described in greater above.
Other modifications and variations to the invention will be apparent to those skilled in the art from the foregoin~ disclosure and teachin~s. Thus, while oniy certain embodiments of the invention have been specifically described herein, it will be apparent that numerous modifications may be made thereto without departing from the spirit and scope of the invention.

Claims (17)

WHAT IS CLAIMED IS:
1. In a communications channel receiving an original image of a first size and producing a reconstructed image of the first size, a method for compressing the original image into an intermediate sized image, said method comprising the steps of:
(a) simultaneously horizontally and vertically reducing the original image by a factor of m responsive to maximum differences between rows and columns and respective references with pixel segments of size /to thereby generate a first stage compressed image;
(b) vertically reducing said first stage compressed image by additional factor of n using inclusive ORing so as to combine n of said row into one row thereby generating a second stage compressed image; and (c) horizontally reducing said second stage compressed image by an additional factor of n using inclusive ORing so as to combine n of said columns into one column thereby generating the intermediate sized image, wherein l, m, and n are all integers greater than or equal to 1.
2. The method as recited in claim 1, wherein said first stage compressed image is formed by pixels occupying intersection points between said row and said columns.
3. The method of claim 1, wherein / ~ m.
4. In a communications channel receiving an original image of a first size and producing a reconstructed image of the first size, a method for compressing the original image into an asymmetrical intermediate sized image, said method comprising the steps of:
(a) vertically reducing the original image by a factor of my responsive to maximum differences between rows and a respective first reference with pixel segments of size 1 to thereby generate a first stage compressed image;
(b) horizontally reducing said first stage compressed image by an additional factor of mx responsive to maximum differences between columns and a respective second reference with pixel segments of size 1 thereby generating the asymmetrical intermediate sized image, wherein m, x and y are all integers greater than or equal to 1.
5. The method as recited in claim 4, wherein said my factor is not equal to said mx factor and wherein horizontal and vertical resolution of said original image are unequal.
6. The method as recited in claim 4, wherein said asymmetrical intermediate sized image is a fractional intermediate sized image, wherein said step (a) comprises the step of vertically reducing the original image by a factor of myn/myd responsive to maximum differences between said rows and said first reference with pixel segments of size 1 to thereby generate a first stage compressed image, and wherein said step (b) comprises the step of horizontally reducing said first stage compressed image by an additional factor of mxn/mxd responsive to maximum differences between said columns and said second reference with pixel segments of size 1 thereby generating the fractional intermediate sized image, wherein d, m, n, x and y are all integers greater than or equal to 1.
7. The method as recited in claim 6, wherein a pixel removal criteria applied to a present one of said segments of size 1 preferentially a respective pixel removal criteria used in pixel replacement of an adjacent one of said segments of size 1.
8. The method as recited in claim 4, wherein a pixel removal criteria applied to a present one of said segments of size 1 preferentially a respective pixel removal criteria used in pixel replacement of an adjacent one of said segments of size 1.
9. In a communications channel receiving an original image of a first size and producing a reconstructed image of the first size, a method for expanding an intermediate sized image into the reconstructed image, said method comprising the steps of:
(a) simultaneously horizontally and vertically expanding the intermediate sized image by a factor of m to thereby generate a plurality of first and second sets of m x m pixels; and (b) logically ANDing respective ones of said first and second sets to thereby generate a resultant set of m x m pixels to thereby generate said reconstructed image, wherein m is an integer greater than or equal to 1.
10. The method of claim 9, wherein said step (a) is performed using a predetermined set of interpolation rules as follows:
if a0 = white, then b1 = b2 = b3 = b4 = white if a0=black, then if a2=a5, then b1=b2=b3=b4=black .
if a2~a5, then if a2 = black, then if (a4 = black) OR (a6 = black), then b1=b2=b3=b4=black otherwise, b1=b2=black, b3=b4=white if a5=black, then if (a1=black) OR (a3=black), then b1=b2=b3=b4=black otherwise, b1=b2=white, b3=b4=black / ~ m.
where a0 is the pixel of interest, a1-a6 are pixels adjacent to pixel a0 and b1-b4 are resultant pixels replacing pixel a0.
11. In a communications channel receiving an original image of a first size and producing a reconstructed image of the first size, the communications channel having a lookup table for storing m x n expansion data, a method for expanding an intermediate sized image into the reconstructed image, said method comprising the steps of:
(a) supplying an address identifying a pixel of interest and a predetermined number of pixels adjacent thereto to the lookup table; and (b) producing an addressed one of said m x n expansion data so as to generate said reconstructed image, where m and n are integers greater than or equal to 1.
12. In a communications channel receiving an original image of a first size and producing a reconstructed image of the first size, the communications channel having a lookup table for storing first and second expansion data, a method for expanding an intermediate sized image into the reconstructed image, said method comprising the steps of:
(a) expanding each mxd pixel horizontal segment in said intermediate sized image by replacing said mxd pixel horizontal segment by a larger mxn-pixel segment included in the first expansion data to thereby produce a first stage expanded image; and (b) expanding each myd-pixel vertical segment is said first stage replacing said myd-pixel vertical segment by a larger myd-pixel segment included in the second expansion data to thereby produce the reconstructed image.
13. The method as recited in claim 12, wherein the first and the second expansion data are based on statistical probabilities that m pixels and pixels adjacent thereto would produce n pixels and said pixels adjacent thereto during a corresponding fractional maximum differences encoding operation.
14. The method as recited in claim 13, wherein the first and the second expansion data wherein errors in the reconstructed image are weighted towards black pixel errors.
15. A communications channel comprising:
first means for recovering an original image from a Group 3 facsimile signal;
second means for encoding said original image using maximum differences between row and column segments and corresponding segments of respective reference pixel strings to thereby produce an intermediate compressed image;
third means receiving said first intermediate compressed image for performing optimized run-length coding to thereby provide transmission data;
fourth means receiving said transmission data for performing optimized run-length decoding thereon so as to produce a second intermediate compressed image:
fifth means receiving said second intermediate compressed image for simultaneously horizontally and vertically expanding said second intermediate compressed image by a factor of m to thereby generate a plurality of first and second sets of m x m pixels and for logically ANDing respective ones of said first and second sets to thereby generate a resultant set of m x m pixels, where m is an integer greater than or equal to 1, so as to produce a recovered image; and sixth means receiving said recovered image for generating a signal corresponding to said Group 3 facsimile signal.
16. A communications channel comprising:
first means for recovering an original image from a Group 3 facsimile signal;
second means for vertically reducing said original image by a factor of my responsive to maximum differences between rows and a respective first reference with pixel segments of size 1 to thereby generate a first stage compressed image and for horizontally reducing said first stage compressed image by an additional factor of mx responsive to maximum differences between columns and a respective second reference with pixel segments of size 1, where m, x and y are all integers greater than or equal to 1, to thereby produce an asymmetrical intermediate compressed image;
third means receiving said asymmetrical intermediate compressed image for performing optimized run-length coding to thereby provide transmission data;
fourth means receiving said transmission data for performing optimized run-length decoding thereon so as to produce a second intermediate compressed image;
fifth means receiving said second intermediate compressed image for entering a look up table using an address formed by pixels of interest and surrounding pixels and for producing a pixel block larger than said pixels of interest so as to produce a recovered image; and sixth means receiving said recovered image for generating a signal corresponding to said Group 3 facsimile signal.
17. The communication channel as recited in claim 16, wherein said asymmetrical intermediate compressed image is a fractional intermediate compressed image, wherein said second means comprises:
seventh means for vertically reducing the original image by a factor of myn/myd responsive to maximum differences between said rows and said first reference with pixel segments of size 1 to thereby generate a first stage compressed image; and eighth means for horizontally reducing said first stage compressed image by an additional factor of mxn/mxd responsive to maximum differences between said columns and said second reference with pixel segments of size 1 thereby generating said fractional intermediate compressed image, wherein d, m, n, x and y are all integers greater than or equal to 1.
CA 2206426 1994-12-02 1995-12-01 Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor Expired - Lifetime CA2206426C (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US08/352,836 1994-12-02
US08/352,836 US6181825B1 (en) 1994-12-02 1994-12-02 Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor
PCT/US1995/014945 WO1996017469A1 (en) 1994-12-02 1995-12-01 Methods performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus

Publications (2)

Publication Number Publication Date
CA2206426A1 CA2206426A1 (en) 1996-06-06
CA2206426C true CA2206426C (en) 2000-08-22

Family

ID=29552573

Family Applications (1)

Application Number Title Priority Date Filing Date
CA 2206426 Expired - Lifetime CA2206426C (en) 1994-12-02 1995-12-01 Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor

Country Status (1)

Country Link
CA (1) CA2206426C (en)

Also Published As

Publication number Publication date
CA2206426A1 (en) 1996-06-06

Similar Documents

Publication Publication Date Title
US6008847A (en) Temporal compression and decompression for video
US5363219A (en) Image processing method and apparatus
EP0392701A2 (en) Image reduction apparatus
Ono et al. JBIG2-the ultimate bi-level image coding standard
US6181825B1 (en) Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor
JPS60153264A (en) Transmission system of half tone picture
JPH03165181A (en) High quality compression method of binary text picture
US5271072A (en) Image reduction of binary images using filtering processing
US5729625A (en) Image processing method and apparatus which expand a pixel into multiple pixels, with a change in the number of gray levels
US7710605B2 (en) Print system and printer
US4558370A (en) Image processing method for graphics images
US6728412B1 (en) Method and apparatus for on-the-fly image coding
JPH0197064A (en) Processing device and method of pel signol of original inage
US5577134A (en) Method and apparatus for encoding a segmented image without loss of information
JPH08125868A (en) Method and device for processing picture
US8023756B2 (en) Image encoding apparatus and method of controlling same
US5293251A (en) Encoding/decoding system with two-stages of encoding/decoding
CA2206426C (en) Methods for performing 2-dimensional maximum differences coding and decoding during real-time facsimile image compression and apparatus therefor
JPH0789618B2 (en) Image coding method
KR100275690B1 (en) Image processing apparatus using mask
Kafashan et al. New rectangular partitioning methods for lossless binary image compression
Sathappan et al. Block based prediction with Modified Hierarchical Prediction image coding scheme for Lossless color image compression
JP4919577B2 (en) One-dimensional compression technology
JP3459723B2 (en) Image processing apparatus and method
US6272256B1 (en) Fast compression of periodic halftoned bitonal images

Legal Events

Date Code Title Description
EEER Examination request
MKEX Expiry

Effective date: 20151201