EP0411076A1 - Systeme d'imagerie video - Google Patents

Systeme d'imagerie video

Info

Publication number
EP0411076A1
EP0411076A1 EP19900902024 EP90902024A EP0411076A1 EP 0411076 A1 EP0411076 A1 EP 0411076A1 EP 19900902024 EP19900902024 EP 19900902024 EP 90902024 A EP90902024 A EP 90902024A EP 0411076 A1 EP0411076 A1 EP 0411076A1
Authority
EP
European Patent Office
Prior art keywords
image
representation
pin
reduced
transmission system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP19900902024
Other languages
German (de)
English (en)
Other versions
EP0411076A4 (en
Inventor
Yehoshua Y. Zeevi
Oliver Hilsenrath
Ran Ginosar
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
ISIGHT, INC.
Original Assignee
iSight Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from IL88969A external-priority patent/IL88969A0/xx
Priority claimed from IL89065A external-priority patent/IL89065A0/xx
Application filed by iSight Inc filed Critical iSight Inc
Publication of EP0411076A1 publication Critical patent/EP0411076A1/fr
Publication of EP0411076A4 publication Critical patent/EP0411076A4/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/59Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/136Incoming video signal characteristics or properties
    • H04N19/14Coding unit complexity, e.g. amount of activity or edge presence estimation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/42Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by implementation details or hardware specially adapted for video compression or decompression, e.g. dedicated software implementation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression

Definitions

  • the present invention relates to video imaging systems generally and more particularly to video imaging systems which are image content responsive.
  • Video imaging systems typically incorporate video cameras with image transmission, processing, storage, and display subsystems. These systems, some of which are described in Computer Vision, by Ballard and Brown, are designed to enhance images and to analyze them such that a computer, or other processing device, can perform other operations using information found in the image.
  • the lowest resolution decimated image is transmitted first followed by the higher resolution images.
  • the image is reconstructed in stages where initially the lowest resolution image is displayed and subsequently, as pixels from the higher resolution images arrive, they are integrated into the image.
  • Such a pyramidal scheme is known as a Progressive Resolution scheme and it has also been applied to storage and retrieval systems.
  • Bit reduction in the transmission of moving images is achieved by only transmitting the changes which occur from frame to frame. To do so, a first image is transmitted. Subsequently, successive frames are compared and any differences are extracted and transmitted. The receiving end maintains image storage and the received differences are added to the storage to produce an updated image.
  • U.S. Patent 4,716,462 to Wargo et al discloses a motion detector used to control video signal processing and functionality which operates as described hereinabove. An additional motion detector, produced as an integrated circuit, is disclosed in a publication entitled "A Correlating Optical Motion Detector" by John E.
  • the detector derives the spatio-temporal derivative of the image at the very focal plane of the image sensor and produces an image consisting of the temporal differences only.
  • the present invention seeks to provide a novel image content responsive system for image acquisition, processing, storage and retrieval, transmission and display.
  • a dynamic image representation system comprising apparatus for sensing a dynamic scene and apparatus for providing a pixel count reduced dynamic digital representation of the scene having pixel count reductions in portions of the scene not fulfilling predetermined spatial criteria and pixel count reductions in portions of the scene not undergoing change within predetermined temporal criteria.
  • an image transmission system including apparatus for acquiring an image including apparatus for providing a digital image representation having a reduced information bit content, apparatus for transmitting the reduced information bit content digital image representation over a communications link, apparatus for receiving the reduced information bit content digital image representation over the communications link and apparatus for interpolating the received reduced information bit content digital image representation to provide a visible digital image representation.
  • the reduced bit content digital image representation comprises an initial low bit content digital image representation and a plurality of subsequent low bit content additions to the initial representation for enhancing the image quality thereof.
  • the apparatus for interpolating is operative to provide an initial display based on the initial digital image representation and subsequently to modify the initial display based on the plurality of additions.
  • the apparatus for interpolating is operative for enhancing the resolution of the display in accordance with the plurality of additions.
  • the apparatus for acquiring includes apparatus for identifying temporal variations in the image and the apparatus for transmitting is operative to transmit an initial digital image representation and thereafter information representing temporal variations in the scene.
  • the apparatus for acquiring includes a raster scanning camera, a frame buffer and an encoder.
  • the raster scanning camera acquires full resolution images.
  • the encoder reduces the full resolution images to a hierarchy of lower resolution images wherein the lowest resolution image is retained in its entirety and the remaining higher resolution images contain areas of the image containing high information content.
  • the apparatus for acquiring includes a camera for acquiring reduced information bit content images and a camera controller to direct the acquisition of the reduced information bit content images.
  • the camera controller initially directs the camera for acquiring reduced information bit content images to acquire a low resolution image.
  • the camera controller analyzes the low resolution image for high information content areas and directs the camera for acquiring reduced information bit content images to acquire the high information areas at a higher resolution.
  • the present invention seeks to provide a complete image content responsive system, comprising image acquisition, processing, storage and retrieval, transmission and display elements, that transmits or store a reduced image representation which, nevertheless, retains the high information content level areas of the original image and thus, enables quick recognition of the features of the original image once the reduced image representation is expanded upon display.
  • the reduced image representation preferably is formed according to the following steps: 1) a hierarchy of different resolution images is created; 2) at each level of resolution, features of the image which are potentially 'interesting' are identified, where 'interesting' is defined by the system user; and 3) the resulting 'interesting' features at higher resolutions, as well as the lowest resolution image, are combined together, typically from the lowest resolution to the highest resolution.
  • the result is a pixel count reduced non-spatially consecutive representation for transmission, processing and storage, herein referred to as the pixel count reduced representation, containing only the most 'interesting' information.
  • the pixel count reduced representation is lossy, although the information lost is relatively insignificant to the understanding of the image.
  • the pixel count reduced representation is advantageous for image storage, for image processing and for image data transmission.
  • Fig. 1 is a block diagram illustration of an image content responsive image acquisition, processing, transmission, storage/retrieval and display system constructed and operative in accordance with a preferred embodiment of the present
  • Fig. 2 is a block diagram illustration of an image acquisition unit useful in the system of Fig. 1;
  • Fig. 3 is a block diagram illustration of the structure and operation of an image encoder for producing pixel count reduced dynamic digital representations of images useful in the image acquisition unit of Fig. 2;
  • Fig. 4 is a block diagram illustration of an alternative embodiment of an image acquisition unit useful in the system of Fig. 1;
  • Figs. 5A, 5B, 5C and 5D are block diagram illustrations of alternative embodiments of an image acquisition unit for extracting temporal variations useful in the system of Fig. 1;
  • Figs. 6A and 6B are electronic circuit diagrams of image reconstructors useful in the system of Fig. 1;
  • Figs. 7A and 7B are electronic circuit diagrams of elements of the encoder of Fig. 3;
  • Fig. 8A is an electronic circuit diagram of a combiner for creating the pixel count reduced representation useful in the image acquisition unit of Fig. 3;
  • Fig. 8B is a state diagram of a controller for the combiner of Fig. 8A;
  • Fig. 9A is a state diagram of the logic of a coordinate generator useful in the image acquisition unit of Fig. 4;
  • Fig. 9B is an electronic circuit diagram for implementing a logic pattern similar to that of Fig. 9A;
  • Fig. 10 is a block diagram of an alternate embodiment of the coordinate generator of Fig. 4;
  • Figs. 11A and 11B are electronic circuit diagram of alternate embodiments of image reconstructors for images with temporal variations useful in the system of Fig. 1;
  • Figs. 12A through 12L are illustrations of an image as it passes through the system of Fig. 1.
  • Fig. 1 a block diagram illustration of a preferred embodiment of the system.
  • the system comprises an image content responsive image acquisition unit 10 which acquires a dynamic image in such a manner as to produce an image content responsive pixel count reduced digital representation of the image, a signal transmitter 14, such as a modem, which transmits the pixel count reduced image representation as a series of pixels, an optional processing unit 15 for performing image processing on the pixel count reduced image representation, a vision computer for computer vision applications, and a communications link 17, typically a limited-bandwidth communication medium.
  • the system also comprises a signal receiver 18, such as a modem, to receive the signal transmitted by transmitter 14 over communications link 17, an image storage and retrieval system 19, such as a personal computer, and an image reconstructor 20 which interpolates the pixel count reduced image representation and transforms it into an approximation of the original image.
  • An optional additional processing unit 21 may be included for performing additional processing on the received pixel count reduced image representation.
  • a display unit 22 typically a conventional raster-scan display device, such as a CRT display, may be provided for displaying the reconstructed image.
  • image acquisition unit 10 can additionally extract temporal variations by identifying the locations in a series of images where changes have occurred and producing from them a pixel count reduced representation of the series of images as a series of changes.
  • This option may operate if a particular application requires time-varying information, such as in videotelephone or teleconferencing.
  • the processing units 15 and 21 may operate if the image requires some form of enhancement, either before or after transmission or storage/retrieval, or if for a particular computer vision application, the vision computer 16 requires preprocessing of the image where one example of preprocessing is feature extraction.
  • the pixel count reduced representation may be transmitted via the transmission elements 14, 17, and 18, or it may be stored and retrieved via unit 19.
  • image content responsive image acquisition unit 10 comprises a raster scan imager 24, such as a CCD camera, such as the WV-CD22 from Panasonic, an A/D converter 26, a frame buffer 28 , such as a matrix of static or dynamic RAM chips, and an encoder 30.
  • the raster scan imager 24 and the A/D converter 26 together produce a digital raster scan image, although other image generation systems, such as synthetic or "computer graphics" images, or a previously processed, acquired or stored image, can be used.
  • Frame buffer 28 stores the full resolution digital image thus produced.
  • Image encoder 30 encodes images into the pixel count reduced representation according to a set of operations described hereinbelow. Reference is now made to Fig.
  • An order N image I that is, an (N ⁇ N) image
  • an order N low-pass filter 32 such as a linear convolver whose convolution kernel is typically a 3 ⁇ 3 matrix whose every element contains 1/9.
  • Such a convolution kernel produces a new image where the value of each pixel in the new image is the average of the values, in the previous image, of its 8 neighbors and itself.
  • Such a convolver may comprise an off-the-shelf, special purpose VLSI chip, such as Zoran ZR33481 or ZR33881 Digital Filter Chips and some additional components, for example, delay lines and delay buffers, such as described in Zoran Technical Note No. Z.T.N. 03 entitled “Real Time Spatial Filtering with a Zoran DFP", the teaching of which is incorporated herein by reference.
  • special purpose VLSI chip such as Zoran ZR33481 or ZR33881 Digital Filter Chips and some additional components, for example, delay lines and delay buffers, such as described in Zoran Technical Note No. Z.T.N. 03 entitled "Real Time Spatial Filtering with a Zoran DFP", the teaching of which is incorporated herein by reference.
  • decimator 34 typically comprises a digital gate 35 to select a pixel, a latch 37 to enable a pixel intensity to be included in the decimated image, if and only if the gate 35 is open, and control circuitry, such as flip-flops, to repeatedly open and close the gate 35.
  • the result of the low-pass filtering and decimation is a predetermined number of decimated images where each decimated image I k is of a lower resolution than the previous one and where k indicates the number of decimations performed to produce I k . It will be appreciated that the lower the resolution of the image, the higher its decimation level is.
  • the encoder 30 reaches a detection stage where an image processing operator, such as an edge detector 46, typically comprising a convolver similar to the convolver of the low-pass filters 32, 36 and 40, operates on the lowest resolution image, I 3 in this example. Any operator which can be implemented and which detects some desired feature in an image can be applied in place of, or in addition to, the edge detector 46.
  • an image processing operator such as an edge detector 46, typically comprising a convolver similar to the convolver of the low-pass filters 32, 36 and 40, operates on the lowest resolution image, I 3 in this example.
  • Any operator which can be implemented and which detects some desired feature in an image can be applied in place of, or in addition to, the edge detector 46.
  • operators may include point processors, such as lookup tables for the detection of certain intensity levels, neighborhood processors, such as convolvers for operations such as edge detection, morphological operators, histogram processors, such as the L64250 from LSI Logic, or various transforms, such as the Hough transform for the detection ⁇ xf straight lines, such as the L64250 from LSI Logic.
  • point processors such as lookup tables for the detection of certain intensity levels
  • neighborhood processors such as convolvers for operations such as edge detection, morphological operators
  • histogram processors such as the L64250 from LSI Logic
  • various transforms such as the Hough transform for the detection ⁇ xf straight lines, such as the L64250 from LSI Logic.
  • Operators to detect other desired features, such as texture, color, or motion can also be implemented, as well as more complex operations comprising multiple operation steps.
  • the detection stage ends with a thresholder 48, typically built from one comparator, which typically -transforms the processed image into a binary image. It will be appreciated that more than one comparators can comprise thresholder 48.
  • the binary image E 3 thus produced, is of the same order as decimated image I 3 .
  • the pixels of E 3 contain values of '1' if the corresponding pixels in I 3 belongs to edges and the remaining pixels of E 3 contain values of '0'.
  • E 3 is expanded, by a factor-2 expander 50, to produce binary image Vo of order twice the order of E 3 .
  • Factor-2 expander 50 reproduces each pixel value e ⁇ y of E 3 four times, copying it to pixels 2x,2y, 2x+1,2y, 2x,2y+1, and 2x+1,2y+1 of V 3 , thus performing approximately the reverse operation of decimator 42. It will be appreciated that the pixel in the upper left corner of the image is typically defined as pixel 0,0.
  • a simple factor-2 expander 50 is shown in Fig.
  • the factor-2 expander 50 operates as follows. A row of pixel intensities is received by the selector 51 as well as being stored, in the order it is received, into storage unit 53. Selector 51 also accepts input from the storage unit 53, but only upon receiving a signal to do so from the row counter 55.
  • a pixel clock signal indicating the coordinate positions of the row of pixel intensities in E 3 , is also received, by the frequency multiplier 57, which produces two clock pulses, or coordinate positions, for every pixel which is received by the selector 51.
  • the row counter 55 increases its count by one and the selector 51 selects the input from the storage unit 53. The process repeats again.
  • the frequency multiplier 57 produces two clock pulses, thus reproducing a second row of pixels identical to the first row. Each pixel is thus reproduced 4 times.
  • a dilator 52 such as SNAP from DataCube Inc of Peabody Massachusetts USA, dilates V 3 , producing W 3 , of the same order as Vo. For each continuous segment of pixels with values '1', dilator 52 changes the values of the pixels in a ring around the segment to '1', thus increasing the size of each segment by one pixel in each direction. Wider dilations are also possible.
  • a masker 54 masks I 2 by W 3 to produce masked image M 2 , an image of the same order as I 2 .
  • the masker 44 comprises a digital many-bit two-to-one selector, such as the 74LS157 from Texas Instruments, which, according to a binary control signal, produces either one input or the other at the output.
  • the two input signals are I 2 and a constant null value
  • the binary control signal is W 3 .
  • pixel x,y will contain the value I ⁇ ,y of I 2 if and only if w x , y of W 3 , as the control signal, is '1'. Otherwise, pixel x,y contains a null value, which is not necessarily '0'.
  • Operators 56-62 similar to edge detector 46, thresholder 48, factor-2 expander 50, and dilator 52, but of higher orders, operate on M 2 to produce W 2 .
  • a masker 64 similar to masker 54, but of higher order, masks I 1 by W 2 , yielding masked image M 1 .
  • the process is repeated as many times as is necessary, typically one more time with operators 66-74, to produce M 0 , the masked version of the original image I.
  • a combiner 76 receives the lowest resolution image, I 3 , and the non-null values of the masked images, M 2 through M 0 , and produces a single stream of pixels, the pixel count reduced representation, which is preceded by a header code to indicate the beginning of an image representation.
  • each pixel is defined as a triplet consisting of the value of the pixel identified, or tagged, by its coordinates in the original image.
  • An alternative embodiment of the invention produces a higher order tuple which comprises, in addition to the elements of the triplet described hereinabove, additional information about the pixel, such as the decimation level k.
  • the combiner 76 typically comprises k + 1, FIFO buffers 78, 80, 82 and 84 of various sizes, where in the current example, k is equal to 3 and where each FIFO buffer has associated therewith a pair of x and y counters 85 and 86, 87 and 88, 89 and 90 or 91 and 92.
  • combiner 76 typically comprises a controller 93, whose state-diagram is described hereinbelow, and a selector 94, such as the 74LS157 from Texas Instruments .
  • Decimated image I 3 is stored in FIFO buffer 78 in raster-scan order due to the fact that it is a full matrix.
  • the counters 85 and 86 are reset to 0 and FIFO buffer 78 begins releasing intensities of I 3 .
  • the controller 93 signals the selector 94 to begin accepting the triplets (x,y,I 3 ).
  • the counters are stepped forward, together with the intensities of I 3 , such that for each successive pixel in the same row, the x counter 85 is increased by 8 and every time a new row begins, as is predetermined in hardware, the y counter 86 is increased by 8.
  • the counters are counting modulo 8 since decimated image I 3 is the result of three successive decimations (ie 8 times smaller) and since the coordinate position of the x,y pixel of I 3 in a full resolution image is 8x,8y.
  • the selector 94 receives, at each time step, an x coordinate value from counter 85, a y coordinate value from counter 86, and an intensity from FIFO buffer 78, and it produces the triplet in the order, x value, y value, intensity.
  • the y counter 86 reaches its maximum value, it sends a signal to the controller 93, indicating the end of the I 3 values.
  • the controller then sends two signals, one to the selector 94 indicating that data from the I 3 lines should no longer be accepted and that data from the M 2 lines should be, and one to reset the counters 87 and 88.
  • the counters 87 and 88 operate and are connected similarly to counters 85 and 86; however, they count modulo 4.
  • the values of the counters 87 and 88 are increased as each element of M 2 arrives. If the pixel received is not null valued, as checked by comparator 95, the values of the counters and the current intensity value are stored in FIFO buffer 80. Otherwise, nothing is stored and the succeeding element of M 2 is received.
  • FIFO buffer 80 When FIFO buffer 80 receives a read signal from the controller 93, it sends to selector 94 the triplets stored in it, in the order they were entered. FIFO buffer 80 indicates to controller 93 its current status and, upon an indication of being empty, controller 93 switches selector 94 to accepting data from the M 1 lines. Combiner 76 processes M 1 in a manner similar to its processing of M 2 ; following the processing of M 1 , it processes M 0 , and following M 0 it produces a header code and then returns to the beginning and processes I 3 .
  • Fig. 8B shows a five-state state diagram for controller 93.
  • the controller is at state 5.
  • I 3 arrives, as indicated by a standard video control signal such as the horizontal and vertical sync and the pixel clock, the controller 93 switches to state 1 where it generates the control signals to indicate to selector 94 to select I 3 and its X and Y coordinates.
  • the controller switches to state 2 where it issues a series of clocked READ signals to the FIFO buffer 80 and control signals to the selector 94 to select M 2 and its X and Y coordinates.
  • an EMPTY signal is received by the controller, causing it to switch to state 3.
  • the controller 93 will produce the desired information and send it to the selector 94 to be included in the tuple. For example, should the decimation level be required, the controller 93 will calculate it based upon the image being received by the selector 94. This is indicated in Fig. 8A by a dashed line from the controller 93 to the selector 94.
  • combiner 76 can produce a pixel count reduced representation comprising a stream of windows, rather than a stream of pixels.
  • the window-based pixel count reduced representation similarly to the pixel-based representation, begins with a header code to indicate the start of a pixel count reduced description.
  • the first window typically comprises the complete I 3 which contains no null values and which serves as the lowest resolution background for I r .
  • Each additional window comprises a complete rectangular area, within any of the masked images, M 2 through M 0 , which contains no null valued pixels.
  • a one pixel rectangular area is acceptable and series of them are used to define a one-pixel wide diagonal line.
  • the coordinate tagging of each individual pixel is replaced by a single tag for the entire window comprising a set of coordinates of the upper-left corner, x ul ,y ul , a length, 1, and an height, h, of the window, and a decimation level k denoting the masked ' image, M k , from which the window was taken.
  • a decimation level k denoting the masked ' image, M k , from which the window was taken.
  • Following the window tag is a stream of the values of the pixels in the window organized in row-major form, that is, the first 1 values belong to the first row, the second 1 values belong to the second row, etc, for h rows.
  • the pixel count reduced representation thus produced comprises differing levels of resolution, with the areas containing the most 'interesting' information described at high resolution and the less 'interesting' areas described at lower levels of resolution, where 'interesting' is defined by the system user.
  • the hierarchy of levels of resolution is known as a pyramidal resolutions scheme and it will be appreciated that alternate pyramidal schemes can replace the one described hereinabove.
  • An alternate embodiment of the image content responsive image acquisition unit 10, shown in Figure 4 comprises a realtime image content responsive camera controller 100 controlling an image content responsive camera 102 comprising an Intelligent Scan image sensor 103.
  • the combination of camera 102 and image sensor 103 is disclosed in European patent application 88111330.2, the teaching of which is incorporated by reference, and is commonly owned by the owners of the present application.
  • the actively controlled image acquisition unit 10 identifies 'interesting' features in an image, typically first acquired at the lowest resolution possible, and utilizes the results to direct the Intelligent Scan image sensor 103 to acquire the interesting areas of the image which require the next higher resolution . The process is repeated until the highes t resolution is reached, at which point a new image, at the lowest resolution, is acquired.
  • the pixel count reduced representation is typically built in a manner similar to that of encoder 30. This embodiment of unit 10 differs from that of the previous embodiment, discussed hereinabove and illustrated in Figs.
  • the acquisition system creates the pixel count reduced representation in the process of acquiring the image, whereas the previous embodiment first acquires the entire image and only afterwards encodes it into the pixel count reduced representation; 2) the acquisition begins with the lowest resolution image and adds to it sections of the image of increasingly higher resolution, thus, never needing to acquire the entire image as does the previous embodiment; and 3) the acquisition system operates at a higher frame rate since it does not need to acquire the entire image.
  • the low resolution image is acquired in a manner different from that of the encoder, as described hereinbelow.
  • the camera 102 receives from controller 100 the coordinates at which an intensity value is desired.
  • a decimation level k is also typically included with the coordinate value, to ensure consistency within the system, although camera 102 does not require it.
  • the camera 102 acquires the intensity value at the requested coordinates I x,y and returns it, along with its coordinates x,y and the present decimation level k, to controller 100.
  • the tagged pixel is part of the pixel count reduced image representation, and as such, is also the output of the acquisition system.
  • Controller 100 typically comprises a pixel count reduced edge detector 104, such as a convolver for the pixel count reduced format as described in Israel Patent Application 87310, filed August 2, 1988, the teaching of which is incorporated herein by reference, a series of comparators 106 to convert the results of edge detector 104 into a binary format, a pixel count reduced dilator 108, such as described in Israel patent application 87310, a 1-selector 110 to select only those pixels with a non-null value, a coordinate generator 112 to generate the coordinates to be sent to the camera 102, and a FIFO memory 116, such as the IDT7201A from Integrated Device Technology of Santa Clara, California, to store the coordinates generated by the coordinate generator 112 and to send them, one at a time, to the camera 102.
  • a pixel count reduced edge detector 104 such as a convolver for the pixel count reduced format as described in Israel Patent Application 87310, filed August 2, 1988, the teaching of which is incorporated herein
  • the image processing operator detects desired features in the image. To do so, it stores the incoming tagged pixels until the entirety of pixels necessary to perform the indicated operation, as taught in Israel Patent Application 87310.
  • the tagged pixel for whom the operation is performed is sent to the comparators 106 which then compare the pixel intensity to a threshold, or series of thresholds.
  • the comparators 106 convert the pixel intensity to a binary value.
  • the image processing operators chosen depend on the definition of a desired feature and are as detailed for encoder 30.
  • the result of the image processing operations followed by the dilator 108 are binary tagged pixels indicating the coordinates of the chosen feature.
  • 1-selector 110 eliminates null-valued pixels and only sends to coordinate generator 112 1-valued pixels. For each pixel reaching the coordinate generator 112, four pixel coordinates, to be acquired at the next highest resolution (i.e. the next lowest decimation level), are generated.
  • PAL Programmable Array Logic
  • the coordinate generator 112 before any pixel intensities have been acquired, begins in state 6 and generates and sends to the output a complete set of triplets for the low resolution image, typically with a decimation level k equal to 3 and with coordinate values in increments of 8, covering the entire field of view of the image sensor 103.
  • the coordinate generator 112 proceeds to state 7 and generates and sends to the output a dummy triplet, (T,T,3), known as a trailer, whose purpose is to mark the end of a decimation level.
  • the generator 112 proceeds to state 1 and waits for triplets to arrive from the 1-selector.
  • the decimation level k for each triplet is decreased by 1, denoting the next highest resolution.
  • the trailer arrives after all the triplets at a given level have been processed and is not processed in the manner of the other triplets. If the decimation level k is greater than 0, the trailer is sent to the output with its final component, that of the decimation level, reduced by 1.
  • the coordinate generator 112 If, however, the decimation level k of the trailer is 0, marking the end of the triplets to be acquired at the highest resolution or, in other words, indicating that all the necessary information in image has been acquired, the coordinate generator 112 returns to state 6 and begins the process of acquiring the succeeding image by generating a set of triplets for the low resolution image.
  • FIG. 9B An electronic circuit which embodies the state diagram of Fig. 9A is shown in Fig. 9B. It comprises a programmable controller 115, such as a PAL chip, whose program listing is given in Annex AA, three selectors 117, 119, and 121, two adders 123 and 125, a subtracter 127. and three comparators 129, 131 and 133.
  • a programmable controller 115 such as a PAL chip, whose program listing is given in Annex AA
  • three selectors 117, 119, and 121 three selectors 117, 119, and 121, two adders 123 and 125, a subtracter 127.
  • three comparators 129, 131 and 133 three comparators 129, 131 and 133.
  • the circuit of Fig. 9B operates as follows.
  • a triplet (x,y,k) arrives and each element is compared, by the comparators 129, 131 and 133, to the expected values of the trailer triplet (T,T,0).
  • the resulting values, XET, YET and DGZ, respectively, are sent to the programmable controller 115 which checks whether or not the trailer has arrived. If the trailer has arrived and its decimation level k is not 0 (ie DGZ > 0) then the programmable controller 115 remains in state 1 but instructs the selector 117, 119 and 121 to output the trailer (T,T,k-1).
  • the programmable controller 115 proceeds to state 6 and enables an X counter 135 and a Y counter 137 for N/8 steps each, where N is the order of the original image.
  • the programmable controller 115 instructs the selectors 117, 119 and 121 to output the values of the X and Y counters 135 and 137 as well as the value 3, producing the triplet (x,y,3).
  • the programmable controller 115 proceeds to state 7 where the selectors 117, 119 and 121 are instructed to output the trailer (T,T,3).
  • the programmable controller 115 then proceeds to state 1.
  • the programmable controller 115 steps through states 2-5, instructing the selectors 117, 119 and 121 to output the triplets (x+p,y+q,k-1) as described hereinabove, and returns to state 1.
  • the present embodiment of the coordinate generator 112 typically comprises a frame buffer 118 and a comparator 120, in addition to the PAL chip 113 of the previous embodiment.
  • Frame buffer 118 stores in each x,y position a number L x, y specifying the minimum level of decimation (indicating the maximum resolution) which is allowed for the pixel residing at the respective coordinate position.
  • Frame buffer 118 accepts external input from an external processor, such as a personal computer, defining the elements
  • the external updating occurs only after the first low resolution image has been acquired.
  • the present embodiment begins at state 6 and generates the low resolution matrix of coordinates, moves to state 7 and generates the trailer, and only thereafter allows the external processor to affect its operation.
  • the coordinate generator 112 operates as follows. For each incoming triplet
  • the triplet is accepted and is sent to the PAL chip 113 where it is operated upon as described hereinabove. Otherwise, the triplet is discarded.
  • the four triplets produced by the coordinate generator 112 are sent to a FIFO buffer 116 to be stored there until the previously stored coordinates, produced from previous inputs to the coordinate generator 112, are released. Consequently, the four triplets are received by the camera 102, one at a time, and are used to define the next coordinates of the image to be acquired.
  • An additional embodiment of the image acquisition unit 10 acquires a set of images of a moving scene and extracts the temporal variations for efficient storage and transmission.
  • the embodiments described hereinabove are modified slightly to incorporate motion detection as follows and as shown in Figs. 5A-5D.
  • Fig. 5A One embodiment of the temporal variations extraction method of image acquisition unit 10 is shown in Fig. 5A and comprises a raster-scan method camera, referenced 150, a read-modify-write frame buffer 152 which initially contains all null values, a subtracter 154, and an image encoder 156.
  • Raster-scan camera 150 produces a full resolution current image I t , where t denotes the present time, and the frame buffer 152 stores the previous full resolution image, I t-1 .
  • the value of I t-1 at the latest location x,y, according to the raster-scan format is read out of the frame buffer 154 and subtracted from the value of the latest element in the raster-scan of I t . It will be appreciated that the two values share the same coordinate location x,y.
  • the value of the element of I t is written into the x,y location of frame buffer 154.
  • the differences Dt are sent to encoder 156 which is similar to encoder 30 but without edge detectors 46, 56 and 66. In place of the edge .detection, encoder 156 typically has a single comparator, typically based on absolute intensity levels, to determine when the difference level is large enough to indicate the location of 'interesting' changes in the image.
  • FIG. 5B An alternate embodiment of the temporal variations extraction method of image acquisition unit 10 which encodes only the largest differences is shown in Fig. 5B.
  • This embodiment of unit 10 comprises the raster-scan camera 150, the subtracter 154 and the encoder 156, as in the previous embodiment, as well as two frame buffers 158 and 160 which initially contain null values.
  • the current image I t is simultaneously subtracted from the image stored in frame buffer l60 while being stored into frame buffer 158.
  • the full resolution image of differences is encoded by encoder 156 as described hereinabove; however, according to the present embodiment, the combiner of the encoder produces the images in the reverse order, that is, it produces the pixel count reduced representation with the highest resolution masked image, typically M 0 , first, followed by the lower resolution masked images and with the lowest resolution image I 3 last.
  • the controller 93 logic must be changed as follows. The controller must wait until all of the masked images have been produced and then it must select them and the low resolution image I 3 in the order described hereinabove. Thus, in the situation of the transmission of the pixel count reduced representation, the largest differences are transmitted first.
  • each triplet is transmitted, its coordinates are sent to the frame buffers 158 and l60 and the corresponding pixel from frame buffer 158 is copied into frame buffer 160.
  • a transmission medium is typically fairly slow and it is to be expected that within one frame time, typically 30 milliseconds, the entire pixel count reduced representation will not be transmitted; however, this embodiment ensures that, at least, the largest differences are transmitted.
  • a signal is sent from the camera 150 to the encoder 156 and, as soon as data from the new image arrives at the encoder 156, it stops producing the pixel count reduced representation.
  • the subtraction is now between the new image I t+1 and whatever is stored in frame buffer 160, typically most of the previous image I t but containing information from earlier images also.
  • the old value is retained in frame buffer 160, as mentioned hereinabove. If movements continue in the image, then the difference between the old value at x,y and the current value will increase such that, eventually, the difference is large enough to be positioned early enough in the pixel count reduced representation to be transmitted.
  • FIG. 5C A third embodiment of the temporal variations extraction method of image acquisition unit 10 is shown in Fig. 5C.
  • This embodiment typically comprises the same elements and the same initial values as per the previous embodiment; however, the output is a stream of the latest intensity values of the pixels at which movement occurred, rather than the differences at those pixels.
  • the current image I t is subtracted from the image stored in frame buffer 160 and the difference is sent to an encoder 162.
  • Encoder 162 operates on the image of differences as per the second embodiment with the exception that it produces only the coordinate positions of the pixel count reduced representation.
  • the coordinates produced are sent as control signals to the frame buffers 158 and 160 and the intensity value of I t at the x,y location in frame buffer 158 is copied into frame buffer l6 ⁇ .
  • the intensity value of I t is combined with the coordinate values and thus, a pixel count reduced representation of the latest image intensities is produced.
  • transmission is stopped shortly after the camera begins acquiring a new image.
  • FIG. 5D A fourth embodiment of the temporal variations extraction method of image acquisition unit 10 is shown in Fig. 5D.
  • This embodiment is similar to that of the first embodiment in that it transmits the entire image of differences; however, it utilizes, with no modifications, the image content responsive camera 102 and camera controller 100 to acquire the images.
  • the resultant pixel count reduced representation which has incorporated in it the
  • the intensity value I x,y of the triplet (x,y,I x,y ) is stored at position x,y in FIFO frame buffer 164, writing over the value that was previously there.
  • the I x,y of a pixel in the pixel count reduced representation is subtracted, in subtracter 154, from the value it will overwrite in frame buffer 164.
  • Three types of differences are produced. If both the I x,y and the stored value are non-null values, the difference is computed. If the stored value is a null value, then I x,y is presented as the difference value.
  • the difference value produced depends on the type of interpolating done at the receiving end, as described hereinbelow, and is either a null value, for a zero-order updating interpolator, or an instruction to the camera controller 100 to produce an actual reading of the pixel intensity at some later time. In either case, the pixel intensity is nullified in the frame buffer 164.
  • the difference produced by the subtracter 154 is combined with the coordinate values produced from the controller 100, thus producing the output of this embodiment of unit 10.
  • the differences can control which pixels, stored in frame buffer 164, combined with the coordinate values, are produced as the output.
  • the differences without any thresholding, can so control the output since the camera-controller system 100 operates at high speed, thus producing few differences between images, and any differences thus produced are, by definition, the important ones.
  • the FIFO frame buffer 164 flushes out the locations which have not been updated since the last flushing.
  • a fifth embodiment of the temporal variations extraction method of image acquisition unit 10 incorporates a sensor cell 165 designed by John E. Tanner and Carver Mead, as discussed in the Background of the Invention, the referenced publication of which is incorporated herein by reference, into the image acquisition unit 10.
  • the image sensor 166 as described in European Patent Application 88111330.2, contains an array of sensor cells for detecting light impinging upon it and a system for selectively reading the information stored in the cells.
  • image sensor 166 comprises an array of sensor cells I65 and thus, acquires and stores the difference image according to the directions it receives from a controller 168, similar to controller 100 except that the image processing element is not an edge detector 104 but an absolute intensity thresholding device.
  • the output is a stream of tagged pixels of the image of differences.
  • Pixel count reduced representations are organized in a manner which is conducive to efficient storage and to communication where slow transmission is common.
  • transmitter 14 can transmit them in real time over communication link 17 to be received by receiver 18.
  • Image reconstructor 20 receives the pixel count reduced representations from receiver 18 and rebuilds them, displaying the resulting image on display 22.
  • a preferred embodiment of the image reconstructor 20 utilizes a zero-order interpolative reconstruction to produce the reconstructed pixel image B, an approximation of the original image I, from its pixel count reduced representation A.
  • the method is as follows. If a pixel is present in A, it is copied directly into its corresponding pixel in B. All other pixels in B correspond to areas of the original image I which did not contain 'interesting' information and therefore, need to be interpolated from their neighbors which were stored in A.
  • the x,y pixel of B assumes the value of pixel m,n of A where m,n is a member of IJ and achieves the minimum of ((x-i) 2 + (y-j) 2 ) over all i,j in IJ.
  • This zero-order interpolation is tantamount to a sample-and-hold process in two spatial dimensions.
  • a neighborhood generator (NG), referenced 300 receives a pixel of the pixel count reduced representation, referenced 302 , as the triplet
  • NG 300 will generate a sequence of coordinates x,y constituting a neighborhood around i,j of predetermined size D.
  • the coordinates x,y are generated according to concentric circles centered around i,j and in order of increasing distance d x,y from i,j.
  • d x,y is the Euclidian distance between i,j and x,y.
  • d x,y is fixed for the entire square 'circle' and only increases from 'circle' to 'circle'.
  • a frame buffer 304 such as a matrix of RAM chips, is accessed during the read subcycle of a read-modify-write memory cycle, and a value r x,y , the stored value at the x,y position of the frame buffer, is retrieved.
  • the value r x,y is the distance from coordinates x,y to the coordinates of the input pixel which last determined the intensity value at x,y.
  • a comparator 306 compares d x,y to r x,y . If d x,y is less than or equal to r x,y , then the following activities occur:
  • a controller 310 sets an 'update' flag.
  • Controller 310 saves the current distance d x,y in a register. Every time a new distance and x,y pair are generated, the controller 310 compares the new d x,y with the currently stored one. If d x,y grows by 1, the controller 310 examines the
  • the controller 310 thus insures that updating proceeds while there exist pixels which are closer to the current input pixel 302 than to any previous ones. However, the updating process terminates when the neighborhood around the input pixel 302 has been completely updated.
  • the neighborhood size D is incorporated in order to increase the efficiency of the interpolation process.
  • the limit D is most important during the early stages of the interpolation process since at that point, the frame buffers 304 and 308 are mostly empty and the early pixels will, if not contained, interpolate onto wide areas which are likely to be overlaid with later pixels.
  • D is defined as 2 m , where m one less than the lowest decimation level.
  • a video controller 312 accesses frame buffer 308, which holds the interpolated image B concurrently with its being updated.
  • video controller 312 comprises counters to generate the raster-scan order row and column addresses of the most recently interpolated image and a video generator, such as ZNA 134J from Ferranti of Commack, New York, for generating vertical and horizontal sync signals for the display 22 .
  • the video controller 312 thus displays first an interpolation of the lowest resolution of the image and subsequently improving the reproduced image with the higher resolution 'interesting' features.
  • the image reconstructor 20 in this embodiment, comprises four major elements, an interface 313 to a processor, such as an IBM PC, a controller 314, a data path 316, and a video controller 320.
  • the video controller 320 is an interface to display 22 and operates every other clock time, after image reconstructor 20 has interpolated the input signal.
  • the operation of the circuit board begins through the interface 313 as an input triplet (i,j,a i,j ) is downloaded to it from the processor.
  • Interface 313 writes the i, j, and intensity a i,j values into X-CARD COUNT, Y-CARD COUNT, and VALUE PORT, respectively, where X-CARD COUNT and Y-CARD COUNT are typically counters, such as the 20X8, and VALUE PORT typically comprises an 8-bit register, such as the 20R8.
  • VALUE PORT typically has a synchronous clear line for initialization purposes.
  • the interface 313 performs the operation of unit 302 of Fig. 6A.
  • it checks the read status of the present embodiment of interpolator 20 and produces another set of values as soon as the interpolator 20 is ready for it.
  • Interface 313 is incorporated into chips U34 and U35.
  • the controller 314 typically comprises three PAL chips, typically the 16R4 with a 10 nsec clock, and performs the operations of the controller 310 and the neighborhood generator 300.
  • Control chip U13 the main controller, controls the other two chips, thus exerting control over the entire image reconstructor 20.
  • the array logic incorporated into control chip U13 is included in Annex A.
  • Coordinate chip Ul4 produces the coordinates in the neighborhood of the current pixel i,j such that at every read or write stage the proper address is available.
  • the neighborhood is typically defined as a square spiral beginning at the pixel position directly above the current pixel; thus the coordinates of the neighborhood elements are calculated as follows: move up one square from the current pixel, move right one square, move down two squares, move over two squares, move up three squares, etc.
  • Coordinate chip Ul4 performs the operations of neighborhood, generator 300; its array logic is included in Annex B.
  • Initialization chip U15 clears display 22 and initializes a frame buffer R-buffer, such as chips 1120-22 built from the 7187, to the maximum radius D.
  • D is defined in hardware by the positioning of a dip switch, SW1, attached to a counter R-count.
  • R-buffer performs the same operation as frame buffer 304, specifically, it stores the distance of each pixel in the frame to the most recently available input pixel closest to it.
  • Data path 316 typically comprises a memory, typically built from two static RAM frame buffers, R-buffer and I-buffer, 64K ⁇ 8 and 64K ⁇ 3 blocks respectively, two counters R-count and 2R-count, instantiated as chips U10 and Ul6, VALUE PORT, and a write-disable system, built into chip U19.
  • I-buffer similar to frame buffer 308, stores the current state of interpolated image B and is instantiated in chips U23-U30.
  • R-buffer operates as mentioned hereinabove.
  • Fig. 6B(2) shows the connections of both R-buffer and I-buffer in more detail.
  • R-count and 2R-count produce the neighborhood spiral as follows.
  • R-count counts R the number of squares in the spiral, up to a maximum D as defined in hardware.
  • R-count increases by 1 when the square is finished and the next square is begun.
  • the array logic for R-count is included in Annex D.
  • 2R-count counts the number of pixels per side of the current square of radius R and as such, counts modulo 2R - 1. For example, if R is equivalent to 3, 2R-count counts modulo 5 « 2R-count announces to coordinate chip U14 when a side of the square.
  • the array logic for 2R-count is included in Annex E.
  • the array logic for WRITE DISABLE is included in Annex F .
  • the video controller 320 produces synchronization pulses to be sent to display 20, to a D/A chip and to counters, Y-VIDEO COUNT and X-VIDEO COUNT.
  • X-VIDEO COUNT and Y-VIDEO COUNT indicate the location in I-buffer that the video controller 320 will read, and as such, count in a raster-scan order. For each clock cycle during which the video controller 320 operates, it reads the entirety of elements of I-buffer, thus producing on the display 22 the most current version of the reconstructed image B.
  • an approximation B of the original image is displayed to a viewer on display 22.
  • original image I rather than its approximation B.
  • the original image I is necessary, but the viewer does not wish to wait for the entire image to be displayed before recognizing it and responding to it.
  • the pixel count reduced representation, with modifications, is ideally suited for display in the abovementioned situations since pixels are allocated earlier to
  • the modifications necessary to display the original image I in its entirety typically comprise incorporating an additional controller 324, comprising a frame buffer 326, into the encoder 30 of image acquisition unit 10 to operate as follows.
  • a minimal, pixel count reduced representation, of the kind described hereinabove, is first produced and a '1' is stored in the locations of the frame buffer 326 corresponding to each pixel of the minimal representation.
  • the additional controller 324 then reduces the threshold level of the thresholder utilized in the chosen embodiment of image acquisition unit 10 to allow more pixels to be incorporated in the representation.
  • the additional controller 324 checks each pixel of the thus generated representation, via a comparator 325, and incorporates only those pixels for whom a null value was stored in frame buffer 326.
  • Additional controller 324 then stores a '1' in the locations of frame buffer 326 corresponding to the pixels thus incorporated. The process continues until all the pixels of the original image have been acquired, producing an image content responsive sequence of the pixels of the original image.
  • the original image is reproduced, after transmission over communications link 17, or from storage, via either of the reconstruction methods discussed hereinabove.
  • most pixels are interpolated over large areas, constrained only by the D limit. Due to the spatial responsiveness of the image acquisition method, after less than 10% of the pixels have been displayed, the reproduced image resembles the original image enough to allow a quick identification of the image.
  • the interpolated pixels are replaced by actual, received ones, resulting in a final image which is the original image I.
  • interpolator 20 typically comprises a read-modify-write frame buffer, referenced 350 and initially containing null values, an adder 352, and a display controller 356, typically comprised of counters for the generation of row and column addresses in raster-scan order and a video generator, such as ZNA 134J from Ferranti of Commack, NY, for the generation of vertical and horizontal sync signals.
  • a read-modify-write frame buffer referenced 350 and initially containing null values
  • an adder 352 typically comprised of counters for the generation of row and column addresses in raster-scan order
  • a video generator such as ZNA 134J from Ferranti of Commack, NY, for the generation of vertical and horizontal sync signals.
  • the input signal is a stream of tagged difference pixels, the pixel count reduced representation of an image of differences.
  • the x,y coordinates of each new tagged difference pixel are sent as control signals to the frame buffer 350 which extracts the x,y pixel intensity I x,y and sends it to adder 352.
  • Adder 352 adds I x,y to Dt x,y , the difference value at x,y, to create the updated version of I x,y which is then stored in the x,y location of frame buffer 350.
  • Display controller 356 simultaneously displays, in raster-scan order, the entire contents of frame buffer 350.
  • interpolator 20 is typically combined with the first, second, fourth and fifth embodiments of the image acquisition unit 10 for temporal variations extraction.
  • interpolator 20 for motion reconstruction which decreases the effect of pixels with age by checking their time stamps and is known as a zero-order updating interpolator, is shown in Fig. lib.
  • the circuit is similar to that of Fig. 6A, the zero-order hold interpolator, with the addition of a third frame buffer 360, a clock 362 for generating time stamps t, two multipliers 364 and 366 and an adder 368 whose only output is the most significant bit, denoting, in a two's-complement system, whether or not the result is negative.
  • the elements of the circuit which remain the same as in Fig. 6B are numbered as they were in that figure.
  • This modified zero-order interpolator operates as follows.
  • a triplet (i,j,a i,j ) is received, where a i,j is the value at a location in a new image where there was a large difference from the previous image.
  • the Neighborhood Generator (NG) 300 generates a neighborhood around i,j, as well as the distance d x, y , as described hereinabove.
  • Frame buffer 304 stores the distances r x,y which are the distances from the coordinates x,y generated by the
  • Frame buffer 308 stores the current intensity values for all pixels and frame buffer 360 stores the time stamps t x,y denoting the clock time at which the intensity I x,y was last stored into frame buffer 308.
  • a given neighborhood pixel x,y will assume the value of the input pixel intensity a i,j if and only if the following relationship holds true.
  • This embodiment ensures that all pixels are updated as they get older and is useful for the third and fourth embodiments of the image acquisition unit 10 for temporal variations.
  • Figs. 12A - 12L a series of images, which detail the processing of an image as it proceeds through the system of Fig. 1 as well as compare the pixel count reduced representation with a standard raster scan bit reduction technique.
  • Fig. 12A is the original image which shows a woman wearing a hat of feathers standing in front of a fairly nondescript background.
  • Fig. 12B is a bit mapping of the locations of the pixels which are chosen for a three level decimation pixel count reduced representation, such as described hereinabove. The areas of lowest concentration indicates the low resolution image I 3 and, it should be noted, are concentrated mainly in the background areas.
  • the pixels belonging to M 2 mark the interesting features in the background, as well as parts of the face and hat.
  • Pixels belonging to M 1 being the dark grey areas of the figure, are concentrated along the edges of the woman's face, body and hat. Finally, the black areas of the figure indicate the highest decimation level, belonging to M 1 , and are near the pupils, along the edges of the hat and in the feathers of the hat. It will be appreciated that, even from the bit mapping of the pixel count reduced representation, the basic elements of the image are visible.
  • Fig. 12C is an illustration of the pixel count reduced representation itself, in image form.
  • the pixels shown have the intensity values corresponding to the location they are in.
  • the black areas are pixels which do not belong to the pixel count reduced representation and as such, do not have any intensity values. Once again it is possible to see the basic elements of the image.
  • Fig. 12D shows a reconstruction of 1 . 5% of the entirety of pixels in the image, when reduced via either the raster scan technique and utilizing only every 8th pixel in both directions, or according to the teaching of the present invention and reproducing only the lowest resolution image portion of the pixel count reduced representation.
  • Figs. 12E and 12F show reconstructions of 2 . 8% of the entirety of pixels in the image, where Fig. 12E is from a reduction via the raster scan method, which utilizes only every 6th pixel in both directions and where Fig. 12F is from the first 2 . 8% of the pixel count reduced representation. It will be appreciated that Fig. 12F more clearly defines the area around the eyes and along the edges of the hat.
  • Figs. 12G and 12H are similar to Figs. 12E and 12F but with 6.25% of the image reproduced.
  • the raster scan reduced image the raster scan utilized every fourth pixel.
  • the reproduction of the pixel reduced count representation produces more definition, specifically in the areas around the eyes, nose and mouth, along the edges of the hat and among the feathers of the hat.
  • Figs. 121 and 12J are similar to Figs. 12E and 12F but with 11% of the image reproduced. In the case of Fig. 121, every third pixel was utilized. Although the image in Fig. 121 is perfectly recognizable, the image in Fig. 12J is better defined.
  • Figs. 12K and 12L show a reproduction of the image using 25% of the pixels in the image, where, for the raster-scan reduction, every second pixel was incorporated.
  • Fig. 12L the reproduction of the pixel count reduced representation, has much better defined edges.
  • Pin 1 CLK /* RISING EDGE 20 MHZ */
  • Pin 3 START /* ***ACTIVE LOW *** */
  • Pin 4 WR_FIN /* ACTIVE HIGH */
  • Pin 5 R_TC /* ACTIVE HIGH */
  • NC1 !Q0 & !Q1 & !Q2 & Q3 ;
  • NC2 RESET & !Q3
  • NC3 Q0 & Q3 # Q1 & Q3
  • Pin 1 CLK /* FALLING EDGE 20 MHZ */
  • Pin 4 TC_2R /* TC OF 2*R COUNTER */
  • NCI !START # Q1 & Q3 & ENE ;
  • Pin 1 CLK ; /* FALLING EDGE 20 mHZ */
  • Pin 6 X7 ; /* * /
  • Pin 7 Y7 ; /* */
  • Pin 10 GND ; /* */
  • Pin 20 VCC ; /* */
  • NC1 !Q0 & Q1 & !Q2 & !Q3 # Q0 & Q1 & Q2 & !Q3 & (TC_Y # !TC_Y & IY7)
  • Q3.D BEGIN & ( Q0 & Q1 & Q2 & !Q3 & !TC_Y & Y7 # !Q0 & !Q1 & !Q2 & Q3 );
  • Pin 1 CLK ; /* RISING EDGE */
  • Pin 2 RESET ; /* "0" FOR RESET COUNT */
  • Pin 3 D0 ; /* */
  • Pin 8 ENDFROMINIT ; /* END FROM INIT MACHINE */
  • Pin 13 IOE ; /* "0" FOR OUTPUT ENE (active allwayes)
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */
  • Pin 17 ENDTOCONT ; /* REGISTERED END TO THE CONTROLLER*/
  • Pin 9 OE_ENE ; /* SHOULD BE CONECTED TO PIN 12
  • ENE_R.D Q0 & !Q1 & Q2 & Q3 & OE_ENE & !ENE ;
  • NOWRITE.D !EOPRESET & ( !EOPSET # EOPSET & NOWRITE ) ; /* EOPRESET
  • Pin 1 CLK ; /* FALLING EDGE 20 MHZ */
  • Pin 2 CP10 ; /* */
  • Pin 3 CPS ; /* */
  • Pin 6 LOADX ; /* ACTIVE LOW */
  • Pin 7 LOADY ; /* ACTIVE LOW */
  • Pin 8 X6 ; /* */
  • Pin 9 Y6 ; /* */
  • Pin 11 IOE ; /* "0" FOR OUTPUT ENEBLE (allwayes active) */
  • Pin 10 GND ; /* */
  • Pin 20 VCC ; /* */
  • Pin 1 CLK ; /* CP-20 RISING EDGE */
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */
  • Pin 21 !RESET ; /** ACTIVE LOW - ADD H300-301 **/
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */
  • Pin 20 Q2 ; /* */
  • Pin 1 CLK ; /* LINE DRIVE */
  • Pin 3 DO ; /* */
  • Pin 8 D5 ; /* */
  • Pin 13 IOE ; /* "0" FOR OUTPUT ENE */
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */
  • Pin 20 Q2 ; /* */
  • RCO ! LOAD & Q0 & Q1 & Q2 & Q3 & Q4 & Q5 & Q6 & Q7;
  • Pin 1 A0; /* FROM U5 */
  • Piri 20 VCC /** Outputs **/
  • Pin 1 CLK ; /*- RISING EDGE 20 MHZ */
  • Pin 2 OE_CARD ; /* ACTIVE LOW */
  • Pin 3 WR ; /* FROM CONT ( active low ) */
  • Pin 10 GND ; /* */
  • Pin 20 VCC ; /* */
  • Pin 19 IMEMWR ; /* MEMORY WRITE */
  • Pin 18 TO_WR_FIN ; /* CONECTES TO WR_FIN D-FF */
  • Pin 17 CP10 ; /* */
  • Pin 16 CP5 ; /* */
  • Pin 15 CP2_5 ; /* */
  • Pin 14 LATCH_DTOA ; /* ACTIVE HIGH */
  • MEMWR WR_DIS & !WR & CP5 & !CP10 & IOE_CARD ;
  • CP2_5.D CP5 & CP10 & !CP2_5 # CP2_5 & !( CP5 & CP10 ) ;
  • Pin 1 CLK ; /* FALLIND EDGE */
  • Pin 2 LOAD ; /* "0" FOR LOAD COUNT */
  • Pin 3 D0 ; /* */
  • Pin 8 D5 ; /* */
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */
  • Pin 20 Q2 ; /* */
  • Pin 1 CLK ; /* RISING EDGE */
  • Pin 2 LOAD ; /* "0" FOR LOAD VALUE */
  • Pin 3 DO ; /* */
  • Pin 8 D5 ; /* */
  • Pin 11 CLR ; /* "1" FOR CLEAR */
  • Pin 13 IOE ; /* "0" FOR OUTPUT ENEBLE */
  • Pin 12 GND ; /* */
  • Pin 24 VCC ; /* */

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Processing (AREA)

Abstract

Un système de représentation dynamique d'image comprend un appareil pour détecter une scène dynamique et un appareil pour permettre une représentation numérique dynamique réduite du nombre de pixels de la scène présentant des réductions du nombre de pixels dans des parties de la scène ne répondant pas à des critères spatiaux prédéterminés, et des réductions du nombre de pixels dans des parties de la scène ne subissant pas de modification dans le cadre de critères temporels prédéterminés.
EP19900902024 1989-01-16 1990-01-16 Video imaging system Withdrawn EP0411076A4 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
IL88969 1989-01-16
IL88969A IL88969A0 (en) 1989-01-16 1989-01-16 Dynamic image representation system
IL89065 1989-01-25
IL89065A IL89065A0 (en) 1989-01-25 1989-01-25 Dynamic image representation system

Publications (2)

Publication Number Publication Date
EP0411076A1 true EP0411076A1 (fr) 1991-02-06
EP0411076A4 EP0411076A4 (en) 1993-05-05

Family

ID=26321890

Family Applications (1)

Application Number Title Priority Date Filing Date
EP19900902024 Withdrawn EP0411076A4 (en) 1989-01-16 1990-01-16 Video imaging system

Country Status (2)

Country Link
EP (1) EP0411076A4 (fr)
WO (1) WO1990008443A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4369464A (en) * 1979-07-09 1983-01-18 Temime Jean Pierre Digital video signal encoding and decoding system
EP0181321A2 (fr) * 1984-11-08 1986-05-14 MUPID COMPUTER GESELLSCHAFT m.b.H. Procédé et dispositif pour la transmission du signal d'un point d'image
EP0330455A2 (fr) * 1988-02-22 1989-08-30 Kabushiki Kaisha Toshiba Dispositif de codage d'images

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5836090A (ja) * 1981-08-27 1983-03-02 Kokusai Denshin Denwa Co Ltd <Kdd> テレビジョン信号の中央値予測符号化方式
JPS61253993A (ja) * 1985-05-07 1986-11-11 Nippon Hoso Kyokai <Nhk> 立体テレビジョン画像信号の伝送方法

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4369464A (en) * 1979-07-09 1983-01-18 Temime Jean Pierre Digital video signal encoding and decoding system
EP0181321A2 (fr) * 1984-11-08 1986-05-14 MUPID COMPUTER GESELLSCHAFT m.b.H. Procédé et dispositif pour la transmission du signal d'un point d'image
EP0330455A2 (fr) * 1988-02-22 1989-08-30 Kabushiki Kaisha Toshiba Dispositif de codage d'images

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
IEEE/IEICE GLOBAL TELECOMMUNICATION CONFERENCE 1987, Tokyo, 15th - 18th November 1987, vol. 1, pages 416-420; M. TANIMOTO et al.: "The TAT system for high quality image compression" *
See also references of WO9008443A1 *

Also Published As

Publication number Publication date
WO1990008443A1 (fr) 1990-07-26
EP0411076A4 (en) 1993-05-05

Similar Documents

Publication Publication Date Title
US5420637A (en) Dynamic image representation system
EP0358498B1 (fr) Méthode et appareil pour générer des images animées
US6593955B1 (en) Video telephony system
CA2294388C (fr) Procede de creation d&#39;images-objets destine a des systemes de codage bases sur les objets et utilisant des masques et une moyenne d&#39;arrondi
CA2432741C (fr) Optimisation de blocs de transformation
US5799113A (en) Method for expanding contracted video images
US5854856A (en) Content based video compression system
US5907626A (en) Method for object tracking and mosaicing in an image sequence using a two-dimensional mesh
AU593394B2 (en) Interpolator for television special effects system
EP0581594B1 (fr) Dispositif de commande d&#39;affichage
Lievin et al. Lip features automatic extraction
US5982452A (en) Analog video merging system for merging N video signals from N video cameras
JPH05316491A (ja) 顔画像符号化方式
McLean Structured video coding
US11895308B2 (en) Video encoding and decoding system using contextual video learning
EP0411076A1 (fr) Systeme d&#39;imagerie video
GB2265783A (en) Bandwidth reduction employing a DATV channel
Ribas-Corbera et al. Interframe interpolation of cinematic sequences
JP2828977B2 (ja) 動画像符号化装置
KR100289702B1 (ko) 모델 기반 부호화 시스템에서 파노라마 영상을 이용한 얼굴영상 정합 방법
Tekalp et al. Two-versus three-dimensional object-based video compression
JPH06187455A (ja) 動画像の顔領域抽出装置
Komatsu et al. Global motion segmentation representation for advanced digital moving image processing
Weiman Video compression by matching human perceptual channels
JPS63197186A (ja) 画像符号化伝送方法

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 19901005

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AT BE CH DE DK ES FR GB IT LI LU NL SE

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: ISIGHT, INC.

RBV Designated contracting states (corrected)

Designated state(s): DE FR GB IT

RIN1 Information on inventor provided before grant (corrected)

Inventor name: HILSENRATH, OLIVER

Inventor name: ZEEVI, YEHOSHUA Y.

Inventor name: GINOSAR, RAN

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN

A4 Supplementary search report drawn up and despatched

Effective date: 19930315

AK Designated contracting states

Kind code of ref document: A4

Designated state(s): AT BE CH DE DK ES FR GB IT LI LU NL SE

18W Application withdrawn

Withdrawal date: 19930422