EP1415402A2 - Method and apparatus for video communication over a limited bandwidth medium - Google Patents

Method and apparatus for video communication over a limited bandwidth medium

Info

Publication number
EP1415402A2
EP1415402A2 EP02757055A EP02757055A EP1415402A2 EP 1415402 A2 EP1415402 A2 EP 1415402A2 EP 02757055 A EP02757055 A EP 02757055A EP 02757055 A EP02757055 A EP 02757055A EP 1415402 A2 EP1415402 A2 EP 1415402A2
Authority
EP
European Patent Office
Prior art keywords
frames
image frame
frame
image
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP02757055A
Other languages
German (de)
French (fr)
Inventor
John W. Callaci
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
JPS International Inc
Original Assignee
JPS International Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/927,132 external-priority patent/US6700601B1/en
Application filed by JPS International Inc filed Critical JPS International Inc
Publication of EP1415402A2 publication Critical patent/EP1415402A2/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • H04N7/15Conference systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/537Motion estimation other than block-based
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/577Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/58Motion compensation with long-term prediction, i.e. the reference frame for a current frame not being the temporally closest one
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/12Systems in which the television signal is transmitted via one channel or a plurality of parallel channels, the bandwidth of each channel being less than the bandwidth of the television signal

Definitions

  • the invention relates generally to a communications system that communicate data over narrow or limited bandwidth medium. More particularly, the invention relates to a method and apparatus for transmission of video over narrow-band transmission channels, such as plain old telephone system (POTS) lines.
  • POTS plain old telephone system
  • Video telephones have existed within the marketplace for several years with little commercial success. The lack of success of videophones is attributable in part to the fact that they do not work very well. It has been problematic to transmit sound and video of acceptable quality across existing phone lines.
  • Video conferencing systems produce acceptable video and audio quality, and have met with some commercial success. Those video conferencing systems depend on wide bandwidth communication connections such as leased lines, ISDN (Integrated Services Digital
  • DSL Digital Subscriber Lines
  • the high bandwidth is necessary to produce acceptable audio and video quality.
  • the apparatus comprises a video device that is configured to capture a first image frame of an object, and further configured to capture a second image frame of the object after a defined period of time from the time the first image frame is captured.
  • the apparatus further comprises a memory, electrically connected to the video device, the memory being configured to store pixel information related to the first and second image frames.
  • the apparatus further comprises a processor, elecfrically connected to the memory, that is configured to identify a portion of the second image frame, provided that the identified portion was changed during the period of time.
  • the processor is further configured to determine a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
  • the apparatus comprises first means for capturing a first image frame of an object.
  • the apparatus further comprises second means for capturing a second image frame of the object after a defined period of time from the time the first image frame is captured.
  • the apparatus further comprises means for identifying a portion of the second image frame, provided that the identified portion was changed during the period of time.
  • the apparatus further comprises means for determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
  • the invention provides a method of creating at least one intermediate image frame based on at least two image frames in a video communications system.
  • the method comprises capturing a first image frame of an object, and capturing a second image frame of the object after a defined period of time from the time the first image frame is captured.
  • the method further comprises identifying a portion of the second image frame, provided that the identified portion was changed during the period of time.
  • the method further comprises determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
  • Figure 1 shows a standard telephone system communicating voice over telephone lines.
  • Figure 2 shows a standard telephone system that is configured to communicate voice and video over telephone lines.
  • FIG 3 illustrates graphically transmitted and created frames in accordance with one embodiment of the invention.
  • Figure 4 illustrates graphically how intermediate frames are created between two successive frames.
  • Figure 5 shows one embodiment of a math and distribution processor used in accordance with the invention.
  • Figure 6 illustrates determining motion in image portions of successive frames in accordance with the invention.
  • Figure 7 illustrates creation of intermediate frames using unchanged image portions of successive frames.
  • Figure 8 is shows one method of creating motion equations in accordance with the invention.
  • Figure 9 shows creation of intermediate frames in accordance with the invention.
  • the present invention comprises methods and apparatus for reducing bandwidth requirements for video transmission, and at the same time producing bi-directional audio and video transmission across a standard phone line.
  • Embodiments of the present invention also minimize the objectionable jerkiness which has plagued many previous video communication systems.
  • the first step in creating an acceptable bi-directional video transmission is to select a standardized frame rate.
  • a 5 frame per second (in each direction) transmission rate is chosen.
  • Five frames per second sill allow for bidirectional video and audio transmission across standard telephone lines. Since only 5 frames are sent per second at the sending end, intermediate frames are recreated at the receiving end. If five frames per second are received at the receiving end and additional 25 frames per second must be created in order to meet a 30 frame per second target frame rate.
  • the first step in creating intermediate frames is to examine a subset of successive frames to determine which elements (pixels) within the successive frames do not change.
  • the elements, which do not change need only be transmitted once, and then the areas, which do not change, may be used to recreate successive frames, without having to retransmit the same information for each frame.
  • By identifying the elements that do no change conversely the elements, which do change, are identified.
  • a picture element, which does change can have its position tracked across several successive frames. Once the changing position is known across several frames, an equation for the motion of that element across several frames can be developed. Generally it is easier to compute the motion across successive frames using a polar coordinate system. This is so because objects more naturally travel in arcs and curved motion than in rectangular coordinate type motion.
  • Figure 1 represents a standard telephone system.
  • 101 is an analog telephone line connecting a first telephone 103 to a second telephone 105.
  • telephone companies insert programmable filters 107 within telephone lines.
  • the function of these analog filters is to control the bandwidth of the telephone line.
  • a phone line may be controlled to a bandwidth of 18 KHZ.
  • the bandwidth of this type of filter, 107 can be programmed by sending a tone across the telephone line thereby commanding the filter to a certain bandwidth. In such a manner the bandwidth of the telephone line can be controlled.
  • the maximum bandwidth that is available on an analog telephone line is typically in the neighborhood of 35-48 Kilohertz.
  • the filters 107 are often used to limit the telephone line bandwidth to be compatible with older existing systems.
  • FIG 2 is an illustration of an exemplary system according to an embodiment of the present invention.
  • two telephones 203 and 205 are in communication across a telephone line 201.
  • the telephone line 201 has programmable analog filters 207 which may be commanded, via tones, to the maximum bandwidth.
  • Digital to analog converters 209 and 211 are used to convert the digital video from cameras 217 and 219 into analog telephone signals to the analog telephone line 201.
  • analog to digital converters (not shown) are used to convert the video from an analog representation, as may be transmitted across phone lines, to a video representation, in order to be processed by the system and displayed on video monitors 215 and 215.
  • 9 to 15 frames per second can be transmitted across an analog phone line, such as 201 by using current compression technology.
  • 9 to 15 frames per second generally represents all the telephone line 210 bandwidth that is available for the transmission of bi-directional video and sound. Because 9 to 15 frames of video are available an embodiment of the present invention will allow for the transmission of 5 frames per second simultaneously in both directions.
  • a protocol may be adopted whereby unchanging portions of the video are not continuously retransmitted. That is if a portion of the video does not change over N frames it need be transmitted only once. In this way bandwidth can be further conserved. Studies have shown that the actual motion across several frames comprises worst case 40% of the frame and typically far less than 40%.
  • FIG. 3 is a graphical depiction of a receiving portion of the video system. As shown in
  • FIG 3 5 frames per second 301 are received by the system. 5 frames per second are then inserted into a buffer 321, by a microprocessor or equivalent digital circuit. The frames are stored in buffer 321 as illustrated in Figure 3. Buffer 321 represents the actual video that will be displayed. In Figure 3 buffer 321 is divided into frames which are transmitted (which are preceded with the letter "t") and frames which are reconstructed (which are preceded with the letter "r"). So for example t305 is a transmitted frame, received by the system as part of the 5 frames per second received by block 301. T311 is also a frame received by the system in block 301. Intermediate frames r306.
  • r307, r308, r309, and r310 are the frames that are recreated, and these frames represent frames intermediate in time between frame t305 and frame t311.
  • two successive frames t401 and t402 are received.
  • the system then inserts t401 and t402 into buffer 411 as shown.
  • the intermediate frames r403 through r407 are then recreated using frame t401 and t402.
  • the intermediate frames, r403 through r407 may also be recreated using equations of motion derived from successive frames, as described later.
  • Figure 5 is an illustration showing an exemplary complete system.
  • a ten second input FIFO (First In First Out) buffer 501 is shown. Because in the exemplary system in Figure 5 ten seconds of input video are captured, a grand total of 50 frames will have to be saved.
  • the 10-second input FIFO buffer 501 accepts 5 frames per second from a video source. Mechanisms for generating 5 frames per second for transmission to systems, such as shown in Figure 5 are well known in the art, and the five frame per second video can be easily generated on the transmission side.
  • the video when received into the 10 second input FIFI buffer is then coupled into the programmable math and distribution logic circuit 503.
  • the programmable math and distribution logic circuit 503 is the circuitry, which actually recreates the complete frames, as well as the recreated frames, which are not broadcast. For example the programmable math and distribution logic will place a first frame 505 into video RAM. The programmable math and distribution logic circuit 503 will also place a second frame 507 into frame 2. Frame 505 may be a complete frame, along with which is transmitted an indicator of which portions of frame 1 which do not change between frame 1 and frame 4 for example. Then any portion in frame 1, which does not change between frame 1, and frame 4 can then be written into frames 2, 3, and 4, without having to be retransmitted. The programmable math and distribution logic circuit 503 can then insert the received portions of transmitted frames 1, 2, 3, and 4 the portion of the video, which is changing. Once the unchanging portions of the video have been combined with the changing portions of the video in the transmitted frame, the transmitted frames are completely rebuilt with no loss. Once frames 1, 2, 3, and 4 have been rebuilt, frames nl through nl5 can be recreated.
  • FIG. 6 is an illustration of 4 successively transmitted frames 602, 604, 606, and 608 along with circuitry to determine which portion of those frames have changed.
  • Each pixel within frames 602, 604, 606, and 608 is compare with every other pixel in the same position within the frame. For example pixel 601, is compared with pixel 603 in comparator 609.
  • the comparator 609 may be a simple hardware circuit such an exclusive OR gate. If pixel 601 is different than pixel 603 the output of comparator 609 will be a high level, whereas if pixel 601 is not different than pixel 603 the output of comparator 609 will be a low level.
  • comparator 611 Similarly if pixel 603 and pixel 605 are different comparator 611 will have a high output and a low output if the pixels are the same. Similarly if pixel 605 and pixel 607 are different comparator 613 will have a high output and a low output if the pixels are the same.
  • the output of comparators 609, 611, and 613 are then summed in a circuit 615, which may be an OR gate.
  • the output of 615 will be high if any difference between pixels 601, 603, 605 and 607 are detected.
  • the result then can be placed in a matrix 619. Within matrix 619 the first position will represent whether that portion of the video has changed in frame 602, 604, 606 or 608.
  • FIG. 7 illustrates a mechanism by which the unchanging portions of adjacent frames are used to recreate frames on the receiving end. For example if frames 701 and 702 are successively transmitted frames and a portion of frame 705 is repeated the repeated portion can simply be written into successive frames as illustrated in Figure 7.
  • the unchanging portion of 705 can be written into frame 702 and also can be written into transmitted frame 702, which is transmitted without the unchanging portion 705 and also can be written into frame 702 through 707. Because the data, which is unchanging need only be broadcast once, considerable bandwidth is saved.
  • the data, which has not changed may be data, which has not changed over several frames, or it may be material, which has not changed over successive frames.
  • a first frame may be sent in its entirety.
  • the second transmitted frame may contain only the pixels, which have changed from the first frame.
  • the third transmitted frame may contain only pixels that have changed from the second frame, etc.
  • data over several frames can be examined simultaneously. Then the data, which has not changed over several frames, can be sent in an initial frame. Both methods can yield acceptable results and either method can be implemented with minimum computation.
  • a process which may be used over 4 transmitted frames is as follows: The initial requirement is that the frame rate being received is a fixed rate, for example 5 frames per second. All received frames will then go into a stack. The 4 frames can then be inserted into video buffers. Data that doesn't change between frames can then be copied into the four incoming frames. The recreated frames must be recreated between transmitted frames. The output may be sent out to the display one frame at a time.
  • Figure 8 is a graphical illustration of developing the motion equation(s) that may be used in fabricating the recreated frames between the transdmitted frames.
  • Motion within any video frame does not proceed at random.
  • a motion that is perceived in video frames is probably more easily described in terms of polar coordinates than in terms of rectangular coordinates. Although either coordinates system may be used.
  • Motion within frames typically involves humans moving in arcs, circles and linearly. For example taking four successive frames 801, 803, 805, and 807 a point 800 is identified within those frames. Because the frames 801, 803, 805, and 807 are successive, the process of identifying a point is simplified because, unless something is moving at very high speed, point 800 will not have moved very far between frames. The position of point 800 in frames 801, 803, 805, and 807 is determined.
  • frames 901 and 913 represent transmitted frames which have been received by the system. Portions of the frames, which are not changing, have been filled in, as discussed earlier. Given the motion equation 811, frames 901 and 913 are compared. When 901 and 913 are compared the predicted location of points which are changing, from frame 901 are then calculated and inserted into frame 907. In this way frame 907 is recreated. Once frame 907 has been recreated frame 909 can be fabricated by comparing frame 907 with frame 913 and using the motion equation to compute 907.
  • frame 907 Once frame 907 has been recreated it can be compared with frame 913 using the motion equation 811 and frame 909 can be calculated. Similarly once frame 909 has been recreated frame 911 can be recreated by comparing frame 909 with frame 913. In each case the new frame is fabricated by using the motion equation 811.
  • frame 907 has been recreated frame 905 can be recreated. It can be recreated by extrapolating back from frame 907 and extrapolating forward from frame 901 as shown in Figure 9 can create it. Additionally the two extrapolation, i.e. from frame 907 and 901 may be averaged in order to create frame 905. In other words a recreated frame may be created from a l ⁇ iown frame, or several known frames by averaging the extrapolations from the several l ⁇ iown frames.
  • frames 901 and 905 individually or together may be examined in order to create frame 903.
  • intermediate recreated frames which comprise a smooth transition between transmitted and recreated frames can be fabricated.
  • Recreated frames then, by design, can display a smooth transmission between the received frames. Because the transition is smooth the video is perceived as high quality, even though a minute inspection of the successively displayed video frames may reveal that small errors are actually present when compared with actual motion.
  • Such distortions are typically inherent in the majority of cathode ray tube displays. For example in a common television picture when bright objects are displayed the picture raster invariably distorts. Television pictures therefore continually display minor changing distortions as a part of the picture. Just as these minor changing distortions are not perceived with television pictures, they are also ignored in the present described system.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A method and apparatus for creating intermediate image frame (r403-r407) based on two or more successive captured image frames for transmission over a limited bandwidth medium, such as a telephone line (201). The intermediate image frames (r403-r407) are constructed based on pixel information of portions that did change between two successive frames. The intermediate frames are constructed without the necessity of creating portions of the successive image frames (t401,t402) that did not move or change. The created intermediate frames may be combined with originally captured frames to convey video information to a user.

Description

METHOD AND APPARATUS FOR VIDEO COMMUNICATION OVER A LIMITED BANDWIDTH MEDIUM
Background of the Invention I. Field of the Invention
The invention relates generally to a communications system that communicate data over narrow or limited bandwidth medium. More particularly, the invention relates to a method and apparatus for transmission of video over narrow-band transmission channels, such as plain old telephone system (POTS) lines.
H. Description of the Related Art
Video telephones have existed within the marketplace for several years with little commercial success. The lack of success of videophones is attributable in part to the fact that they do not work very well. It has been problematic to transmit sound and video of acceptable quality across existing phone lines.
Some available video conferencing systems produce acceptable video and audio quality, and have met with some commercial success. Those video conferencing systems depend on wide bandwidth communication connections such as leased lines, ISDN (Integrated Services Digital
Network), DSL (Digital Subscriber Lines) and the like. The high bandwidth is necessary to produce acceptable audio and video quality.
The available bandwidth on standard telephone lines has been too low to support industry standard 30 frames per second video. Currently, using compression, the best performance obtainable on standard U.S. phone lines is approximately 15 video frames per second in one direction. Because 15 video frames per second is less than the persistence of the human eye, which is generally about 24 frames per second, the 15 video frames per second results in a jerky unacceptable video quality. Even with expensive compression hardware, the quality of the resultant video may be unacceptable.
There is therefore a need for video communication systems, which do not depend on expensive compression hardware and yet yield an acceptable video display when transmitted bi- directionally across standard analog phone lines.
Summary of the Invention The invention provides an apparatus for creating at least one intermediate image frame based on at least two image frames in a video communications system. In one embodiment, the apparatus comprises a video device that is configured to capture a first image frame of an object, and further configured to capture a second image frame of the object after a defined period of time from the time the first image frame is captured. The apparatus further comprises a memory, electrically connected to the video device, the memory being configured to store pixel information related to the first and second image frames. The apparatus further comprises a processor, elecfrically connected to the memory, that is configured to identify a portion of the second image frame, provided that the identified portion was changed during the period of time. The processor is further configured to determine a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame. In another embodiment, the apparatus comprises first means for capturing a first image frame of an object. The apparatus further comprises second means for capturing a second image frame of the object after a defined period of time from the time the first image frame is captured. The apparatus further comprises means for identifying a portion of the second image frame, provided that the identified portion was changed during the period of time. The apparatus further comprises means for determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
In another embodiment, the invention provides a method of creating at least one intermediate image frame based on at least two image frames in a video communications system. The method comprises capturing a first image frame of an object, and capturing a second image frame of the object after a defined period of time from the time the first image frame is captured. The method further comprises identifying a portion of the second image frame, provided that the identified portion was changed during the period of time. The method further comprises determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
Brief Description of the Drawings The features, objectives, and advantages of the invention will become more apparent from the detailed description set forth below when taken in conjunction with the drawings wherein like parts are identified with like reference numerals throughout, and wherein:
Figure 1 shows a standard telephone system communicating voice over telephone lines. Figure 2 shows a standard telephone system that is configured to communicate voice and video over telephone lines.
Figure 3 illustrates graphically transmitted and created frames in accordance with one embodiment of the invention. Figure 4 illustrates graphically how intermediate frames are created between two successive frames.
Figure 5 shows one embodiment of a math and distribution processor used in accordance with the invention. Figure 6 illustrates determining motion in image portions of successive frames in accordance with the invention.
Figure 7 illustrates creation of intermediate frames using unchanged image portions of successive frames.
Figure 8 is shows one method of creating motion equations in accordance with the invention.
Figure 9 shows creation of intermediate frames in accordance with the invention.
Detailed Description of the Invention The present invention comprises methods and apparatus for reducing bandwidth requirements for video transmission, and at the same time producing bi-directional audio and video transmission across a standard phone line.
Embodiments of the present invention also minimize the objectionable jerkiness which has plagued many previous video communication systems.
The first step in creating an acceptable bi-directional video transmission is to select a standardized frame rate. For the purposes of illustration of the inventive concepts herein a 5 frame per second (in each direction) transmission rate is chosen. Five frames per second sill allow for bidirectional video and audio transmission across standard telephone lines. Since only 5 frames are sent per second at the sending end, intermediate frames are recreated at the receiving end. If five frames per second are received at the receiving end and additional 25 frames per second must be created in order to meet a 30 frame per second target frame rate.
The first step in creating intermediate frames is to examine a subset of successive frames to determine which elements (pixels) within the successive frames do not change. The elements, which do not change, need only be transmitted once, and then the areas, which do not change, may be used to recreate successive frames, without having to retransmit the same information for each frame. By identifying the elements that do no change, conversely the elements, which do change, are identified. A picture element, which does change, can have its position tracked across several successive frames. Once the changing position is known across several frames, an equation for the motion of that element across several frames can be developed. Generally it is easier to compute the motion across successive frames using a polar coordinate system. This is so because objects more naturally travel in arcs and curved motion than in rectangular coordinate type motion. Of course since one coordinate system can be transformed to another, either coordinate system may be used. Once the equation for motion across several different received frames has been computed, it can be utilized to project the motion between successive frames. Because a point does not move by itself, the point can be used to trace the motion of objects containing that point. Various approximations can be used to track the motion of objects between received frames. Once the equation of motion has been found over several frames, the portions of the transmitted image that do change may be recreated between transmitted frames using the motion equation. Because the motion between received frames is being governed by a mathematical equation, jerkiness that is commonly associated with the transmission of individualized frames is greatly reduced.
Even if the motion that is predicted by the motion equations is slightly different than the actual motion, the amount of error between successive frames is small and is generally not perceived by the eye. In addition the mathematical relationship between successively created frames has a tendency to assure that the motion is smooth and the resultant video will appear acceptable without the jerlcy artifacts commonly associated with such video across phone line transmissions.
Figure 1 represents a standard telephone system. 101 is an analog telephone line connecting a first telephone 103 to a second telephone 105. Typically telephone companies insert programmable filters 107 within telephone lines. The function of these analog filters is to control the bandwidth of the telephone line. Typically a phone line may be controlled to a bandwidth of 18 KHZ. The bandwidth of this type of filter, 107, can be programmed by sending a tone across the telephone line thereby commanding the filter to a certain bandwidth. In such a manner the bandwidth of the telephone line can be controlled. The maximum bandwidth that is available on an analog telephone line is typically in the neighborhood of 35-48 Kilohertz. The filters 107, are often used to limit the telephone line bandwidth to be compatible with older existing systems.
Figure 2 is an illustration of an exemplary system according to an embodiment of the present invention. In the system in Figure 2, two telephones 203 and 205 are in communication across a telephone line 201. The telephone line 201 has programmable analog filters 207 which may be commanded, via tones, to the maximum bandwidth. Additionally Digital to analog converters 209 and 211 are used to convert the digital video from cameras 217 and 219 into analog telephone signals to the analog telephone line 201. In addition analog to digital converters (not shown) are used to convert the video from an analog representation, as may be transmitted across phone lines, to a video representation, in order to be processed by the system and displayed on video monitors 215 and 215. As an example, with a 56K modem, approximately 9 to 15 frames per second can be transmitted across an analog phone line, such as 201 by using current compression technology. 9 to 15 frames per second generally represents all the telephone line 210 bandwidth that is available for the transmission of bi-directional video and sound. Because 9 to 15 frames of video are available an embodiment of the present invention will allow for the transmission of 5 frames per second simultaneously in both directions. In order to conserve the limited bandwidth, a protocol may be adopted whereby unchanging portions of the video are not continuously retransmitted. That is if a portion of the video does not change over N frames it need be transmitted only once. In this way bandwidth can be further conserved. Studies have shown that the actual motion across several frames comprises worst case 40% of the frame and typically far less than 40%. By transmitting only the changing part of a frame, statistically more than 60% of the bandwidth can be saved. Therefore 5 frames per second can be transmitted bi-directionally, with bandwidth to spare. The 5 frames per second that is transmitted is then rebuilt to 30 frames per second on the receiving end. Figure 3 is a graphical depiction of a receiving portion of the video system. As shown in
Figure 3, 5 frames per second 301 are received by the system. 5 frames per second are then inserted into a buffer 321, by a microprocessor or equivalent digital circuit. The frames are stored in buffer 321 as illustrated in Figure 3. Buffer 321 represents the actual video that will be displayed. In Figure 3 buffer 321 is divided into frames which are transmitted (which are preceded with the letter "t") and frames which are reconstructed (which are preceded with the letter "r"). So for example t305 is a transmitted frame, received by the system as part of the 5 frames per second received by block 301. T311 is also a frame received by the system in block 301. Intermediate frames r306. r307, r308, r309, and r310 are the frames that are recreated, and these frames represent frames intermediate in time between frame t305 and frame t311. In Figure 4 two successive frames t401 and t402 are received. The system then inserts t401 and t402 into buffer 411 as shown. The intermediate frames r403 through r407 are then recreated using frame t401 and t402. The intermediate frames, r403 through r407 may also be recreated using equations of motion derived from successive frames, as described later. Several frames are used to develop a motion equation, representing the motion across the several frames, and then the motion equation is utilized, along with adjacent frames such as t401 and t402 in order to regenerate or recreate intermediate frames r403 through r407. The frames can then be streamed to a display 409 out of buffer 411 at a rate of 30 frames per second.
Figure 5 is an illustration showing an exemplary complete system. In Figure 5 a ten second input FIFO (First In First Out) buffer 501 is shown. Because in the exemplary system in Figure 5 ten seconds of input video are captured, a grand total of 50 frames will have to be saved. The 10-second input FIFO buffer 501 accepts 5 frames per second from a video source. Mechanisms for generating 5 frames per second for transmission to systems, such as shown in Figure 5 are well known in the art, and the five frame per second video can be easily generated on the transmission side. The video when received into the 10 second input FIFI buffer is then coupled into the programmable math and distribution logic circuit 503. The programmable math and distribution logic circuit 503 is the circuitry, which actually recreates the complete frames, as well as the recreated frames, which are not broadcast. For example the programmable math and distribution logic will place a first frame 505 into video RAM. The programmable math and distribution logic circuit 503 will also place a second frame 507 into frame 2. Frame 505 may be a complete frame, along with which is transmitted an indicator of which portions of frame 1 which do not change between frame 1 and frame 4 for example. Then any portion in frame 1, which does not change between frame 1, and frame 4 can then be written into frames 2, 3, and 4, without having to be retransmitted. The programmable math and distribution logic circuit 503 can then insert the received portions of transmitted frames 1, 2, 3, and 4 the portion of the video, which is changing. Once the unchanging portions of the video have been combined with the changing portions of the video in the transmitted frame, the transmitted frames are completely rebuilt with no loss. Once frames 1, 2, 3, and 4 have been rebuilt, frames nl through nl5 can be recreated.
The recreated frames along with the rebuilt transmitted frames can then be used to produce full motion video. Figure 6 is an illustration of 4 successively transmitted frames 602, 604, 606, and 608 along with circuitry to determine which portion of those frames have changed. Each pixel within frames 602, 604, 606, and 608 is compare with every other pixel in the same position within the frame. For example pixel 601, is compared with pixel 603 in comparator 609. The comparator 609 may be a simple hardware circuit such an exclusive OR gate. If pixel 601 is different than pixel 603 the output of comparator 609 will be a high level, whereas if pixel 601 is not different than pixel 603 the output of comparator 609 will be a low level. Similarly if pixel 603 and pixel 605 are different comparator 611 will have a high output and a low output if the pixels are the same. Similarly if pixel 605 and pixel 607 are different comparator 613 will have a high output and a low output if the pixels are the same. The output of comparators 609, 611, and 613 are then summed in a circuit 615, which may be an OR gate. The output of 615 will be high if any difference between pixels 601, 603, 605 and 607 are detected. The result then can be placed in a matrix 619. Within matrix 619 the first position will represent whether that portion of the video has changed in frame 602, 604, 606 or 608. In such a manner portions of the frames which need not be retransmitted can be identified. Conversely the portions which are changing and must be transmitted with each frame are also identified. Figure 7 illustrates a mechanism by which the unchanging portions of adjacent frames are used to recreate frames on the receiving end. For example if frames 701 and 702 are successively transmitted frames and a portion of frame 705 is repeated the repeated portion can simply be written into successive frames as illustrated in Figure 7. The unchanging portion of 705 can be written into frame 702 and also can be written into transmitted frame 702, which is transmitted without the unchanging portion 705 and also can be written into frame 702 through 707. Because the data, which is unchanging need only be broadcast once, considerable bandwidth is saved. Not only is considerable bandwidth saved, but because the unchanging portion of the video is merely written into successive frames very little processing is involved. The data, which has not changed, may be data, which has not changed over several frames, or it may be material, which has not changed over successive frames. For example a first frame may be sent in its entirety. The second transmitted frame may contain only the pixels, which have changed from the first frame. The third transmitted frame may contain only pixels that have changed from the second frame, etc. Alternatively data over several frames can be examined simultaneously. Then the data, which has not changed over several frames, can be sent in an initial frame. Both methods can yield acceptable results and either method can be implemented with minimum computation.
For example a process which may be used over 4 transmitted frames is as follows: The initial requirement is that the frame rate being received is a fixed rate, for example 5 frames per second. All received frames will then go into a stack. The 4 frames can then be inserted into video buffers. Data that doesn't change between frames can then be copied into the four incoming frames. The recreated frames must be recreated between transmitted frames. The output may be sent out to the display one frame at a time.
Figure 8 is a graphical illustration of developing the motion equation(s) that may be used in fabricating the recreated frames between the transdmitted frames.
Motion within any video frame does not proceed at random. A motion that is perceived in video frames is probably more easily described in terms of polar coordinates than in terms of rectangular coordinates. Although either coordinates system may be used. Motion within frames typically involves humans moving in arcs, circles and linearly. For example taking four successive frames 801, 803, 805, and 807 a point 800 is identified within those frames. Because the frames 801, 803, 805, and 807 are successive, the process of identifying a point is simplified because, unless something is moving at very high speed, point 800 will not have moved very far between frames. The position of point 800 in frames 801, 803, 805, and 807 is determined. Once the position of point 800 is determined in frames 801, 803, 805, and 807 computation mechanism 809 can produce a motion equation 811, which can describe the motion of point 800. The motion equation, which has been developed over 4 transmitted frames may then be used to recreate the motion off the point 800 in the frames to be recreated between transmitted frames.
In Figure 9 the process of recreating frames between successive received frames is illustrated. In Figure 9 frames 901 and 913 represent transmitted frames which have been received by the system. Portions of the frames, which are not changing, have been filled in, as discussed earlier. Given the motion equation 811, frames 901 and 913 are compared. When 901 and 913 are compared the predicted location of points which are changing, from frame 901 are then calculated and inserted into frame 907. In this way frame 907 is recreated. Once frame 907 has been recreated frame 909 can be fabricated by comparing frame 907 with frame 913 and using the motion equation to compute 907.
Once frame 907 has been recreated it can be compared with frame 913 using the motion equation 811 and frame 909 can be calculated. Similarly once frame 909 has been recreated frame 911 can be recreated by comparing frame 909 with frame 913. In each case the new frame is fabricated by using the motion equation 811.
There are several ways in which the motion equation may be used to generate new frames. For example since frame 907 has been recreated frame 905 can be recreated. It can be recreated by extrapolating back from frame 907 and extrapolating forward from frame 901 as shown in Figure 9 can create it. Additionally the two extrapolation, i.e. from frame 907 and 901 may be averaged in order to create frame 905. In other words a recreated frame may be created from a lαiown frame, or several known frames by averaging the extrapolations from the several lαiown frames.
Once frame 905 has been recreated, in like manner frames 901 and 905 individually or together may be examined in order to create frame 903. In such a way, using the motion equation, intermediate recreated frames, which comprise a smooth transition between transmitted and recreated frames can be fabricated. Recreated frames then, by design, can display a smooth transmission between the received frames. Because the transition is smooth the video is perceived as high quality, even though a minute inspection of the successively displayed video frames may reveal that small errors are actually present when compared with actual motion.
Such small discrepancies are inconsequential and are typically not perceived, because the eve perceives smooth motion. The eye does not perceive that portions of the video may be displaced and minor distortion is present.
Such distortions are typically inherent in the majority of cathode ray tube displays. For example in a common television picture when bright objects are displayed the picture raster invariably distorts. Television pictures therefore continually display minor changing distortions as a part of the picture. Just as these minor changing distortions are not perceived with television pictures, they are also ignored in the present described system.
The invention may be embodied in other specific forms without departing from the spirit or essential characteristics of the invention. The described embodiments are to be considered in all respects only illustrative and not restrictive and the scope of the invention is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning of equivalency of the claims are to be embraced within their scope.

Claims

WHAT IS CLAIMED IS:
1. An apparatus for creating at least one intermediate image frame based on at least two image frames in a video communications system, the apparatus comprising: a video device that is configured to capture a first image frame of an object, and further configured to capture a second image frame of the object after a defined period of time from the time the first image frame is captured; a memory, electrically connected to the video device, the memory being configured to store pixel information related to the first and second image frames; and a processor, electrically connected to the memory, that is configured to identify a portion of the second image frame, provided that the identified portion was changed during the period of time, wherein the processor is further configured to determine a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
2. The apparatus of Claim 1, wherein the processor is configured to identify a portion of the second image frame that remained unchanged during the predetermined period of time.
3. The apparatus of Claim 2, wherein the processor is configured to place the unchanged portion in an intermediate image frame.
4. The apparatus of Claim 3, wherein the processor is configured to place the intermediate image portion with the unchanged portion in the intermediate image frame.
5. The apparatus of Claim 4, further comprising a transmitter that communicates the first image frame to a receiver.
6. The apparatus of Claim 5, wherein the transmitter communicates to the receiver an indicator that identifies the corresponding portion of the image that was changed.
7. The apparatus of Claim 5, wherein the transmitter communicates to the receiver pixel information related to the identified portion.
8. The apparatus of Claim 5, wherein each of the transmitter and receiver comprises a telephone unit.
9. A method of creating at least one intermediate image frame based on at least two image frames in a video communications system, the method comprising: capturing a first image frame of an object; capturing a second image frame of the object after a defined period of time from the time the first image frame is captured; identifying a portion of the second image frame, provided that the identified portion was changed during the period of time; and determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
10. The method of Claim 9, wherein identifying the second image portion includes identifying a portion of the second image frame that remained unchanged during the predetermined period of time.
11. The method of Claim 10, further comprising placing the unchanged portion in an intermediate image frame.
12. The method of Claim 11, further comprising placing the intermediate image portion with the unchanged portion in the intermediate image frame.
13. The method of Claim 12, further comprising communicating the first image frame from a transmitter to a receiver.
14. The method of Claim 13, further comprising communicating to the receiver an indicator that identifies the corresponding portion of the image that was changed.
15. The method of Claim 13, further comprising communicating pixel information related to the second image portion to the receiver.
16. An apparatus for creating at least one intermediate image frame based on at least two image frames in a video communications system, the apparatus comprising: first means for capturing a first image frame of an object; second means for capturing a second image frame of the object after a defined period of time from the time the first image frame is captured; means for identifying a portion of the second image frame, provided that the identified portion was changed during the period of time; and means for determining a location of an intermediate image portion based on at least the average respective locations of the identified portion and its corresponding portion in the first frame.
17. The apparatus of Claim 16, wherein the identifying means includes means for identifying a portion of the second image frame that remained unchanged during the predetermined period of time.
18. The apparatus of Claim 17, further comprising means for placing the unchanged portion in an intermediate image frame.
19. The apparatus of Claim 18, further comprising means for placing the intermediate image portion with the unchanged portion in the intermediate image frame.
20. The apparatus of Claim 19, further comprising means for communicating the first image frame from a transmitter to a receiver.
EP02757055A 2001-08-10 2002-08-07 Method and apparatus for video communication over a limited bandwidth medium Withdrawn EP1415402A2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US09/927,132 US6700601B1 (en) 2000-08-10 2001-08-10 Method and apparatus for video communication over a limited bandwidth medium
US927132 2001-08-10
PCT/US2002/025356 WO2003015293A2 (en) 2001-08-10 2002-08-07 Method and apparatus for video communication over a limited bandwidth medium

Publications (1)

Publication Number Publication Date
EP1415402A2 true EP1415402A2 (en) 2004-05-06

Family

ID=25454237

Family Applications (1)

Application Number Title Priority Date Filing Date
EP02757055A Withdrawn EP1415402A2 (en) 2001-08-10 2002-08-07 Method and apparatus for video communication over a limited bandwidth medium

Country Status (8)

Country Link
EP (1) EP1415402A2 (en)
JP (1) JP2005517320A (en)
KR (1) KR20040035727A (en)
AU (1) AU2002323095B2 (en)
BR (1) BR0211811A (en)
CA (1) CA2455501A1 (en)
MX (1) MXPA04001139A (en)
WO (1) WO2003015293A2 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100826355B1 (en) * 2005-10-28 2008-05-02 삼성전기주식회사 Improvement apparatus of image quality in 3 panel type display system using diffraction modulation and method thereof
KR100832631B1 (en) * 2005-10-28 2008-05-27 삼성전기주식회사 Improvement apparatus of image quality in 1 panel type display system using diffraction modulation and method thereof

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01272280A (en) * 1988-04-23 1989-10-31 Miyuukomu:Kk Picture data transmitting system and displaying method
JP2001177837A (en) * 1999-12-21 2001-06-29 Matsushita Electric Ind Co Ltd Image transmission system

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of WO03015293A2 *

Also Published As

Publication number Publication date
AU2002323095B2 (en) 2007-09-13
BR0211811A (en) 2004-09-08
CA2455501A1 (en) 2003-02-20
KR20040035727A (en) 2004-04-29
WO2003015293A2 (en) 2003-02-20
WO2003015293A3 (en) 2003-12-04
JP2005517320A (en) 2005-06-09
MXPA04001139A (en) 2005-02-17

Similar Documents

Publication Publication Date Title
Haskell et al. Interframe coding of videotelephone pictures
US8089502B2 (en) Adaptive video telephone system
Steinbach et al. Standard compatible extension of H. 263 for robust video transmission in mobile environments
JP2767652B2 (en) Video teleconferencing system using aspect ratio conversion
US5389965A (en) Video telephone station having variable image clarity
US5453780A (en) Continous presence video signal combiner
US6466248B1 (en) Videoconference recording
US8237764B1 (en) Local video feedback for videoconferencing
US20040179591A1 (en) System and method for optimal transmission of a multitude of video pictures to one or more destinations
JP2004536529A (en) Method and apparatus for continuously receiving frames from a plurality of video channels and alternately transmitting individual frames containing information about each of the video channels to each of a plurality of participants in a video conference
JP3844490B2 (en) Method and apparatus for detecting and detecting errors in video images
CN111654660B (en) Video conference system coding transmission method based on image segmentation
WO1991010324A1 (en) Video telephone systems
AU2002323095B2 (en) Method and apparatus for video communication over a limited bandwidth medium
US6700601B1 (en) Method and apparatus for video communication over a limited bandwidth medium
Russell Multimedia networking performance requirements
AU2002323095A1 (en) Method and apparatus for video communication over a limited bandwidth medium
JPS62200994A (en) Animation communication system
Haskell et al. A low-bit-rate interframe coder for videotelephone
Anastassiou et al. Series/1-based videoconferencing system
US8107525B1 (en) Variable bit rate video CODEC using adaptive tracking for video conferencing
Whybray et al. A DSP based videophone for the hearing impaired using valledge processed pictures
KR960002515B1 (en) Moving picture telephone
JPH0927950A (en) Video scramble method and video scrambler
KR20030066029A (en) System and method for real time transmitting a compression moving picture based on the internet-network

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20040128

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LI LU MC NL PT SE SK TR

AX Request for extension of the european patent

Extension state: AL LT LV MK RO SI

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20090202