WO2006003611A2 - Video processing - Google Patents

Video processing Download PDF

Info

Publication number
WO2006003611A2
WO2006003611A2 PCT/IB2005/052135 IB2005052135W WO2006003611A2 WO 2006003611 A2 WO2006003611 A2 WO 2006003611A2 IB 2005052135 W IB2005052135 W IB 2005052135W WO 2006003611 A2 WO2006003611 A2 WO 2006003611A2
Authority
WO
WIPO (PCT)
Prior art keywords
depth
image signal
data
signal
video processing
Prior art date
Application number
PCT/IB2005/052135
Other languages
French (fr)
Other versions
WO2006003611A3 (en
Inventor
Richard P. Kleihorst
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to EP05752034A priority Critical patent/EP1766558A2/en
Priority to US11/570,945 priority patent/US20080279285A1/en
Priority to JP2007518789A priority patent/JP2008505522A/en
Publication of WO2006003611A2 publication Critical patent/WO2006003611A2/en
Publication of WO2006003611A3 publication Critical patent/WO2006003611A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/17Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being an image region, e.g. an object
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/124Quantisation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/102Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
    • H04N19/132Sampling, masking or truncation of coding units, e.g. adaptive resampling, frame skipping, frame interpolation or high-frequency transform coefficient masking
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/46Embedding additional information in the video signal during the compression process
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Definitions

  • the invention relates to a video processing apparatus and method, and in particular to a video compression apparatus and method.
  • Video compression techniques are commonly used for transmitting video signals more efficiently over communication channels having a limited bandwidth.
  • region based coding is proposed to enable different regions in the scene to be coded with different qualities.
  • the main objective of this technique is to send important objects with high quality, while less important regions of the scene are transmitted with lower quality.
  • the aim of the present invention is to provide an improved video processing.
  • the invention is defined by the independent claims.
  • the dependent claims define advantageous embodiments.
  • a video processing apparatus for processing an image signal having one or more regions of interest.
  • the apparatus comprises depth estimation means for determining the depth of a region in the image signal, and providing a corresponding depth signal.
  • a data compressor receives the image signal and the depth signal, and is configured to compress the image data in a particular region based on the corresponding depth signal received from the depth estimation means.
  • the invention has the advantage of being able to compress a region of the image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal.
  • a mobile communications device comprising a first imaging means for taking a first image signal, and a second imaging means for taking a second image signal.
  • the first and second imaging means are arranged to point in substantially the same direction.
  • the communications device has the advantage of being able to determine depth information in the image signal being viewed, which can then be used to dynamically compress different regions in the image signal as described above.
  • a method of processing an image signal having one or more regions of interest comprises the steps of determining the depth of a region in the image signal to provide a corresponding depth signal.
  • the depth signal is used by a data compressor for compressing the image signal, such that the image data for a particular region is compressed based on the corresponding depth of that region in the image signal.
  • Fig. 1 shows a video processing apparatus according to the present invention
  • Fig. 2 shows a typical scene
  • Figs. 3A and 3B show the images obtained in the first and second cameras of Fig. 1;
  • Fig. 4 shows a simple compression engine
  • Fig. 5 shows an alternative embodiment of the present invention.
  • Fig. 1 discloses a video processing apparatus according to the present invention.
  • a first camera 1 produces a first image signal 9, and a second camera 3 produces a second image signal 11.
  • the first image signal 9 and the second image signal 11 are offset versions of the same scene, for example relating to "right” and “left” versions of the scene as viewed through the first and second cameras, respectively.
  • a depth estimator 5 receives the first and second image signals 9, 11, and produces a depth signal 13.
  • a data compressor 7 receives an image signal from one of the cameras, for example the first camera 1, and compresses the video data in the image signal to produce a compressed image signal 14.
  • the data compression level is based on the depth signal 13 received from the depth estimator 5.
  • the apparatus can be configured to compress image data based on the assumption that the objects that are closer to the camera are more important than the objects in the background.
  • the depth signal 13 is determined based on the first image signal 9 and second image signal 11 received by the depth estimator 5.
  • the first image signal 9 and the second image signal 11 are used to determine the disparity between corresponding pixels for the same object in the left and right images.
  • the disparity is translated into a depth signal per pixel, which is used to control the degree of quantization in the data compressor 7 when compressing the normal image.
  • objects that are closer to the cameras are coded with high quality, i.e. high quantization, while objects that are further away from the cameras are subjected to lower coding, i.e. lower quantization resulting in a lower bandwidth requirement.
  • the data compressor 7 can be configured to insert data that is more easily coded in place of the true background information.
  • a flag or indicator can be inserted, which causes a receiver to insert pixel data at the receiver side.
  • Fig. 2 shows a typical scene S in which the main object 15 is found in the foreground, at a distance of about one to two meters away from the first and second cameras 1, 3.
  • the less important objects 17 are found in the background of the scene, for example at a depth of about three to four meters away from the cameras 1, 3.
  • Figs. 3A and 3B show the image signals that are seen by the first and second cameras.
  • Fig. 3A shows the image signal seen by the second camera 3, i.e. the " left” camera in the embodiment
  • Fig. 3B shows the image signal seen by the first camera 1, i.e. the "right” camera in the embodiment.
  • the disparity is inversely proportional to the distance of the object to the cameras.
  • Disparity of a specific object in a stereoscopic image is the difference in pixels between the position of the object at the left image and the position of the same object at the right image.
  • the disparity between the images seen by the first and second cameras 1, 3 will be small if the pixel relates to an object that is far away from the cameras, while the disparity will be large if the pixel relates to an object that is close to the cameras.
  • the pixel data will appear in almost the same position in both the image frames when that pixel data relates to an object that is far away from the cameras 1, 3.
  • the pixel data will appear in significantly different positions in the image frame when the pixel data relates to an object that is close to the cameras 1, 3.
  • the background object 17 is located in almost the same frame position in both image signals.
  • each pixel in the image signal is provided a depth signal, which is used to provide the quantization value for the data compressor when compressing the normal image.
  • Fig. 4 shows a simplified compression engine according to the present invention.
  • the compression engine 40 receives incoming pixel data (pixel (i, j) ; n ) from one of the cameras, and a depth signal (depth (i, j)j n ) from the depth estimator 5.
  • the incoming pixel data is quantized depending upon the depth signal for that pixel, to provide an output pixel data (pixel (i, J) 0 Ut)-
  • each pixel is compressed depending on the depth of the associated object from the cameras.
  • known variable length coding means 43 can be used to take advantage of the compressed range of these values, to provide compressed output data 45.
  • the invention is particularly suited for applications in which video data must be compressed for transmission over a communication channel having a limited bandwidth.
  • the invention is particularly suited for use in a mobile telephone.
  • a mobile telephone having first and second cameras, the first and second cameras being arranged to point in substantially the same direction.
  • the cameras can be used to determine depth information, for use in providing a depth signal for the data compression, as described above.
  • the video processing apparatus could be used to reduce the amount of video data to be stored, for example in a mobile telephone or video camera.
  • the depth value can also be measured using other means, such as "time of flight of light" from an object in a scene, or other focusing techniques for determining the depth of an object.
  • the depth of an object in a scene can be determined from successive frames of the same scene, provided an object is moving between the respective frames.
  • the invention can also be used in reverse, whereby objects in the background are treated as the more important objects, for example in securing applications in which a background scene is being monitored.
  • the invention could be used to provide the best quality at a predetermined depth from the cameras, for example if the cameras are used in a fixed location, and intended to monitor a scene that is at a predetermined distance away from the cameras.
  • Fig. 5 shows a further embodiment for realizing the invention.
  • the embodiment of Fig. 5 has first and second lenses 51, 53.
  • the first and second lenses are spaced apart along a direction perpendicular to the line of sight, and direct light to a periscopic mirror arrangement 55.
  • the periscopic mirror arrangement 55 acts to direct light from the spaced apart lenses 51, 53 to a single sensor or camera 57.
  • a "calibration" is performed to match the middle of the sensor to the mirrors.
  • the invention described in the embodiments above has the advantage of being able to compress a region of an image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal.
  • the word 'comprising' does not exclude the presence of elements or steps other than those listed in a claim.
  • any reference signs placed between parentheses shall not be construed as limiting the claim.
  • the word "a” or "an” preceding an element does not exclude the presence of a plurality of such elements.
  • the invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer.
  • the device claim enumerating several means several of these means may be embodied by one and the same item of hardware.
  • the mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

A video processing apparatus comprises a first camera (1) for producing a first image signal (9), and a second camera (3) for producing a second image signal (11). The first image signal (9) and the second image signal (11) are offset versions of the same scene, for example relating to 'right' and 'left' versions of the scene as viewed through the first and second cameras, respectively. A depth estimator (5) receives the first and second image signals (9, 11), and produces a depth signal (13) for a region in the scene. A data compressor (7) receives an image signal from one of the cameras, for example the first camera (1), and compresses the video data in the image signal to produce a compressed image signal (14). The data compression for a particular region is performed based on the depth signal (13) received from the depth estimator (5) for that region. The apparatus can be configured to compress image data for objects in the foreground with a higher resolution than objects located in the background.

Description

Video processing
FIELD OF THE INVENTION
The invention relates to a video processing apparatus and method, and in particular to a video compression apparatus and method.
BACKGROUND OF THE INVENTION
Video compression techniques are commonly used for transmitting video signals more efficiently over communication channels having a limited bandwidth. In modern day video compression techniques, such as MPEG4, region based coding is proposed to enable different regions in the scene to be coded with different qualities. The main objective of this technique is to send important objects with high quality, while less important regions of the scene are transmitted with lower quality.
"Region based Video Coding using Mathematical Morphology", Philippe Salembier et al, Proceedings of the IEEE, Vol. 83, No. 6, June 1995, discloses a region based coding in which regions in an image are segmented based on intensity, color and grey value. This has the disadvantage that it is not clear which is the significant object in the scene. Often, the significant object will be the moving object in the image.
SUMMARY OF THE INVENTION
The aim of the present invention is to provide an improved video processing. The invention is defined by the independent claims. The dependent claims define advantageous embodiments.
According to a first aspect of the present invention, there is provided a video processing apparatus for processing an image signal having one or more regions of interest.
The apparatus comprises depth estimation means for determining the depth of a region in the image signal, and providing a corresponding depth signal. A data compressor receives the image signal and the depth signal, and is configured to compress the image data in a particular region based on the corresponding depth signal received from the depth estimation means. The invention has the advantage of being able to compress a region of the image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal.
According to another aspect of the invention, there is provided a mobile communications device comprising a first imaging means for taking a first image signal, and a second imaging means for taking a second image signal. The first and second imaging means are arranged to point in substantially the same direction.
The communications device according to this aspect of the invention has the advantage of being able to determine depth information in the image signal being viewed, which can then be used to dynamically compress different regions in the image signal as described above.
According to another aspect of the invention, there is provided a method of processing an image signal having one or more regions of interest. The method comprises the steps of determining the depth of a region in the image signal to provide a corresponding depth signal. The depth signal is used by a data compressor for compressing the image signal, such that the image data for a particular region is compressed based on the corresponding depth of that region in the image signal.
BRIEF DESCRIPTION OF THE DRAWINGS For a better understanding of the invention, and to show more clearly how it may be carried into effect, reference will now be made, by way of example only, to the following drawings in which:
Fig. 1 shows a video processing apparatus according to the present invention;
Fig. 2 shows a typical scene; Figs. 3A and 3B show the images obtained in the first and second cameras of Fig. 1;
Fig. 4 shows a simple compression engine; and
Fig. 5 shows an alternative embodiment of the present invention.
DETAILED DESCRIPTION OF A PREFERRED EMBODIMENT OF THE PRESENT INVENTION
Fig. 1 discloses a video processing apparatus according to the present invention. A first camera 1 produces a first image signal 9, and a second camera 3 produces a second image signal 11. The first image signal 9 and the second image signal 11 are offset versions of the same scene, for example relating to "right" and "left" versions of the scene as viewed through the first and second cameras, respectively. A depth estimator 5 receives the first and second image signals 9, 11, and produces a depth signal 13.
A data compressor 7 receives an image signal from one of the cameras, for example the first camera 1, and compresses the video data in the image signal to produce a compressed image signal 14. The data compression level is based on the depth signal 13 received from the depth estimator 5.
For example, the apparatus can be configured to compress image data based on the assumption that the objects that are closer to the camera are more important than the objects in the background.
The depth signal 13 is determined based on the first image signal 9 and second image signal 11 received by the depth estimator 5. The first image signal 9 and the second image signal 11 are used to determine the disparity between corresponding pixels for the same object in the left and right images. Preferably, the disparity is translated into a depth signal per pixel, which is used to control the degree of quantization in the data compressor 7 when compressing the normal image.
Thus, according to the invention, objects that are closer to the cameras are coded with high quality, i.e. high quantization, while objects that are further away from the cameras are subjected to lower coding, i.e. lower quantization resulting in a lower bandwidth requirement.
Optionally, a decision can be made to completely ignore pixels relating to the insignificant parts of the scene. In such pixels, the data compressor 7 can be configured to insert data that is more easily coded in place of the true background information. Alternatively, a flag or indicator can be inserted, which causes a receiver to insert pixel data at the receiver side.
Fig. 2 shows a typical scene S in which the main object 15 is found in the foreground, at a distance of about one to two meters away from the first and second cameras 1, 3. The less important objects 17 are found in the background of the scene, for example at a depth of about three to four meters away from the cameras 1, 3.
Figs. 3A and 3B show the image signals that are seen by the first and second cameras. Fig. 3A shows the image signal seen by the second camera 3, i.e. the " left" camera in the embodiment, while Fig. 3B shows the image signal seen by the first camera 1, i.e. the "right" camera in the embodiment. As can be seen from the Figs., there is a disparity between the image signals seen by the right and left cameras. It is noted that the disparity is inversely proportional to the distance of the object to the cameras.
Disparity of a specific object in a stereoscopic image is the difference in pixels between the position of the object at the left image and the position of the same object at the right image. In other words, for a given pixel relating to a particular object, the disparity between the images seen by the first and second cameras 1, 3 will be small if the pixel relates to an object that is far away from the cameras, while the disparity will be large if the pixel relates to an object that is close to the cameras. Thus, the pixel data will appear in almost the same position in both the image frames when that pixel data relates to an object that is far away from the cameras 1, 3. Conversely, the pixel data will appear in significantly different positions in the image frame when the pixel data relates to an object that is close to the cameras 1, 3.
For example, in Figs. 3 A and 3B, the background object 17 is located in almost the same frame position in both image signals. On the other hand, there is greater disparity between the positions of the object 15, which is located in the foreground of the scene.
Various techniques for calculating the depth of an object from the images obtained from two cameras are known per se, and will not be described in greater detail in this application. These techniques include the steps of taking a specific pixel from a first image and finding the corresponding pixel in the second image. If the corresponding pixel is found, then the disparity is calculated, and a depth value assigned to that pixel.
From the above it will be seen that each pixel in the image signal is provided a depth signal, which is used to provide the quantization value for the data compressor when compressing the normal image. Fig. 4 shows a simplified compression engine according to the present invention. The compression engine 40 receives incoming pixel data (pixel (i, j) ;n) from one of the cameras, and a depth signal (depth (i, j)jn) from the depth estimator 5. The incoming pixel data is quantized depending upon the depth signal for that pixel, to provide an output pixel data (pixel (i, J)0Ut)- Thus, each pixel is compressed depending on the depth of the associated object from the cameras. Afterwards, known variable length coding means 43 can be used to take advantage of the compressed range of these values, to provide compressed output data 45.
The invention is particularly suited for applications in which video data must be compressed for transmission over a communication channel having a limited bandwidth. For example, the invention is particularly suited for use in a mobile telephone. According to this aspect of the invention, there is provided a mobile telephone having first and second cameras, the first and second cameras being arranged to point in substantially the same direction. The cameras can be used to determine depth information, for use in providing a depth signal for the data compression, as described above.
Alternatively, the video processing apparatus could be used to reduce the amount of video data to be stored, for example in a mobile telephone or video camera.
Although the preferred embodiment has been described in relation to the cameras providing "right" and "left" versions of the scene, it will be appreciated that any orientation will be possible, providing the two cameras are in a fixed position relative to one another. In addition, the depth value can also be measured using other means, such as "time of flight of light" from an object in a scene, or other focusing techniques for determining the depth of an object. Furthermore, when used with video cameras, the depth of an object in a scene can be determined from successive frames of the same scene, provided an object is moving between the respective frames. Although such an embodiment relies on knowledge about the size of the objects in a scene, it can nevertheless be useful for determining which object is in front of the other objects, thereby enabling the closest object (and hence the most important object) to be determined.
In addition, although the preferred embodiment has been described on the basis that objects in the foreground are more important than objects in the background, it will be readily appreciated that the invention can also be used in reverse, whereby objects in the background are treated as the more important objects, for example in securing applications in which a background scene is being monitored. Alternatively, the invention could be used to provide the best quality at a predetermined depth from the cameras, for example if the cameras are used in a fixed location, and intended to monitor a scene that is at a predetermined distance away from the cameras.
Fig. 5 shows a further embodiment for realizing the invention. Instead of having two separate cameras or sensors as shown in Figs. 1 and 2, the embodiment of Fig. 5 has first and second lenses 51, 53. The first and second lenses are spaced apart along a direction perpendicular to the line of sight, and direct light to a periscopic mirror arrangement 55. The periscopic mirror arrangement 55 acts to direct light from the spaced apart lenses 51, 53 to a single sensor or camera 57. Thus, the left part of the image will come from the left lens, while the right part will come from the right lens . A "calibration" is performed to match the middle of the sensor to the mirrors. The invention described in the embodiments above has the advantage of being able to compress a region of an image signal, for example relating to a particular object, based on the depth of that region in the image signal, and hence the importance of the region within the overall image signal. It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design many alternative embodiments without departing from the scope of the appended claims. The word 'comprising' does not exclude the presence of elements or steps other than those listed in a claim. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the device claim enumerating several means, several of these means may be embodied by one and the same item of hardware. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.

Claims

CLAIMS:
1. A video processing apparatus for processing an image signal having one or more regions of interest, the apparatus comprising: depth estimation means (5) for determining the depth of a region in the image signal, and providing a corresponding depth signal (13); a data compressor (7) for receiving the image signal and the depth signal (13); wherein the data compressor (7) is configured to compress the image data in a particular region based on the corresponding depth signal (13) received from the depth estimation means (5).
2. A video processing apparatus as claimed in claim 1, further comprising first and second cameras (1, 3), the first and second cameras (1,3) providing first and second image signals (9, 11) to the depth estimation means (5), for determining the depth of a region in the image signal.
3. A video processing apparatus as claimed in claim 2, wherein the depth estimation means (5) is configured to determine the depth of a region based on the disparity between the first and second image signals (9, 11).
4. A video processing apparatus as claimed in claim 1 , wherein the data compressor is adapted to vary the quantization of the data compression (7) according to the depth signal (13).
5. A video processing apparatus as claimed in claim 4, wherein the data compressor (7) is adapted to apply high quantization to a region having a small value depth signal, and lower quantization to a region having a high value depth signal.
6. A video processing apparatus as claimed in claim 1, wherein the data compression and depth is determined on a per pixel basis.
7. A video processing apparatus as claimed in claim 6, wherein the data compressor (7) is arranged to code non-significant pixels in a predetermined manner.
8. A video processing apparatus as claimed in claim 7, wherein the data compressor (7) is arranged to omit pixel data relating to non-significant pixels.
9. A video processing apparatus as claimed in claim 7, wherein the data compressor (7) is arranged to code pixel data for non-significant pixels with data requiring less bandwidth.
10. A video processing apparatus as claimed in claim 7, wherein the data compressor (7) is arranged to code pixel data for non-significant pixels with a flag for causing predetermined data to be inserted at a receiver.
11. A mobile communications device having video processing apparatus as defined in claim 1.
12. A mobile communications device as claimed in claim 11, further comprising: first imaging means (1; 51) for taking a first image signal; second imaging means (3; 53) for taking a second image signal; wherein the first and second imaging means are arranged to point in substantially the same direction.
13. A mobile communication device as claimed in claim 12, wherein the first and second imaging means comprise first and second lenses (51, 53), respectively, the first and second lenses being spaced apart along a direction perpendicular to the line of sight.
14. A mobile communications device as claimed in claim 12, wherein the first and second imaging means comprise first and second cameras (1, 3), respectively.
15. A mobile communications device as claimed in claim 12, wherein the first and second image signals are used for determining the depth of an object in the image signal.
16. A method of processing an image signal having one or more regions of interest, the method comprising the steps of: determining the depth of a region in the image signal to provide a corresponding depth signal; providing a data compressor for compressing the image signal; and compressing the data in a particular region based on the corresponding depth signal.
17. A method as claimed in claim 16, further comprising the step of providing first and second cameras (1, 3), the first and second cameras providing first and second image signals for determining the depth of a region in the image signal.
PCT/IB2005/052135 2004-07-02 2005-06-28 Video processing WO2006003611A2 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP05752034A EP1766558A2 (en) 2004-07-02 2005-06-28 Video processing
US11/570,945 US20080279285A1 (en) 2004-07-02 2005-06-28 Video Processing
JP2007518789A JP2008505522A (en) 2004-07-02 2005-06-28 Video processing

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP04103122.0 2004-07-02
EP04103122 2004-07-02

Publications (2)

Publication Number Publication Date
WO2006003611A2 true WO2006003611A2 (en) 2006-01-12
WO2006003611A3 WO2006003611A3 (en) 2006-12-21

Family

ID=35783223

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2005/052135 WO2006003611A2 (en) 2004-07-02 2005-06-28 Video processing

Country Status (5)

Country Link
US (1) US20080279285A1 (en)
EP (1) EP1766558A2 (en)
JP (1) JP2008505522A (en)
CN (1) CN1981295A (en)
WO (1) WO2006003611A2 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2442577A3 (en) * 2010-10-15 2012-05-16 Sony Corporation Information processing apparatus, informaton processing method and program
US8620099B2 (en) 2007-12-21 2013-12-31 Samsung Electronics Co., Ltd. Method, medium, and apparatus representing adaptive information of 3D depth image
EP2763420A1 (en) * 2013-02-04 2014-08-06 Sony Corporation Depth based video object coding
US9406132B2 (en) 2010-07-16 2016-08-02 Qualcomm Incorporated Vision-based quality metric for three dimensional video

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8296662B2 (en) * 2007-02-05 2012-10-23 Brother Kogyo Kabushiki Kaisha Image display device
JP5303399B2 (en) * 2009-08-18 2013-10-02 日本放送協会 Moving picture bit depth reduction apparatus and program
KR101636539B1 (en) 2009-09-10 2016-07-05 삼성전자주식회사 Apparatus and method for compressing three dimensional image
JP4764516B1 (en) * 2010-06-14 2011-09-07 シャープ株式会社 Multi-view image encoding device
WO2012050808A1 (en) * 2010-09-29 2012-04-19 Dolby Laboratories Licensing Corporation Region based asymmetric coding for 3d video compression
EP2485493A3 (en) * 2011-02-03 2013-01-02 Broadcom Corporation Method and system for error protection of 3D video

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6104837A (en) * 1996-06-21 2000-08-15 U.S. Philis Corporation Image data compression for interactive applications

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2681941B2 (en) * 1987-09-14 1997-11-26 ソニー株式会社 Image processing device
JP2701393B2 (en) * 1988-12-13 1998-01-21 日本電気株式会社 Video encoding device
JPH03230691A (en) * 1990-02-05 1991-10-14 Minolta Camera Co Ltd Digital electronic still camera
US6055330A (en) * 1996-10-09 2000-04-25 The Trustees Of Columbia University In The City Of New York Methods and apparatus for performing digital image and video segmentation and compression using 3-D depth information
JPH11112844A (en) * 1997-09-30 1999-04-23 Canon Inc Image processor, image processing method and computer readable storage medium
US7203356B2 (en) * 2002-04-11 2007-04-10 Canesta, Inc. Subject segmentation and tracking using 3D sensing technology for video compression in multimedia applications

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6104837A (en) * 1996-06-21 2000-08-15 U.S. Philis Corporation Image data compression for interactive applications

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8620099B2 (en) 2007-12-21 2013-12-31 Samsung Electronics Co., Ltd. Method, medium, and apparatus representing adaptive information of 3D depth image
US9406132B2 (en) 2010-07-16 2016-08-02 Qualcomm Incorporated Vision-based quality metric for three dimensional video
EP2442577A3 (en) * 2010-10-15 2012-05-16 Sony Corporation Information processing apparatus, informaton processing method and program
US9001188B2 (en) 2010-10-15 2015-04-07 Sony Corporation Information processing apparatus, information processing method and program
EP2763420A1 (en) * 2013-02-04 2014-08-06 Sony Corporation Depth based video object coding
CN103974067A (en) * 2013-02-04 2014-08-06 索尼公司 Enhanced video encoding using depth information

Also Published As

Publication number Publication date
CN1981295A (en) 2007-06-13
US20080279285A1 (en) 2008-11-13
JP2008505522A (en) 2008-02-21
WO2006003611A3 (en) 2006-12-21
EP1766558A2 (en) 2007-03-28

Similar Documents

Publication Publication Date Title
US20080279285A1 (en) Video Processing
KR100918480B1 (en) Stereo vision system and its processing method
EP0707428B1 (en) Motion vector differential coding method with median prediction
US9729818B2 (en) Adaptive post-processing for mobile video calling system
US6091767A (en) System for improving efficiency of video encoders
US9854167B2 (en) Signal processing device and moving image capturing device
US20080240549A1 (en) Method and apparatus for controlling dynamic depth of stereo-view or multi-view sequence images
US20100259631A1 (en) Data compression apparatus, data compression program and image-taking apparatus
JPH10257527A (en) Optimum disparity estimation for stereo video encoding
US20120154541A1 (en) Apparatus and method for producing 3d images
CN103139446A (en) Received video stabilization
US20070047642A1 (en) Video data compression
US20110129012A1 (en) Video Data Compression
KR100951637B1 (en) Stereo vision system and its video processing method
KR20180015248A (en) A video encoding method, a video decoding method, a video encoder and a video decoder
Nur Yilmaz A no reference depth perception assessment metric for 3D video
Froehlich et al. Content aware quantization: Requantization of high dynamic range baseband signals based on visual masking by noise and texture
CN111801948B (en) Image processing apparatus and method, imaging element, and imaging apparatus
Bijl et al. Visual acuity, contrast sensitivity, and range performance with compressed motion video
KR100243865B1 (en) Motion Vector
JPH0730888A (en) Moving image transmitter and moving image receiver
KR100446528B1 (en) Apparatus and method for processing stereoscopic image in portable phone
JP4764516B1 (en) Multi-view image encoding device
WO2008145560A1 (en) Method for selecting a coding data and coding device implementing said method
JPH1028274A (en) Stereoscopic image encoding device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

WWE Wipo information: entry into national phase

Ref document number: 2005752034

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2007518789

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 11570945

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Ref document number: DE

WWE Wipo information: entry into national phase

Ref document number: 200580022599.8

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 438/CHENP/2007

Country of ref document: IN

WWP Wipo information: published in national office

Ref document number: 2005752034

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 2005752034

Country of ref document: EP