WO2001091467A1 - Method and device for encoding image - Google Patents
Method and device for encoding image Download PDFInfo
- Publication number
- WO2001091467A1 WO2001091467A1 PCT/JP2001/001828 JP0101828W WO0191467A1 WO 2001091467 A1 WO2001091467 A1 WO 2001091467A1 JP 0101828 W JP0101828 W JP 0101828W WO 0191467 A1 WO0191467 A1 WO 0191467A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- encoding
- shape
- objects
- coding
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
- G06T9/20—Contour coding, e.g. using detection of edges
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/20—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
- H04N19/21—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with binary alpha-plane coding for video objects, e.g. context-based arithmetic encoding [CAE]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/20—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
- H04N19/29—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding involving scalability at the object level, e.g. video object layer [VOL]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
- H04N19/31—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability in the temporal domain
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/40—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
Definitions
- the present invention relates generally to encoding and transcoding multiplexed video objects, and more particularly to a system for controlling the encoding and transcoding of multiplexed video objects using variable time resolution.
- FDIS 14496-2 (MPEG4 Visual), see January 1998, can encode and decode objects of any shape as separate video object planes (VOPs) .
- Objects can be visual, audio, natural, synthetic, primitive, composite, or a combination thereof.
- Video objects are configured to form composite objects or "scenes.”
- the emerging emerging MPEG-4 standard is intended to enable multimedia applications, such as interactive video, where natural and synthetic materials are integrated and universally accessible. MPEG-4 allows for content-based interactions. For example, you may want to “cut and paste” a person or object that moves from one video to another. In this type of application, it is assumed that the objects in the multimedia content have been identified using some type of segmentation process.
- a network may represent a wireless channel or the Internet.
- networks are limited in capacity and contention because resources must be decomposed when content needs to be transmitted.
- Rate control is used to allocate the number of bits per coding time. Rate control ensures that the bitstream generated by the encoder satisfies the buffer constraints.
- the rate control process provides a constant bit rate while trying to maximize the quality of the encoded signal.
- frames such as MPEG-2
- US Pat. No. 5, 847, 76 issued to Uz et al. No. 1 “Method for performing rate control in a video encoder which, provides a bit budget for each frame while employing virtual buffers and virtual buffer verifiersj.
- Method for encoding based on an object such as MPEG-4 see 1 U.S. Pat. No. 5,969,764 issued to Sun and Vetro on Oct. 19, 1999
- Bitstream conversion or “transcoding” can be categorized into bitrate conversion, resolution conversion, and syntax conversion. Bit rate conversion involves bit rate scaling and conversion between a constant bit rate (CBR) and a variable bitrate (VBR). The basic function of beat scaling is to take an incoming bit stream and generate a scaled output bit stream that meets the new load constraints of the receiver.
- the bitstream scaler is a transcoder or filter that matches the source bitstream with the reception load. As shown in FIG. 7, usually, scaling can be accomplished by the transcoder 100.
- the transcoder has a decoder 110 and an encoder 120.
- the compressed input bit stream 101 is completely decoded at the input rate R in and is encoded at the new output rate R out 102 to generate the output bit stream 103.
- the output rate is lower than the input rate
- encoding the decoded bitstream is very complex. Due to the complexity, full decoding and full encoding at the transcoder is not performed, but instead transcoding is performed on the compressed or partially decoded bit stream.
- Figure 8 shows an exemplary method, in which the video bitstream is simply partially decoded.
- the macroblocks of the input bit stream 201 are subjected to variable length decoding (VLD) 210.
- the input bit stream is also delayed 220 and inverse quantized (IQ) 230
- VLD variable length decoding
- IQ inverse quantized
- the partially decoded data is analyzed 240 and a new quantizer is generated.
- the set is applied to the DCT macroblocks with the code 250.
- VLC variable length coding
- the receiver may configure the object so that all pixels in the reconstructed scene are defined. Undefined pixels in the scene can result from background and foreground objects. Alternatively, overlapping objects are sampled at different temporal resolutions, and "ho 1 es" appear in the reconstructed scene. Therefore, it was important to maintain synchronization when changing the temporal resolution of multiple objects during encoding or transcoding. To further illustrate this point, consider a scene with relatively stationary background objects (eg, blank walls) and more active foreground objects, such as moving people. Background can be encoded at a relatively low temporal resolution (eg, 10 frames per second). Foreground objects are encoded with a higher temporal resolution of 30 frames per second.
- relatively stationary background objects eg, blank walls
- Foreground objects are encoded with a higher temporal resolution of 30 frames per second.
- MPEG-7 The main application of MPEG-7 is expected to be search and retrieval applications. See "MPEG-7 Applicationsj ISO / IEC N2861, July 1999. For simple applications, the user specifies some attributes of a particular object. In this low-level representation, these attributes are May include descriptors that describe the texture, motion, and shape of a particular object.A method for representing and comparing shapes was submitted by Lin et al. On June 4, 1999. No. 09 / 326,759, “Method for Ordering Image Spaces to Represent Object Shapes”, a method for describing motion activity was submitted by Divakaran et al. On September 27, 1999. U.S. patent application Ser. No.
- these properties may represent look-ahead information that the transcoder was supposed to be inaccessible to.
- the encoder or transcoder has access to these properties because Only if the content is initially obtained from the content, i.e., the content is pre-processed and stored in a database with an associated media database.
- the information itself has syntax or semantics. Syntax information describes the physical and logical signaling aspects of content. However, semantic information refers to the conceptual meaning of content. For video sequences, syntax elements can be related to the color, shape, and motion of a particular object, while semantic elements are Can refer to information that cannot be extracted from low-level descriptors, such as time and place, names of persons in a video sequence, etc.
- Object-based encoder or transformer for video objects in scenes with variable temporal resolution It is desirable to maintain synchronization in the coder, and that such changes are desired to be identified using the video content menu.
- the present invention provides a video coding apparatus and method. Coding according to the present invention can be accomplished by an encoder or transcoder.
- the video is first divided into video objects. In the case of an encoder, this division is performed using a segmentation plane, and in the case of a transcoder, a demultiplexer is used. Used. Over time, shape features are extracted from each object. The shape feature can be obtained by measuring how the shape of each object evolves temporarily. Hamming or Hausdorff distance measurements may be used.
- the extracted shape features are combined at a rate or transcoder control unit and the temporal resolution is determined over time for each object. Temporal resolution is used to encode various video objects. If necessary, motion features and coding complexity can also be considered and make trade-offs in temporal resolution determination.
- FIG. 1 is a block diagram of a scene reconstructed from two video objects
- FIG. 2 is a block diagram of a scene reconstructed from two video objects having different temporal resolutions.
- FIG. 3 is a block diagram of the encoder according to the present invention.
- FIG. 4 is a block diagram of a transcoder according to the present invention.
- FIG. 5 is a flowchart of the encoding method according to the present invention.
- FIG. 6 is a flowchart of an exemplary encoding method used by the method of FIG. 5
- FIG. 7 is a block diagram of a conventional transcoder
- FIG. 8 is a block diagram of a conventional partial decoder / encoder. BEST MODE FOR CARRYING OUT THE INVENTION
- Temporal resolution controllers enable encoding, transcoding, and reconstruction of objects with variable and different temporal resolutions.
- One of the main advantages of object-based coding schemes is that both the spatial and temporal resolution of an object can vary independently. It is desirable to provide higher spatial quality for more interesting objects such as human faces. The same applies to temporal resolution. However, there are significant subtleties in temporal resolution. That is, synchronization between objects in the scene must be maintained so that all pixels in the reconstructed scene are defined.
- the video reconstruction of compressed video is specified by the normative part of most video standards (MPEG-1 / 2/4) and is processed by conventional decoders. Therefore, the decoder is not described herein.
- the methods and apparatus described herein are applicable to object-based encoding and transcoding systems, as well as non-real-time and real-time applications.
- the input video is not compressed during encoding, but is compressed during transcoding.
- the output video is compressed during encoding and transcoding.
- the mechanisms and approaches described herein can be seamlessly integrated into the architecture of conventional devices.
- Figure 1 shows a scene 303 divided into two video objects, a foreground object 301 and a background object 302.
- a scene can be reconstructed by combining two objects.
- the foreground object is a moving person
- the background object is a stationary wall.
- the pixels of the background object and in the initial frame define all the pixels in the scene. If these two objects are encoded with the same temporal resolution, there is no problem with the object composition during image reconstruction in the receiver. All the pixels in the reconstructed scene 303 are defined.
- problems arise when objects are encoded at different temporal resolutions.
- the background is coded at a frame rate of 15 Hz
- the foreground is coded at a frame rate of 30 Hz, twice the first rate.
- foreground objects can also be relatively stationary, but have higher internal motion than background objects.
- the foreground is rich in texture and has moving eyes, lips, and other moving facial features, while the background is a blank wall. Therefore, it is desirable to encode foreground with higher spatial and temporal resolution than background.
- the foreground object is moving with respect to the background. In the sequences 401 to 403, time elapses from left to right.
- sequence 4 0 1 is a background object encoded at a relatively low temporal resolution
- sequence 4 0 2 is a foreground object encoded at a relatively high resolution
- sequence 4 0 3 is It is a reconstructed scene.
- Sequence 403 has a hole 404 in every other frame. These holes are caused by the movement of one object if you do not update adjacent or duplicate objects.
- a hole is an uncovered area of the scene that cannot be associated with any object, and has no defined pixels. The hole disappears when the object is resynchronized (eg, every other frame).
- Shape distortion metrics A method and apparatus for controlling and making decisions on the temporal resolution of an object according to the present invention indicates the amount of shape change (distortion) in a scene.
- one shape feature measures the temporal shape difference of an object.
- the encoder may determine the amount of temporal resolution used for each object during encoding or transcoding
- the shape difference for each object is measured over time.
- the shape difference is inversely proportional to the amount of variation in temporal resolution between objects. For a fixed amount of time, a small difference indicates a larger variation, while a large difference indicates a smaller variation. If the duration during which the objects are resynchronized is greater, the stored bits can be allocated to objects that need better quality.
- the method for optimally synthesizing a time metric object works as follows. The video is sampled periodically, and the differences between the shapes of each object are found over time. If the shape difference of the object is small over time, increase the sampling period for measuring the difference. Continue increasing the sampling period until the difference is greater than the predetermined threshold D.
- a frame is output to resynchronize the video object with the difference or to determine a new frequency at which the object should be synthesized.
- the frequency may be based on an average, minimum, or intermediate time interval between synchronization frames. This frequency can be used to determine the optimal time rate for each of the various video objects.
- Shape features based on differences For simplicity, only between two scenes, ie from one frame to the next Consider the difference in shape features to. However, such features may also be associated with scenes in various cue pelels. Kurepel is defined in US Patent Application No. 09 / 546,717, filed April 11, 2000 by Vetro et al., In the Adap table Bitstream Video Delivery Systemj.
- the time controller can provide various ways to achieve the time resolution of the objects in the scene. These methods are applicable to both encoders and transcoders Hamming distance
- the first difference considered in this application is the well-known Hamming distance
- the Hamming distance is the difference between the two shapes. Measure the number of pixels First consider the binary shape, that is, the case where the segmentation (alpha) value can simply be zero or one. Refers to transparent pixels in the segmentation surface, 1 refers to the opaque pixels in the segment integrators one Chillon surface.
- the Hamming distance d is defined by the following equation,
- Hausdorff distance Another widely used shape difference measurement is the Hausdorff distance.
- the Hausdorff distance is defined as the maximum function between two sets of pixels.
- h (A, B) max x ⁇ min ⁇ d (a, b) ⁇
- a and b are the sets of two video objects A and B respectively
- d (a, b) is the Euclidean distance between these pixels.
- the above metric indicates the maximum distance of a pixel in set A to the closest pixel in set B. Because this metric is not symmetric. That is, h (A, B) is not equal to h (B, A), and a more general definition is given by:
- H (A, B) max ⁇ h (A, B), h (B, A) ⁇
- the measurement of these differences is the most accurate when calculated in a pixel-domain, but an approximation from the compressed domain Note that data overnight can also be used in the above calculations. Pixel-domain data can be easily obtained in the encoder, but decoding of shape data cannot be realized by calculation for a transcoder. Instead, the data can be approximated in some computationally efficient way.
- Shape features based on macro-programs For example, in MPEG-4, shapes are coded in various different modes and are performed at the macroblock level. For example, within a mode, a shape macroblock is coded as an opaque macroblock, a transparent macroblock, or a border macroblock.
- FIG. 3 shows an object-based encoder 500 according to the present invention.
- Encoders include switch 510, shape coder 520, motion estimator 530, motion compensator 540, motion coder 550, texture coder 560, VOP memory 570, multiplexer (MUX) 580, output buffer 590, and memory. It has an evening storage unit 591.
- the encoder also has a rate control unit (RCU) 592 for performing a QP texture analyzer, a time analyzer, a shape analyzer, and a data analyzer 593-596.
- the input to the encoder 500 is a video (input) 501 based on the object.
- a video is composed of image sequence data and a segmentation (alpha) plane that defines the boundary (shape) of each video object.
- Encoder Operation The shape coder 520 processes the shape of each object and writes the result of the shape coding to the output bit stream (output) 509 via the MUX 580 and the buffer 590.
- Shape data may also be used for motion estimator 530, motion compensator 540, and texture coder 560.
- shape data is used to extract the shape characteristics of each object.
- the objects, and their associated shape and motion features, are stored in the memory 570.
- motion estimator 530 a motion vector is determined for each macroblock.
- the motion vectors are also coded and written to the output bitstream via MUX and buffer.
- a motion compensated prediction is formed from the video object data stored in the VOP memory 570. This prediction is subtracted 541 from the input object to generate a set of residual macroblocks.
- These residual macroblocks are subjected to a texture coder 560 and the corresponding data is written to the output bitstream. Texture coding follows the QP control signal provided by the RCU.
- the RCU 592 quantization parameters (QP) is responsible for selecting the appropriate quantization parameters QP for each video object. This is done using the model This is done by estimating the corresponding quantization parameter QP according to the assigned rate budget.
- the time analysis is described in detail below. Briefly, temporal analysis involves controlling the temporal resolution of each object during coding and transcoding. In the prior art, as described above with reference to FIG. 8, in order to avoid a configuration problem, the time resolution of all video objects is the same. Therefore, in the prior art, the time resolution for various objects has not been considered independently. Also, in the prior art, the temporal analysis provided a signal to skip all video objects if the output buffer was at risk of overflow. The present invention provides a better solution.
- Shape analysis 5 955 extracts the shape features used by temporal analysis to see if variable time resolution can be achieved without problems, i.e. avoid holes even if the time coding rates of different objects are different Involved in deciding if they can do it. Shape analysis can work in real-time coding mode.
- the data is obtained from the VOP memory 570.
- FIG. 4 shows a high-level block diagram of an object-based transcoder 600 according to another embodiment of the present invention.
- the input video is already compressed.
- the transcoder 600 has a demultiplexer (DE-MUX) 601, a multiplexer (MUX) 602, and an output buffer 603.
- DE-MUX demultiplexer
- MUX multiplexer
- the transcoder 600 also has a transcoder 630 based on one or more objects operated by a transcoding control unit (TCU) 610 according to the control information 604.
- the unit TCU has a shape analyzer, a QO texture analyzer, a time analyzer, and a data analyzer 611-614.
- the compressed input bit stream 605 is divided by a demultiplexer into elementary bit streams based on one or more objects.
- the bitstream based on the object can be serial or parallel.
- the total bit rate of bit stream 605 is R in .
- the compressed output bit stream 606 from the transcoder 600 is full bit rate R. have ut , r. ut ⁇ Rin.
- the demultiplexer 601 provides one or more elementary bitstreams to each of the object-based transcoders 630, and the object-based transcoder provides the object decoder 607 to the TCU 610.
- the transcoder scales the elementary bitstream.
- the scaled bit stream is formed by the multiplexer 602 before being passed to the output buffer 603, from where it is passed to the receiver.
- Output buffer 603 also provides rate feedback information 608 to the TCU.
- the control information 604 passed to each of the transcoders is
- FIG. 5 shows the steps of a method 700 for encoding and transcoding a video input 700 according to the present invention.
- the video input 701 used in this method is an uncompressed video in the case of the encoder 500, and is a compressed video in the case of the transcoder 600.
- the video input 701 is divided into objects 711.
- shape features 721 are extracted with time from each object.
- Shape extraction may be based on distances or macroblocks, as described above.
- the motion features are selectively extracted from each object over time.
- Other features that can be extracted and considered to determine the optimal temporal resolution include coding complexity, eg, spatial complexity, DCT complexity, texture complexity, and so on.
- the extracted features are combined to determine the temporal resolution 740 to be used while encoding or transcoding the various objects 711 in step 750. Is done.
- Exemplary Encoding Scenarios FIG. 6 shows some exemplary encoding scenarios based on analyzing the evolution of a video object over time.
- the inputs are the first and second extracted object sequences 81-802.
- 0 plots shape features, for example, time-dependent (t) shape differences ( ⁇ ). Note that the object shape between times t and 2 is relatively constant. Graphs 811 and 821 selectively plot the internal motion characteristics of each of the effects over time. The first object has very high internal motion Note that the internal motion of the second object is very high, while small.
- Combiner 850 (RCU 592 or TCU 610) considers the features extracted, possibly using a maximum, sum, comparison, or other combinational function, and considers how the resulting bits should be during actual coding. Decide whether to best distribute it across the various objects. In scenario 831, the first object is not coded at all in the event [tt 2 ], and all resulting bits are allocated to the second object.
- the core of the object-based transcoder of the present invention is the adaptation of the MPEG-12 transcoder described above.
- the main difference is that the shape information is included in the bit stream, and for texture coding, a rule is provided to predict DC and AC for the inside of the block. It is also important to note that the transcoding of the texture is actually dependent on the geometry. In other words, shape data cannot simply be analyzed and ignored.
- the syntax of the standard bit stream depends on the decoding shape data. Obviously, input and output bit streams based on the object of the present invention
- MPEG-2 also does not allow for dynamic frame skipping.
- the G0P structure and the reference frame are usually fixed.
- the content 651 and the corresponding content descriptor 652 are stored in the database 650 overnight.
- the content descriptor is generated from a feature extractor 640, which receives a bitstream 605 based on the input object.
- the input bit stream is provided to the demultiplexer 601 and the transcoder, as described above.
- the message is sent to the message analyzer 614 in the TCU.
- Temporal analysis functionality The main purpose of the time controller in an object-based encoder or transcoder is to avoid the configuration problems described above with reference to Figure 2 while maintaining the quality of the configuration scene at the receiver. Is to maximize. In order to maximize the quality under these constraints, it is necessary to utilize the time redundancy in the signal as much as possible. According to most video coding schemes, temporal redundancy is eliminated in the motion compensation process. However, specifying the motion vector for every coding unit or macroblock can be more than is actually needed. In addition to the bits for the motion vector, the rest of the motion compensation difference must also be coded. The important point is that not all objects need to be coded hourly to maximize quality. Thus, these stored bits can be used at different times for still other important objects.
- the time controller uses the shape distortion metrics to indicate the amount of movement in the shape in the scene.
- This measurement may relate to scenes at various cue pelels as defined in US patent application Ser. No. 09 / 546,711.
- the time controller may provide various ways to impact the time resolution of the objects in the scene. These methods are applicable to encoders and transcoders.
- the time controller works similarly. However, observations are limited due to potential limitations, so only causality is considered. Thus, the time coding decision is made immediately.
- the extraction of the shape distortion metric can be performed on either the pixel or the compressed domain.
- tolerances can be introduced into the time control decision process. In other words, if the gain in the defined area is significant, some applications may tolerate a small amount of undefined area. In this case, a weight between [0, 1] is defined. Here, 0 means that there is no movement at the shape boundary, and 1 means that the shape boundary is completely different.
- the weights are a function of the shape distortion metrics defined above and may correspond to percentages or normalized values. On the other hand, in applications that do not consider the construction problem at all, this weighting does not exist. Rather, only overweighting (ie, 0 or 1) is valid
- the time controller provides the following effects and advantages. Determines the instant at which an object can be encoded or transcoded using a variable temporal resolution. Assign a fixed, non-uniform frame rate to the video segment object. Extracts or headlines keyframes and enables content summarization. Improve bit allocation or save bits for parts of the image (frames) where the shape of the object changes significantly. Such frames require more bits than needed for shape information. Additional bits may be needed to maintain the quality of the texture information.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2001586925A JP4786114B2 (ja) | 2000-05-26 | 2001-03-08 | 映像をコード化するための方法及び装置 |
EP01912202A EP1289301B1 (en) | 2000-05-26 | 2001-03-08 | Method and device for encoding image |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US09/579,889 US6650705B1 (en) | 2000-05-26 | 2000-05-26 | Method for encoding and transcoding multiple video objects with variable temporal resolution |
US09/579,889 | 2000-05-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2001091467A1 true WO2001091467A1 (en) | 2001-11-29 |
Family
ID=24318760
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2001/001828 WO2001091467A1 (en) | 2000-05-26 | 2001-03-08 | Method and device for encoding image |
Country Status (5)
Country | Link |
---|---|
US (1) | US6650705B1 (ja) |
EP (1) | EP1289301B1 (ja) |
JP (1) | JP4786114B2 (ja) |
CN (1) | CN1199467C (ja) |
WO (1) | WO2001091467A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2842983A1 (fr) * | 2002-07-24 | 2004-01-30 | Canon Kk | Transcodage de donnees |
WO2008123568A1 (ja) * | 2007-04-04 | 2008-10-16 | Nec Corporation | コンテンツ配信システム、コンテンツ配信方法及びそれらに用いる変換装置 |
JP2009501476A (ja) * | 2005-07-13 | 2009-01-15 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | ビデオ時間アップコンバージョンを用いた処理方法及び装置 |
JP2009253764A (ja) * | 2008-04-08 | 2009-10-29 | Fujifilm Corp | 画像処理システム、画像処理方法、およびプログラム |
US8054888B2 (en) | 2003-12-24 | 2011-11-08 | Lg Electronics Inc. | Apparatus and method for converting a codec of image data |
US8447128B2 (en) | 2008-04-07 | 2013-05-21 | Fujifilm Corporation | Image processing system |
Families Citing this family (68)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6711278B1 (en) * | 1998-09-10 | 2004-03-23 | Microsoft Corporation | Tracking semantic objects in vector image sequences |
US7260826B2 (en) * | 2000-05-31 | 2007-08-21 | Microsoft Corporation | Resource allocation in multi-stream IP network for optimized quality of service |
GB0014671D0 (en) * | 2000-06-15 | 2000-08-09 | Seos Displays Ltd | Head slaved area of interest (HSAOI) using framestore demultiplexing |
US7155067B2 (en) * | 2000-07-11 | 2006-12-26 | Eg Technology, Inc. | Adaptive edge detection and enhancement for image processing |
US7020335B1 (en) * | 2000-11-21 | 2006-03-28 | General Dynamics Decision Systems, Inc. | Methods and apparatus for object recognition and compression |
JP4534106B2 (ja) * | 2000-12-26 | 2010-09-01 | 日本電気株式会社 | 動画像符号化システム及び方法 |
WO2002071736A2 (en) * | 2001-03-05 | 2002-09-12 | Intervideo, Inc. | Systems and methods of error resilience in a video decoder |
US7321624B1 (en) * | 2001-03-16 | 2008-01-22 | Objectvideo, Inc. | Bit-rate allocation system for object-based video encoding |
US6925501B2 (en) * | 2001-04-17 | 2005-08-02 | General Instrument Corporation | Multi-rate transcoder for digital streams |
US7734997B2 (en) * | 2001-05-29 | 2010-06-08 | Sony Corporation | Transport hint table for synchronizing delivery time between multimedia content and multimedia content descriptions |
US6757648B2 (en) * | 2001-06-28 | 2004-06-29 | Microsoft Corporation | Techniques for quantization of spectral data in transcoding |
US20040013198A1 (en) * | 2001-08-31 | 2004-01-22 | Haruo Togashi | Encoding apparatus and method for encoding |
US6950464B1 (en) * | 2001-12-26 | 2005-09-27 | Cisco Technology, Inc. | Sub-picture level pass through |
KR100850705B1 (ko) * | 2002-03-09 | 2008-08-06 | 삼성전자주식회사 | 시공간적 복잡도를 고려한 적응적 동영상 부호화 방법 및그 장치 |
US8214741B2 (en) | 2002-03-19 | 2012-07-03 | Sharp Laboratories Of America, Inc. | Synchronization of video and data |
US7224731B2 (en) * | 2002-06-28 | 2007-05-29 | Microsoft Corporation | Motion estimation/compensation for screen capture video |
US7085420B2 (en) * | 2002-06-28 | 2006-08-01 | Microsoft Corporation | Text detection in continuous tone image segments |
US7072512B2 (en) * | 2002-07-23 | 2006-07-04 | Microsoft Corporation | Segmentation of digital video and images into continuous tone and palettized regions |
US7421129B2 (en) * | 2002-09-04 | 2008-09-02 | Microsoft Corporation | Image compression and synthesis for video effects |
US7292574B2 (en) * | 2002-09-30 | 2007-11-06 | Intel Corporation | Automated method for mapping constant bit-rate network traffic onto a non-constant bit-rate network |
US7558320B2 (en) * | 2003-06-13 | 2009-07-07 | Microsoft Corporation | Quality control in frame interpolation with motion analysis |
US7408986B2 (en) * | 2003-06-13 | 2008-08-05 | Microsoft Corporation | Increasing motion smoothness using frame interpolation with motion analysis |
KR100612852B1 (ko) * | 2003-07-18 | 2006-08-14 | 삼성전자주식회사 | GoF/GoP의 질감 표현 방법과, 이를 이용한GoF/GoP 검색 방법 및 장치 |
DE10335009A1 (de) * | 2003-07-23 | 2005-02-10 | Atmel Germany Gmbh | Verfahren zur drahtlosen Datenübertragung zwischen einer Basisstation und einem Transponder |
US7016409B2 (en) * | 2003-11-12 | 2006-03-21 | Sony Corporation | Apparatus and method for use in providing dynamic bit rate encoding |
US20050175099A1 (en) * | 2004-02-06 | 2005-08-11 | Nokia Corporation | Transcoder and associated system, method and computer program product for low-complexity reduced resolution transcoding |
EP1719346A1 (en) * | 2004-02-20 | 2006-11-08 | Koninklijke Philips Electronics N.V. | Method of video decoding |
US7983835B2 (en) | 2004-11-03 | 2011-07-19 | Lagassey Paul J | Modular intelligent transportation system |
US20050232497A1 (en) * | 2004-04-15 | 2005-10-20 | Microsoft Corporation | High-fidelity transcoding |
US7818444B2 (en) | 2004-04-30 | 2010-10-19 | Move Networks, Inc. | Apparatus, system, and method for multi-bitrate content streaming |
KR101042623B1 (ko) * | 2004-11-17 | 2011-06-20 | 삼성전자주식회사 | 필드 가변분할방식을 이용한 디인터레이싱방법 및 장치 |
US20060233258A1 (en) * | 2005-04-15 | 2006-10-19 | Microsoft Corporation | Scalable motion estimation |
US20060291412A1 (en) | 2005-06-24 | 2006-12-28 | Naqvi Shamim A | Associated device discovery in IMS networks |
US7724753B2 (en) | 2005-06-24 | 2010-05-25 | Aylus Networks, Inc. | Digital home networks having a control point located on a wide area network |
US7864936B2 (en) | 2005-06-24 | 2011-01-04 | Aylus Networks, Inc. | Method of avoiding or minimizing cost of stateful connections between application servers and S-CSCF nodes in an IMS network with multiple domains |
US20070011718A1 (en) * | 2005-07-08 | 2007-01-11 | Nee Patrick W Jr | Efficient customized media creation through pre-encoding of common elements |
JP2007206644A (ja) * | 2006-02-06 | 2007-08-16 | Seiko Epson Corp | 画像表示システム,画像表示方法,画像表示プログラム,記録媒体,データ処理装置,画像表示装置 |
US20070197227A1 (en) * | 2006-02-23 | 2007-08-23 | Aylus Networks, Inc. | System and method for enabling combinational services in wireless networks by using a service delivery platform |
US8155195B2 (en) * | 2006-04-07 | 2012-04-10 | Microsoft Corporation | Switching distortion metrics during motion estimation |
US8494052B2 (en) * | 2006-04-07 | 2013-07-23 | Microsoft Corporation | Dynamic selection of motion estimation search ranges and extended motion vector ranges |
US9026117B2 (en) * | 2006-05-16 | 2015-05-05 | Aylus Networks, Inc. | Systems and methods for real-time cellular-to-internet video transfer |
US8611334B2 (en) | 2006-05-16 | 2013-12-17 | Aylus Networks, Inc. | Systems and methods for presenting multimedia objects in conjunction with voice calls from a circuit-switched network |
US8432899B2 (en) | 2007-02-22 | 2013-04-30 | Aylus Networks, Inc. | Systems and methods for enabling IP signaling in wireless networks |
US20070268964A1 (en) * | 2006-05-22 | 2007-11-22 | Microsoft Corporation | Unit co-location-based motion estimation |
US9094686B2 (en) * | 2006-09-06 | 2015-07-28 | Broadcom Corporation | Systems and methods for faster throughput for compressed video data decoding |
US8380864B2 (en) * | 2006-12-27 | 2013-02-19 | Microsoft Corporation | Media stream slicing and processing load allocation for multi-user media systems |
KR100968204B1 (ko) * | 2007-01-11 | 2010-07-06 | 전자부품연구원 | 다시점 비디오 코덱에서의 영상 예측 방법 및 이를 위한프로그램을 기록한 컴퓨터로 판독 가능한 기록매체 |
US7856226B2 (en) | 2007-04-17 | 2010-12-21 | Aylus Networks, Inc. | Systems and methods for IMS user sessions with dynamic service selection |
US8457958B2 (en) | 2007-11-09 | 2013-06-04 | Microsoft Corporation | Audio transcoder using encoder-generated side information to transcode to target bit-rate |
BRPI0820720A2 (pt) * | 2007-12-11 | 2015-06-16 | Thomson Licensing | Métodos e sistemas para transcodificação dentro da cadeia de distribuição |
WO2009109940A1 (en) * | 2008-03-06 | 2009-09-11 | Nxp B.V. | Temporal fallback for high frame rate picture rate conversion |
US8164862B2 (en) * | 2008-04-02 | 2012-04-24 | Headway Technologies, Inc. | Seed layer for TMR or CPP-GMR sensor |
FR2932055B1 (fr) * | 2008-06-03 | 2010-08-13 | Thales Sa | Procede d'adaptation du debit de transmission de flux videos par pretraitement dans le domaine compresse et systeme en oeuvre le procede |
US8311115B2 (en) | 2009-01-29 | 2012-11-13 | Microsoft Corporation | Video encoding using previously calculated motion information |
US8396114B2 (en) | 2009-01-29 | 2013-03-12 | Microsoft Corporation | Multiple bit rate video encoding using variable bit rate and dynamic resolution for adaptive video streaming |
US20100309987A1 (en) * | 2009-06-05 | 2010-12-09 | Apple Inc. | Image acquisition and encoding system |
US8270473B2 (en) | 2009-06-12 | 2012-09-18 | Microsoft Corporation | Motion based dynamic resolution multiple bit rate video encoding |
US8848802B2 (en) * | 2009-09-04 | 2014-09-30 | Stmicroelectronics International N.V. | System and method for object based parametric video coding |
US10178396B2 (en) | 2009-09-04 | 2019-01-08 | Stmicroelectronics International N.V. | Object tracking |
US8705616B2 (en) | 2010-06-11 | 2014-04-22 | Microsoft Corporation | Parallel multiple bitrate video encoding to reduce latency and dependences between groups of pictures |
US9094685B2 (en) * | 2010-09-21 | 2015-07-28 | Dialogic Corporation | Efficient coding complexity estimation for video transcoding systems |
US20120281748A1 (en) * | 2011-05-02 | 2012-11-08 | Futurewei Technologies, Inc. | Rate Control for Cloud Transcoding |
EP2716041A4 (en) * | 2011-05-31 | 2014-10-15 | Dolby Lab Licensing Corp | VIDEO COMPRESSION WITH RESOLUTION COMPENSATION AND OPTIMIZATION |
US9591318B2 (en) * | 2011-09-16 | 2017-03-07 | Microsoft Technology Licensing, Llc | Multi-layer encoding and decoding |
US11089343B2 (en) | 2012-01-11 | 2021-08-10 | Microsoft Technology Licensing, Llc | Capability advertisement, configuration and control for video coding and decoding |
EP3420726B1 (en) | 2016-02-26 | 2021-05-19 | Versitech Limited | Shape-adaptive model-based codec for lossy and lossless compression of images |
US10847048B2 (en) * | 2018-02-23 | 2020-11-24 | Frontis Corp. | Server, method and wearable device for supporting maintenance of military apparatus based on augmented reality using correlation rule mining |
EP3808086A1 (en) * | 2018-08-14 | 2021-04-21 | Huawei Technologies Co., Ltd. | Machine-learning-based adaptation of coding parameters for video encoding using motion and object detection |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS63157579A (ja) * | 1986-12-22 | 1988-06-30 | Nippon Telegr & Teleph Corp <Ntt> | 疑似3次元撮像装置 |
JPH01228384A (ja) * | 1988-03-09 | 1989-09-12 | Kokusai Denshin Denwa Co Ltd <Kdd> | 領域分割を用いた動画像符号化方式 |
JPH0232688A (ja) * | 1988-07-22 | 1990-02-02 | Hitachi Ltd | 適応型変換符号化装置 |
JPH047989A (ja) * | 1989-08-02 | 1992-01-13 | Fujitsu Ltd | 画像信号符号化制御方式 |
JPH04321391A (ja) * | 1990-09-29 | 1992-11-11 | Victor Co Of Japan Ltd | 画像符号化装置 |
JPH04354489A (ja) * | 1991-05-31 | 1992-12-08 | Fujitsu Ltd | 画像符号化装置 |
JPH05111015A (ja) * | 1991-10-17 | 1993-04-30 | Sony Corp | 動き適応画像符号化装置 |
JPH0622292A (ja) * | 1992-06-30 | 1994-01-28 | Sony Corp | ディジタル画像信号の伝送装置 |
JPH07222145A (ja) * | 1994-01-31 | 1995-08-18 | Mitsubishi Electric Corp | 画像符号化装置 |
JPH07288806A (ja) * | 1994-04-20 | 1995-10-31 | Hitachi Ltd | 動画像通信システム |
JPH1185966A (ja) * | 1997-07-18 | 1999-03-30 | Sony Corp | 画像信号多重化装置および方法、画像信号逆多重化装置および方法、並びに伝送媒体 |
JPH11196411A (ja) * | 1997-10-27 | 1999-07-21 | Mitsubishi Electric Corp | 画像符号化装置、画像符号化方法、画像復号化装置、及び画像復号化方法 |
JP2000050254A (ja) * | 1998-07-17 | 2000-02-18 | Mitsubishi Electric Inf Technol Center America Inc | 改良された適応性のあるビデオ符号化方法 |
JP2000078572A (ja) * | 1998-08-31 | 2000-03-14 | Toshiba Corp | オブジェクト符号化装置およびオブジェクト符号化装置のコマ落し制御方法およびプログラムを記録した記憶媒体 |
JP2000092489A (ja) * | 1998-09-09 | 2000-03-31 | Toshiba Corp | 画像符号化装置および画像符号化方法およびプログラムを記録した媒体 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5686963A (en) | 1995-12-26 | 1997-11-11 | C-Cube Microsystems | Method for performing rate control in a video encoder which provides a bit budget for each frame while employing virtual buffers and virtual buffer verifiers |
JP3263807B2 (ja) * | 1996-09-09 | 2002-03-11 | ソニー株式会社 | 画像符号化装置および画像符号化方法 |
US5969764A (en) | 1997-02-14 | 1999-10-19 | Mitsubishi Electric Information Technology Center America, Inc. | Adaptive video coding method |
US6005980A (en) * | 1997-03-07 | 1999-12-21 | General Instrument Corporation | Motion estimation and compensation of video object planes for interlaced digital video |
AU748947C (en) * | 1998-05-04 | 2003-01-30 | General Instrument Corporation | Method and apparatus for inverse quantization of MPEG-4 video |
US6167084A (en) * | 1998-08-27 | 2000-12-26 | Motorola, Inc. | Dynamic bit allocation for statistical multiplexing of compressed and uncompressed digital video signals |
US6295371B1 (en) * | 1998-10-22 | 2001-09-25 | Xerox Corporation | Method and apparatus for image processing employing image segmentation using tokenization |
US6192080B1 (en) * | 1998-12-04 | 2001-02-20 | Mitsubishi Electric Research Laboratories, Inc. | Motion compensated digital video signal processing |
US6411724B1 (en) * | 1999-07-02 | 2002-06-25 | Koninklijke Philips Electronics N.V. | Using meta-descriptors to represent multimedia information |
-
2000
- 2000-05-26 US US09/579,889 patent/US6650705B1/en not_active Expired - Fee Related
-
2001
- 2001-03-08 WO PCT/JP2001/001828 patent/WO2001091467A1/ja active Application Filing
- 2001-03-08 JP JP2001586925A patent/JP4786114B2/ja not_active Expired - Fee Related
- 2001-03-08 CN CN01802111.5A patent/CN1199467C/zh not_active Expired - Fee Related
- 2001-03-08 EP EP01912202A patent/EP1289301B1/en not_active Expired - Lifetime
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS63157579A (ja) * | 1986-12-22 | 1988-06-30 | Nippon Telegr & Teleph Corp <Ntt> | 疑似3次元撮像装置 |
JPH01228384A (ja) * | 1988-03-09 | 1989-09-12 | Kokusai Denshin Denwa Co Ltd <Kdd> | 領域分割を用いた動画像符号化方式 |
JPH0232688A (ja) * | 1988-07-22 | 1990-02-02 | Hitachi Ltd | 適応型変換符号化装置 |
JPH047989A (ja) * | 1989-08-02 | 1992-01-13 | Fujitsu Ltd | 画像信号符号化制御方式 |
JPH04321391A (ja) * | 1990-09-29 | 1992-11-11 | Victor Co Of Japan Ltd | 画像符号化装置 |
JPH04354489A (ja) * | 1991-05-31 | 1992-12-08 | Fujitsu Ltd | 画像符号化装置 |
JPH05111015A (ja) * | 1991-10-17 | 1993-04-30 | Sony Corp | 動き適応画像符号化装置 |
JPH0622292A (ja) * | 1992-06-30 | 1994-01-28 | Sony Corp | ディジタル画像信号の伝送装置 |
JPH07222145A (ja) * | 1994-01-31 | 1995-08-18 | Mitsubishi Electric Corp | 画像符号化装置 |
JPH07288806A (ja) * | 1994-04-20 | 1995-10-31 | Hitachi Ltd | 動画像通信システム |
JPH1185966A (ja) * | 1997-07-18 | 1999-03-30 | Sony Corp | 画像信号多重化装置および方法、画像信号逆多重化装置および方法、並びに伝送媒体 |
JPH11196411A (ja) * | 1997-10-27 | 1999-07-21 | Mitsubishi Electric Corp | 画像符号化装置、画像符号化方法、画像復号化装置、及び画像復号化方法 |
JP2000050254A (ja) * | 1998-07-17 | 2000-02-18 | Mitsubishi Electric Inf Technol Center America Inc | 改良された適応性のあるビデオ符号化方法 |
JP2000078572A (ja) * | 1998-08-31 | 2000-03-14 | Toshiba Corp | オブジェクト符号化装置およびオブジェクト符号化装置のコマ落し制御方法およびプログラムを記録した記憶媒体 |
JP2000092489A (ja) * | 1998-09-09 | 2000-03-31 | Toshiba Corp | 画像符号化装置および画像符号化方法およびプログラムを記録した媒体 |
Non-Patent Citations (1)
Title |
---|
See also references of EP1289301A4 * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
FR2842983A1 (fr) * | 2002-07-24 | 2004-01-30 | Canon Kk | Transcodage de donnees |
US7260264B2 (en) | 2002-07-24 | 2007-08-21 | Canon Kabushiki Kaisha | Transcoding of data |
US8054888B2 (en) | 2003-12-24 | 2011-11-08 | Lg Electronics Inc. | Apparatus and method for converting a codec of image data |
JP2009501476A (ja) * | 2005-07-13 | 2009-01-15 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | ビデオ時間アップコンバージョンを用いた処理方法及び装置 |
WO2008123568A1 (ja) * | 2007-04-04 | 2008-10-16 | Nec Corporation | コンテンツ配信システム、コンテンツ配信方法及びそれらに用いる変換装置 |
JPWO2008123568A1 (ja) * | 2007-04-04 | 2010-07-15 | 日本電気株式会社 | コンテンツ配信システム、コンテンツ配信方法及びそれらに用いる変換装置 |
JP5013141B2 (ja) * | 2007-04-04 | 2012-08-29 | 日本電気株式会社 | コンテンツ配信システム、コンテンツ配信方法及びそれらに用いる変換装置 |
US8447128B2 (en) | 2008-04-07 | 2013-05-21 | Fujifilm Corporation | Image processing system |
JP2009253764A (ja) * | 2008-04-08 | 2009-10-29 | Fujifilm Corp | 画像処理システム、画像処理方法、およびプログラム |
Also Published As
Publication number | Publication date |
---|---|
EP1289301A1 (en) | 2003-03-05 |
US6650705B1 (en) | 2003-11-18 |
JP4786114B2 (ja) | 2011-10-05 |
CN1386376A (zh) | 2002-12-18 |
EP1289301A4 (en) | 2009-06-17 |
CN1199467C (zh) | 2005-04-27 |
EP1289301B1 (en) | 2011-08-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4786114B2 (ja) | 映像をコード化するための方法及び装置 | |
JP4601889B2 (ja) | 圧縮ビットストリームを変換するための装置及び方法 | |
KR100763181B1 (ko) | 기초계층과 향상계층의 데이터를 바탕으로 예측 정보를코딩하여 코딩율을 향상시키는 방법 및 장치 | |
US6404814B1 (en) | Transcoding method and transcoder for transcoding a predictively-coded object-based picture signal to a predictively-coded block-based picture signal | |
JP4650868B2 (ja) | 圧縮ビデオのトランスコーディング方法 | |
JP4576783B2 (ja) | データ処理方法及びデータ処理装置 | |
CN101189882B (zh) | 用于视频压缩的编码器辅助帧率上变换(ea-fruc)的方法和装置 | |
JP3979897B2 (ja) | 映像圧縮ビットストリームのトランスコーディング方法 | |
CN105721880B (zh) | 用于降低视频编码和解码中的延迟的方法和系统 | |
US6490320B1 (en) | Adaptable bitstream video delivery system | |
KR100781525B1 (ko) | 가중 평균합을 이용하여 fgs 계층을 인코딩 및디코딩하는 방법 및 장치 | |
CN110740318A (zh) | 用于视频处理和视频译码的自动自适应长期参考帧选择 | |
US7088777B2 (en) | System and method for low bit rate watercolor video | |
JP2005526457A (ja) | ビデオ・トランスコーダ | |
JP2003533067A (ja) | 基準レイヤ符号化情報の使用による改良された精細スケーラブル・ビデオに関するシステム及び方法 | |
WO2003028380A1 (en) | Method for generating a scalable encoded video bitstream with constant quality | |
JP2001112006A (ja) | レート−歪み特性推定方法 | |
Moura et al. | Retrieving quality video across heterogeneous networks. Video over wireless | |
GB2371434A (en) | Encoding and transmitting video data | |
KR20040048289A (ko) | 트랜스코딩 장치 및 방법, 이에 사용되는 타겟비트할당 및픽처 복잡도 예측 장치 및 방법 | |
JP4421734B2 (ja) | 符号化データ変換方法,及びデータ記録媒体 | |
KR20230053243A (ko) | 분산형 병렬 인코딩 방법 및 장치 | |
KR20030000310A (ko) | 영상 변환 부호화 장치 | |
JP2000092489A (ja) | 画像符号化装置および画像符号化方法およびプログラムを記録した媒体 | |
JP3652889B2 (ja) | 映像符号化方法、映像符号化装置、記録媒体、及び映像通信システム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
ENP | Entry into the national phase |
Ref country code: JP Ref document number: 2001 586925 Kind code of ref document: A Format of ref document f/p: F |
|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CN JP |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2001912202 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 018021115 Country of ref document: CN |
|
WWP | Wipo information: published in national office |
Ref document number: 2001912202 Country of ref document: EP |