WO2003049450A2 - Methods for multimedia content repurposing - Google Patents

Methods for multimedia content repurposing Download PDF

Info

Publication number
WO2003049450A2
WO2003049450A2 PCT/IB2002/005091 IB0205091W WO03049450A2 WO 2003049450 A2 WO2003049450 A2 WO 2003049450A2 IB 0205091 W IB0205091 W IB 0205091W WO 03049450 A2 WO03049450 A2 WO 03049450A2
Authority
WO
WIPO (PCT)
Prior art keywords
content
constructs
video
video content
images
Prior art date
Application number
PCT/IB2002/005091
Other languages
French (fr)
Other versions
WO2003049450A3 (en
Inventor
Radu S. Jasinschi
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/011,883 external-priority patent/US20030105880A1/en
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to EP02785800A priority Critical patent/EP1459552A2/en
Priority to JP2003550509A priority patent/JP2005512215A/en
Priority to KR10-2004-7008696A priority patent/KR20040071176A/en
Priority to AU2002351088A priority patent/AU2002351088A1/en
Publication of WO2003049450A2 publication Critical patent/WO2003049450A2/en
Publication of WO2003049450A3 publication Critical patent/WO2003049450A3/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
    • H04N21/4402Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
    • H04N21/440236Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by media transcoding, e.g. video is transformed into a slideshow of still pictures, audio is converted into text
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/186Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a colour or a chrominance component
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/20Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding
    • H04N19/21Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video object coding with binary alpha-plane coding for video objects, e.g. context-based arithmetic encoding [CAE]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/40Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using video transcoding, i.e. partial or full decoding of a coded input stream followed by re-encoding of the decoded output stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/436Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
    • H04N21/4363Adapting the video stream to a specific local network, e.g. a Bluetooth® network
    • H04N21/43637Adapting the video stream to a specific local network, e.g. a Bluetooth® network involving a wireless protocol, e.g. Bluetooth, RF or wireless LAN [IEEE 802.11]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/442Monitoring of processes or resources, e.g. detecting the failure of a recording device, monitoring the downstream bandwidth, the number of times a movie has been viewed, the storage space available from the internal hard disk
    • H04N21/44227Monitoring of local network, e.g. connection or bandwidth variations; Detecting new devices in the local network
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4621Controlling the complexity of the content stream or additional data, e.g. lowering the resolution or bit-rate of the video stream for a mobile client with a small screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics

Definitions

  • the present invention is directed, in general, to multimedia content transcoding and, more specifically, to intra- and inter-modality multimedia content transcoding for use under resource constraints of mobile devices.
  • Multimedia content may take the form of one of the three distinct modalities of audio, visual, and textual, or any combination thereof.
  • Content "re-purposing” refers generally and theoretically to re-formatting, re-scaling, and/or transcoding content by changing the content representation within a given domain, such as: from video to video, video to still graphic images, or natural pictures to cartoons in the visual domain; from natural to synthetic sound in the audio domain; and from full text to summaries in the textual domain.
  • content may be re-purposed by changing from one domain to another, such as from video to text or from audio to text.
  • a primary use of content re-purposing is to enable the processing, storage, transmission and display of multimedia information on mobile (e.g., wireless) devices. Such devices typically have very stringent limitations on processing, storage, transmission/reception and display capabilities.
  • content re-purposing a mobile device user may have constant access to multimedia information with variable quality depending upon the circumstances, and by using the best available multimedia modality.
  • Current content re-purposing implementations include primarily speech-to- text, where spoken sounds are analyzed to transform them into vowels and consonants for translation into text to be employed, for example, in answering or response (dial-in) systems. Summarization, which deals almost exclusively with textual information, is also employed.
  • the constructs are content operators that represent 2D image regions and/or 3D volumetric regions for objects within the sequence and characterized by various visual attributes, and are extracted from the video sequence by segmentation utilizing video processing techniques.
  • the constructs are employed for intra- and inter-modality transformation to accommodate resource constraints of the mobile device.
  • FIG. 1 depicts a data processing system network employing content re- purposing according to one embodiment of the present invention
  • Figs. 2A through 2C illustrate infra-modality visual content re-purposing according to one embodiment of the present invention
  • Fig. 3 illustrates inter-modality content re-purposing utilizing compact information according to one embodiment of the present invention.
  • Fig. 1 depicts a data processing system network employing content re- purposing according to one embodiment of the present invention.
  • the data processing system network 100 includes a server system 101 and a client system 102.
  • the server 101 and client 102 are wirelessly coupled and interoperable.
  • the server 101 maybe any system, such as a desktop personal computer (PC), a laptop, a "super-computer,” or any other system including a central processing unit (CPU), a local memory system, and a set of dedicated chips that perform specific signal processing operations such as convolutions, etc.
  • Data processing system 100 may include any type of wireless communications network, including video, data, voice/audio, or some combination thereof.
  • Mobile (or fixed wirelessly connected) device 102 may be, for example, a telephone, a personal digital assistant (PDA), a computer, a satellite or terrestrial television and/or radio reception system, or a set top box.
  • PDA personal digital assistant
  • Figs. 2 A through 2C illustrate infra-modality visual content re-purposing according to one embodiment of the present invention.
  • server 101 is capable of video sequence and/or static image re-purposing for content delivered to client 102.
  • a video sequence 201 is fransformed into constructs by construct generator 202.
  • the constructs describe elements of a compact video sequence representation, allowing (a) access to video sequence content information 203, synthesis of the original input video sequence 204 (or the creation of a new video sequence), and (c) compression of the video sequence 205.
  • the constructs are each a compact representation of video content information, with a small number of constructs capable of representing long video sequences.
  • a video sequence is represented by frames or fields in their uncompressed form or by video streams in their compressed form.
  • the atomic units are pixels or fields (frames) in the uncompressed form and packages in the compressed form, with the representation being unstructured with respect to video content information.
  • Video content information is mid-level visual content information given by "objects” such as two dimensional (2D) image regions or three dimensional (3D) volumetric regions characterized by various visual attributes (e.g., color, motion, shape).
  • objects such as two dimensional (2D) image regions or three dimensional (3D) volumetric regions characterized by various visual attributes (e.g., color, motion, shape).
  • the information must be segmented from the video sequence, which requires use of various image processing and/or computer vision techniques. For example, edge/shape segmentation, motion analysis (2D or 3D), or color segmentation may be employed for the segmentation process.
  • the compact representation of the segmented video content information is also important.
  • Fig. 2B illustrates segmentation and compaction, in which the input video sequence 201 is processed by segmentation and compaction units 206 and 207 to generate compact video content operators 208.
  • the content operators 208 form part of the video content construct set.
  • Another type of video content constructs is layered mosaics 209, generated by: (i) determining the relative depth information between different mosaics; and (ii) incrementally combining the relative depth information with individual frame from the input source, partial mosaics, and content operators as illustrated in Fig. 2C.
  • 2C constitute video constructs which, together with video content segmentation and compaction units 206 and 207, represent the construct generator 202 of Fig. 2A.
  • the 3D world is composed of rigid objects; those objects are distributed at different depth levels forming the scene background, which is static (or at least slowly varying) while the foreground comprises a collection of independently moving (rigid) objects; the objects have a local surface which may be approximated as a plane; and the overall scene illumination is uniform.
  • Other suitable models include the 8-para- meter perspective model. In any case, the result of registering image I k _ to image I k is image / _, .
  • image velocity is estimated for the registered images I k _ x and I k , utilizing one of many techniques including energy-based and gradient-based.
  • the resulting image velocity determines the pixel velocity of regions associated with 3D rigid objects moving in a uniform manner, and correspond to the foreground 3D objects and associated 2D image regions.
  • image regions are then segmented to determine the parts associated with the foreground objects. This results in image regions that may be appropriately post-processed to fill in gaps, with associated Alpha maps.
  • a compact set of shape templates may be generated via computational geometry techniques.
  • a simple representation is in terms of rectangular shape approximations.
  • mosaics are extended planar images encoding non- redundant information about the video sequence, coming in layers according to the associated relative depth of world regions and generated incrementally through recursive algorithms. At each step of such algorithms, comparison of the last previously- generated mosaic with the current video sequence image generates the new instance of the mosaic.
  • the generation of layered mosaics begins with a video sequence ⁇ I ,...,I N ⁇ made up of N successive frames each having an associated compact Alpha map a within ⁇ a x ,...,a N ⁇ .
  • the result of video construct generation is a set of compact video content operators, a set of layered mosaics, and ancillary information.
  • Image re-purposing is directed to reducing the complexity of the images. For example, the image may be transformed into regions of smooth value of color, brightness, texture, motion, etc.
  • equation (3) determines the "error" between the actual image and the smooth image
  • the second term determines the "smoothness” term
  • the third term is proportional to the boundary length
  • equation (3) should be appropriately discretized— i.e., approximated by a sum of terms.
  • /( ⁇ ,•) and I M (;-) denote the visual attribute being smoothed. For example, if smoothing image velocity V(; •) , then
  • the cartoonification of /(•, •) creates regions with a constant value for a given attribute. A full cartoonification is accomplished when the region boundaries are marked in black.
  • the cartoon image I c is a very simplified version of the original image that keeps the main characteristics of the original image I .
  • Visual information transformation from natural to synthetic is one important application of content re-purposing.
  • 3D meshes may be employed for transforming natural 3D objects to synthetic 3D objects; a combination of perspective and projective transformations with 2D meshes may be employed for transforming natural 3D objects to synthetic 2D objects; and 2D meshes and computational geometry tools may be employed for transforming natural 2D objects to synthetic 2D objects.
  • Audio re-purposing includes speech-to-text transformation according to known techniques, with phonemes being generated by speech recognition and then transformed from phonemes to text.
  • the phonemes should be regarded as a compact set of basic elements by which text information is generated utilizing a dictionary as described in further detail below.
  • Inter-modality content re-purposing corresponds to re-purposing multimedia information between different modalities.
  • the framework for inter-modality content re-purposing includes (i) multimedia content segmentation, (ii) template/pattern matching; (iii) use of cross-modality translation dictionaries.
  • transformations across these different modalities should follow the flow defined in equation (7). While not necessarily dictated as a content hierarchy, this patterned is necessitated by the bits required to represent the content within the various modalities.
  • One common technique for re-purposing content according to the flow defined by equation (7) is to transform all visual and audio information into textual description. Video to still image transformation is commonly performed by sub-sampling frames of a video sequence, with transformation of content information with respect to point-of-view (or perspective) being less common.
  • a description of the compact video content (video constructs) is given in the textual domain.
  • compact image content is transformed to textual description.
  • FIG. 3 illustrates inter-modality content re-purposing utilizing compact information according to one embodiment of the present invention.
  • content re- purposing across multimedia modalities is performed in the present invention using compact information (e.g., video constructs, image cartoons). Transformation between compact y elements representing a given modality utilizes a compact information format, which is important in transformation from video frames/fields to static frames or text.
  • Compact constructs 305-308 are generated as described above, with inter-modality content re-purposing employing a set of dictionaries (not separately depicted), which translate information between sets of compact content elements in different modalities.
  • Across-modality dictionaries define how the compact content information is described in a given modality, and may be textual and/or based on metadata of a either a proprietary form or employing an agreed standard (e.g., MPEG-7, TV- Anytime, and/or SIMPTE).
  • the present invention may be implemented on a continuous access content server containing content within a database, to re-purpose content for mobile access of such content.
  • the content may be re-purposed prior to any request for such content by a mobile device (e.g., when the content is loaded for access from the server) or in response to a specific request from a particular device, customizing the content to the resources available within the mobile device.
  • the present invention may be advantageously employed within wireless communications utilizing Transmission Convergence Protocol (TCP) or Radio Transmission Protocol (RTP) to provide Internet access to customized PDAs, mini-laptops, etc.
  • TCP Transmission Convergence Protocol
  • RTP Radio Transmission Protocol
  • machine usable mediums include: nonvolatile, hard-coded type mediums such as read only memories (ROMs) or erasable, electrically programmable read only memories (EEPROMs), recordable type mediums such as floppy disks, hard disk drives and compact disc read only memories (CD-ROMs) or digital versatile discs (DVDs), and transmission type mediums such as digital and analog communication links.
  • ROMs read only memories
  • EEPROMs electrically programmable read only memories
  • CD-ROMs compact disc read only memories
  • DVDs digital versatile discs

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Computer Graphics (AREA)
  • Processing Or Creating Images (AREA)
  • Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)

Abstract

Comprehensive multimedia content re-purposing relating to wireless communications employs content constructs that are compact representations of the content information. For video content, the constructs are content operators that represent 2D image regions and/or 3D volumetric regions for objects within the sequence and characterized by various visual attributes, and are extracted from the video sequence by segmentation utilizing video processing techniques. The constructs are employed for intra- and inter-modality transformation to accommodate resource constraints of the mobile device.

Description

Methods for multimedia content repurposing
This application claims priority as a continuation-in-part of U.S. Patent Application Serial No. 10/011,883 entitled distributed processing, storage and transmission of multimedia information filed December 4, 2001. the content of the above-identified application(s) is incorporated herein by reference.
The present invention is directed, in general, to multimedia content transcoding and, more specifically, to intra- and inter-modality multimedia content transcoding for use under resource constraints of mobile devices.
Multimedia content may take the form of one of the three distinct modalities of audio, visual, and textual, or any combination thereof. Content "re-purposing" refers generally and theoretically to re-formatting, re-scaling, and/or transcoding content by changing the content representation within a given domain, such as: from video to video, video to still graphic images, or natural pictures to cartoons in the visual domain; from natural to synthetic sound in the audio domain; and from full text to summaries in the textual domain. In addition, content may be re-purposed by changing from one domain to another, such as from video to text or from audio to text.
A primary use of content re-purposing is to enable the processing, storage, transmission and display of multimedia information on mobile (e.g., wireless) devices. Such devices typically have very stringent limitations on processing, storage, transmission/reception and display capabilities. Through content re-purposing, a mobile device user may have constant access to multimedia information with variable quality depending upon the circumstances, and by using the best available multimedia modality. Current content re-purposing implementations include primarily speech-to- text, where spoken sounds are analyzed to transform them into vowels and consonants for translation into text to be employed, for example, in answering or response (dial-in) systems. Summarization, which deals almost exclusively with textual information, is also employed. There is, therefore, a need in the art for improved techniques for content re- purposing directed to more general uses. To address the above-discussed deficiencies of the prior art, it is a primary object of the present invention to provide, for use in wireless communications system, comprehensive multimedia content re-purposing employing content constructs that are compact representations of the content information. For video content, the constructs are content operators that represent 2D image regions and/or 3D volumetric regions for objects within the sequence and characterized by various visual attributes, and are extracted from the video sequence by segmentation utilizing video processing techniques. The constructs are employed for intra- and inter-modality transformation to accommodate resource constraints of the mobile device.
The foregoing has outlined rather broadly the features and technical advantages of the present invention so that those skilled in the art may better understand the detailed description of the invention that follows. Additional features and advantages of the invention will be described hereinafter that form the subject of the claims of the invention. Those skilled in the art will appreciate that they may readily use the conception and the specific embodiment disclosed as a basis for modifying or designing other structures for carrying out the same purposes of the present invention. Those skilled in the art will also realize that such equivalent constructions do not depart from the spirit and scope of the invention in its broadest form.
Before undertaking the detailed description of the invention below, it maybe advantageous to set forth definitions of certain words or phrases used throughout this patent document: the terms "include" and "comprise," as well as derivatives thereof, mean inclusion without limitation; the term "or" is inclusive, meaning and/or; the phrases "associated with" and "associated therewith," as well as derivatives thereof, may mean to include, be included within, interconnect with, contain, be contained within, connect to or with, couple to or with, be communicable with, cooperate with, interleave, juxtapose, be proximate to, be bound to or with, have, have a property of, or the like; and the term "controller" means any device, system or part thereof that controls at least one operation, whether such a device is implemented in hardware, firmware, software or some combination of at least two of the same. It should be noted that the functionality associated with any particular controller may be centralized or distributed, whether locally or remotely. Definitions for certain words and phrases are provided throughout this patent document, and those of ordinary skill in the art will understand that such definitions apply in many, if not most, instances to prior as well as future uses of such defined words and phrases.
For a more complete understanding of the present invention, and the advantages thereof, reference is now made to the following descriptions taken in conjunction with the accompanying drawings, wherein like numbers designate like objects, and in which:
Fig. 1 depicts a data processing system network employing content re- purposing according to one embodiment of the present invention; Figs. 2A through 2C illustrate infra-modality visual content re-purposing according to one embodiment of the present invention; and
Fig. 3 illustrates inter-modality content re-purposing utilizing compact information according to one embodiment of the present invention.
Figs. 1 through 3, discussed below, and the various embodiments used to describe the principles of the present invention in this patent document are by way of illustration only and should not be construed in any way to limit the scope of the invention. Those skilled in the art will understand that the principles of the present invention may be implemented in any suitably arranged device.
Fig. 1 depicts a data processing system network employing content re- purposing according to one embodiment of the present invention. The data processing system network 100 includes a server system 101 and a client system 102. In the example shown, the server 101 and client 102 are wirelessly coupled and interoperable. The server 101 maybe any system, such as a desktop personal computer (PC), a laptop, a "super-computer," or any other system including a central processing unit (CPU), a local memory system, and a set of dedicated chips that perform specific signal processing operations such as convolutions, etc. Data processing system 100 may include any type of wireless communications network, including video, data, voice/audio, or some combination thereof. Mobile (or fixed wirelessly connected) device 102 may be, for example, a telephone, a personal digital assistant (PDA), a computer, a satellite or terrestrial television and/or radio reception system, or a set top box.
Those skilled in the art will recognize that the complete construction and operation of a data processing system network is not depicted in the drawings or described herein. Instead, for simplicity and clarity, only so much of the details of the construction and operation of a data processing system as are either unique to the present invention or necessary for an understanding of the present invention are shown and described. The remainder of the system may be constructed and operate in accordance with conventional practices. Figs. 2 A through 2C illustrate infra-modality visual content re-purposing according to one embodiment of the present invention. In the exemplary embodiment, server 101 is capable of video sequence and/or static image re-purposing for content delivered to client 102.
For video re-purposing, illustrated in Fig. 2 A, a video sequence 201 is fransformed into constructs by construct generator 202. The constructs describe elements of a compact video sequence representation, allowing (a) access to video sequence content information 203, synthesis of the original input video sequence 204 (or the creation of a new video sequence), and (c) compression of the video sequence 205. The constructs are each a compact representation of video content information, with a small number of constructs capable of representing long video sequences.
Construct use goes far beyond just video compression and the like. When transforming a video sequence into a set of constructs, the video sequence is actually redesigned into a new set of building blocks. In video coding, for example, a video sequence is represented by frames or fields in their uncompressed form or by video streams in their compressed form. In this representation, the atomic units are pixels or fields (frames) in the uncompressed form and packages in the compressed form, with the representation being unstructured with respect to video content information.
Video content information is mid-level visual content information given by "objects" such as two dimensional (2D) image regions or three dimensional (3D) volumetric regions characterized by various visual attributes (e.g., color, motion, shape). To generate video content information, the information must be segmented from the video sequence, which requires use of various image processing and/or computer vision techniques. For example, edge/shape segmentation, motion analysis (2D or 3D), or color segmentation may be employed for the segmentation process. Moreover, the compact representation of the segmented video content information is also important. Fig. 2B illustrates segmentation and compaction, in which the input video sequence 201 is processed by segmentation and compaction units 206 and 207 to generate compact video content operators 208. The content operators 208 form part of the video content construct set. Another type of video content constructs is layered mosaics 209, generated by: (i) determining the relative depth information between different mosaics; and (ii) incrementally combining the relative depth information with individual frame from the input source, partial mosaics, and content operators as illustrated in Fig. 2C. The compact video content operators 208 and the layered mosaics 209 in Fig.
2C constitute video constructs which, together with video content segmentation and compaction units 206 and 207, represent the construct generator 202 of Fig. 2A.
In providing an example of construct generation below, the following visual conditions are assumed: the 3D world (scene) is composed of rigid objects; those objects are distributed at different depth levels forming the scene background, which is static (or at least slowly varying) while the foreground comprises a collection of independently moving (rigid) objects; the objects have a local surface which may be approximated as a plane; and the overall scene illumination is uniform.
Given two successive frames Ik_ϊ and Ik taken from a video sequence at instants k — \ and k , respectively, the compact video content operators are generated as follows:
First, register images Ik_x and Ik by comparing the image intensities at each pixel. If Ik_x = .I(xJt_1,Λ_1) = JA<Λ> Λ) > where (**-ι>JΑ-ι) ∞d fa*,?*) represent the x and y coordinate image pixels at instants k - \ and k , respectively, then images Ik_x and Ik are registered by computing the nine elements of a 3X3 matrix R(-,-) such that:
4-1 R(2,0) x xk +R(2,ϊ) yk +R(2,2) '
_ i?(l,0) x ^ +i?(l,l) x y, + i?(l,2) yk" R(2, 0) x xk + R(2, l) x yk +R(2,2) '
Matrix R(; •) may be calculated in different ways, such as by using the (6 parameter) affine model with R(2, 0) = R(2, 1) = 0 , R(2, 2) = 1 , R(0, 0) = sx , R(0, \) = rx , R(l, 0) = ry > and iJ(l,l) = sy , where sx > sy , rx and ry denote the x and y components of the scaling and rotation (2D) image vectors s and , respectively. Other suitable models include the 8-para- meter perspective model. In any case, the result of registering image Ik_ to image Ik is image / _, .
Next, image velocity is estimated for the registered images Ik_x and Ik , utilizing one of many techniques including energy-based and gradient-based. The resulting image velocity determines the pixel velocity of regions associated with 3D rigid objects moving in a uniform manner, and correspond to the foreground 3D objects and associated 2D image regions.
Based on the results of image velocity estimation and other visual attributes, image regions are then segmented to determine the parts associated with the foreground objects. This results in image regions that may be appropriately post-processed to fill in gaps, with associated Alpha maps.
From the image regions, a compact set of shape templates may be generated via computational geometry techniques. A simple representation is in terms of rectangular shape approximations. For example, mosaics are extended planar images encoding non- redundant information about the video sequence, coming in layers according to the associated relative depth of world regions and generated incrementally through recursive algorithms. At each step of such algorithms, comparison of the last previously- generated mosaic with the current video sequence image generates the new instance of the mosaic. In general terms, the generation of layered mosaics begins with a video sequence {I ,...,IN} made up of N successive frames each having an associated compact Alpha map a within {ax,...,aN} . Each
Alpha map is obtained from the compact video content operator by filing in the interior of mosaic regions, and is a binary image with ones in the interior region and zeros elsewhere. Assuming that information about relative depth—that is, the relative order of each foreground object and also of all of the foreground objects with respect to the background image-has been employed and that the discrimination between each mosaic plane within L levels is possible, a set of L mosaics (Φ1 ,...,ΦL} , where the i h mosaic Φ' is computed in an initial step r = l by Φ[ = a[Ix and in subsequent steps r = 2,...,N by recursively combining the set of Alpha maps {a2',..., N' } with {7,,...,/^} , thereby generating Φ^. for each step r . Finally, a determination is made of any ancillary information complementing the image regions and shape templates and describing image velocity and other visual attributes necessary to fully represent the video content information.
The result of video construct generation is a set of compact video content operators, a set of layered mosaics, and ancillary information. Image re-purposing is directed to reducing the complexity of the images. For example, the image may be transformed into regions of smooth value of color, brightness, texture, motion, etc. One possible general technique for this task is to minimize the cost function: E(I,T) = -IM(x,y)fώ dy + dxdy + v\r\, (3)
Figure imgf000008_0001
Figure imgf000008_0002
where /(-,) denotes the image region R to be estimated, IM(-,-) denotes the actual (raw) image, and
Figure imgf000008_0003
Actually, the image region R = U,-^ + T , and the overall boundary T encircles the entire region R . The first term in equation (3) determines the "error" between the actual image and the smooth image, the second term determines the "smoothness" term, and the third term is proportional to the boundary length |r| , where v is a constant. For actuation implementation, equation (3) should be appropriately discretized— i.e., approximated by a sum of terms. In analyzing equation (3), it should be noted that /(,•) and IM(;-) denote the visual attribute being smoothed. For example, if smoothing image velocity V(; •) , then
1 (;-) ≡ ¥(; ) , etc.
The image may alternatively be transformed into a cartoon image Ic by utilizing a simplified version of equation (3) in which /(-, •) is restricted to piecewise constant values /(•,•) >K . More precisely, for each region R. the value of /(•,•) is approximated as /,■(,) = Ks , Kt has a constant real value inside region Rt . If μ is a constant, then equation (3) may be approximated by -2E(/, r) = , (5)
Figure imgf000008_0004
y where v = —r . It can be seen that μ
\\lM(x,y)dxdy K, = mean (IM) = . (6)
Figure imgf000008_0005
The cartoonification of /(•, •) creates regions with a constant value for a given attribute. A full cartoonification is accomplished when the region boundaries are marked in black. The cartoon image Ic is a very simplified version of the original image that keeps the main characteristics of the original image I . Visual information transformation from natural to synthetic is one important application of content re-purposing. 3D meshes may be employed for transforming natural 3D objects to synthetic 3D objects; a combination of perspective and projective transformations with 2D meshes may be employed for transforming natural 3D objects to synthetic 2D objects; and 2D meshes and computational geometry tools may be employed for transforming natural 2D objects to synthetic 2D objects. Audio re-purposing includes speech-to-text transformation according to known techniques, with phonemes being generated by speech recognition and then transformed from phonemes to text. In the present invention, the phonemes should be regarded as a compact set of basic elements by which text information is generated utilizing a dictionary as described in further detail below. Inter-modality content re-purposing corresponds to re-purposing multimedia information between different modalities. In general, the framework for inter-modality content re-purposing includes (i) multimedia content segmentation, (ii) template/pattern matching; (iii) use of cross-modality translation dictionaries. When dealing with multimedia information, an overall hierarchy with respect to complexity exists in the three constituent elements (visual, audio and textual) as follows:
Visual (video) — Visual (pictures) -» Audio - Text. (7)
Therefore, transformations across these different modalities should follow the flow defined in equation (7). While not necessarily dictated as a content hierarchy, this patterned is necessitated by the bits required to represent the content within the various modalities. One common technique for re-purposing content according to the flow defined by equation (7) is to transform all visual and audio information into textual description. Video to still image transformation is commonly performed by sub-sampling frames of a video sequence, with transformation of content information with respect to point-of-view (or perspective) being less common. In the present invention, when transforming video to text, a description of the compact video content (video constructs) is given in the textual domain. Similarly, compact image content is transformed to textual description. In video to image transformation, specific regions (information) of the video constructs are accessed by applying compact image content operators to those regions. Fig. 3 illustrates inter-modality content re-purposing utilizing compact information according to one embodiment of the present invention. Generally, content re- purposing across multimedia modalities is performed in the present invention using compact information (e.g., video constructs, image cartoons). Transformation between compact y elements representing a given modality utilizes a compact information format, which is important in transformation from video frames/fields to static frames or text.
Within system 300, separate video, audio and textual inputs 301-303 are employed, with an additional input 304 employed for still images from either an independent input or sub-sampled from video input 301. Compact constructs 305-308 are generated as described above, with inter-modality content re-purposing employing a set of dictionaries (not separately depicted), which translate information between sets of compact content elements in different modalities. Across-modality dictionaries define how the compact content information is described in a given modality, and may be textual and/or based on metadata of a either a proprietary form or employing an agreed standard (e.g., MPEG-7, TV- Anytime, and/or SIMPTE). Translation between elements of different modalities should be performed using these descriptions, which are particularly suited for transforming from video to images. When transforming from video, images or audio to text, the descriptions represent explanations that may be realized at different levels of details. The structure and functionality of dictionaries of this type are described in greater detail in the cross-referenced application identified above and incorporated herein by reference.
The present invention may be implemented on a continuous access content server containing content within a database, to re-purpose content for mobile access of such content. The content may be re-purposed prior to any request for such content by a mobile device (e.g., when the content is loaded for access from the server) or in response to a specific request from a particular device, customizing the content to the resources available within the mobile device. In particular, the present invention may be advantageously employed within wireless communications utilizing Transmission Convergence Protocol (TCP) or Radio Transmission Protocol (RTP) to provide Internet access to customized PDAs, mini-laptops, etc.
It is important to note that while the present invention has been described in the context of a fully functional system, those skilled in the art will appreciate that at least portions of the mechanism of the present invention are capable of being distributed in the form of a machine usable medium containing instructions in a variety of forms, and that the present invention applies equally regardless of the particular type of signal bearing medium utilized to actually carry out the distribution. Examples of machine usable mediums include: nonvolatile, hard-coded type mediums such as read only memories (ROMs) or erasable, electrically programmable read only memories (EEPROMs), recordable type mediums such as floppy disks, hard disk drives and compact disc read only memories (CD-ROMs) or digital versatile discs (DVDs), and transmission type mediums such as digital and analog communication links.
Although the present invention has been described in detail, those skilled in the art will understand that various changes, substitutions, variations, enhancements, nuances, gradations, lesser forms, alterations, revisions, improvements and knock-offs of the invention disclosed herein may be made without departing from the spirit and scope of the invention in its broadest form.

Claims

CLA S:
1. A system 100 for multimedia content re-purposing comprising a controller 200, 300 generating content constructs 208 that are compact representations of content information and include content operators for video content, wherein the content constructs 208 are suitable for transformation of the content within a modality or between modalities.
2. The system 100 according to claim 1, wherein visual content operators 306 for the video content are formed by registering consecutive images within a sequence, estimating image velocity for the registered images, segmenting image regions to identify foreground objects and generating shape templates.
3. The system 100 according to claim 1, wherein the content constructs 208 include layered mosaics 209 for the video content that are derived from Alpha maps for consecutive images recursively combined by cut and paste operations.
4. The system 100 according to claim 1, wherein the content constructs 208 for the video content are employed to transform images within the video content into regions of smooth value for one or more of color, brightness, texture, and motion.
5. The system 100 according to claim 1, wherein the content constructs 208 for the video content are employed to fransform images within the video content into cartoonlike images.
6. A system 100 for multimedia content re-purposing comprising:
- a mobile device 102 capable of selectively accessing multimedia content; and - a server 101 containing multimedia content for transmission to mobile devices, the server 101 comprising a controller 200 generating content constructs 208 that are compact representations of content information and include content operators for video content, wherein the content constructs 208 are suitable for transformation of the content within a modality or between modalities.
7. A method of multimedia content re-purposing comprising generating content constructs 208 that are compact representations of content information and include content operators for video content, wherein the content constructs 208 are suitable for transformation of the content within a modality or between modalities.
8. The method according to claim 7, wherein the step of generating the content operators for video content further comprises:
- registering consecutive images within a sequence; - estimating image velocity for the registered images;
- segmenting image regions to identify foreground objects and generating shape templates.
9. The method according to claim 7, wherein the step of generating the content constructs 208 for video content further comprises recursively combining Alpha maps for consecutive images by cut and paste operations to form layered mosaics 209.
10. The method according to claim 7, further comprising employing the content constructs 208 for the video content to transform images within the video content into regions of smooth value for one or more of color, brightness, texture, and motion.
11. The method according to claim 7, further comprising employing the content constructs 208 for the video content to transform images within the video content into cartoon-like images.
12. A signal comprising multimedia information generated from content constructs 208 that are compact representations of content information and include content operators for video content, wherein the content constructs 208 are suitable for transformation of the content within a modality or between modalities.
13. The signal according to claim 12, wherein visual content operators for the video content are formed by registering consecutive images within a sequence, estimating image velocity for the registered images, segmenting image regions to identify foreground objects and generating shape templates.
14. The signal according to claim 12, wherein the content constructs 208 include layered mosaics for the video content are derived from Alpha maps for consecutive images recursively combined by cut and paste operations.
15. The signal according to claim 12, wherein the content constructs 208 for the video content are employed to transform images within the video content into regions of smooth value for one or more of color, brightness, texture, and motion.
16. The signal according to claim 12, wherein the content constructs 208 for the video content are employed to transform images within the video content into cartoons.
PCT/IB2002/005091 2001-12-04 2002-12-02 Methods for multimedia content repurposing WO2003049450A2 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
EP02785800A EP1459552A2 (en) 2001-12-04 2002-12-02 Methods for multimedia content repurposing
JP2003550509A JP2005512215A (en) 2001-12-04 2002-12-02 Multimedia content re-purpose processing method
KR10-2004-7008696A KR20040071176A (en) 2001-12-04 2002-12-02 Methods for multimedia content repurposing
AU2002351088A AU2002351088A1 (en) 2001-12-04 2002-12-02 Methods for multimedia content repurposing

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US10/011,883 US20030105880A1 (en) 2001-12-04 2001-12-04 Distributed processing, storage, and transmision of multimedia information
US10/011,883 2001-12-04
US10/265,582 2002-10-07
US10/265,582 US7305618B2 (en) 2001-12-04 2002-10-07 Methods for multimedia content repurposing

Publications (2)

Publication Number Publication Date
WO2003049450A2 true WO2003049450A2 (en) 2003-06-12
WO2003049450A3 WO2003049450A3 (en) 2003-11-06

Family

ID=26682897

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2002/005091 WO2003049450A2 (en) 2001-12-04 2002-12-02 Methods for multimedia content repurposing

Country Status (5)

Country Link
EP (1) EP1459552A2 (en)
JP (1) JP2005512215A (en)
CN (1) CN1600032A (en)
AU (1) AU2002351088A1 (en)
WO (1) WO2003049450A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2048887A1 (en) * 2007-10-12 2009-04-15 Thomson Licensing Encoding method and device for cartoonizing natural video, corresponding video signal comprising cartoonized natural video and decoding method and device therefore
US11218530B2 (en) 2016-10-12 2022-01-04 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112887733A (en) * 2021-01-25 2021-06-01 中兴通讯股份有限公司 Volume media processing method and device, storage medium and electronic device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2205704A (en) * 1987-04-01 1988-12-14 Univ Essex Reduced bandwidth video transmission
US6061462A (en) * 1997-03-07 2000-05-09 Phoenix Licensing, Inc. Digital cartoon and animation process

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2205704A (en) * 1987-04-01 1988-12-14 Univ Essex Reduced bandwidth video transmission
US6061462A (en) * 1997-03-07 2000-05-09 Phoenix Licensing, Inc. Digital cartoon and animation process

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
MOHAN R ET AL: "Adapting multimedia Internet content for universal access" IEEE TRANSACTIONS ON MULTIMEDIA, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 1, no. 1, March 1999 (1999-03), pages 104-114, XP002159629 ISSN: 1520-9210 *
SZU SHENG CHEN ET AL: "New view generation from a video sequence" CIRCUITS AND SYSTEMS, 1998. ISCAS '98. PROCEEDINGS OF THE 1998 IEEE INTERNATIONAL SYMPOSIUM ON MONTEREY, CA, USA 31 MAY-3 JUNE 1998, NEW YORK, NY, USA,IEEE, US, 31 May 1998 (1998-05-31), pages 81-84, XP010289421 ISBN: 0-7803-4455-3 *

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2048887A1 (en) * 2007-10-12 2009-04-15 Thomson Licensing Encoding method and device for cartoonizing natural video, corresponding video signal comprising cartoonized natural video and decoding method and device therefore
WO2009047349A1 (en) * 2007-10-12 2009-04-16 Thomson Licensing Encoding method and device for cartoonizing natural video, corresponding video signal comprising cartoonized natural video and decoding method and device therefore
US11218530B2 (en) 2016-10-12 2022-01-04 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11283850B2 (en) 2016-10-12 2022-03-22 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Spatially unequal streaming
US11489900B2 (en) 2016-10-12 2022-11-01 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11496538B2 (en) 2016-10-12 2022-11-08 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E. V. Spatially unequal streaming
US11496541B2 (en) 2016-10-12 2022-11-08 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11496540B2 (en) 2016-10-12 2022-11-08 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11496539B2 (en) 2016-10-12 2022-11-08 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11516273B2 (en) 2016-10-12 2022-11-29 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11539778B2 (en) 2016-10-12 2022-12-27 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming
US11546404B2 (en) 2016-10-12 2023-01-03 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Spatially unequal streaming

Also Published As

Publication number Publication date
CN1600032A (en) 2005-03-23
JP2005512215A (en) 2005-04-28
EP1459552A2 (en) 2004-09-22
WO2003049450A3 (en) 2003-11-06
AU2002351088A1 (en) 2003-06-17

Similar Documents

Publication Publication Date Title
US7305618B2 (en) Methods for multimedia content repurposing
US11436780B2 (en) Matching mouth shape and movement in digital video to alternative audio
JP4138459B2 (en) Low resolution image production method and apparatus
US7224731B2 (en) Motion estimation/compensation for screen capture video
US7123774B2 (en) System and method for coding data
US7889949B2 (en) Joint bilateral upsampling
US6285794B1 (en) Compression and editing of movies by multi-image morphing
US8553782B2 (en) Object archival systems and methods
EP1641275B1 (en) Interactive design process for creating stand-alone visual representations for media objects
EP1641282B1 (en) Techniques for encoding media objects to a static visual representation
Kaufmann et al. Finite element image warping
EP1641281A1 (en) Techniques for decoding and reconstructing media objects from a still visual representation
CN113869138A (en) Multi-scale target detection method and device and computer readable storage medium
JP2001197507A (en) Method and system for processing image in image compression/expansion system employing hierarchical coding
JP2007141107A (en) Image processor and its method
WO2003049450A2 (en) Methods for multimedia content repurposing
US20220301523A1 (en) Method and apparatus for efficient application screen compression
CN116403142A (en) Video processing method, device, electronic equipment and medium
Masmoudi et al. Adaptive block-wise alphabet reduction scheme for lossless compression of images with sparse and locally sparse histograms
JP2005184062A (en) Image data conversion apparatus and image data conversion program
CN113365072B (en) Feature map compression method and device, computing equipment and storage medium
US20040101205A1 (en) Position coding system and method
Sri Geetha et al. Enhanced video articulation (eva)—a lip-reading tool
US20230360376A1 (en) Semantic Image Fill at High Resolutions
JPH0837664A (en) Moving picture encoding/decoding device

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SC SD SE SG SK SL TJ TM TN TR TT TZ UA UG UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2002785800

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2003550509

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 20028240332

Country of ref document: CN

WWE Wipo information: entry into national phase

Ref document number: 1020047008696

Country of ref document: KR

WWP Wipo information: published in national office

Ref document number: 2002785800

Country of ref document: EP