CN101208929A - Apparatus, system, method, and article of manufacture for automatic context-based media transformation and generation - Google Patents

Apparatus, system, method, and article of manufacture for automatic context-based media transformation and generation Download PDF

Info

Publication number
CN101208929A
CN101208929A CNA200680022788XA CN200680022788A CN101208929A CN 101208929 A CN101208929 A CN 101208929A CN A200680022788X A CNA200680022788X A CN A200680022788XA CN 200680022788 A CN200680022788 A CN 200680022788A CN 101208929 A CN101208929 A CN 101208929A
Authority
CN
China
Prior art keywords
data
information
medium
media
client device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CNA200680022788XA
Other languages
Chinese (zh)
Inventor
拉利特·S·沙纳
大卫·M·威斯伍德
葛瑞哥里·L·路特
邓坚永
丹尼尔·F·吉斯
崔辛山缪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Vidiator Enterprises Inc
Original Assignee
Vidiator Enterprises Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Vidiator Enterprises Inc filed Critical Vidiator Enterprises Inc
Publication of CN101208929A publication Critical patent/CN101208929A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/80Responding to QoS
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/756Media network packet handling adapting media to device capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/762Media network packet handling at the source 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L65/00Network arrangements, protocols or services for supporting real-time applications in data packet communication
    • H04L65/60Network streaming of media packets
    • H04L65/75Media network packet handling
    • H04L65/764Media network packet handling at the destination 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • H04L67/565Conversion or adaptation of application format or content
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/02Services making use of location information
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/2866Architectures; Arrangements
    • H04L67/30Profiles
    • H04L67/303Terminal profiles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/2866Architectures; Arrangements
    • H04L67/30Profiles
    • H04L67/306User profiles
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/52Network services specially adapted for the location of the user terminal
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/53Network services using third party service providers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/50Network services
    • H04L67/56Provisioning of proxy services
    • H04L67/561Adding application-functional data or data for application control, e.g. adding metadata

Abstract

Embodiments provide automatic media generation or other transformation (including enhancement) of multiple media inputs for delivery to user devices in a device-agnostic manner. The transformation process is further automatically enhanced based on contextual information. The contextual information provides customized content enhancement to the media delivered to a user device. The contextual information is automatically integrated with the transformed and generated media to provide highly customized rich-media content delivery in a device-agnostic manner.

Description

Be used for automatically device, system, method and goods according to the media conversion and the generation of content
The related application of cross reference
The application requires following priority according to 35U.S.C § 119 (e): application number is 60/693,381, denomination of invention is the U.S. Provisional Patent Application on June 23rd, 2005 for " device, system, method and the goods that are used for contextual media conversion of foundation automatically and generation ", the applying date, this application and the application belong to same patentee, and the mode by the reference reference integrally is combined among the present invention.
Technical field
The present invention relates to transmit medium, especially but not limited to relating to by the generation and the conversion of communication network to the medium of client device transmission by communication network.
Background technology
The media format conversion that various ways is arranged in the market.For example, there is many companies a kind of telephone voice recognition system of customary use to enter data in client's service server end application software that speech data can be converted to lteral data, so that handle.Media transformation techniques includes but not limited to: text conversion becomes voice, image transitions to become 3D to describe and image transitions becomes video.Also have multiple prior art to be used to be passed to different subscriber equipmenies according to the capacity of equipment customization data.For example, in the technique known, normally the information that provides according to some requesting services before being passed to client device, at different demonstrations and audio capability, is made amendment data in the server end application software.The example of this respect comprises at the display of portable phone or PDA size revises web page contents with respect to full-scale PC display.
Yet, media format conversion that these are basic and media transformation techniques still have a lot of deficiencies, especially in present environment, the user constantly increases for the demand of the media content of the number of different types in the client device that can easily be passed to number of different types.For the user that the different user devices of wishing them can maximize the use, these media format conversion and media transformation techniques too fundamental sum are original, can't satisfy these high-tech user's needs.
Summary of the invention
According to an aspect of the present invention, a kind of method, comprise and obtain first data, analyze described first data, thereby obtain contextual information, the contextual information that employing obtains is determined second data, replenishes described first data with second data of determining and obtains data converted, and data converted is passed at least one client device.
Description of drawings
With reference to the following drawings non-limiting and non-exhaustive specific embodiment is described, wherein, unless otherwise indicated, reference numerals identical among each figure is represented identical part.
Fig. 1 is the structure chart according to the system of an embodiment;
Fig. 2 is the structure chart according to the system of another embodiment;
Fig. 3 is the foundation structure chart of the system of another embodiment again;
Fig. 4 is according to the structure chart of the system of another embodiment more again.
Embodiment
Here be described according to the media conversion of context relation and the specific embodiment of generation automatic.In below describing, will provide many specific details to help to understand fully specific embodiment.Yet, those skilled in the art will appreciate that described specific embodiment can lack one or more specific detail, or adopt under the situation of additive method, element, material etc. and implement.In addition, well-known structure, material or operation are no longer concrete to be shown or description, in order to avoid blured the emphasis aspect of specific embodiment.
Referring to whole part of specification, specific characteristics, structure or the feature relevant with this specific embodiment that " specific embodiment " or " specific embodiment " expression is described is contained at least one specific embodiment.Therefore, term " in a specific embodiment " or " in one embodiment " that appears in the whole part specification everywhere might not all be meant same specific embodiment.In addition, described specific characteristics, structure or feature can be attached in one or more specific embodiments with any suitable mode.
As mentioned above, the technology formerly of more existing at present media format conversion and media conversion.Yet these formerly in the technology, also can provide automatic media to produce or conversion without any system or method, so that input of a plurality of medium and equipment are exported and can be carried out simultaneously.Especially, can provide media conversion and medium to strengthen according to context parameters without any system and method.
Generally speaking, the embodiment of the invention provides a kind of system, and the automatic media that is used for multiple medium pattern of the input produces or conversion (comprising enhancing), is passed to subscriber equipment in the mode of employing equipment the unknown.In a specific embodiment, described transfer process based on contextual information further strengthens automatically.Described contextual information provides the content of customization to strengthen to any medium that are passed to subscriber equipment.Described contextual information automatically with conversion and the medium integrating that produces, provide the rich media contents transmission that highly customizes in the mode of employing equipment the unknown.
Fig. 1 shows first system 10 that is used to realize a specific embodiment.Fig. 1 shows the media customization system 100 of an automation, and this media customization system 100 has a plurality of medium inputs (101,102,103 etc.) and the medium output (111 that offers the customization of one or more subscriber equipmenies (110,120,130 etc.), 121,131 etc.).Input medium (101,102,103 etc.) can be any form, as audio frequency, video, 3D describe, figure, Internet/network content for global, on-the-spot broadcasting audio or video, archives, animation or such as written report and for example satellite image and sports reports etc. such as stock quotes, news and weather forecasts.
The different medium inputs (101,102,103 etc.) of input medium custom-built system 100 can comprise the medium of making such as by the terminal use, such as contents such as word message, audio message, image and audio fragments.This medium input can also comprise predefined list information, such as fixing word message, picture or other paints, image and theme.The medium input can be provided by third party's information service firm, for example news report, weather forecast, physical culture top news and stock quotes etc.Can understand, also can comprise not at these other medium kind/forms of mentioning especially in the medium inputs (101,102,103 etc.).
One specific embodiment of automation media customization system 100 provides at different medium inputs (101,102,103 etc.) dynamic format conversion also can produce medium according to context parameters, exports (111 so that combine the medium that form customization separately with input medium, 121,131 etc.), export each subscriber equipment to, and/or one or more equipment (110,120,130 etc.) can receive more than one medium output.
According to a specific embodiment, a user of client device can select a communicate priority for one or more target client device (110,120,130 etc.).System 100 can be configured to the combination of unit client-side technology or client-server technology.In a specific embodiment, the medium of customization result from the server (for example, media customization system 100 is positioned on the server), and are passed to client device (110,120,130 etc.).Can also go up the medium that produce some or all at equipment (110,120,130 etc.) according to the instruction and data of media customization system 100.Medium produce the available resources of preferably using independent user account number and/or equipment, and the available components in media customization system.
The context component 107 of a specific embodiment is determined contextual information and/or contextual elements is integrated into an abundant media formats, to be passed to client device.Described contextual information can preestablish or draw according to a plurality of factor inferences, and these factors comprise the position of application type, medium type, client device, terminal use's input and terminal use's demographic statistics etc.Final context-enriched media and/or other rich-medias can be produced by medium producing component 106, with the capacity or the above-mentioned factor of coupling client device.In a specific embodiment, medium producing component 106 and context component 107 can be made up of identity element.
In a specific embodiment, context component 107 comprises the context input, priority such as user's input of sending from a client device, or the user account data from be stored in User Information Database 109 or other memory cell, such as service request historical record, position data, billing data, suitable consensus data, user's geography/position data etc., the context data of reasoning out, User Information Database 109 or other memory cell can be separated with media customization system 100 and/or be integrated in wherein.In a specific embodiment, other contextual elements can be by context component 107 by decomposing the medium input and resolving its content and determine, utilize any contextual elements of having determined then, such as the emotional factor in literal and voice, the onomatopoeia change of tuneization and figure release, and location-based context, such as the rain sound that is used for the rainy day weather forecast, produce new customized media.
In one embodiment, media customization system 100 comprises code conversion element 105, and in order to change the medium that will be passed to client device (110,120,130 etc.), the mode of conversion adopts the mode to these equipment and/or tunnel condition the best.For example, code conversion element 105 can be according to the characteristic of the client device of wanting receiving media and/or according to the tunnel condition of this client device, described tunnel condition may change in whole communication period, dynamically changes bit rate, frame frequency, resolution, coded format, color form or other and will be with the medium that the optimized media mode is transmitted relevant parameter.Application No. is 09/502,390, denomination of invention is " being used to change the computer program of stream video data ", the applying date is on February 10th, 2000, the patentee is with in the identical patent of the application, disclose some code conversion element 105 operable technology and module instance, this patent merges among the application in the mode of reference.
System 100 obtains different medium inputs, and the customized media of similar and different medium type is changed or produced to based on contextual information then.Described customized media can be passed to the client device with video demonstration and/or audio capability subsequently.Pass through mechanism includes but not limited to crossfire, Email, multimedia messaging service, MMS (MMS), WAP PUSH and download link.Transmission channels includes but not limited to wireless and cable network, the relief cable uploading channel.
Fig. 2 and Fig. 3 show other specific embodiments of the system 100 among Fig. 1.For example, Fig. 2 shows a kind of media codes converting system (or other converting systems) 200, and wherein, context component 107 comprises medium producing component 106.In a further specific embodiment, User Information Database 109 can comprise the information relevant with capacity of equipment or additional user information or other information.In the example as shown in Figure 2, different client devices can comprise mobile phone 210, wireless device (such as PDA or blackberry Blackberry) 220, kneetop computer 230 etc.
User Information Database 109 in specific embodiment can comprise priority and user's historical summary.The user can select image or customized image or other paints, and this information can be stored in client device or the User Information Database 109.For example, the user may want with predetermined " Maria Bartiromo " or people's such as " Larry Kudlow " image stock news to be passed on their client device.The user can also provide digital photograph to be used as the basis of new custom images.As mentioned above, User Information Database 109 can also comprise the capacity of equipment database, and media capability is associated with specific user's account number.If user has the equipment more than one, then can be according to coming cross-reference from the service request of equipment and dynamically determining the identification code of MIN, SID or other equipment.
According to different specific embodiments, one or more steps are arranged, comprising:
First step is to use medium producing component 106 and context component 107 analysis input mediums, derives different contextual informations, and this contextual information is combined from the available information of user's information database 109 with other Tathagata.Media customization system configuration among Fig. 1-3 is used to provide audio analysis, literal analysis, video analysis, graphical analysis, location-based information analysis, user demographic analysis and user behavior historical analysis.
Second step comprises characteristic and other information determined of analyzing of utilizing, described other information such as owner priority and capacity of equipment, produce medium type new and/or that revise, in order to this information is passed to client device with attractive and infusive mode.The customization output medium that obtains can comprise: image, 3D animation, 2D animation, video+audio frequency, pure video, pure audio output and other different medium output types.But the content of customized media can have the new feature time spent, adjusts and upgrades.Characteristics of one specific embodiment are the automatic decomposition (if suitably) according to intelligence, and the original media input is adjusted into the medium of customization, based on context parameter analysis and produce the medium of user/device customizing.
Characteristics of this system are the conversions of using " intelligence " medium generation and conversion or other intelligence.The intelligence of this system produces and conversion element provides as obtaining the original media data source and also is converted into the service of 3D rendering animation or other paints automatically.Each of this system is used the common functions of sharing some.For example, literal changes into voice, voice and changes into animation, the 3D server end is described, video is described and 3D is changed into to the 3D client basic capacity technology elements such as coding.
In order to make up the medium output of a customization, a specific embodiment has been used a plurality of context component, as template (for example, the context component 307 in the media customization system among Fig. 3 300).Can adopt specific context medium configuration template 337 to define the position of medium in an image animation fragment; For example, template 337 can be configured in weather map in the background, and with the left side of image configurations at screen.
Can also adopt a specific context media animation template 347 to define animation flow process in time; For example, audio frequency is to the conversion of animation, and the general flow of application software screen.Template 347 can be configured to come instruction animation synthetic according to time sequencing; For example, show and to appear screen suddenly, show that in the middle of device screen weather map introduces weather, the weatherman is shifted to the left side and weather map and the audio frequency that forecast in 24 hours are associated.Can also adopt the specific context database 317 that relates to User Information Database, media object is become audio frequency with text conversion, so that media object and event matches.
Context component 307 can comprise that the context medium produce engine 3 27, and these context medium produce engine 3 27 and have the function that for example scene produces and animation produces." intelligence " context medium among embodiment produce engine 3 27 and at first obtain input medium and optional user priority, and utilization is stored in the information in the customer data base 109, decide in scene which media resource to need to change, further utilize configuration template 337 then, correct time in media fragment, those media resources are configured to position correct in the scene.With the weather forecast is example, and its input may be text strings " 60F, a precipitation probability 80% ".Context component 307 can be resolved these literal and be decided key contextual parameters, such as " precipitation ", then with from the contrast of the contextual information of context database 317, so that obtain presenting the media object that the wet weather of high rainfall probability occurs, for example, the woman weatherman changes and wears raincoat, and the place of weather forecast may be out of doors and show with the 3D scene animation and to rain.
Medium producing component 306 among Fig. 3 utilizes animation template and the contextual information that is stored in the User Information Database 109 to produce animation in a similar fashion.The weather forecast of literal can convert the audio frequency forecast to, and wherein, the audio frequency forecast is broadcasted according to the script of animation template.In one embodiment, animation engine 346 can be used for providing animation for media object.This animation or be made into video and transmit again then by describing engine 3 36, or with the 3D data passes to client device, or transmit with other media formats.An example describing engine 3 36 is the 3D batch processor, and it utilizes hardware-accelerated OpenGL (Open GraphicsLibrary) engine to describe the 3D animation frame, and described hardware-accelerated OpenGL engine can be used for describing video content.
These frames can be passed to patented video coding element (VeeCoder) or other video engine 316 subsequently.This VeeCoder element or other video engine 316 can encode video into the multiple different video format that client device can be supported.Also can utilize audio engine 326 to produce or provide suitable audio frequency for this video.If this segment is transmitted with the 3D form, then its basic iconic model (as coming from image template 356), structure and all animation datas are understood packaged and are sent to client device, and describe engine at client device with client and play.
Client is described the canonical dissection that engine is an application software, for example meteorological application software.Medium producing component 306 can comprise the image recognition element, and the image recognition element can be used for multiple image adjustment and character produces.The image input can be used for changing into according to template, user's input, literal the video output of voice or next mutagenic image of other predefined image reinforced effects or modification by analysis then.
In a specific embodiment of media customization system shown in Figure 3 300, device capability information can be stored in the database 350.In addition, User Information Database 109 can be stored personal images 309, and described personal images 309 is the related image of medium that is used for and transmits user priority or selection.
In order to further specify the effect of one or more specific embodiments, the non-limiting example application of some media customization system is described now.
An example of customized media system is about producing automatically and the distribution video according to the literal input.In this application, the kinds of words input such as weather forecast, quotations on the stock market and news, is converted into abundant medium type, such as 3D or video weather forecast etc.Visual scene can be according to the physical location of client device, utilizes location-based template or visual cues to make or strengthen.
For example, if a client device is relevant with San Francisco, then can be as a setting with the photo of Gold Gate Bridge, be shown in as literal, 3D describe, other guides such as animation or video below.Giving one example, can the user application priority be the customization medium again.For example, the user can select sound, video image, and perhaps the user can also provide image to produce personalized image or other paints.User's demographic statistics also can be used for customized media.For example, can derive user's hobby according to specific age/gender/income group, to customize the content of bold and unrestrained, sharp-pointed or more conservative professional type.Capacity of equipment also can be used to the client device customized media.For example, can produce the scene that meets capacity of equipment especially, wherein, capacity of equipment comprises video, 3D and the audio format of support.
An example of specific embodiment comprises the literal input identification of intelligence.In this example, the medium of customization produce according to the data that are contained in the literal input.With the weather information is example, and the GPS position of client device can be used as the condition of inquiry weather information database.Can describe the meteorology returned and analyze, be used to virtual portrait to make sound then, it is rainy to express possibility nimbus to occur on the map simultaneously, can be with the graphical representation of trembling and be lower than the temperature of 32 degree.In this mode, can be that specific client device customizes group media input according to context parameters as position etc.In another related example, can come transforming message by media customization system, as instant message, so that word message is converted to the animation of sound.In particular, can be with the birthday greetings conversion and the reinforcement medium of the literal in the instant message, so that the sound animation birthday greetings of repeat playing birthday song or original character message to be provided.
Another related example of one specific embodiment comprises traffic returns or the intelligent travel information that produces intelligence.Can utilize gps system to obtain positional information.For example, the several seconds of being separated by is obtained two GPS measurement results, can determine current road and direction, and the traffic information that can obtain and continue to walk along this road for client device customization.Generally speaking, have the client device of GPS ability and can be during general communication repeatedly send gps data to the server of media customization system or connection.Therefore, system can determine whether the user has one or more conventional route.Then according to request, and according to the route information of deriving, the traffic of the remainder of the route that may select along the user can be reported by system.In the time of can also determining communication with the time data is morning or evening.If server can obtain real-time traffic incident (accident), media customization system can send alarm to client device when blocking up generation as if having moved on along common itinerary.In case derive user's route information, just can be used for producing animation information with road-map, traffic data picture or 3D rendering, reconfigure report or other customized parameters according to other contextual informations.
The tourism related service can adopt similar customization mode to provide, be position, formerly user's hobby and demographic statistics equally according to client device, and available input medium passage, as restaurant guides (Zagat), have the strategic alliance etc. of the charge of hotel and amusement upgrade service in addition, but provide local information about dining room, hotel, club and entertainment options.
Other use the media customization application of the different specific embodiments of method and system of the present invention to comprise the content conversion and produce, so that the transmission of the message between the corporate users to be provided, such as Friendster, Tribe, Myspace, Dodgeball, Match.com etc.Message can produce with minimum denominator form (pure words: instant message, SMS, Email).Yet, also support other inputs, such as Voice ﹠ Video.In conversion and production process, can be according to destination device capability and additional context parameters, the image of selecting such as the user or make, repeat playing sound, location-based scenery etc. come input medium is carried out code conversion.The message of sending from a user will just can send to another user by center " personal information " engine of group, should form end-to-end path by " personal information " engine, and hide two users' end points (telephone number, e-mail address etc.) information.Have only two users' personal information code name can allow the other side see, the identity of its " real world " can not show.Can utilize " textimation " (text conversion becomes animation) conversion like this is more high-end form with literal or SMS message " upgrading ", to offer the user who has the equipment/player that can play them.This class example comprise SMS to image have literal to the conversion of voice and generation, use sender's image (from their personal information, extracting) and the SMS shorthand of using specific SMS context engine to compress and to shorten be extended to have extraction from the mood nuance of Emoticon clue, punctuation mark more complete uttered sentence.
In a specific embodiment, different media elements need not " destructing ".For example, if offer this system should (etc.) element is an audio track, then before this audio track is incorporated other media elements into and/or is converted, need not its destructing.
In different specific embodiments, may surmount based on contextual conversion.For example, merge different artistic elements (for example, the 3D generating writing pattern adds sound) and make that " different things "-this is a kind of transfer process of making derivative works or other results' works.Be not only to obtain the difference input of different-format and convert them to same format that (for example, 3GPP), the situation that this specific embodiment comprises is whole summation greater than each several part.In another specific embodiment, element notice or influence other elements-for example, the animation (moving) of audio track " driving " lip, eyes etc.
According to a specific embodiment, at least some features described here can be implemented in software or other are stored in machine readable instructions on one or more machine-readable mediums.These machine readable instructions can be carried out by one or more processors, so that feature described here and other functions to be provided.For example, the different engine shown in Fig. 1 to 3, template, element etc. are embodied as the software module of being carried out by one or more processors.
Fig. 4 shows another specific embodiment.This specific embodiment provides based on contextual medium producing component (module 106 and 107), has the ability that generation can strengthen the side information of video coding speed and quality.Code conversion element 105 (or some other conversion elements) utilize described side information generate video (such as, generate the 3GPP file from the 3D frame).The video compression that code conversion element 105 uses needs to move between estimated frame and the coding of predicated error.3D medium generator 106 can provide the interframe mobile message.There has been accurate mobile message can reduce the video coding time (for example, need not estimate move) and reduce predicated error, because the variation between moving of having taken place and a frame and another frame is known.
This mobile message or other side informations are passed to code conversion element 105 as the cuetrack 151 that is used to encode.Predicated error reduces to make compression quality to improve the file size that also can obtain higher video quality under fixed bit rate or reduce the video file of fixed mass.Should be appreciated that mobile data is the example that is used for improving video encoding quality and speed that 106 of 3D medium generators can provide.As additional nonrestrictive example, medium generator 106 can also provide the information about interframe brightness or other brightness changes, interframe structural change.Give one example, a specific embodiment of 3D medium generator 106 can also send the information of forming about frame when being covered by literal again, or which part scene should be the focus of multidigit unit more about.
In one embodiment, can be according to capacity of equipment with the form of 3GPP video file (for example) or 3D video file with communicate to the terminal use.This information is to get under the assistance of device capabilities detection engine 109.In addition, video or 3D medium can customize, with the matching unit ability.Different equipment may have different 3D files and describe ability, therefore may carry out 3D content optimization at this equipment.Produce the file that engine can produce the matching unit ability based on contextual medium.In a similar fashion, code conversion element 105 can produce the video file with suitable end user device attribute.Be encoded in advance at medium under the situation of a plurality of 3D files of all having different characteristic and a plurality of 3GPP files, delivery engine can utilize facility information and file property to select only file for end user device.
Delivery engine in specific embodiment utilizes capacity of equipment to come select File, rather than utilizes network condition or status of equipment.In case communication period is ready, the ability of delivery engine just can be coupled with dynamic bandwidth adaptation technology, as Application No. is 10/452,035, denomination of invention is " method and a device that dynamic bandwidth is regulated ", the applying date is on May 30th, 2003, and the patentee is with disclosed in the identical patent of the application, and this patent merges among the application in the mode of reference.The disclosed regulation technology of application that code conversion element (105) in specific embodiment also can utilize this to own together.
Above-mentioned all mention and/or be listed in United States Patent (USP), U.S. Patent Application Publication text, U.S. Patent application, foreign patent, foreign patent application and non-patent publications in the request for data table in this manual, all the mode with reference merges among the application.
More than to the description of specific embodiment, comprise the description in the summary, be not the most detailedly maybe will limit the present invention and only be disclosed precise forms.Certain specific embodiments described herein and example are for illustrative purpose, but still might do the modification of different equivalences within the scope of the invention, and do not break away from the spirit and scope of the present invention.
Can make these and other modifications according to above detailed description.Employed wording should not be construed to the present invention is limited to disclosed certain specific embodiments in specification and the claim in the following claim.By following claim decision, should according to principle that every claim propose be understood fully by claim for protection scope of the present invention.

Claims (22)

1. a method is characterized in that, comprising:
Obtain first data;
Obtain contextual information by analyzing described first data;
Utilize the contextual information that obtains to determine second data, the described second definite data comprise individual's paint and the media object that at least one is different from described paint, and all the content with described first data is relevant for both;
Utilize the described paint and the described media object of described definite second data of first data-driven, obtain translation data, wherein, described driving comprises makes animation transmitting the described content of described first data with described paint, and by described paint utilization isolated described media object from described transmission the described content of described first data is carried out relevance and describe;
Obtain cuetrack information based on second data of determining with based on contextual information;
The cuetrack information that utilization obtains is optimized the code conversion of translation data at least one client device or for the form of communication port the best of cordless communication network; And
Translation data with described optimised form on the described communication port of cordless communication network passes at least one client device.
2. the method for claim 1, it is characterized in that, described acquisition first data comprise the acquisition lteral data, described analysis first data comprise by resolving this lteral data and obtain contextual information, and described at least one media object comprises that audio frequency, video, another paint, one or more image or relevance describe other abundant media elements of this lteral data.
3. the method for claim 1 is characterized in that, described second data of utilizing first data-driven to determine further comprise at least two described first data and described second data are merged, obtain translation data.
4. the method for claim 1 is characterized in that, described acquisition contextual information comprises from described first data determines the mood nuance.
5. the method for claim 1 is characterized in that, further comprises:
According to the characteristic of client device or according to the situation of the communication port that is used for transmitting described translation data, convert this translation data or second data to another form from first fotmat code; And
According to the change of the characteristic or the tunnel condition of client device, upgrade described translation data in that communication period is dynamic.
6. the method for claim 1 is characterized in that, further comprise utilize template in the picture frame of selecting the position and present described paint and described media object according to time sequencing.
7. the method for claim 1 is characterized in that, wherein:
Second data that described basis is determined or based on contextual information obtain obtaining emotion information before cuetrack information is included in code conversion, brightness changes information and structural change information; And
Described this cuetrack information of utilizing comprises described code conversion optimization, utilizes the described emotion information that obtains, brightness to change information and the described code conversion of structural change information and executing.
8. a media customization system is characterized in that, comprising:
Be used to obtain the device of first data;
Be used for by analyzing the device that described first data obtain contextual information;
Be used to utilize the contextual information of acquisition to determine the device of second data relevant with described first data;
Be used for replenishing the device that described first data produce translation data with second data of determining;
Be used for based on second data of determining and the device that obtains cuetrack information based on contextual information;
Be used to utilize the cuetrack information that obtains that the code conversion of translation data is optimized at least one client device or for the device of the form of communication port the best of cordless communication network; And
Be used on described communication port, the translation data of this optimised form being passed to the device of described at least one client device.
9. system as claimed in claim 8 is characterized in that, further comprises:
Be used for described translation data is carried out the device of code conversion; And
Be used to store the device of the client device data, user profile or the pictorial information that can be used to produce translation data.
10. system as claimed in claim 8 is characterized in that, the described device that is used for analyzing first data comprises at least one of context database, context engine, configuration template and animation template.
11. system as claimed in claim 8 is characterized in that, the described device that is used for replenishing comprises video engine, audio engine, describes engine, animation engine and graphic template at least one, can be used for is that translation data produces medium.
12. system as claimed in claim 8, it is characterized in that, describedly be used to utilize cuetrack information that the optimized device of code conversion is comprised, be used for emotion information, brightness according to the described cuetrack information of conduct that obtains by the described device that is used for obtaining in advance described code conversion and change the parts that information and structural change information strengthen video coding.
13. system as claimed in claim 8, it is characterized in that, further comprise: be used to transmit the translation data that has obtained according to capacity of equipment as video file, described video file comprises as described translation data, described individual paint is made animation transmitting the described content of described first data, and by described paint isolated at least one media object that is different from described paint from described transmission the described content of described first data is carried out relevance and describe.
14. goods is characterized in that, comprising:
Machine-readable medium store and can be carried out to finish the instruction of following processing by at least one processor:
Obtain contextual information by analyzing first data;
Utilize the contextual information that obtains to determine second data, described second data of determining comprise that individual's paint is different from the media object of described paint with at least one, and all the content with described first data is relevant for both;
Utilize the described paint and the described media object of described definite second data of this first data-driven, obtain translation data, wherein said driving comprises makes animation transmitting the described content of described first data with described paint, and by described paint utilization isolated described media object from described transmission the described content of described first data is carried out relevance and describe; And
On cordless communication network, this translation data is passed at least one client device.
15. goods as claimed in claim 14 is characterized in that, described machine-readable medium further comprise and storing in order to carry out the instruction of following processing:
Obtain cuetrack information based on second data of determining or based on contextual information; And
Utilize described cuetrack information that the code conversion of translation data is optimized for client device or for the form of the communication port the best that is connected to this client device.
16. goods as claimed in claim 14 is characterized in that, the instruction that first data that described analysis obtains obtain contextual information comprise from described first data obtain the mood nuance instruction.
17. goods as claimed in claim 14, it is characterized in that, the contextual information that described utilization obtains determines that the instruction of second data comprises according at least one or a plurality of owner priority, device characteristics or media content that can be associated with first data, discerns second data relevant with first data.
18. a device is characterized in that, comprising:
The input terminal that is used for the receiving media input;
Context component with this input terminal coupling, is used for the analyzing medium input with the acquisition contextual information, and utilizes the contextual information that obtains to determine the medium that are associated with medium input for the medium input and replenish;
The medium producing component, with described context component coupling, be used for replenishing the supplementing media input with the medium of determining, to obtain convert media, the described medium of determining replenish and comprise that individual paint and at least one are different from the media object of described paint, all the content with described medium input is relevant for both, described paint is made into animation transmitting the described content of described medium input, and by described paint isolated described media object from described transmission the described content of described first data is carried out relevance and describe; And
Outlet terminal is with described medium producing component coupling, in order to give at least one client device with the communicate of described conversion on cordless communication network.
19. device as claimed in claim 18 is characterized in that, further comprises:
The code conversion element with the coupling of described medium producing component, is paired in client device or for the form of communication port the bests of the medium that can be used for transmitting and converting that are connected to client device in order to the media conversion with described conversion; And
At least one memory cell is used to store the information relevant with client device, comprises device capability information, user profile, priority information or image template.
20. device as claimed in claim 18, it is characterized in that, further comprise the code conversion element, with described medium producing component coupling, be used for that media conversion with described conversion is paired in client device or for the form of communication port the bests of the medium that can be used for transmitting and converting that are connected to client device, described code conversion element further is coupled and is used for receiving from the medium producing component or from context component cuetrack information, described cuetrack information can be used for the conversion of optimization convert media, and described cuetrack information comprises the emotion information that is obtained by described code conversion element before coding, brightness change information and structural change information.
21. device as claimed in claim 18 is characterized in that, described context component comprises context database, context engine, configuration template and animation template.
22. device as claimed in claim 18 is characterized in that, described medium producing component comprise can be used for for the medium of conversion produce object video engine, audio engine, describe engine, animation engine and graphic template.
CNA200680022788XA 2005-06-23 2006-06-23 Apparatus, system, method, and article of manufacture for automatic context-based media transformation and generation Pending CN101208929A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US69338105P 2005-06-23 2005-06-23
US60/693,381 2005-06-23

Publications (1)

Publication Number Publication Date
CN101208929A true CN101208929A (en) 2008-06-25

Family

ID=37000058

Family Applications (1)

Application Number Title Priority Date Filing Date
CNA200680022788XA Pending CN101208929A (en) 2005-06-23 2006-06-23 Apparatus, system, method, and article of manufacture for automatic context-based media transformation and generation

Country Status (5)

Country Link
EP (1) EP1908255A1 (en)
JP (1) JP2008544412A (en)
CN (1) CN101208929A (en)
TW (1) TW200718134A (en)
WO (1) WO2007002448A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102244788A (en) * 2010-05-10 2011-11-16 索尼公司 Information processing method, information processing device, scene metadata extraction device, loss recovery information generation device, and programs
CN101640627B (en) * 2008-07-29 2011-12-28 优必达科技有限公司 System and method for transmitting multimedia data stream
CN104919813A (en) * 2012-12-19 2015-09-16 微软技术许可有限责任公司 Computationally generating turn-based game cinematics

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8521842B2 (en) * 2008-02-29 2013-08-27 Microsoft Corporation Delivering composite media to a client application
EP2099198A1 (en) * 2008-03-05 2009-09-09 Sony Corporation Method and device for personalizing a multimedia application
US8578259B2 (en) 2008-12-31 2013-11-05 Microsoft Corporation Media portability and compatibility for different destination platforms
US9084079B2 (en) * 2010-11-19 2015-07-14 Qualcomm Incorporated Selectively formatting media during a group communication session
US20150340067A1 (en) * 2014-05-22 2015-11-26 Idomoo Ltd. System and Method to Generate a Video on the Fly
US10631070B2 (en) 2014-05-22 2020-04-21 Idomoo Ltd System and method to generate a video on-the-fly
US10038757B2 (en) 2015-04-29 2018-07-31 Microsoft Technology Licensing, Llc Providing personalized greetings on a digital assistant
US10123090B2 (en) 2016-08-24 2018-11-06 International Business Machines Corporation Visually representing speech and motion

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030163315A1 (en) * 2002-02-25 2003-08-28 Koninklijke Philips Electronics N.V. Method and system for generating caricaturized talking heads

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101640627B (en) * 2008-07-29 2011-12-28 优必达科技有限公司 System and method for transmitting multimedia data stream
CN102244788A (en) * 2010-05-10 2011-11-16 索尼公司 Information processing method, information processing device, scene metadata extraction device, loss recovery information generation device, and programs
CN102244788B (en) * 2010-05-10 2015-11-25 索尼公司 Information processing method, information processor and loss recovery information generation device
CN104919813A (en) * 2012-12-19 2015-09-16 微软技术许可有限责任公司 Computationally generating turn-based game cinematics

Also Published As

Publication number Publication date
WO2007002448A1 (en) 2007-01-04
JP2008544412A (en) 2008-12-04
TW200718134A (en) 2007-05-01
EP1908255A1 (en) 2008-04-09

Similar Documents

Publication Publication Date Title
CN101208929A (en) Apparatus, system, method, and article of manufacture for automatic context-based media transformation and generation
CN101120348B (en) Method and system for providing news information by using three dimensional character for use in wireless communication network
US7142645B2 (en) System and method for generating and distributing personalized media
Giachetti et al. Evolution of firms' product strategy over the life cycle of technology-based industries: A case study of the global mobile phone industry, 1980–2009
WO2018054181A1 (en) Virtual resource presentation method, client, and plug-in
JP2008529345A (en) System and method for generating and distributing personalized media
CN106534875A (en) Barrage display control method and device and terminal
CN105069737A (en) Mobile service terminal, mobile service system and data processing method for airport services
JP2008518326A (en) System and method for mobile 3D graphical messaging
CN103117057A (en) Application method of special human voice synthesis technique in mobile phone cartoon dubbing
CN102137047B (en) Multiparameter media adapter gateway and adaption method thereof
US20060019636A1 (en) Method and system for transmitting messages on telecommunications network and related sender terminal
JP2004088335A (en) Mail server, program to realize the same, and portable terminal
CN110225398A (en) Multimedia object playback method, device and equipment and computer storage medium
CN114946193A (en) Customized video production service providing system using cloud-based voice integration
CN110012059B (en) Electronic red packet implementation method and device
Moll Mobile web design
KR20060098178A (en) Method and system for registering voice message with community site
CN116980538A (en) Video generation method, device, equipment, medium and program product
CN101854610B (en) Method for uploading information data of mobile phone
CN101430694A (en) Method for assisting blind men to using network search service
KR20130074886A (en) Automated system for remote producing and transmitting the image and message for multi purpose/customer/media
CN104125305A (en) Method and system for dynamic push of calendar information
KR20040013071A (en) Voice mail service method for voice imitation of famous men in the entertainment business
CN101312549A (en) Method for converting text information into stream media of multimedia and further receiving by terminal

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Open date: 20080625