WO2010045736A1 - Rendu à latence réduite pour un système texte vers film - Google Patents

Rendu à latence réduite pour un système texte vers film Download PDF

Info

Publication number
WO2010045736A1
WO2010045736A1 PCT/CA2009/001521 CA2009001521W WO2010045736A1 WO 2010045736 A1 WO2010045736 A1 WO 2010045736A1 CA 2009001521 W CA2009001521 W CA 2009001521W WO 2010045736 A1 WO2010045736 A1 WO 2010045736A1
Authority
WO
WIPO (PCT)
Prior art keywords
text
initial
modified
sub
video
Prior art date
Application number
PCT/CA2009/001521
Other languages
English (en)
Inventor
Herve Lange
Original Assignee
Xtranormal Technology Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xtranormal Technology Inc. filed Critical Xtranormal Technology Inc.
Publication of WO2010045736A1 publication Critical patent/WO2010045736A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T13/00Animation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/131Fragmentation of text files, e.g. creating reusable text-blocks; Linking to fragments, e.g. using XInclude; Namespaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/194Calculation of difference between files
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/222Studio circuitry; Studio devices; Studio equipment
    • H04N5/262Studio circuits, e.g. for mixing, switching-over, change of character of image, other special effects ; Cameras specially adapted for the electronic generation of special effects

Definitions

  • the present disclosure relates to the field of digital animation techniques. More specifically, it relates to the field of text-to-movie/video rendering techniques.
  • a Text-To-Movie (TTM) or Text-To-Animation (TTA) system converts a text inputted by a user into a movie or an animation.
  • the last step of the process for creating the movie is the rendering of the movie or animation from the input text, which consists in generating the images constituting the movie.
  • the rendering process is a very computationally consuming step.
  • To increase the speed of rendering a project it is often broken up into jobs and the jobs are sent to a corresponding node of the rendering farm in order to be processed in parallel.
  • a rendering farm refers to a computer cluster for rendering computer generated images. Each computer or node of the rendering farm renders a video from a corresponding received job and the different rendered videos are then combined together in order to obtain a single video representative of the input text.
  • a method for creating a video comprises: receiving an initial text divided into at least two initial parts; rendering initial sub-videos, each one of the initial sub-videos being a visual representation of each one of the at least two initial parts of the initial text; combining the initial sub-videos together to generate the video; receiving a modified text of the initial text, the modified text comprising a modification; comparing the modified text with the at least two initial parts of the initial text to determine a modified part of the modified text corresponding to one of the at least two initial parts of the initial text, the modified part comprising the modification; rendering a modified sub-video for the modified part; combining the modified sub-video with at least one of the initial sub- videos corresponding to an unmodified part of the initial text, to generate a modified version of the video; and displaying the modified version of the video to present a visually representation of the modification in the modified text.
  • a system for creating a video comprises a text analyzer for receiving an initial text divided into at least two initial parts, and sending rendering jobs each comprising one of the at least two initial parts; a rendering farm in operative communication with the text analyzer for receiving and distributing the rendering jobs, and for rendering initial sub-videos, each one of the initial sub-videos visually representing each one of the at least two initial parts; and a sub-video combiner in operative communication with the rendering farm, for combining the initial sub-videos together to generate the video, the video visually representing the initial text; wherein the text analyzer is adapted to: receive a modified text of the initial text, the modified text comprising a modification; compare the modified text to the initial text to identify a modified part comprising the modification, and at least one unmodified part corresponding to one of the at least two initial parts; and send a new rendering job comprising the modified part, to the rendering farm; wherein the rendering farm is adapted to output a modified sub-
  • a method for modifying a video initially rendered from a text comprises receiving a modification to the text, the text comprising at least two parts, each of the at least two parts being associated to at least two respective sub-videos forming the video; determining a modified part and an unmodified part amongst the at least two parts, the modified part comprising the modification; rendering a modified sub-video for the modified part; combining the modified sub-video with one of the at least two respective sub-videos corresponding to the unmodified part, to generate a modified final video; and displaying the modified final video to visually represent the modification made to the text.
  • a system for modifying a video initially rendered from a text comprising: a processor; and a memory in operative communication with the processor, the memory comprising instructions for implementing the processor to: receive a modification to the text, the text comprising at least two parts, each of the at least two parts being associated to at least two respective sub-videos forming the video; determine a modified part and an unmodified part amongst the at least two parts, the modified part comprising the modification; render a modified sub-video for the modified part; combine the modified sub-video with one of the at least two respective sub-videos corresponding to the unmodified part, to generate a modified final video; and output the modified final video which visually represents the modification made to the text.
  • video should be understood as any form of motion pictures.
  • a video can be a film, a 2D animation, a 3D animation, an animated cartoon, and the like.
  • a video can have an audio frame or be silent.
  • modified part of text should be understood as a part of a text that has been directly and/or indirectly modified by a user or by any text modification tool, and that the TTM system considers as being modified.
  • unmodified part of text is a part of text that has not been directly and/or indirectly modified, and that the TTM system considers as being unmodified.
  • an unmodified part of text can be a part of text which has been directly modified by the user, but which is considered or tagged as unmodified by the TTM system.
  • sub-video is used to describe a video which corresponds to a part of a text.
  • a sub-video is a visual representation of its corresponding part of text.
  • a sub-video may have an audio track or be silent.
  • Sub-videos corresponding to different parts of the text are combined together in order to obtain a final video.
  • the final video is a visual representation of the whole text.
  • FIG. 1 is a flow chart illustrating a method for creating a video from a written text, in accordance with an embodiment
  • FIG. 2A illustrates an input of an initial text divided into three parts, in a TTM system, in accordance with an embodiment
  • Fig. 2B illustrates an input of a modified text of which only one part has been modified, in a TTM system, in accordance with an embodiment
  • Fig. 2C illustrates an input of only one part of text that has been modified, in a TTM system, in accordance with an embodiment
  • FIG. 3 is a block diagram illustrating a system for creating a video from an input text, in accordance with an embodiment
  • FIG. 4 is a block diagram illustrating a text analyzer, in accordance with an embodiment.
  • Figure 1 illustrates an embodiment of a method 300 for creating a video according to an input text.
  • the rendering of the video occurs in the context of a TTM or TTA system.
  • the first step 302 of the method is the reception of an initial text inputted by a user.
  • This initial text is the basis for the creation of the video.
  • the video to be rendered is a visual representation of the text and can comprise an audio frame.
  • the initial text can be a written text or an oral text subsequently converted into a written text by the system. As such audio data can be received and then converted into text data representative of the initial text.
  • the second step 304 of the method consists in dividing the initial input text into at least two parts.
  • Each part of the text is considered to be a job (also referred to herein as a rendering job).
  • Numerous strategies may be used to divide the initial text into parts.
  • each sentence of the initial text may be considered as a job.
  • the system analyses the initial text and identifies actions.
  • Each action is then considered as being a job.
  • the sentence "John walks to his bed and lies down” comprises two actions, namely the action "walking" and the action "lying down”.
  • the sentence is then divided into two parts, a first part corresponding to the action " walking” and the second part corresponding to the action "lying down".
  • Each of the two parts is considered by the system as being a job. Any strategy or method for dividing a text into parts known by a person skilled in the art may be used.
  • the second step 304 of the method namely dividing the initial text into parts of text, is performed prior to the reception of the text.
  • the user of the method 300 is in charge of the division of the text into parts; step 304 is optional.
  • the user may insert separator markers within the text in order to divide the text and thus define a position where each text part starts and ends. For example, markers such as "/", "•", and the like, can be inserted within the text by the user to create the different parts of text.
  • a user interface used by a user to input the text offers a natural division of the text, whereby individual data entry fields are presented to the user for entering text. In this way, delimitations as per the data entry fields provide for the separator markers.
  • a text box may be associated with a chapter or a scene of the story, for example.
  • the user describes a first scene by writing text in a first text box and describes a second scene by writing text in a second text box.
  • Each text box of the interface is dedicated to a corresponding scene or chapter of the story to be animated.
  • each text box may have a limited space in which the text is entered.
  • a limited number of characters may be entered in each text box, for example. In this case, the user starts writing a story in the first text box and if the limit of characters of the first box is reached, the user continues writing the story in the second box, etc. It should be understood that any method for dividing a text before being sent to the TTM or TTA system may be used.
  • step 306 an initial sub-video is rendered for each part of the initial text.
  • Each job corresponding to a part of the initial text is then sent to a corresponding node on a rendering farm and an initial sub-video is rendered for each job.
  • Each initial sub-video is a visual representation of a corresponding part of the initial text.
  • the initial sub-videos are stored in a memory and then combined together in order to create a final video at step 308.
  • the final video is a visual representation of the whole initial text, i.e. an animation or movie.
  • the final video is then made available to the user. After watching the final video, the user may decide to modify the story by modifying the initial text.
  • the generated initial sub-videos are silent videos.
  • An audio frame is generated separately from the sub-videos.
  • the audio frame is then combined with the sub-videos during the creation of the final video.
  • each initial sub-video may comprise an audio frame.
  • An audio frame of a sub-video is referred to herein as a sub-audio frame.
  • the next step 310 of the method illustrated in Figure 1 is the reception of a modified text from the user.
  • the modified text corresponds to the initial text plus the modifications directly made by the user.
  • the system divides the modified text into several parts at step 312, in accordance with the parts of the initial text, and compares each part of the modified text to its corresponding part of the initial text in order to identify the parts of the modified text which contain modifications directly made by the user.
  • the reception of a modified text may include the reception of the whole text or the reception of only the parts of text which have been directly modified by the user.
  • the whole initial text 10 is sent to the TTM or TTA system 12.
  • the whole initial text 10 is divided into three initial parts 14, 16, and 18 before being sent to the TTM system 12.
  • the whole modified text 20 is sent to the TTM system 12, as illustrated in Figure 2B.
  • the modified text 20 is also divided into three parts 14, 18, and 22 according to the parts of the initial text 10.
  • only the second part 22 of the modified text 20 has been modified by the user.
  • the whole modified text 20 comprising the parts of text 14, 18, and 22, is sent to the system 12.
  • only the modified part of text 22 is sent to the system 12, as illustrated in Figure 2C.
  • the next step is the determination of the parts of the text that have been directly modified by the user. If the whole modified text is received, this is done by comparing it to the previously received initial text stored in memory. If only the modified parts of text are received, then all of the received parts of text are considered as being directly modified by the user and the missing parts of text are retrieved from the initial text stored in memory.
  • each part of the modified text comprising modifications directly made by the user is tagged as being a modified part of text.
  • each part of the modified text which comprises no modification made by the user is tagged as an unmodified part of text.
  • the method 300 comprises the step of analyzing the importance of the modifications made to the initial text in order to determine if they will have impact on the video.
  • the method also comprises the step of determining if the modifications have an impact on parts of the text that have not been directly modified by the user. For example, if the user has changed the name of a character in his story from "John" to "Peter” in a part of the text, it may have no impact on the visual representation of the entity in the final video.
  • the system is configured in accordance with a given pre-set representation parameter to represent any male character with black hair, naming the character "John” or "Peter” will not impact the visual representation of the character.
  • a part of text in which the user has entered no modification may be tagged as a modified part of text if it is indirectly modified by a change made by the user in another part of text. For example, in a first part of the initial text, there is the sentence "Peter is sleeping on his bed” and in a second part of the initial text, there is the sentence "Peter wakes up and reads a book”. Watching the final video resulting from this text, the user sees Peter sleeping, waking up and then reading a book while lying on his bed.
  • the system If the user consequently replaces the first sentence by the sentence "Peter is sleeping on his chair” in the first part of the text but does not modify the second sentence, namely "Peter wakes up and reads a book", the system considers both the first and the second parts of the text as being modified even if the user has not directly modified the second sentence. The system determines and understands that the action "Peter wakes up and reads a book” occurs while Peter is sitting on a chair and not while he is lying on his bed. In this case, the second part of the text while being unmodified by the user is tagged as a modified part of text.
  • a graphic structure of the inputted text is generated in order to determine if a part of the text directly modified by the user should be tagged as unmodified and if parts of the text that have not been directly modified by the user should be tagged as modified.
  • a graphic structure represents the interconnections between the actions of the entities in the story and the state of the animation world. If a modification directly made by the user in the text does not affect the graphic structure, then the part of the text comprising the modification is tagged as being unmodified. If a modification directly made by the user implies a modification to the graphic structure, the system analyses the whole graphic structure in order to determine if this modification will give rise to additional modifications to the remainder of the graphic structure.
  • the method may take into account synchronization rules.
  • the synchronization rules express the temporal relations that can exist between two or more actions.
  • action 1 and action 2 always occur simultaneously, action 1 initiates action 2, action 1 terminates action 2, action 1 and action 2 are mutually exclusive, action 1 is a sub-action of action 2, action 1 always precedes action 2, action 1 always follows action 2, action 1 clips action 2, action 1 unclips action 2, action 1 and action 2 have to be fully synchronized, and the like.
  • a modified sub-video is rendered for each one of the parts of text tagged as being modified.
  • the modified sub-videos are then stored in memory. There is no rendering of a sub-video for the parts of text that are tagged as being unmodified since they have already be rendered during the creation of the final video associated with the initial text.
  • the last step 314 of the method 300 illustrated in Figure 1 is the generation of a modified final video by combining the modified sub-videos corresponding to modified parts of text with the previously stored initial sub-videos corresponding to unmodified parts of text. By only re-rendering the sub-videos from the modified parts of text, the period of time required to create the final video is shortened.
  • the modified final video is made available to the user which can modify the text again.
  • the same process occurs for each modification to the text entered by the user in an iterative manner. For each iteration, only the parts of text considered as modified with respect to the text of the previous iteration lead to a rendering process to obtain modified sub-videos. These sub-videos replace the sub-videos which correspond to the same parts of text and which are stored in memory.
  • Figure 3 illustrates one embodiment of a system 50 for generating a video from a text.
  • the system comprises a text analyzer 52, a memory 54 (aka memory device), a rendering farm 56, and a sub-video combiner 58.
  • the system 50 is connected to a user interface 60 comprising a display unit 62.
  • the user interface 60 is used by a user of the system 50 to input an initial text and any subsequent modifications to the initial text.
  • the user interface 60 is in operative communication with the text analyzer 52.
  • the text analyzer 52 comprises a text divider 70, a text modifications locator 72, a tag generator 74, and a job generator 76, as illustrated in Figure 4.
  • the text divider 70 receives a single piece of text from the user interface 60 and is adapted to divide the initial text received as a single piece of text into multiple parts.
  • the text analyzer does not contain any text divider 70 and the text is divided into parts before being sent to the text analyzer 52.
  • the user interface 60 is adapted to send the whole initial text inputted by the user and divided into several parts to the text analyzer 52. After the user has made modifications to the initial text, the whole text divided into parts is sent again to the text analyzer 52.
  • the user interface 60 first sends the whole initial text but subsequently it only sends the parts of text that have been modified by the user.
  • separator markers, text boxes or other means may be used to divide the text.
  • the text analyzer 52 saves the parts of the initial text in the memory 54 and also creates a job per part of initial text.
  • the text analyzer 52 sends the jobs to the rendering farm 56 which associates each job to a respective node (aka rendering node) (not shown). Hence there may be multiple (at least two) rendering nodes.
  • Each job corresponding to a part of the initial text is sent to its respective node in the rendering farm 56.
  • the rendering farm 56 outputs a rendered initial sub-video for each part of the initial text.
  • Each initial sub-video is saved in the memory 54 and also sent to the sub-video combiner 58.
  • the sub-video combiner 58 is adapted to combine the initial sub-videos together to generate a final video being a visual representation of the initial text.
  • the text analyzer 52 is further adapted to receive modifications to the initial text from the user interface 60.
  • the modifications should be understood as being the whole undivided initial text comprising modifications, the initial text divided into parts of which a certain number have been modified, or only the parts of the initial text that have been modified by the user.
  • the text modifications locator 72 compares the modified text to the initial text saved into memory 54 in order to identify the modifications to the text directly made by the user.
  • the text divider 70 i s adapted to divide the modified text according to the divisions previously made to the initial text. This results in directly modified and unmodified parts of text.
  • the text analyzer 52 further comprises the tag generator 74 which is adapted to tag each received part of text as being modified or unmodified.
  • the parts of text tagged as modified are saved in the memory 54 in which they can be substituted for their corresponding parts of initial text.
  • the tag generator 74 can be adapted to perform the steps above described with respect to Figure 1 in order to tag the modified parts of text.
  • the tag generator 74 can be adapted to generate graphic structures and to use synchronization rules in order to tag the parts of text.
  • the text analyzer 52 does not save the received text in memory 54 but it only stores the graphic structures corresponding to the received text.
  • the job generator 76 of the text analyzer 52 creates a job for each one of the parts of text tagged as being modified and then sends them to the rendering farm 56.
  • the rendering farm 56 assigns each job to a respective node where a modified sub- video is generated for each part of text tagged as modified.
  • These modified sub- videos are saved in memory 54 and sent to the sub-video combiner 58.
  • the sub- video combiner 58 retrieves the initial sub-videos corresponding to the parts of text tagged as being unmodified from the memory 58 and combines them with the modified sub-videos in order to create a modified final video.
  • the sub-video combiner 58 may also combine the sub-videos with an audio frame.
  • the modified final video is sent to the display unit 62 of the user. Alternatively, it can be sent and stored in the memory of the user's computer. [0048] While in the present description, a job is associated with a single part of text, it should be understood that a job can be associated with more the one parts of text. For example, two parts of text can be regrouped and constitute a job.
  • the present disclosure can be carried out as a method, can be embodied in a system, a computer readable medium or an electrical or electro-magnetic signal.

Abstract

L'invention concerne un procédé et un système de création de vidéo. Le procédé comprend les étapes suivantes : recevoir un texte initial divisé en au moins deux parties initiales ; effectuer un rendu des vidéos initiales, chaque vidéo initiale étant une représentation visuelle de chacune des deux parties initiales au moins du texte initial ; combiner ensemble les sous-vidéos initiales pour générer la vidéo ; recevoir une version modifiée du texte initial, le texte modifié comportant une modification ; comparer le texte modifié aux deux parties initiales au moins du texte initial pour déterminer une partie modifiée du texte modifié correspondant à l'une des deux parties initiales au moins du texte initial, la partie modifiée comportant la modification ; effectuer un rendu d'une sous-vidéo modifiée pour la partie modifiée ; combiner la sous-vidéo modifiée à au moins une des sous-vidéos initiales correspondant à une partie non modifiée du texte initial afin de générer une version modifiée de la vidéo.
PCT/CA2009/001521 2008-10-22 2009-10-22 Rendu à latence réduite pour un système texte vers film WO2010045736A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10756808P 2008-10-22 2008-10-22
US61/107,568 2008-10-22

Publications (1)

Publication Number Publication Date
WO2010045736A1 true WO2010045736A1 (fr) 2010-04-29

Family

ID=42118889

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CA2009/001521 WO2010045736A1 (fr) 2008-10-22 2009-10-22 Rendu à latence réduite pour un système texte vers film

Country Status (1)

Country Link
WO (1) WO2010045736A1 (fr)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8731339B2 (en) * 2012-01-20 2014-05-20 Elwha Llc Autogenerating video from text
US8937620B1 (en) * 2011-04-07 2015-01-20 Google Inc. System and methods for generation and control of story animation
US9237322B2 (en) 2013-02-07 2016-01-12 Cyberlink Corp. Systems and methods for performing selective video rendering
CN111885313A (zh) * 2020-07-17 2020-11-03 北京来也网络科技有限公司 一种音视频的修正方法、装置、介质及计算设备
CN111885416A (zh) * 2020-07-17 2020-11-03 北京来也网络科技有限公司 一种音视频的修正方法、装置、介质及计算设备
GB2596414A (en) * 2020-06-01 2021-12-29 Nvidia Corp Content animation using one or more neural networks
CN115336247A (zh) * 2020-06-10 2022-11-11 Jvc建伍株式会社 影像处理装置以及影像处理系统

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060227142A1 (en) * 2005-04-06 2006-10-12 Microsoft Corporation Exposing various levels of text granularity for animation and other effects
US20080107398A1 (en) * 2003-10-04 2008-05-08 Samsung Electronics Co., Ltd. Information storage medium storing text-based subtitle, and apparatus and method for processing text-based subtitle

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080107398A1 (en) * 2003-10-04 2008-05-08 Samsung Electronics Co., Ltd. Information storage medium storing text-based subtitle, and apparatus and method for processing text-based subtitle
US20060227142A1 (en) * 2005-04-06 2006-10-12 Microsoft Corporation Exposing various levels of text granularity for animation and other effects

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8937620B1 (en) * 2011-04-07 2015-01-20 Google Inc. System and methods for generation and control of story animation
US10402637B2 (en) 2012-01-20 2019-09-03 Elwha Llc Autogenerating video from text
US9036950B2 (en) 2012-01-20 2015-05-19 Elwha Llc Autogenerating video from text
US9189698B2 (en) 2012-01-20 2015-11-17 Elwha Llc Autogenerating video from text
US9552515B2 (en) 2012-01-20 2017-01-24 Elwha Llc Autogenerating video from text
US8731339B2 (en) * 2012-01-20 2014-05-20 Elwha Llc Autogenerating video from text
US9237322B2 (en) 2013-02-07 2016-01-12 Cyberlink Corp. Systems and methods for performing selective video rendering
GB2596414A (en) * 2020-06-01 2021-12-29 Nvidia Corp Content animation using one or more neural networks
GB2596414B (en) * 2020-06-01 2024-01-03 Nvidia Corp Content animation using one or more neural networks
CN115336247A (zh) * 2020-06-10 2022-11-11 Jvc建伍株式会社 影像处理装置以及影像处理系统
CN115336247B (zh) * 2020-06-10 2024-03-08 Jvc建伍株式会社 影像处理装置以及影像处理系统
CN111885313A (zh) * 2020-07-17 2020-11-03 北京来也网络科技有限公司 一种音视频的修正方法、装置、介质及计算设备
CN111885416A (zh) * 2020-07-17 2020-11-03 北京来也网络科技有限公司 一种音视频的修正方法、装置、介质及计算设备
CN111885416B (zh) * 2020-07-17 2022-04-12 北京来也网络科技有限公司 一种音视频的修正方法、装置、介质及计算设备

Similar Documents

Publication Publication Date Title
US10880598B2 (en) Video data generation method, computer device, and storage medium
CN108010112B (zh) 动画处理方法、装置及存储介质
WO2010045736A1 (fr) Rendu à latence réduite pour un système texte vers film
JP6355800B1 (ja) 学習装置、生成装置、学習方法、生成方法、学習プログラム、および生成プログラム
KR102054548B1 (ko) 다시점 오디오 및 비디오 대화형 재생
JP4430882B2 (ja) 複合メディアコンテンツの変換装置及び変換方法並びに複合メディアコンテンツ変換プログラム
US20150046797A1 (en) Document format processing apparatus and document format processing method
CN108769733A (zh) 视频剪辑方法和视频剪辑装置
JP2005513831A (ja) 多数の異種装置に配信するためのマルチメディアデータの変換
JP4833216B2 (ja) 映像作成装置および映像作成方法
CN114331820A (zh) 图像处理方法、装置、电子设备及存储介质
US20180143741A1 (en) Intelligent graphical feature generation for user content
Kaneko et al. Sign language animation using TVML
CN103986938A (zh) 基于视频播放的预览的方法和系统
CN108495174B (zh) 一种h5页面效果生成视频文件的方法及系统
CN115357755B (zh) 视频生成方法、视频展示方法和装置
US9396574B2 (en) Choreography of animated crowds
CN112348932A (zh) 口型动画录制方法及装置、电子设备、存储介质
CN109359687B (zh) 视频风格转换处理方法和装置
CN114513706B (zh) 视频生成方法和装置、计算机设备、存储介质
JP6946898B2 (ja) 表示態様決定装置、表示装置、表示態様決定方法及びプログラム
JP2017192080A (ja) 画像圧縮装置、画像復号装置、画像圧縮方法及び画像圧縮プログラム
CN117376660A (zh) 字幕元素的渲染方法、装置、设备、介质及程序产品
CN113905254A (zh) 视频合成方法、装置、系统与可读存储介质
JPH10134030A (ja) マルチメディアデータ・プレゼンテーションシステムおよび方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09821486

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 09821486

Country of ref document: EP

Kind code of ref document: A1