WO2022063093A1 - 用于生成文字模式的视频的方法、装置、设备和介质 - Google Patents

用于生成文字模式的视频的方法、装置、设备和介质 Download PDF

Info

Publication number
WO2022063093A1
WO2022063093A1 PCT/CN2021/119438 CN2021119438W WO2022063093A1 WO 2022063093 A1 WO2022063093 A1 WO 2022063093A1 CN 2021119438 W CN2021119438 W CN 2021119438W WO 2022063093 A1 WO2022063093 A1 WO 2022063093A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
generating
user
text
information sharing
Prior art date
Application number
PCT/CN2021/119438
Other languages
English (en)
French (fr)
Inventor
吴怡颖
孙辉
王道裕
Original Assignee
北京字跳网络技术有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京字跳网络技术有限公司 filed Critical 北京字跳网络技术有限公司
Priority to KR1020237002186A priority Critical patent/KR102613143B1/ko
Priority to EP21871463.2A priority patent/EP4171047A4/en
Priority to JP2023506273A priority patent/JP7450112B2/ja
Publication of WO2022063093A1 publication Critical patent/WO2022063093A1/zh
Priority to US18/087,566 priority patent/US11922975B2/en
Priority to US18/429,190 priority patent/US20240170026A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/475End-user interface for inputting end-user data, e.g. personal identification number [PIN], preference data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/816Monomedia components thereof involving special video data, e.g 3D video
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/036Insert-editing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/165Management of the audio stream, e.g. setting of volume, audio stream path
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234336Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements by media transcoding, e.g. video is transformed into a slideshow of still pictures or audio is converted into text
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/472End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
    • H04N21/47205End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for manipulating displayed content, e.g. interacting with MPEG-4 objects, editing locally
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8126Monomedia components thereof involving additional data, e.g. news, sports, stocks, weather forecasts
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/8146Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
    • H04N21/8153Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/27Server based end-user applications
    • H04N21/274Storing end-user multimedia data in response to end-user request, e.g. network recorder
    • H04N21/2743Video hosting of uploaded data from client
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/44Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs

Definitions

  • Various implementations of the present disclosure relate to the field of computers, and in particular, to a video method, apparatus, device, and computer storage medium for generating text patterns.
  • a method for generating a text-mode video in an information sharing application is provided.
  • a request for generating a video is received from a user of an information sharing application.
  • an initial page for generating a video is displayed, and the initial page includes a prompt for entering text.
  • text input from the user is obtained.
  • an apparatus for generating a text-mode video in an information sharing application includes: a receiving module configured to receive a request for generating a video from a user of an information sharing application; a display module configured to display an initial page for generating a video in the information sharing application, the initial page including an initial page for generating a video in the information sharing application a prompt for entering text; an obtaining module configured to obtain textual input from a user in response to detecting a user's touch in the area where the initial page is located; and a generating module configured to generate a video based on the textual input for use in the Published in information sharing applications.
  • an electronic device comprises: a memory and a processor; wherein the memory is for storing one or more computer instructions, wherein the one or more computer instructions are executed by the processor to implement the method according to the first aspect of the present disclosure.
  • a computer-readable storage medium having stored thereon one or more computer instructions, wherein the one or more computer instructions are executed by a processor to implement the method according to the first aspect of the present disclosure .
  • a user can directly generate a corresponding video based on text input inside the information sharing application. In this way, the complexity of the user's operation can be reduced, and richer published content can be provided to the user.
  • FIG. 1 schematically shows a block diagram of an application environment according to an exemplary implementation of the present disclosure
  • FIG. 2 schematically illustrates a block diagram of a user interface for generating text-mode video according to an exemplary implementation of the present disclosure
  • FIG. 3 schematically shows a flowchart of a method for generating a text-mode video according to an exemplary implementation of the present disclosure
  • FIG. 4 schematically illustrates a block diagram of a user interface for entering text according to an exemplary implementation of the present disclosure
  • Figure 5 schematically illustrates a block diagram of a user interface for selecting a video background according to an exemplary implementation of the present disclosure
  • FIG. 6 schematically illustrates a block diagram of a user interface for editing video according to an exemplary implementation of the present disclosure
  • FIG. 7 schematically illustrates a block diagram of a user interface for previewing a video according to an exemplary implementation of the present disclosure
  • FIG. 8 schematically shows a block diagram of an apparatus for generating a text-mode video according to an exemplary implementation of the present disclosure.
  • FIG. 9 illustrates a block diagram of a computing device capable of implementing various implementations of the present disclosure.
  • the term “comprising” and the like should be understood as open-ended inclusion, ie, “including but not limited to”.
  • the term “based on” should be understood as “based at least in part on”.
  • the terms “one implementation” or “the implementation” should be understood to mean “at least one implementation”.
  • the terms “first”, “second”, etc. may refer to different or the same objects. Other explicit and implicit definitions may also be included below.
  • FIG. 1 schematically shows a block diagram of an application environment 100 according to an exemplary implementation of the present disclosure.
  • a user may view and/or post a video through application 110 .
  • the application 110 can push the video 120 to the user, and the user can watch the favorite video by searching, scrolling down, turning pages, and the like.
  • the user can press the "publish" button 130 to publish the video.
  • a variety of video publishing modes have been developed. For example, users can publish videos by taking pictures, shooting in sections, snapshots, uploading from albums, and so on. Each user can choose the way they like to publish the video. Some users may wish to publish a video produced based on text, for example, the user expects to input wishes such as "Happy Mid-Autumn Festival", "Happy Birthday”, etc., and a corresponding video is generated for publication.
  • a method for generating a text-mode video in an information sharing application is proposed.
  • a request for generating a video may be received from a user of an information sharing application, and a method for generating a text-mode video may then be initiated.
  • FIG. 2 schematically shows a block diagram of a user interface 200 for generating text-mode video according to an exemplary implementation of the present disclosure.
  • the user interface 200 shown in FIG. 2 may be entered.
  • the user may select the text mode 220 in the menu at the bottom of the user interface 200 in order to initiate a generation method according to an exemplary implementation of the present disclosure.
  • an initial page 210 for generating a video may be displayed.
  • the initial page 210 may include a prompt for entering text: "Tap to enter text".
  • the user can enter corresponding text in the initial page 210 .
  • the user may perform a touch operation in the area where the initial page 210 is located in order to initiate the process of entering text.
  • the application 110 obtains textual input from the user and generates a video including the textual input for publication.
  • the page layout shown in FIG. 2 is only schematic, and other page layouts may be adopted according to the exemplary implementation of the present disclosure, as long as the method according to the exemplary implementation of the present disclosure can be implemented.
  • the user does not need to call the video editing application separately, but can directly generate the corresponding video based on the text input inside the information sharing application. In this way, the complexity of the user's operation can be reduced, errors that may be caused by the user during switching between multiple applications can be avoided, and the user can be provided with richer published content.
  • FIG. 3 schematically illustrates a flowchart of a method 300 for generating a text-mode video according to an exemplary implementation of the present disclosure.
  • a request to generate a video is received from a user of the information sharing application.
  • a user may swipe the menu at the bottom of the user interface 200 as shown in FIG. 2 and select the text mode 220 from a variety of video modes.
  • an initial page 210 for generating a video is displayed, where the initial page 210 includes prompts for entering text.
  • the input prompt may be displayed at a prominent position in the initial page. The user can enter the desired text according to the prompts, for example, the user can invoke an input dialog by touching any blank area in the initial page 210 to enter text.
  • a textual input from the user is obtained.
  • the user may touch any blank area in the initial page 210 in order to enter text, more details of which will be described below with reference to FIG. 4 .
  • 4 schematically illustrates a block diagram of a user interface 400 for entering text in accordance with an exemplary implementation of the present disclosure.
  • an input box 410 may pop up for receiving text input.
  • the user may enter the plain text content "Happy Mid-Autumn Festival".
  • the text input may include text and emojis.
  • the user can also input an emoticon such as "smiley".
  • the emoticons herein may be emoticons drawn by an operating system on the mobile terminal, and each emoticon may have a unique code.
  • the emoji images drawn can vary from operating system to operating system. For example, in the "smiley faces" drawn by the two operating systems, the degree to which the corners of the mouth are raised can vary.
  • a video is generated based on the textual input for publication in the information sharing application. Having obtained the textual input, a video including the textual input can be generated for publication.
  • text input is the most basic element for generating a video
  • a default length video may be generated based on a default video background.
  • the application 110 may select a moonlit background based on the content of the text and generate a video including the text "Happy Mid-Autumn Festival".
  • the initial page 210 may include more options. In the following, returning to FIG. 2, more details about the initial page 210 are described.
  • the initial page 210 may further include an option 234 for selecting a video background.
  • the user can click on option 234 to select the desired video background, for example, one or more of images, videos, emoticons, and emoticon animations can be selected as the background.
  • the video may be generated based on the video background selected by the user. Assuming the user selects an image of a mooncake, the background of the resulting video will include the mooncake pattern.
  • the image position, the number of images, the motion trajectory of the image, and the like can be further specified. More details regarding video backgrounds are described with reference to FIG. 5 , which schematically illustrates a block diagram of a user interface 500 for selecting a video background in accordance with an exemplary implementation of the present disclosure.
  • the user can choose a mooncake image as the background, and can specify to include 3 images randomly distributed in the video. At this point, images 510, 520, and 530 will be included in the resulting video. Further, the image can be specified to move in a certain direction.
  • motion trajectories may be predefined, eg, straight lines, curves, etc.; alternatively and/or additionally, motion trajectories may be randomly generated.
  • additional rules may be set: for example, it may be specified that in the case of displaying multiple images, collisions between the images should be avoided; for another example, it may be specified to change the direction of movement when the image reaches the display boundary, etc.
  • a video may be selected as a background, and a portion of a certain time period in the video may be specified to be used (eg, the start time and end time of the specified time period), a certain time period in the video may be selected region (for example, specifying a portion of a window to use), and so on.
  • an emoticon or an emoticon animation may be selected as a video background.
  • the initial page 210 may further include a speaking option 230 for speaking a text input.
  • the user can activate or deactivate the automatic reading function by clicking the operation.
  • the application 110 can automatically read the text input by the user based on artificial intelligence technology, and generate a video based on the read audio.
  • the generated video may include audio read aloud; alternatively and/or additionally, the generated video may include both textual and audio content.
  • the reading options may further include at least any one of the following: gender, age, voice style, and speaking rate of the speaker.
  • gender gender
  • age age
  • voice style and speaking rate of the speaker.
  • various sound styles can be provided to meet the needs of different users.
  • sound styles may include, but are not limited to: rich, sweet, lively, and the like. Users can choose different speech rates of high, medium or low to support personalized settings for the reading effect.
  • the user can cancel the read aloud option, and the generated video only includes text content at this time.
  • a user can be provided with various materials for generating a video, thereby providing a richer media representation.
  • the content of the initial page 210 has been described above with reference to the accompanying drawings.
  • the user can make settings in the initial page 210 to define various parameters for generating the video.
  • the "next" button 232 may be clicked to display the edit page.
  • FIG. 6, schematically shows a block diagram of a user interface 600 for editing a video according to an exemplary implementation of the present disclosure.
  • the user may operate in the editing page 610, and the application 110 may generate a corresponding video based on the user's user operation on the editing page 610.
  • the edit page 610 may include at least any of the following: an option 620 for editing speaking settings, an option 622 for editing text input, and an option 624 for editing a video background.
  • the user can enable or disable the automatic reading function via the option 620; the user can edit the text that has been entered via the option 622, and can set the font, font size, color, display position, etc. of the text; Option 624 to edit an already selected background, reselect a background or add a new background, etc.
  • the user may press the "Next" button 640 to generate a corresponding video based on the edited options specified by the user in the edit page 610 .
  • the edit page 610 may provide the user with the function of modifying various parameters. In this way, the user is provided with an opportunity to modify the previous settings when they are not satisfied, thereby facilitating the user's operation and generating a satisfactory video.
  • the editing page 610 may further include an option 630 for selecting to add background sound to the video.
  • the background sound here may include background music and/or other sounds such as character narration.
  • the user may press option 630 to select background music or other audio for the video.
  • the user may enter a narration, eg, the user may read a poem about the Mid-Autumn Festival, and the like.
  • the application 110 may generate a corresponding video based on the background sound specified by the user operation.
  • a user can be allowed to add more diverse sound files to a video, so as to generate richer video content.
  • the edit page 610 may further include an option 632 for selecting to add stickers to the video.
  • the stickers here can include text stickers as well as image stickers.
  • Text stickers can include text, such as common phrases in various artistic fonts, etc.
  • Image stickers can include icons, common expressions, and frames.
  • the user may press option 632 to insert stickers into the video, eg, the user may insert the text sticker "family reunion", and the image sticker "heart”, etc. Further, the user can adjust the position, size and orientation of the sticker by touching, dragging, rotating, zooming and so on.
  • the application 110 may generate a corresponding video based on the sticker specified by the user operation.
  • a user may be allowed to add more personality elements to a video. In this way, the video may be more interesting and provide a richer media representation.
  • the editing page 610 may further include: an option for specifying the length of the video.
  • the video may have a default length of, for example, 3 seconds (or other value).
  • users can customize the length of the video.
  • the user may be allowed to further set the matching relationship between the background sound (or video) and the length of the video.
  • a sound (or video) segment that matches the length of the video can be cut from the background sound (or video). If the length specified by the user is greater than the length of the background sound (or video), the user can set loop playback.
  • the length of the generated video may be set based on the length of the background sound (or video).
  • a corresponding video may be generated based on the length specified by the user operation.
  • the user is allowed to adjust more parameters of video generation, thereby facilitating the user to generate a satisfactory video work.
  • the editing page 610 may further include an option for specifying an animation mode for at least any one of text input and video background.
  • the animation mode here can include various display modes of text input and video background.
  • an animation mode for text input can specify that the text input be displayed in a gradient manner, a motion track manner.
  • the animation mode for the video background may specify the manner in which the background is displayed.
  • the animation mode can specify the display area, number, display method (stretching or tiled display), display track, etc. of the image.
  • the animation mode can specify that a portion of the video is used as the background for the generated video, can specify the relationship between the video background and the resolution of the generated video, and so on.
  • the video background is an emoticon (or an emoticon animation)
  • the number of emoticons included in the generated video, the display position of the emoticon, and the motion trajectory, etc. can be specified.
  • FIG. 7 schematically illustrates a block diagram of a user interface 700 for previewing a video, according to an exemplary implementation of the present disclosure.
  • the text input will move in the direction indicated by arrow 720, reappear in the upper part of the display area after moving out of the lower part of the display area, and so on.
  • the three images 510, 512 and 514 can be moved in a randomly selected straight line direction. For example, image 512 may move in direction 710 and reorient the movement when reaching the boundaries of the display area.
  • a predetermined default animation mode can be provided, in which the user does not have to select parameters related to animation display one by one, but can directly select a static background image to generate a dynamic video.
  • a default animation mode for background images may specify that 3 images are displayed, with the images jumping through the video. At this time, when the user selects the moon cake pattern, the generated video will include the jumping effect of 3 moon cake patterns.
  • another default animation mode may specify that 1 image is displayed and that the image rotate in the video. At this point, the resulting video will include a spinning animation of the mooncake pattern.
  • the default animation mode for text entry may specify that the text entry is displayed at the center of the video.
  • dynamic video pictures can be generated based on static text input. In this way, a richer visual expression can be provided to the user, thereby satisfying the needs of different users.
  • the video is published in the information sharing application.
  • the "next" button 640 may be pressed in order to generate a video.
  • the video herein may be a video file in various formats supported by the application 110 .
  • text-mode videos can be generated and published in a single application. Compared with the prior art solution of switching between a video editing application and an information sharing application, the above-described method can generate and publish a video in a simpler and more efficient manner without switching applications.
  • the code of the emoji may be stored in association with the video. It will be understood that there may be differences in the drawing of emoji when the terminal device adopts different operating systems. Suppose the user inputs the emoji "smiley”, and the code of the emoji is "001", then the code "001" can be stored directly instead of directly adding the video content based on the operating system of the user's terminal device. Emoticons.
  • FIG. 8 schematically shows a block diagram of an apparatus 800 for generating a text-mode video according to an exemplary implementation of the present disclosure.
  • the apparatus 800 includes: a receiving module 810, configured to receive a request for generating a video from a user of an information sharing application; a display module 820, configured to display an initial page for generating a video in the information sharing application , the initial page includes a prompt for entering text; an obtaining module 830 is configured to obtain text input from the user in response to detecting a user's touch in the area where the initial page is located; and a generating module 840 is configured to based on the text Enter Generate video for posting in information sharing applications.
  • a receiving module 810 configured to receive a request for generating a video from a user of an information sharing application
  • a display module 820 configured to display an initial page for generating a video in the information sharing application , the initial page includes a prompt for entering text
  • an obtaining module 830 is configured to obtain text input from the user in response to detecting a user's touch in the area where the initial page is located
  • a generating module 840 is configured to based on the text
  • the initial page further includes an option for selecting a video background; and the generating module 840 is further configured to include: in response to receiving the user-selected video background, generating a video based on the video background, the video background including At least one of the following: image, video, emoji, and emoji animation.
  • the initial page further includes a speaking option for speaking the text input; and the generating module 840 is further configured to include: in response to receiving a user selection of the speaking option, based on the audio of the speaking text input Generate video.
  • the reading options include at least any one of the following: the speaker's gender, age, voice style, and speaking rate.
  • the generating module 840 is further configured to generate a video based on the textual input in response to receiving a user de-selection of the read aloud option.
  • the generating module 840 includes: an editing page display module configured to display an editing page for generating a video in the information sharing application in response to detecting that the user confirms the initial page; and the generating module 840 further It includes: a video generation module, configured to generate a video based on the user's operation on the editing page.
  • the editing page includes: an option for editing at least any one of the text input, the video background, and the read-aloud option; and the video generation module is further configured to: based on the post-editing specified by the user operation option to generate video.
  • the editing page includes: an option for selecting to add background sound to the video; and the video generation module is further configured to: generate the video based on the background sound specified by the user operation.
  • the editing page includes: an option for selecting to add a sticker to the video; and the video generation module is further configured to: generate a video based on the sticker specified by the user operation, the sticker including a text sticker and an image sticker .
  • the editing page includes: an option for specifying a length of the video; and the video generating module is further configured to: generate the video based on the length specified by the user operation.
  • the editing page includes: an option for specifying an animation mode of at least any one of text input and a video background; and the video generation module is further configured to: specify the animation mode based on a user operation or Predetermined animation mode to generate video.
  • the text input includes an emoji
  • the generating module 840 includes an emoji storage module configured to store a code of the emoji in association with the video, for use in accordance with the terminal for playing the video The type of device to display the emoji corresponding to the code.
  • the apparatus 800 further includes: a publishing module configured to publish the video in the information sharing application in response to a request from the user for publishing the video.
  • the units included in the apparatus 800 may be implemented in various manners, including software, hardware, firmware, or any combination thereof.
  • one or more units may be implemented using software and/or firmware, such as machine-executable instructions stored on a storage medium.
  • some or all of the units in apparatus 800 may be implemented, at least in part, by one or more hardware logic components.
  • exemplary types of hardware logic components include field programmable gate arrays (FPGAs), application specific integrated circuits (ASICs), application specific standards (ASSPs), systems on chips (SOCs), complex programmable logic devices (CPLD), etc.
  • FIG. 9 shows a block diagram of a computing device/server 900 in which one or more implementations of the present disclosure may be implemented. It should be understood that the computing device/server 900 shown in FIG. 9 is merely exemplary and should not constitute any limitation on the functionality and scope of the implementations described herein.
  • computing device/server 900 is in the form of a general purpose computing device.
  • Components of computing device/server 900 may include, but are not limited to, one or more processors or processing units 910, memory 920, storage devices 930, one or more communication units 940, one or more input devices 950, and one or more Output device 960.
  • the processing unit 910 may be an actual or virtual processor and can perform various processes according to programs stored in the memory 920 . In a multi-processor system, multiple processing units execute computer-executable instructions in parallel to increase the parallel processing capabilities of the computing device/server 900 .
  • Computing device/server 900 typically includes multiple computer storage media. Such media can be any available media that can be accessed by computing device/server 900, including but not limited to volatile and nonvolatile media, removable and non-removable media.
  • Memory 920 may be volatile memory (eg, registers, cache, random access memory (RAM)), non-volatile memory (eg, read only memory (ROM), electrically erasable programmable read only memory (EEPROM) , Flash) or some combination of them.
  • Storage device 930 may be removable or non-removable media, and may include machine-readable media, such as flash drives, magnetic disks, or any other media that may be capable of storing information and/or data (eg, training data for training). ) and can be accessed within computing device/server 900.
  • Computing device/server 900 may further include additional removable/non-removable, volatile/non-volatile storage media.
  • disk drives for reading or writing from removable, non-volatile magnetic disks eg, "floppy disks"
  • CD-ROM drive for reading or writing.
  • each drive may be connected to a bus (not shown) by one or more data media interfaces.
  • Memory 920 may include a computer program product 925 having one or more program modules configured to perform various methods or actions of various implementations of the present disclosure.
  • the communication unit 940 enables communication with other computing devices through a communication medium. Additionally, the functions of the components of computing device/server 900 may be implemented in a single computing cluster or multiple computing machines capable of communicating over a communication connection. Thus, computing device/server 900 may operate in a networked environment using logical connections to one or more other servers, network personal computers (PCs), or another network node.
  • PCs network personal computers
  • Input device 950 may be one or more input devices, such as a mouse, keyboard, trackball, and the like.
  • Output device 960 may be one or more output devices, such as a display, speakers, printer, and the like.
  • the computing device/server 900 may also communicate with one or more external devices (not shown), such as storage devices, display devices, etc., through the communication unit 940, as needed, with one or more external devices that enable the user to communicate with the computing device/server. 900 interacts with any device (eg, network card, modem, etc.) that enables computing device/server 900 to communicate with one or more other computing devices. Such communication may be performed via an input/output (I/O) interface (not shown).
  • I/O input/output
  • a computer-readable storage medium having stored thereon one or more computer instructions, wherein the one or more computer instructions are executed by a processor to implement the method described above.
  • These computer readable program instructions may be provided to a processing unit of a general purpose computer, special purpose computer or other programmable data processing apparatus to produce a machine that causes the instructions when executed by the processing unit of the computer or other programmable data processing apparatus , resulting in means for implementing the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
  • These computer readable program instructions can also be stored in a computer readable storage medium, these instructions cause a computer, programmable data processing apparatus and/or other equipment to operate in a specific manner, so that the computer readable medium on which the instructions are stored includes An article of manufacture comprising instructions for implementing various aspects of the functions/acts specified in one or more blocks of the flowchart and/or block diagrams.
  • Computer-readable program instructions can also be loaded onto a computer, other programmable data processing apparatus, or other equipment to cause a series of operational steps to be performed on the computer, other programmable data processing apparatus, or other equipment to produce a computer-implemented process , thereby causing instructions executing on a computer, other programmable data processing apparatus, or other device to implement the functions/acts specified in one or more blocks of the flowcharts and/or block diagrams.
  • each block in the flowchart or block diagrams may represent a module, segment, or portion of instructions, which comprises one or more executables for implementing the specified logical function(s) instruction.
  • the functions noted in the blocks may occur out of the order noted in the figures. For example, two blocks in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved.
  • each block of the block diagrams and/or flowchart illustrations, and combinations of blocks in the block diagrams and/or flowchart illustrations can be implemented in dedicated hardware-based systems that perform the specified functions or actions , or can be implemented in a combination of dedicated hardware and computer instructions.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Graphics (AREA)
  • General Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

在本公开中,提供了用于生成文字模式的视频的方法、装置、设备和介质。在一种方法中,从信息共享应用的用户接收用于生成视频的请求。在信息共享应用中,显示用于生成视频的初始页面,初始页面包括用于输入文字的提示。响应于在初始页面所在的区域中检测到用户的触摸,获取来自用户的文字输入。基于文字输入生成视频,以用于在信息共享应用中发布。根据本公开的示例性实现方式,提供了相应的装置、设备和介质。利用本公开的示例性实现方式,在信息共享应用内部,用户可以直接基于文字输入来生成相应的视频。以此方式,可以降低用户操作的复杂度,并且向用户提供更为丰富的发布内容。

Description

用于生成文字模式的视频的方法、装置、设备和介质 技术领域
本公开的各实现方式涉及计算机领域,具体地,涉及用于生成文字模式的视频方法、装置、设备和计算机存储介质。
背景技术
随着信息技术的发展,目前已经提供了多种信息共享应用。用户可以编辑文字、拍摄照片或者视频,并且在信息共享应用中进行发布。由于视频可以包括例如声音、图像、文字等多方面的信息,视频信息已经成为广大用户乐于接受的信息类型。目前已经开发了支持向视频中插入文字的视频编辑应用。然而,当用户希望在信息共享应用中发布文字模式的视频时,不得不首先在视频编辑应用中生成视频并存储,继而在信息共享应用中上传该视频。此时,如何以更为方便并且有效的方式来生成文字模式的视频,成为一个研究热点。
发明内容
在本公开的第一方面,提供了一种用于在信息共享应用中生成文字模式的视频的方法。在该方法中,从信息共享应用的用户接收用于生成视频的请求。在信息共享应用中,显示用于生成视频的初始页面,初始页面包括用于输入文字的提示。响应于在初始页面所在的区域中检测到用户的触摸,获取来自用户的文字输入。基于文字输入生成视频,以用于在信息共享应用中发布。
在本公开的第二方面,提供了一种用于在信息共享应用中生成文字模式的视频的装置。该装置包括:接收模块,配置用于从信息共享应用的用户接收用于生成视频的请求;显示模块,配置用于在 信息共享应用中,显示用于生成视频的初始页面,初始页面包括用于输入文字的提示;获取模块,配置用于响应于在初始页面所在的区域中检测到用户的触摸,获取来自用户的文字输入;以及生成模块,配置用于基于文字输入生成视频,以用于在信息共享应用中发布。
在本公开的第三方面,提供了一种电子设备。该电子设备包括:存储器和处理器;其中存储器用于存储一条或多条计算机指令,其中一条或多条计算机指令被处理器执行以实现根据本公开的第一方面的方法。
在本公开的第四方面,提供了一种计算机可读存储介质,其上存储有一条或多条计算机指令,其中一条或多条计算机指令被处理器执行实现根据本公开的第一方面的方法。
利用本公开的示例性实现方式,用户可以在信息共享应用内部直接基于文字输入来生成相应的视频。以此方式,可以降低用户操作的复杂度,并且向用户提供更为丰富的发布内容。
附图说明
结合附图并参考以下详细说明,本公开各实现方式的上述和其他特征、优点及方面将变得更加明显。在附图中,相同或相似的附图标注表示相同或相似的元素,其中:
图1示意性示出了根据本公开的示例性实现方式的应用环境的框图;
图2示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的用户界面的框图;
图3示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的方法的流程图;
图4示意性示出了根据本公开的示例性实现方式的用于输入文字的用户界面的框图;
图5示意性示出了根据本公开的示例性实现方式的用于选择视 频背景的用户界面的框图;
图6示意性示出了根据本公开的示例性实现方式的用于编辑视频的用户界面的框图;
图7示意性示出了根据本公开的示例性实现方式的用于预览视频的用户界面的框图;
图8示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的装置的框图;以及
图9示出了能够实施本公开的多个实现方式的计算设备的框图。
具体实施方式
下面将参照附图更详细地描述本公开的实现方式。虽然附图中显示了本公开的某些实现方式,然而应当理解的是,本公开可以通过各种形式来实现,而且不应该被解释为限于这里阐述的实现方式,相反提供这些实现方式是为了更加透彻和完整地理解本公开。应当理解的是,本公开的附图及实现方式仅用于示例性作用,并非用于限制本公开的保护范围。
在本公开的实现方式的描述中,术语“包括”及其类似用语应当理解为开放性包含,即“包括但不限于”。术语“基于”应当理解为“至少部分地基于”。术语“一个实现方式”或“该实现方式”应当理解为“至少一个实现方式”。术语“第一”、“第二”等等可以指代不同的或相同的对象。下文还可能包括其他明确的和隐含的定义。
目前已经提供了多种信息共享应用(简称为应用)。由于视频可以包括多方面的信息,广大用户更乐于接受视频类型的媒体信息。为便于描述起见,首先参见图1概述根据本公开的示例性实现方式的应用环境。具体地,该图1示意性示出了根据本公开的示例性实现方式的应用环境100的框图。在图1中,用户可以通过应用110来观看和/或发布视频。例如,应用110可以向用户推送视频120,并且用户可以通过搜索、下滑、翻页等操作来观看喜爱的视频。此 外,用户可以按下“发布”按钮130来发布视频。
已经开发出了多种视频发布模式,例如,用户可以利用拍照、分段拍、快拍、从相册中上传等方式来发布视频。各个用户可以选择自己喜好的方式来发布视频。有些用户可能希望发布基于文字制作的视频,例如,用户期望输入诸如“中秋快乐”、“生日快乐”等祝福语,并且生成相应的视频来用于发布。
目前,已经开发了支持向视频中插入文字的视频编辑应用。然而,应用110的用户不得不首先在视频编辑应用中生成视频并存储,继而在应用110中上传该视频以便发布。上述操作涉及多个应用,导致用户的操作复杂并且难以在例如手机等具有较小屏幕区域的终端设备处实现。此时,如何以更为方便并且有效的方式来为信息共享应用的用户生成文字模式的视频,成为一个研究热点。
为了至少部分地解决已有技术方案中的上述和/或其他不足,根据本公开的示例性实现方式,提出了一种用于在信息共享应用中生成文字模式的视频的方法。在该方法中,可以从信息共享应用的用户接收用于生成视频的请求,继而启动用于生成文字模式的视频的方法。在下文中,将参见图2描述根据本公开的示例性实现方式的概要,该图2示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的用户界面200的框图。
当用户按下图1中的发布按钮130之后,可以进入如图2所示的用户界面200。用户可以在用户界面200底部的菜单中选择文字模式220,以便启动根据本公开的示例性实现方式的生成方法。此时,在应用110中,可以显示用于生成视频的初始页面210。在此初始页面210可以包括用于输入文字的提示:“轻触即可输入文字”。用户可以在初始页面210中输入相应的文字。例如,用户可以在初始页面210所在的区域中执行触摸操作,以便启动输入文字的过程。
继而,应用110从用户获取文字输入,并且生成包括该文字输入的视频以便用于发布。将会理解,图2所示的页面布局仅仅是示意性的,根据本公开的示例性实现方式,可以采用其他页面布局, 只要能够实现根据本公开的示例性实现方式的方法即可。
利用本公开的示例性实现方式,用户不必单独调用视频编辑应用,而是可以在信息共享应用内部直接基于文字输入来生成相应的视频。以此方式,可以降低用户操作的复杂度,避免用户在多个应用之间切换期间可能造成的错误,并且向用户提供更为丰富的发布内容。
在下文中,将参见图3描述根据本公开的示例性实现方式的更多细节。图3示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的方法300的流程图。在框310处,从信息共享应用的用户接收用于生成视频的请求。根据本公开的示例性实现方式,用户可以滑动如图2所示的用户界面200底部的菜单,并且从多种视频模式中选择文字模式220。
在图3的框320处,在信息共享应用中,显示用于生成视频的初始页面210,在此初始页面210包括用于输入文字的提示。根据本公开的示例性实现方式,可以在初始页面中的醒目位置处显示输入提示。用户可以根据提示来输入期望的文字,例如,用户可以通过触摸初始页面210中的任何空白区域来调用输入对话框,以便输入文字。
在框330处,响应于在初始页面210所在的区域中检测到用户的触摸,获取来自用户的文字输入。用户可以触摸初始页面210中的任意空白区域以便输入文字,在下文中将参见图4描述有关文字输入的更多细节。该图4示意性示出了根据本公开的示例性实现方式的用于输入文字的用户界面400的框图。如图4所示,当用户触摸初始页面210的空白区域之后,可以弹出输入框410以用于接收文字输入。例如,用户可以输入纯文本内容“中秋快乐”。
根据本公开的示例性实现方式,文字输入可以包括文本和表情符号。此时,用户还可以输入例如“笑脸”的表情符号。将会理解,在此的表情符号可以是由移动终端上的操作系统所绘制的表情符号,每个表情符号可以具有唯一的代码。对于某个代码而言,绘制 的表情符号图像可以随着操作系统的不同而有所区别。例如,在两种操作系统所绘制的“笑脸”中,嘴角上扬的程度可以有所不同。
在框340处,基于文字输入生成视频,以用于在信息共享应用中发布。在已经获得了文字输入的情况下,可以生成包括该文字输入的视频以便用于发布。将会理解,文字输入是用于生成视频的最基本要素,对于其他要素而言,可以基于默认的视频背景来生成默认长度的视频。例如,应用110可以基于文字的内容来选择月夜背景并生成包括文字“中秋快乐”的视频。
根据本公开的示例性实现方式,初始页面210可以包括更多的选项。在下文中,返回图2描述有关初始页面210的更多细节。根据本公开的示例性实现方式,初始页面210可以进一步包括用于选择视频背景的选项234。用户可以点击选项234来选择期望的视频背景,例如,可以从图像、视频、表情符号以及表情动画中选择一个或者多个作为背景。可以基于用户选择的视频背景,来生成视频。假设用户选择月饼的图像,则生成的视频的背景将包括月饼图案。
根据本公开的示例性实现方式,在选择视频背景的对话框中,可以进一步指定图像位置、图像的数量、图像的运动轨迹,等等。参见图5描述有关视频背景的更多细节,该图5示意性示出了根据本公开的示例性实现方式的用于选择视频背景的用户界面500的框图。用户可以选择月饼图像作为背景,并且可以指定在视频中包括随机分布的3个图像。此时,生成的视频中将包括图像510、520和530。进一步,可以指定图像沿着某个方向运动。例如,可以预先定义运动轨迹,例如,直线、曲线等;备选地和/或附加地,可以随机生成运动轨迹。根据本公开的示例性实现方式,可以设置附加规则:例如,可以指定在显示多个图像的情况下,应当避免图像之间的碰撞;又例如,可以指定当图像到达显示边界时改变运动方向,等等。
根据本公开的示例性实现方式,可以选择视频来作为背景,并且可以指定使用视频中的某个时间段内的部分(例如,指定时间段的开始时间和终止时间),可以选择视频中的某个区域(例如,指 定使用某个窗口范围内的部分),等等。根据本公开的示例性实现方式,可以选择表情符号或者表情动画来作为视频背景。利用本公开的示例性实现方式,可以为生成视频提供更为丰富的素材,进而满足广大用户多方面的需求。
返回图2,继续描述有关初始页面210的更多细节。根据本公开的示例性实现方式,初始页面210可以进一步包括用于朗读文字输入的朗读选项230。用户可以通过点击操作来启动或者取消自动朗读功能。当用户启动自动朗读功能时,应用110可以基于人工智能技术来自动朗读用户输入的文字,并且基于朗读的音频来生成视频。此时,生成的视频可以包括朗读的音频;备选地和/或附加地,生成的视频可以包括文字内容和音频内容两者。
根据本公开的示例性实现方式,朗读选项可以进一步包括以下中的至少任一项:朗读者的性别、年龄、声音风格以及语速。以此方式,用户可以选择不同性别、不同年龄的朗读者。根据本公开的示例性实现方式,可以提供多种声音风格来满足不同用户的需求。例如,声音风格可以包括但不限于:浑厚型、甜美型、活泼型,等等。用户可以选择高、中或者低的不同的语速,以便支持对于朗读效果的个性化设置。
根据本公开的示例性实现方式,用户可以取消朗读选项,此时生成的视频仅包括文字内容。根据本公开的示例性实现方式,可以向用户提供生成视频的多种素材,从而提供更为丰富的媒体表现方式。
上文已经参见附图描述了有关初始页面210的内容。用户可以在初始页面210中进行设置,以便定义生成视频的各项参数。当用户确认初始页面210中的设置之后,可以点击“下一步”按钮232以便显示编辑页面。在下文中,将参见图6描述有关编辑页面的更多细节,该图6示意性示出了根据本公开的示例性实现方式的用于编辑视频的用户界面600的框图。用户可以在编辑页面610中进行操作,并且应用110可以基于用户对编辑页面610的用户操作来生 成相应的视频。
根据本公开的示例性实现方式,编辑页面610可以包括以下中的至少任一项:用于编辑朗读设置的选项620、用于编辑文字输入的选项622、以及用于编辑视频背景的选项624。在编辑页面610中,用户可以经由选项620来启动或者关闭自动朗读功能;用户可以经由选项622来编辑已经输入的文字,可以设置文字的字体、字号、颜色、显示位置,等等;用户可以经由选项624来编辑已经选择的背景、重新选择背景或者添加新的背景,等等。
在已经编辑了希望调整的参数之后,用户可以按下“下一步”按钮640,以便基于用户在编辑页面610中指定的编辑后的选项,生成相应的视频。利用本公开的示例性实现方式,编辑页面610可以向用户提供修改各项参数的功能。以此方式,在用户在不满意先前设置时提供可以修改机会,进而便于用户操作并且生成满意的视频。
根据本公开的示例性实现方式,编辑页面610可以进一步包括:用于选择向视频中添加背景声音的选项630。在此的背景声音可以包括背景音乐和/或例如人物解说等其他声音。例如,用户可以按下选项630来选择视频的背景音乐或者其他音频。备选地和/或附加地,用户可以录入旁白,例如,用户可以朗读有关中秋的诗歌,等等。
在用户已经选择了期望的背景声音之后,应用110可以基于用户操作指定的背景声音,生成相应的视频。利用本公开的示例性实现方式,可以允许用户向视频中加入更为多样的声音文件,以便生成更为丰富的视频内容。
根据本公开的示例性实现方式,编辑页面610可以进一步包括:用于选择向视频中添加贴纸的选项632。在此贴纸可以包括文字贴纸以及图像贴纸。文字贴纸可以包括文字,例如,各种艺术字体的常用语等。图像贴纸可以包括图标、常用表情、以及画框等内容。用户可以按下选项632来向视频中插入贴纸,例如,用户可以插入文字贴纸“阖家团圆”,并且插入图像贴纸“红心”等。进一步,用户可以通过触摸、拖动、旋转、缩放等操作来调整贴纸的位置、大 小和方向,等等。
在用户已经选择了期望的贴纸之后,应用110可以基于用户操作指定的贴纸,生成相应的视频。利用本公开的示例性实现方式,可以允许用户向视频中加入更多的个性元素。以此方式,可以提高视频的趣味性并且提供更为丰富的媒体表现。
根据本公开的示例性实现方式,编辑页面610可以进一步包括:用于指定视频的长度的选项。视频可以具有例如3秒(或者其他数值)的默认长度。为了提供更优质的定制服务,可以允许用户自定义视频的长度。进一步,当用户选择了背景声音(或视频)时,可以允许用户进一步设置背景声音(或视频)和视频长度之间的匹配关系。在默认情况下,可以从背景声音(或视频)中截取与视频长度相匹配的声音(或视频)片段。如果用户指定的长度大于背景声音(或视频)的长度,则用户可以设置循环播放。备选地和/或附加地,可以基于背景声音(或视频)的长度来设置所生成视频的长度。
在用户已经选择了期望的长度之后,可以基于用户操作指定的长度,生成相应的视频。利用本公开的示例性实现方式,允许用户调整视频生成的更多参数,进而便于用户生成满意的视频作品。
根据本公开的示例性实现方式,编辑页面610可以进一步包括:用于指定文字输入和视频背景中的至少任一项的动画模式的选项。在此的动画模式可以包括文字输入和视频背景的多种显示方式。例如,用于文字输入的动画模式可以指定以渐变方式、运动轨迹方式来显示文字输入。
根据本公开的示例性实现方式,用于视频背景的动画模式可以指定显示背景的方式。当视频背景为图像时,动画模式可以指定图像的显示区域、数量、显示方式(伸缩显示或者平铺显示)、显示轨迹,等等。当视频背景为视频时,动画模式可以指定将视频中的某个时间段内的部分用作生成视频的背景,可以指定视频背景与生成视频的分辨率之间的关系,等等。当视频背景为表情符号(或表情动画)时,可以指定在生成视频中包括的表情符号的数量、表情 符号的显示位置以及运动轨迹,等等。
进一步,可以基于用户操作指定的动画模式,生成相应的视频。假设用户指定从屏幕上方向下循环地移动文字输入,背景中包括3个图像,每个图像分别按照随机选择的直线方向运动,并且在到达显示区域的边界时改变运动方向。此时,生成的视频将如图7所示。图7示意性示出了根据本公开的示例性实现方式的用于预览视频的用户界面700的框图。在图7中,文字输入将沿着箭头720所示方向运动,在移出显示区域下部之后重新出现在显示区域的上部,如此循环。3个图像510、512和514可以沿着随机选择的直线方向运动。例如,图像512可以沿着方向710运动,并且在到达显示区域的边界时重新确定运动方向。
根据本公开的示例性实现方式,可以提供预定的默认动画模式,此时用户不必逐一选择有关动画显示的各项参数,而是可以直接选择的静态的背景图像来生成动态的视频。在一个示例中,用于背景图像的一个默认动画模式可以指定显示3个图像,并且图像在视频中跳跃。此时,当用户选择月饼图案时,生成的视频将包括3个月饼图案的跳跃效果。备选地和/或附加地,另一默认动画模式可以指定显示1个图像,并且指定图像在视频中旋转。此时,生成的视频将包括月饼图案的旋转动画。在另一示例中,用于文字输入的默认动画模式可以指定在视频的中心位置处显示文字输入。
利用本公开的示例性实现方式,可以基于静态的文字输入来生成动态的视频画面。以此方式,可以向用户提供更为丰富的视觉表达,进而满足不同用户的需求。
根据本公开的示例性实现方式,如果接收到来自用户的用于发布视频的请求,在信息共享应用中发布视频。根据本公开的示例性实现方式,当用户已经在编辑页面610中完成操作之后,可以按下“下一步”按钮640,以便生成视频。将会理解,在此的视频可以是由应用110所支持的各种格式的视频文件。利用本公开的示例性实现方式,可以在单一应用中生成并且发布文字模式的视频。相对于 在视频编辑应用和信息共享应用之间切换的已有技术方案而言,采用上文描述的方法可以在无需切换应用的情况下,以更为简单并且有效的方式生成并且发布视频。
根据本公开的示例性实现方式,如果用户输入的文字和/或所选择的背景图像包括依赖于终端设备的表情符号,则可以与视频相关联地存储表情符号的代码。将会理解,当终端设备采用不同的操作系统时表情符号的绘制可以存在区别。假设用户输入了表情符号“笑脸”,并且该表情符号的代码为“001”,此时可以直接存储代码“001”,而不是在视频内容中直接加入基于用户的终端设备的操作系统所绘制的表情符号。以此方式,当另一用户播放生成的视频时,则可以基于另一用户的终端设备的操作系统的类型,来在视频中显示相应的“笑脸”。利用本公开的示例性实现方式,可以跨越多种操作系统来为用户提供更多的选择。
上文已经参见图1至图7描述根据本公开的示例性实现方式的方法300的详细信息。根据本公开的示例性实现方式,进一步提供了用于实现上述方法或过程的相应装置。图8示意性示出了根据本公开的示例性实现方式的用于生成文字模式的视频的装置800的框图。具体地,该装置800包括:接收模块810,配置用于从信息共享应用的用户接收用于生成视频的请求;显示模块820,配置用于在信息共享应用中,显示用于生成视频的初始页面,初始页面包括用于输入文字的提示;获取模块830,配置用于响应于在初始页面所在的区域中检测到用户的触摸,获取来自用户的文字输入;以及生成模块840,配置用于基于文字输入生成视频,以用于在信息共享应用中发布。
根据本公开的示例性实现方式,初始页面进一步包括用于选择视频背景的选项;以及生成模块840进一步配置用于包括:响应于接收到用户选择的视频背景,基于视频背景生成视频,视频背景包括以下中的至少任一项:图像、视频、表情符号以及表情动画。
根据本公开的示例性实现方式,初始页面进一步包括用于朗读 文字输入的朗读选项;以及生成模块840进一步配置用于包括:响应于接收到用户对朗读选项的选择,基于朗读文字输入的音频来生成视频。
根据本公开的示例性实现方式,朗读选项包括以下中的至少任一项:朗读者的性别、年龄、声音风格以及语速。
根据本公开的示例性实现方式,生成模块840进一步配置用于:响应于接收到用户取消对朗读选项的选择,基于文字输入来生成视频。
根据本公开的示例性实现方式,生成模块840包括:编辑页面显示模块,配置用于响应于检测到用户确认初始页面,在信息共享应用中显示用于生成视频的编辑页面;以及生成模块840进一步包括:视频生成模块,配置用于基于用户对编辑页面的用户操作,生成视频。
根据本公开的示例性实现方式,编辑页面包括:用于编辑文字输入、视频背景以及朗读选项中的至少任一项的选项;以及视频生成模块进一步配置用于:基于由用户操作指定的编辑后的选项,生成视频。
根据本公开的示例性实现方式,编辑页面包括:用于选择向视频中添加背景声音的选项;以及视频生成模块进一步配置用于:基于用户操作指定的背景声音,生成视频。
根据本公开的示例性实现方式,编辑页面包括:用于选择向视频中添加贴纸的选项;以及视频生成模块进一步配置用于:基于用户操作指定的贴纸,生成视频,贴纸包括文字贴纸以及图像贴纸。
根据本公开的示例性实现方式,编辑页面包括:用于指定视频的长度的选项;以及视频生成模块进一步配置用于:基于用户操作指定的长度,生成视频。
根据本公开的示例性实现方式,编辑页面包括:用于指定文字输入和视频背景中的至少任一项的动画模式的选项;以及视频生成模块进一步配置用于:基于用户操作指定的动画模式或者预定动画 模式,生成视频。
根据本公开的示例性实现方式,文字输入包括表情符号,以及生成模块840包括:表情符号存储模块,配置用于与视频相关联地存储表情符号的代码,以用于按照用于播放视频的终端设备的类型来显示与代码相对应的表情符号。
根据本公开的示例性实现方式,该装置800进一步包括:发布模块,配置用于响应于来自用户的用于发布视频的请求,在信息共享应用中发布视频。
根据本公开的示例性实现方式,装置800中所包括的单元可以利用各种方式来实现,包括软件、硬件、固件或其任意组合。在一些实现方式中,一个或多个单元可以使用软件和/或固件来实现,例如存储在存储介质上的机器可执行指令。除了机器可执行指令之外或者作为替代,装置800中的部分或者全部单元可以至少部分地由一个或多个硬件逻辑组件来实现。作为示例而非限制,可以使用的示范类型的硬件逻辑组件包括现场可编程门阵列(FPGA)、专用集成电路(ASIC)、专用标准品(ASSP)、片上系统(SOC)、复杂可编程逻辑器件(CPLD),等等。
图9示出了其中可以实施本公开的一个或多个实现方式的计算设备/服务器900的框图。应当理解,图9所示出的计算设备/服务器900仅仅是示例性的,而不应当构成对本文所描述的实现方式的功能和范围的任何限制。
如图9所示,计算设备/服务器900是通用计算设备的形式。计算设备/服务器900的组件可以包括但不限于一个或多个处理器或处理单元910、存储器920、存储设备930、一个或多个通信单元940、一个或多个输入设备950以及一个或多个输出设备960。处理单元910可以是实际或虚拟处理器并且能够根据存储器920中存储的程序来执行各种处理。在多处理器系统中,多个处理单元并行执行计算机可执行指令,以提高计算设备/服务器900的并行处理能力。
计算设备/服务器900通常包括多个计算机存储介质。这样的介 质可以是计算设备/服务器900可访问的任何可以获得的介质,包括但不限于易失性和非易失性介质、可拆卸和不可拆卸介质。存储器920可以是易失性存储器(例如寄存器、高速缓存、随机访问存储器(RAM))、非易失性存储器(例如,只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、闪存)或它们的某种组合。存储设备930可以是可拆卸或不可拆卸的介质,并且可以包括机器可读介质,诸如闪存驱动、磁盘或者任何其他介质,其可以能够用于存储信息和/或数据(例如用于训练的训练数据)并且可以在计算设备/服务器900内被访问。
计算设备/服务器900可以进一步包括另外的可拆卸/不可拆卸、易失性/非易失性存储介质。尽管未在图9中示出,可以提供用于从可拆卸、非易失性磁盘(例如“软盘”)进行读取或写入的磁盘驱动和用于从可拆卸、非易失性光盘进行读取或写入的光盘驱动。在这些情况中,每个驱动可以由一个或多个数据介质接口被连接至总线(未示出)。存储器920可以包括计算机程序产品925,其具有一个或多个程序模块,这些程序模块被配置为执行本公开的各种实现方式的各种方法或动作。
通信单元940实现通过通信介质与其他计算设备进行通信。附加地,计算设备/服务器900的组件的功能可以以单个计算集群或多个计算机器来实现,这些计算机器能够通过通信连接进行通信。因此,计算设备/服务器900可以使用与一个或多个其他服务器、网络个人计算机(PC)或者另一个网络节点的逻辑连接来在联网环境中进行操作。
输入设备950可以是一个或多个输入设备,例如鼠标、键盘、追踪球等。输出设备960可以是一个或多个输出设备,例如显示器、扬声器、打印机等。计算设备/服务器900还可以根据需要通过通信单元940与一个或多个外部设备(未示出)进行通信,外部设备诸如存储设备、显示设备等,与一个或多个使得用户与计算设备/服务器900交互的设备进行通信,或者与使得计算设备/服务器900与一 个或多个其他计算设备通信的任何设备(例如,网卡、调制解调器等)进行通信。这样的通信可以经由输入/输出(I/O)接口(未示出)来执行。
根据本公开的示例性实现方式,提供了一种计算机可读存储介质,其上存储有一条或多条计算机指令,其中一条或多条计算机指令被处理器执行以实现上文描述的方法。
这里参照根据本公开实现的方法、装置(系统)和计算机程序产品的流程图和/或框图描述了本公开的各个方面。应当理解,流程图和/或框图的每个方框以及流程图和/或框图中各方框的组合,都可以由计算机可读程序指令实现。
这些计算机可读程序指令可以提供给通用计算机、专用计算机或其他可编程数据处理装置的处理单元,从而生产出一种机器,使得这些指令在通过计算机或其他可编程数据处理装置的处理单元执行时,产生了实现流程图和/或框图中的一个或多个方框中规定的功能/动作的装置。也可以把这些计算机可读程序指令存储在计算机可读存储介质中,这些指令使得计算机、可编程数据处理装置和/或其他设备以特定方式工作,从而,存储有指令的计算机可读介质则包括一个制造品,其包括实现流程图和/或框图中的一个或多个方框中规定的功能/动作的各个方面的指令。
也可以把计算机可读程序指令加载到计算机、其他可编程数据处理装置、或其他设备上,使得在计算机、其他可编程数据处理装置或其他设备上执行一系列操作步骤,以产生计算机实现的过程,从而使得在计算机、其他可编程数据处理装置、或其他设备上执行的指令实现流程图和/或框图中的一个或多个方框中规定的功能/动作。
附图中的流程图和框图显示了根据本公开的多个实现的系统、方法和计算机程序产品的可能实现的体系架构、功能和操作。在这点上,流程图或框图中的每个方框可以代表一个模块、程序段或指令的一部分,模块、程序段或指令的一部分包含一个或多个用于实 现规定的逻辑功能的可执行指令。在有些作为替换的实现中,方框中所标注的功能也可以以不同于附图中所标注的顺序发生。例如,两个连续的方框实际上可以基本并行地执行,它们有时也可以按相反的顺序执行,这依所涉及的功能而定。也要注意的是,框图和/或流程图中的每个方框、以及框图和/或流程图中的方框的组合,可以用执行规定的功能或动作的专用的基于硬件的系统来实现,或者可以用专用硬件与计算机指令的组合来实现。
以上已经描述了本公开的各实现,上述说明是示例性的,并非穷尽性的,并且也不限于所公开的各实现。在不偏离所说明的各实现的范围和精神的情况下,对于本技术领域的普通技术人员来说许多修改和变更都是显而易见的。本文中所用术语的选择,旨在最好地解释各实现的原理、实际应用或对市场中的技术的改进,或者使本技术领域的其他普通技术人员能理解本文公开的各实现。

Claims (16)

  1. 一种用于在信息共享应用中生成文字模式的视频的方法,包括:
    从所述信息共享应用的用户接收用于生成所述视频的请求;
    在所述信息共享应用中,显示用于生成所述视频的初始页面,所述初始页面包括用于输入文字的提示;
    响应于在所述初始页面所在的区域中检测到所述用户的触摸,获取来自所述用户的文字输入;以及
    基于所述文字输入生成所述视频,以用于在所述信息共享应用中发布。
  2. 根据权利要求1所述的方法,其中所述初始页面进一步包括用于选择视频背景的选项;以及
    生成所述视频进一步包括:响应于接收到所述用户选择的视频背景,基于所述视频背景生成所述视频,所述视频背景包括以下中的至少任一项:图像、视频、表情符号以及表情动画。
  3. 根据权利要求1所述的方法,其中所述初始页面进一步包括用于朗读所述文字输入的朗读选项;以及
    生成所述视频进一步包括:响应于接收到所述用户对所述朗读选项的选择,基于朗读所述文字输入的音频来生成所述视频。
  4. 根据权利要求3所述的方法,其中所述朗读选项包括以下中的至少任一项:朗读者的性别、年龄、声音风格以及语速。
  5. 根据权利要求3所述的方法,其中生成所述视频进一步包括:响应于接收到所述用户取消对所述朗读选项的选择,基于所述文字输入来生成所述视频。
  6. 根据权利要求1所述的方法,其中生成所述视频包括:
    响应于检测到所述用户确认所述初始页面,在所述信息共享应用中显示用于生成所述视频的编辑页面;以及
    基于所述用户对所述编辑页面的用户操作,生成所述视频。
  7. 根据权利要求6所述的方法,其中所述编辑页面包括:用于编辑所述文字输入、所述视频背景以及所述朗读选项中的至少任一项的选项;以及
    基于所述用户操作生成所述视频包括:基于由所述用户操作指定的编辑后的选项,生成所述视频。
  8. 根据权利要求6所述的方法,其中所述编辑页面包括:用于选择向所述视频中添加背景声音的选项;以及
    基于所述用户操作生成所述视频包括:基于所述用户操作指定的背景声音,生成所述视频。
  9. 根据权利要求6所述的方法,其中所述编辑页面包括:用于选择向所述视频中添加贴纸的选项;以及
    基于所述用户操作来生成所述视频包括:基于所述用户操作指定的贴纸,生成所述视频,所述贴纸包括文字贴纸以及图像贴纸。
  10. 根据权利要求6所述的方法,其中所述编辑页面包括:用于指定所述视频的长度的选项;以及
    基于所述用户操作生成所述视频包括:基于所述用户操作指定的长度,生成所述视频。
  11. 根据权利要求6所述的方法,其中所述编辑页面包括:用于指定所述文字输入和所述视频背景中的至少任一项的动画模式的选项;以及
    基于所述用户操作生成所述视频包括:基于所述用户操作指定的动画模式或者预定动画模式,生成所述视频。
  12. 根据权利要求1所述的方法,其中所述文字输入包括表情符号,以及
    生成所述视频包括:与所述视频相关联地存储所述表情符号的代码,以用于按照用于播放所述视频的终端设备的类型来显示与所述代码相对应的表情符号。
  13. 根据权利要求1所述的方法,进一步包括:响应于来自所述用户的用于发布所述视频的请求,在所述信息共享应用中发布所 述视频。
  14. 一种用于在信息共享应用中生成文字模式的视频的装置,包括:
    接收模块,配置用于从所述信息共享应用的用户接收用于生成所述视频的请求;
    显示模块,配置用于在所述信息共享应用中,显示用于生成所述视频的初始页面,所述初始页面包括用于输入文字的提示;
    获取模块,配置用于响应于在所述初始页面所在的区域中检测到所述用户的触摸,获取来自所述用户的文字输入;以及
    生成模块,配置用于基于所述文字输入生成所述视频,以用于在所述信息共享应用中发布。
  15. 一种电子设备,包括:
    存储器和处理器;
    其中所述存储器用于存储一条或多条计算机指令,其中所述一条或多条计算机指令被所述处理器执行以实现根据权利要求1至13中任一项所述的方法。
  16. 一种计算机可读存储介质,其上存储有一条或多条计算机指令,其中所述一条或多条计算机指令被处理器执行以实现根据权利要求1至13中任一项所述的方法。
PCT/CN2021/119438 2020-09-25 2021-09-18 用于生成文字模式的视频的方法、装置、设备和介质 WO2022063093A1 (zh)

Priority Applications (5)

Application Number Priority Date Filing Date Title
KR1020237002186A KR102613143B1 (ko) 2020-09-25 2021-09-18 텍스트-비디오 생성 방법, 장치, 설비 및 매체
EP21871463.2A EP4171047A4 (en) 2020-09-25 2021-09-18 METHOD AND APPARATUS FOR GENERATING TEXT VIDEO, DEVICE AND MEDIUM
JP2023506273A JP7450112B2 (ja) 2020-09-25 2021-09-18 文字モードでビデオを生成する方法、装置、機器、および媒体
US18/087,566 US11922975B2 (en) 2020-09-25 2022-12-22 Method, apparatus, device and medium for generating video in text mode
US18/429,190 US20240170026A1 (en) 2020-09-25 2024-01-31 Method, apparatus, device and medium for generating video in text mode

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011027603.2A CN112153475B (zh) 2020-09-25 2020-09-25 用于生成文字模式的视频的方法、装置、设备和介质
CN202011027603.2 2020-09-25

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US18/087,566 Continuation US11922975B2 (en) 2020-09-25 2022-12-22 Method, apparatus, device and medium for generating video in text mode

Publications (1)

Publication Number Publication Date
WO2022063093A1 true WO2022063093A1 (zh) 2022-03-31

Family

ID=73897580

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/119438 WO2022063093A1 (zh) 2020-09-25 2021-09-18 用于生成文字模式的视频的方法、装置、设备和介质

Country Status (6)

Country Link
US (2) US11922975B2 (zh)
EP (1) EP4171047A4 (zh)
JP (1) JP7450112B2 (zh)
KR (1) KR102613143B1 (zh)
CN (1) CN112153475B (zh)
WO (1) WO2022063093A1 (zh)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112153475B (zh) 2020-09-25 2022-08-05 北京字跳网络技术有限公司 用于生成文字模式的视频的方法、装置、设备和介质

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109215655A (zh) * 2018-10-30 2019-01-15 维沃移动通信有限公司 视频中添加文本的方法和移动终端
CN110134920A (zh) * 2018-02-02 2019-08-16 中兴通讯股份有限公司 绘文字兼容显示方法、装置、终端及计算机可读存储介质
WO2020150693A1 (en) * 2019-01-18 2020-07-23 Snap Inc. Systems and methods for generating personalized videos with customized text messages
CN112153475A (zh) * 2020-09-25 2020-12-29 北京字跳网络技术有限公司 用于生成文字模式的视频的方法、装置、设备和介质

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8744239B2 (en) 2010-08-06 2014-06-03 Apple Inc. Teleprompter tool for voice-over tool
US8701020B1 (en) * 2011-02-01 2014-04-15 Google Inc. Text chat overlay for video chat
US20130294746A1 (en) * 2012-05-01 2013-11-07 Wochit, Inc. System and method of generating multimedia content
US20160173960A1 (en) 2014-01-31 2016-06-16 EyeGroove, Inc. Methods and systems for generating audiovisual media items
US10623747B2 (en) 2014-06-20 2020-04-14 Hfi Innovation Inc. Method of palette predictor signaling for video coding
US20160234494A1 (en) 2015-02-10 2016-08-11 Qualcomm Incorporated Restriction on palette block size in video coding
US20170098324A1 (en) * 2015-10-05 2017-04-06 Vitthal Srinivasan Method and system for automatically converting input text into animated video
WO2017218901A1 (en) * 2016-06-17 2017-12-21 Infields, Llc Application for enhancing metadata tag uses for social interaction
EP3538644B1 (en) * 2016-11-10 2021-12-29 Becton, Dickinson and Company Timeline system for monitoring a culture media protocol
GB2555838A (en) 2016-11-11 2018-05-16 Sony Corp An apparatus, computer program and method
CN107145564A (zh) * 2017-05-03 2017-09-08 福建中金在线信息科技有限公司 一种信息发布方法及装置
KR20180125237A (ko) * 2017-05-15 2018-11-23 한경훈 모바일 단말기의 이모티콘 입력방법, 그 방법을 위한 소프트웨어를 저장하는 소프트웨어 분배 서버
KR101950674B1 (ko) * 2017-05-26 2019-05-17 (주)거노코퍼레이션 동영상 편집 방법을 컴퓨터에서 수행하기 위한 앱이 기록된 컴퓨터
CN110062269A (zh) * 2018-01-18 2019-07-26 腾讯科技(深圳)有限公司 附加对象显示方法、装置及计算机设备
CN109120866B (zh) * 2018-09-27 2020-04-03 腾讯科技(深圳)有限公司 动态表情生成方法、装置、计算机可读存储介质和计算机设备
JP2020053026A (ja) 2019-07-24 2020-04-02 株式会社ドワンゴ サーバシステム、アプリケーションプログラム配信サーバ、閲覧用端末、コンテンツ閲覧方法、アプリケーションプログラム、配信方法、アプリケーションプログラム配信方法

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110134920A (zh) * 2018-02-02 2019-08-16 中兴通讯股份有限公司 绘文字兼容显示方法、装置、终端及计算机可读存储介质
CN109215655A (zh) * 2018-10-30 2019-01-15 维沃移动通信有限公司 视频中添加文本的方法和移动终端
WO2020150693A1 (en) * 2019-01-18 2020-07-23 Snap Inc. Systems and methods for generating personalized videos with customized text messages
CN112153475A (zh) * 2020-09-25 2020-12-29 北京字跳网络技术有限公司 用于生成文字模式的视频的方法、装置、设备和介质

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
ANONYMOUS: "How to Make A Text Video with Rolling Captions on TikTok", 9 December 2019 (2019-12-09), XP055914130, Retrieved from the Internet <URL:http://www.coozhi.com/youxishuma/shouji/123723.html> *
ANONYMOUS: "How to Make A Text-to-Speech Rotating Video on ZiShuo App", BAIDU JINGYAN, 2 July 2019 (2019-07-02), XP055914128, Retrieved from the Internet <URL:https://jingyan.baidu.com/article/b87fe19ec678271218356881.html> *
ANONYMOUS: "Learn How to Make A Text Video on TikTok in 1 Minute", ZHIHU, 11 August 2020 (2020-08-11), XP055914132, Retrieved from the Internet <URL:https://zhuanlan.zhihu.com/p/180383015> *
See also references of EP4171047A4 *
YUDAN JUNSHOU TECHNOLOGIES: "It's that simple! Generate Douyin explosion text video in 3 seconds", 31 July 2018 (2018-07-31), CN, pages 1 - 3, XP009535824, Retrieved from the Internet <URL:https://www.sohu.com/a/244381326_100067544> *

Also Published As

Publication number Publication date
US20230130806A1 (en) 2023-04-27
JP2023534757A (ja) 2023-08-10
JP7450112B2 (ja) 2024-03-14
KR20230023804A (ko) 2023-02-17
EP4171047A4 (en) 2023-11-29
CN112153475B (zh) 2022-08-05
CN112153475A (zh) 2020-12-29
US20240170026A1 (en) 2024-05-23
KR102613143B1 (ko) 2023-12-13
EP4171047A1 (en) 2023-04-26
US11922975B2 (en) 2024-03-05

Similar Documents

Publication Publication Date Title
KR102490421B1 (ko) 터치 감응형 이차 디스플레이에서 사용자 인터페이스 제어부들을 동적으로 제공하기 위한 시스템들, 디바이스들, 및 방법들
JP5752708B2 (ja) 電子テキスト処理及び表示
US8559732B2 (en) Image foreground extraction using a presentation application
TWI653545B (zh) 用於即時手寫辨識之方法、系統及非暫時性電腦可讀媒體
US20240107127A1 (en) Video display method and apparatus, video processing method, apparatus, and system, device, and medium
US20160358367A1 (en) Animation based on Content Presentation Structures
US20060253783A1 (en) Story template structures associated with story enhancing content and rules
TWI478043B (zh) 行動裝置應用頁面樣版之產生系統、方法及其記錄媒體
US20230129847A1 (en) Method, apparatus and device for issuing and replying to multimedia content
US20220093132A1 (en) Method for acquiring video and electronic device
US20160267700A1 (en) Generating Motion Data Stories
US20140164371A1 (en) Extraction of media portions in association with correlated input
US11178356B2 (en) Media message creation with automatic titling
US20240170026A1 (en) Method, apparatus, device and medium for generating video in text mode
CN112040142B (zh) 用于移动终端上的视频创作的方法
US20140163956A1 (en) Message composition of media portions in association with correlated text
WO2024082981A1 (zh) 用于特效交互的方法、装置、设备和存储介质
US20230282240A1 (en) Media Editing Using Storyboard Templates
KR20130027412A (ko) 개인 비디오를 제작하는데 사용되는 편집시스템
WO2023246467A1 (zh) 用于视频推荐的方法、装置、设备和存储介质
WO2023217122A1 (zh) 视频剪辑模板搜索方法、装置、电子设备及存储介质
CN115580749A (zh) 展示方法、装置及可读存储介质
TWM437485U (en) Editing system for producing personalized audio/video

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21871463

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 20237002186

Country of ref document: KR

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 2021871463

Country of ref document: EP

Effective date: 20230119

ENP Entry into the national phase

Ref document number: 2023506273

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE