WO2010119181A1 - Video editing system - Google Patents

Video editing system Download PDF

Info

Publication number
WO2010119181A1
WO2010119181A1 PCT/FI2010/050309 FI2010050309W WO2010119181A1 WO 2010119181 A1 WO2010119181 A1 WO 2010119181A1 FI 2010050309 W FI2010050309 W FI 2010050309W WO 2010119181 A1 WO2010119181 A1 WO 2010119181A1
Authority
WO
WIPO (PCT)
Prior art keywords
video
videos
segments
compilation
editing system
Prior art date
Application number
PCT/FI2010/050309
Other languages
French (fr)
Inventor
Sari JÄRVINEN
Onni Ojutkangas
Johannes Peltola
Original Assignee
Valtion Teknillinen Tutkimuskeskus
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Valtion Teknillinen Tutkimuskeskus filed Critical Valtion Teknillinen Tutkimuskeskus
Publication of WO2010119181A1 publication Critical patent/WO2010119181A1/en

Links

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording

Definitions

  • the invention relates to a video editing system according to the preamble of claim 1.
  • the invention also relates to a method for editing videos according to the preamble of claim 16.
  • Video refers to live picture material consisting of successive pictures quickly played back, the duration of which can vary, recorded by means of video cameras, particularly digital video cameras, mobile phones or equivalent devices. Video recorded with a mobile phone can be referred to as mobile video.
  • the subjects of videos, particularly of personal videos, are, inter alia, family events, such as birthdays, and holiday trips, but videos are recorded as well of various everyday situations.
  • the editing of videos refers to the compiling and processing of one or more videos such that its picture material e.g. is rearranged, sections are deleted of it, or the picture material of several videos, i.e. its sections or segments, is combined to a new final video to be stored and shown.
  • a problem of the development of multimedia services in mobile devices is often the small size of their display and the limited data transfer capacity of the mobile network. Services, which offer the user a possibility to search or browse video contents with a mobile phone, often show the result by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down. This is awkward and time-consuming.
  • a problem of the compiling i.e. editing of videos is often defining the compilation basis.
  • An object of the invention is to eliminate problems related to known video editing systems and equivalent methods.
  • An object of the invention is also to provide a novel improved video editing system the principles of which are simple, efficient and cost-effective to implement.
  • a video editing system according to the invention is characterised by what is presented in claim 1.
  • a method for editing videos according to the invention is characterised by what is presented in claim 16.
  • the dependent claims present some advantageous embodiments of the invention.
  • the video editing system according to the invention comprises:
  • features to be defined of the videos include the geographical region where the video has been shot and the time and/or time period when the video has been shot and recorded.
  • the video editing system comprises means for selecting suitable sections, segments, of the videos, which means include a video feature analysing unit which analyses sound, picture and/or motion features calculated of the video (i.e. content analysis metadata) based on which the segments are selected.
  • a video feature analysing unit which analyses sound, picture and/or motion features calculated of the video (i.e. content analysis metadata) based on which the segments are selected.
  • the video editing system comprises means for creating a new video i.e. video compilation of the videos or their segments most advantageously automatically.
  • the selection of video segments is arranged to be implemented by calculating the above features and by giving each segment points based on this and by selecting those segments in the video compilation which comply with the predetermined point limits.
  • the video editing system comprises means for creating video scenes of the selected video segments by clustering the video segments on the basis of the time stamp and/or geographical data.
  • the video editing system comprises means for creating a new video or video presentation i.e. video compilation of the videos or their segments based on the contents of the user and those recommended to the user i.e. the selected segments such that the location, time or content analysis metadata constitute continua between the video segments.
  • the video editing system comprises means according to independent claim 1 for searching and retrieving videos from a multimedia database or equivalent video database.
  • the video editing system comprises means for searching, by means of specific search criteria, desired videos in a multimedia database, of which videos a video compilation is to be created.
  • the video editing system comprises means for combining classified video segments into a new video i.e. video compilation.
  • the video editing system comprises a video database, such as a multivideo database, which is shared by several users and accessible through a data network, such as the Internet.
  • the video editing system comprises means for updating the video database with new videos.
  • the video editing system comprises means for automatically performing video editing according to predetermined features.
  • the video editing system is arranged on a data network, advantageously on the Internet, where it is accessible for the users.
  • the video editing system comprises a music database from which the user can select music to accompany the video compilation.
  • the method according to the invention comprises the following steps:
  • the system searches lower-level metadata related to videos in the multimedia database and upgrades it into higher-level (semantic) metadata.
  • the system creates a video compilation based on the higher-level metadata.
  • the video compilation is created by segmenting the videos into smaller sections (video segmentation), by giving points to the video segments and by selecting the best segments based on the points into the compilation.
  • scenes are created of the video segments based on the context data of the videos. The scenes are combined into one video presentation.
  • the automatic video editing system it is possible to create a service by means of which the user can define specific features for the video compilation.
  • the user can supply the service with the desired video length and define the geographical region and time period for videos to be selected into the compilation.
  • the user is automatically provided with a personal video compilation in which the system selects the most interesting segments of the video contents created by the user.
  • the user can e.g. define that "I want a video compilation of videos shot at the summer house in July".
  • the user is supplied a service by means of which it is possible to define the context (the creation region and time period of the video), based on which the video compilation of the videos shot by the user or user group is created.
  • the context the creation region and time period of the video
  • the automatic video editing system is entered a set of videos selected based on context for analysis, segmentation and compilation.
  • the system creates a video compilation defined by the user and supplies it to the user for presentation.
  • the automatic video compilation system By means of the automatic video compilation system, it is possible to provide a service which facilitates the editing of videos.
  • the system segments and classifies videos automatically recommending interesting video contents for the user to be added into the compilation desired by the user.
  • the user selects the desired segments and the system creates a compilation of them inserting suitable effects between the scenes.
  • the editing of videos is easy also in mobile terminals.
  • the system searches videos in the multimedia database and performs their segmentation and classification.
  • the system gives points to the segments and recommends the user via the service the best segments to be included in the compilation.
  • the user selects the desired segments and the system creates a video compilation of the selected segments.
  • the system utilises the context data (location and time) of videos in creating the video compilation.
  • the video compilation can be shown by means of the service in the user's terminal.
  • the main principle is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation.
  • the segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction).
  • feature extraction The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values.
  • Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering). The combining is performed based on the contents of the user and those recommended to the user such that the location, time or content analysis metadata constitute continua between the segments.
  • the programme can be changed, whereby it is possible to shift to video contents related to a totally different location and time.
  • the channel compiles the contents into a continuous video stream which can be shown on an IPTV set top box, a digital picture frame or a computer display.
  • a problem of the development of multimedia services in mobile devices is often the small size of their display.
  • the result is often shown by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down.
  • the automatic video compilation it is possible to create a compilation of the videos obtained as a search result for the user in which the most interesting segments of various videos are displayed.
  • the user is able to have a general picture of the contents of the videos without scrolling up and down, and it is also possible to offer quicker operating services as the volume of transferable data decreases.
  • a problem of the development of multimedia services in mobile devices is often the small size of their display and the limited data transfer capacity of the mobile network.
  • Services which offer the user a possibility to search or browse video contents with a mobile phone, often show the result by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down.
  • the automatic video compilation it is possible to create a compilation of the videos obtained as a search result for the user in which the most interesting segments of various videos are displayed.
  • the user is able to have a general picture of the contents of the videos without scrolling up and down, and it is also possible to offer quicker operating services as the volume of transferable data possibly decreases (depends naturally on video coding vs. the size and number of keyframe pictures).
  • the user searches or browses the desired contents in the video achieves with his/her mobile phone. These tasks performed by the user generate a search for video contents in the Internet-based system.
  • the result of the search is entered in the service supplying the automatic video compilation which analyses the videos, segments them and creates a compilation of the selected segments. For this, parameters suitable for this service for the creation of the video compilation have been defined for the segmentation and video rendering.
  • the user selects some of the videos created by him/herself for the compilation.
  • the service offers the user some alternatives for background music one of which the user selects for the compilation.
  • the automatic video compilation system analyses, segments and compiles videos considering the music style, piece duration etc. as parameters when creating the compilation.
  • Multimodal video compilation In the video editing system according to the invention, by means of content analysis methods, it is possible to find e.g. the speech of a specific person in the video i.e. video contents or to identify video segments which include music. By combining this information with the context data (mostly location and time) of the user and the video contents, it is possible to create new kinds of video compilations for various services. It is e.g. possible to create a video compilation in which are selected video segments on a specific route taken including the speech of a specific person.
  • the invention enables compiling videos according to novel compilation bases.
  • the compilation bases can be developed based on available content analysis methods and available context data.
  • the system calculates features of the video contents which are used together with the context data of the user and the videos in selecting the videos.
  • a video compilation is created of the selected video segments by combining them based on the context data of the videos.
  • An embodiment of the invention utilises a context-aware service which saves the route taken by the user, offers the user based on this an automatically created video compilation from the route.
  • videos created on the route by all users and, in the creation of the compilation, it is also possible to consider e.g. the creation time of the videos (season, clock time etc.) Then, e.g. the user is supplied in winter a compilation which includes videos created by users in winter.
  • the invention offers the user a possibility to re-experience a previous trip in the form of the video compilation. Earlier, re-experiencing the trips has relied on the user's own recordings. By means of the invention, it is possible to enrich the video compilation with material shot by other users in the same context.
  • the video editing system searches and selects in the multimedia database those videos which correspond the context route of the user (location and time data).
  • the system creates a video compilation based on the user's videos and videos selected from the multimedia database.
  • the video compilation is created by segmenting the videos into smaller sections (video segmentation), by giving points to the video segments according the predetermined criteria and by selecting the best segments based on the points i.e. having obtained most points into the compilation.
  • scenes are created of the video segments based on the context information of the videos. The scenes are combined into one video presentation.
  • a mobile service is implemented in which the user is offered a video compilation created based on the creation date and location of the videos.
  • the user can watch with his/her mobile phone in his/her current position the video contents previously created by the users/user groups as one video compilation in which the most interesting video segments have been selected.
  • the user is thus offered a compact view of videos, such as video clips, created in the location in question.
  • a location-aware service is created in which the user is able to view video contents produced by other users in his/her current position or in its vicinity as a video compilation in which the most interesting segments of the videos have been collected.
  • a location-aware video compilation is automatically created by comparing the creation location of the videos to the present position of the user. The videos thus selected are analysed, segmented and compiled into one video presentation.
  • the automatic video compilation implemented with an advantageous embodiment of the invention compresses the videos shot by the user in a presentation in which are selected the most important segments of the video contents which are most essential in sharing the experience. Additional information for the video is obtained by utilising context data in compiling the video. Thus, the video segments are in the compilation in the chronological order or, if desired, e.g. classified according to location. Studies have shown that the greatest motive for recording pictures and videos is to later share the experiences with others. This is emphasised when studying digital multimedia contents created with a mobile phone.
  • the automatic video compilation compresses the videos shot by the user or user group in a presentation in which are selected the most important segments of the video contents which are most essential in sharing the experience. Additional information for the video is obtained by utilising context data in compiling the video. Thus, the video segments are in the compilation in the chronological order or, if desired, e.g. classified according to location.
  • the aim is at describing the experience of the user or user group as a video compilation.
  • the contents are provided with context data on the video (location and time), the aim is to also include other information on the event/situation. For instance, it is possible to save the contents of a calendar entry of the mobile phone as a part of context data.
  • the service can save information on other users in the same location at the same time by means of user management and thus create a dynamic group for recording the experience.
  • This more extensive context data is utilised in creating the video compilation describing the experience. Particularly by utilising the user data, it is possible to select in the automatic video compilation contents created in the same situation in order to be able to create of the experience a presentation as versatile as possible and describing it well.
  • Systems designed for recording and sharing experiences often include various mobile devices which contain sensors measuring the status and behaviour of the person.
  • the experience is recorded as multimedia contents: pictures, videos and sound.
  • Various collection and sharing services of multimedia contents exist (e.g. Flickr, YouTube), but they lack the active contents combination implemented by the service for describing the experience.
  • context data (inter alia, location and time) related to video contents
  • context data (inter alia, location and time) related to video contents
  • users at a specific period are collected into one group at a specific time and a compilation of the video clips of these users is created to be shared with the group.
  • the principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation.
  • the segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction).
  • feature extraction The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values.
  • Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering).
  • the selection of video segments focuses on contents created by other persons and on combining them to one's own contents.
  • the emphasis of compilation is in the contents others have created in the same location at the same time as the acquirer of the video compilation.
  • the video compilation is provided with shots by other persons who shot the same subjects as the acquirer of the compilation.
  • the libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting.
  • Traditionally the problem has been solved by editing a presentation of the videos by cutting and pasting them together.
  • this requires special tools, time and expertise which an ordinary user does not have.
  • the combining of videos into logical units involves additional challenges for the editing process. Grouping the videos without the context data of the video is extremely challenging.
  • combining contents shot by others for one's own purposes is awkward as the material is difficult to obtain and edit.
  • the principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation.
  • the segmenting of videos is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction).
  • the selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values.
  • Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering).
  • the selection of video segments focuses on the contents created by persons known by the acquirer of the video compilation. Acquaintances can be managed, inter alia, with a social network service, such as Facebook.
  • the emphasis of compilation is in the contents the acquaintances have created in the same location at the same time as the acquirer of the video compilation.
  • the video compilation is provided with shots by acquaintances who shot the same subjects as the acquirer of the compilation.
  • the creation of the video compilation utilises acceleration and bearing sensors connected to the mobile station by means of which a gesture interface is implemented.
  • the user defines e.g. a period from the videos of which contents are selected into the video compilation.
  • the user has a possibility to affect the creation of the video compilation by means of the gesture interface e.g. by changing the rhythm and tempo of the video compilation by shaking or tilting the mobile phone or equivalent.
  • the main principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation.
  • the segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction).
  • feature extraction The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values.
  • Video segments are compiled into scenes (wcfeo scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering).
  • For the creation of the compilation is used a mobile phone provided with sensors and, for combining segments, the use of sensors is utilised.
  • the creation of video editing can be controlled by motion and bearing sensors.
  • the sensors can e.g. limit the parameters used in the creation of the compilation: a short and fast motion creates a fast compilation including a lot of motion or limits the contents of the compilation in terms of geography or time. Equivalently, a slow and extensive motion creates a more peaceful compilation or a compilation selected more widely in terms of geography and time. During the compilation, a tap on the phone forces the segment to change and tilting can affect the effect which is created in the change of the segment.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

The invention relates to a video editing system. According to the invention, the system comprises means for defining features of videos, means for automatically searching in a group of videos those videos or video sections i.e. segments which include the above defined features and means for creating a video compilation of videos and/or segments including the above defined features. In an advantageous embodiment of the invention, features to be defined include the geographical region where the video has been shot and the time and/or time period when the video has been shot and recorded.

Description

Video editing system
The invention relates to a video editing system according to the preamble of claim 1.
The invention also relates to a method for editing videos according to the preamble of claim 16.
Video refers to live picture material consisting of successive pictures quickly played back, the duration of which can vary, recorded by means of video cameras, particularly digital video cameras, mobile phones or equivalent devices. Video recorded with a mobile phone can be referred to as mobile video. The subjects of videos, particularly of personal videos, are, inter alia, family events, such as birthdays, and holiday trips, but videos are recorded as well of various everyday situations. The editing of videos refers to the compiling and processing of one or more videos such that its picture material e.g. is rearranged, sections are deleted of it, or the picture material of several videos, i.e. its sections or segments, is combined to a new final video to be stored and shown.
Managing personal video files and contents is more and more difficult as the volume of the contents increases at high speed. Furthermore, ordinary users have neither time nor often skills or interest in editing their videos afterwards which would improve the watching of the videos considerably later.
The libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting. Traditionally, the problem has been solved by editing a presentation of the videos by cutting and pasting them together. However, this requires special tools, time and expertise which an ordinary user does not have. Furthermore, in extensive video libraries, which include videos shot by several users, the combining of videos into logical units involves additional challenges for the editing process.
A problem of the development of multimedia services in mobile devices is often the small size of their display and the limited data transfer capacity of the mobile network. Services, which offer the user a possibility to search or browse video contents with a mobile phone, often show the result by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down. This is awkward and time-consuming.
A problem of the compiling i.e. editing of videos is often defining the compilation basis.
An object of the invention is to eliminate problems related to known video editing systems and equivalent methods. An object of the invention is also to provide a novel improved video editing system the principles of which are simple, efficient and cost-effective to implement.
A video editing system according to the invention is characterised by what is presented in claim 1. A method for editing videos according to the invention is characterised by what is presented in claim 16. The dependent claims present some advantageous embodiments of the invention.
The video editing system according to the invention comprises:
- means for defining features of videos; - means for automatically searching in a group of videos those videos or video sections i.e. segments which include the above defined features; and
- means for creating a video compilation of videos and/or segments including the above defined features.
In an advantageous embodiment of the invention, features to be defined of the videos include the geographical region where the video has been shot and the time and/or time period when the video has been shot and recorded.
In an advantageous embodiment of the invention, the video editing system comprises means for selecting suitable sections, segments, of the videos, which means include a video feature analysing unit which analyses sound, picture and/or motion features calculated of the video (i.e. content analysis metadata) based on which the segments are selected.
In an advantageous embodiment of the invention, the video editing system comprises means for creating a new video i.e. video compilation of the videos or their segments most advantageously automatically. In an advantageous embodiment of the invention, the selection of video segments is arranged to be implemented by calculating the above features and by giving each segment points based on this and by selecting those segments in the video compilation which comply with the predetermined point limits. In an advantageous embodiment of the invention, the video editing system comprises means for creating video scenes of the selected video segments by clustering the video segments on the basis of the time stamp and/or geographical data.
In an advantageous embodiment of the invention, the video editing system comprises means for creating a new video or video presentation i.e. video compilation of the videos or their segments based on the contents of the user and those recommended to the user i.e. the selected segments such that the location, time or content analysis metadata constitute continua between the video segments. In an advantageous embodiment of the invention, the video editing system comprises means according to independent claim 1 for searching and retrieving videos from a multimedia database or equivalent video database.
In an advantageous embodiment of the invention, the video editing system comprises means for searching, by means of specific search criteria, desired videos in a multimedia database, of which videos a video compilation is to be created.
In an advantageous embodiment of the invention, the video editing system comprises means for combining classified video segments into a new video i.e. video compilation. In an advantageous embodiment of the invention, the video editing system comprises a video database, such as a multivideo database, which is shared by several users and accessible through a data network, such as the Internet.
In an advantageous embodiment of the invention, the video editing system comprises means for updating the video database with new videos. In an advantageous embodiment of the invention, the video editing system comprises means for automatically performing video editing according to predetermined features. In an advantageous embodiment of the invention, the video editing system is arranged on a data network, advantageously on the Internet, where it is accessible for the users.
In an advantageous embodiment of the invention, the video editing system comprises a music database from which the user can select music to accompany the video compilation.
The method according to the invention comprises the following steps:
- defining features of videos;
- automatically searching in a group of videos those videos or video sections i.e. segments which include the above defined features; and
- creating a video compilation of videos and/or segments including the above defined features.
In advantageous embodiments of the method according to the invention, instead of the means of system claims 2-15 are defined equivalent action steps which are implemented with the means in question and, by performing which, the method steps are realised. This principle becomes evident, inter alia, by comparing independent claims 1 and 16.
Next, the invention, its various embodiments and advantages will be described in more detail. Creating video compilation based on semantic metadata
By means of various content analysis methods and collecting context data, it is possible to obtain low-level data on the features of video contents. It is e.g. possible to calculate momentums of the video or to save the GPS coordinates of the creation location of the video as metadata. Of this lower-level metadata, it is possible to develop higher-level semantic metadata. By means of GPS coordinates, for instance, it is possible to define the city or, by utilising the coordinates together with the shooting time, to obtain data on prevailing weather from external services. Then, it is possible to automatically create a compilation of separate video clips, in which the compilation basis is semantic metadata. For example, videos from a specific location shot in sunny weather are selected in the compilation. A problem of the compiling i.e. editing of the videos is often determining the compilation basis. Metadata produced by low-level content analysis methods does not necessarily offer a sensible compilation basis for a video. It is possible to upgrade the low-level metadata into higher-level metadata which is better suited for the compilation basis of videos.
The system searches lower-level metadata related to videos in the multimedia database and upgrades it into higher-level (semantic) metadata. The system creates a video compilation based on the higher-level metadata. The video compilation is created by segmenting the videos into smaller sections (video segmentation), by giving points to the video segments and by selecting the best segments based on the points into the compilation. Finally, scenes are created of the video segments based on the context data of the videos. The scenes are combined into one video presentation.
Automatic compilation of personal videos by means of context data
By means of the automatic video editing system, it is possible to create a service by means of which the user can define specific features for the video compilation. The user can supply the service with the desired video length and define the geographical region and time period for videos to be selected into the compilation. Thus, the user is automatically provided with a personal video compilation in which the system selects the most interesting segments of the video contents created by the user. The user can e.g. define that "I want a video compilation of videos shot at the summer house in July".
Managing personal video files and contents is more and more difficult as the volume of the contents increases at high speed. In addition to digital video cameras, videos are shot with mobile phones, which enables context data (mostly the creation time and location of the video as GPS data). However, ordinary users have neither time nor often skills or interest in editing their videos afterwards which would improve the watching of the videos considerably later. By means of the service created, the users are offered a possibility to create easily and quickly based on context data video compilations which contain essential segments of the original videos. The user can e.g. create a compilation of videos shot in the summer at the summer house by means of the service. The user is supplied a service by means of which it is possible to define the context (the creation region and time period of the video), based on which the video compilation of the videos shot by the user or user group is created. In the automatic video editing system is entered a set of videos selected based on context for analysis, segmentation and compilation. The system creates a video compilation defined by the user and supplies it to the user for presentation.
Semiautomatic video editor
By means of the automatic video compilation system, it is possible to provide a service which facilitates the editing of videos. The system segments and classifies videos automatically recommending interesting video contents for the user to be added into the compilation desired by the user. The user selects the desired segments and the system creates a compilation of them inserting suitable effects between the scenes. By means of the service, the editing of videos is easy also in mobile terminals.
The libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting. Traditionally, the problem has been solved by editing a presentation of the videos by cutting and pasting video clips together. However, this requires special tools, time and expertise which an ordinary user does not have. The invention automatises the hardest stages of video editing: cutting and pasting. In the semiautomatic video editor, the user only has to select the desired video segments and the system takes care of editing.
The system searches videos in the multimedia database and performs their segmentation and classification. The system gives points to the segments and recommends the user via the service the best segments to be included in the compilation. The user selects the desired segments and the system creates a video compilation of the selected segments. The system utilises the context data (location and time) of videos in creating the video compilation. The video compilation can be shown by means of the service in the user's terminal.
"On-line experience video channel"
Many social media services offer their users a possibility to collect their own contents and to recommend suitable multimedia contents for other users. By means of the automatic video compilation, it is possible to create each user his/her own "on-line experience video channel" of video clips the contents of which are automatically updated according to new recommendation or contents. Thus, the user has no need to separately watch each video clip, but he/she is able to receive a general view of the videos by following the contents of the "on-line" channel. The libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting. Traditionally, the problem has been solved by editing a presentation of the videos by cutting and pasting them together. However, this requires special tools, time and expertise which an ordinary user does not have. Furthermore, in extensive video libraries, which include videos recorded by several users, the combining of videos into logical units involves additional challenges for the editing process. Grouping the videos without the context data of the video is extremely challenging.
The main principle is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation. The segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction). The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values. Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering). The combining is performed based on the contents of the user and those recommended to the user such that the location, time or content analysis metadata constitute continua between the segments. At specific intervals, the programme can be changed, whereby it is possible to shift to video contents related to a totally different location and time. The channel compiles the contents into a continuous video stream which can be shown on an IPTV set top box, a digital picture frame or a computer display.
Using geographical data in clustering the video segments enables the combining of videos shot in the same context (location/time) into one unit (a scene in the video compilation). Without the geographical data, combining videos received from various sources based on context is difficult or impossible to automatise. Mobile video compilation
A problem of the development of multimedia services in mobile devices is often the small size of their display. When the user searches videos with a mobile phone, the result is often shown by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down. By means of the automatic video compilation, it is possible to create a compilation of the videos obtained as a search result for the user in which the most interesting segments of various videos are displayed. Thus, the user is able to have a general picture of the contents of the videos without scrolling up and down, and it is also possible to offer quicker operating services as the volume of transferable data decreases.
A problem of the development of multimedia services in mobile devices is often the small size of their display and the limited data transfer capacity of the mobile network. Services, which offer the user a possibility to search or browse video contents with a mobile phone, often show the result by means of keyframe pictures as a list or equivalent format which the user scrolls on the phone up and down. By means of the automatic video compilation, it is possible to create a compilation of the videos obtained as a search result for the user in which the most interesting segments of various videos are displayed. Thus, the user is able to have a general picture of the contents of the videos without scrolling up and down, and it is also possible to offer quicker operating services as the volume of transferable data possibly decreases (depends naturally on video coding vs. the size and number of keyframe pictures).
The user searches or browses the desired contents in the video achieves with his/her mobile phone. These tasks performed by the user generate a search for video contents in the Internet-based system. The result of the search is entered in the service supplying the automatic video compilation which analyses the videos, segments them and creates a compilation of the selected segments. For this, parameters suitable for this service for the creation of the video compilation have been defined for the segmentation and video rendering.
Mobile service for compiling personal videos
Nowadays, mobile phone users shoot great numbers of pictures and videos. The editing of the videos is often omitted, because it is difficult on the phone and transferring to a computer requires extra work. By means of the automatic video compilation system, it is still possible to create a web-based service to offer the user a possibility to create various compilations of his/her videos and to select suitable music to accompany them. The user is supplied e.g. five music pieces by means of a mobile interface and he/she can select some of his/her own videos of which the video compilation is automatically created by considering e.g. the music style, duration and tempo of the piece.
Nowadays, mobile phone users shoot great numbers of pictures and videos. The editing of the videos is often omitted, because it is difficult on the phone and transferring to a computer requires extra work. By means of the automatic video compilation system, it is still possible to create a web-based service to offer the user a possibility to create various compilations of his/her videos and to select suitable music to accompany them. The video compilation is automatically created by considering e.g. the music style, duration and tempo of the piece.
The user selects some of the videos created by him/herself for the compilation. The service offers the user some alternatives for background music one of which the user selects for the compilation. The automatic video compilation system analyses, segments and compiles videos considering the music style, piece duration etc. as parameters when creating the compilation.
Multimodal video compilation In the video editing system according to the invention, by means of content analysis methods, it is possible to find e.g. the speech of a specific person in the video i.e. video contents or to identify video segments which include music. By combining this information with the context data (mostly location and time) of the user and the video contents, it is possible to create new kinds of video compilations for various services. It is e.g. possible to create a video compilation in which are selected video segments on a specific route taken including the speech of a specific person.
Earlier known automatically created video compilations lack factors combining video segments. This weakens the entertainment value and consistency of the video compilation. The invention enables compiling videos according to novel compilation bases. The compilation bases can be developed based on available content analysis methods and available context data. By means of the content analysis methods, the system calculates features of the video contents which are used together with the context data of the user and the videos in selecting the videos. A video compilation is created of the selected video segments by combining them based on the context data of the videos.
Video compilation of user's context route
Various location-aware services have been studied and developed in recent years. The position of a mobile phone can be determined by means of the GPS technology or mobile phone networks very accurately. Yet, the position information has not been widely utilised in the management of multimedia contents. An embodiment of the invention utilises a context-aware service which saves the route taken by the user, offers the user based on this an automatically created video compilation from the route. For the video compilation are advantageously selected videos created on the route by all users and, in the creation of the compilation, it is also possible to consider e.g. the creation time of the videos (season, clock time etc.) Then, e.g. the user is supplied in winter a compilation which includes videos created by users in winter.
The invention offers the user a possibility to re-experience a previous trip in the form of the video compilation. Earlier, re-experiencing the trips has relied on the user's own recordings. By means of the invention, it is possible to enrich the video compilation with material shot by other users in the same context.
The video editing system according to the invention searches and selects in the multimedia database those videos which correspond the context route of the user (location and time data). The system creates a video compilation based on the user's videos and videos selected from the multimedia database. The video compilation is created by segmenting the videos into smaller sections (video segmentation), by giving points to the video segments according the predetermined criteria and by selecting the best segments based on the points i.e. having obtained most points into the compilation. Finally, scenes are created of the video segments based on the context information of the videos. The scenes are combined into one video presentation. Automatic video compilation for location-aware mobile service
In an embodiment of the invention, a mobile service is implemented in which the user is offered a video compilation created based on the creation date and location of the videos. The user can watch with his/her mobile phone in his/her current position the video contents previously created by the users/user groups as one video compilation in which the most interesting video segments have been selected. The user is thus offered a compact view of videos, such as video clips, created in the location in question. By selecting the contents from a specific type of a video database (e.g. travel contents, historic contents, sports), it is possible to create new services by means of the same concept.
The small size of the display in mobile devices creates challenges for the development of multimedia services. New arrangements are required to display the great volume of multimedia contents in the mobile phone. By means of the automatic video compilation based on the invention, a location-aware service is created in which the user is able to view video contents produced by other users in his/her current position or in its vicinity as a video compilation in which the most interesting segments of the videos have been collected. By means of the arrangement, the usability of location-aware mobile video services can be improved. A location-aware video compilation is automatically created by comparing the creation location of the videos to the present position of the user. The videos thus selected are analysed, segmented and compiled into one video presentation.
Sharing experiences by means of video compilation
Studies have shown that the greatest motive for recording pictures and videos is to later share experiences with others. Often, the sharing of experiences as videos suffers from the quality of the contents of the shot video when the videos are not edited. The automatic video compilation implemented with an advantageous embodiment of the invention (i.e. ultimately the video to be shown) compresses the videos shot by the user in a presentation in which are selected the most important segments of the video contents which are most essential in sharing the experience. Additional information for the video is obtained by utilising context data in compiling the video. Thus, the video segments are in the compilation in the chronological order or, if desired, e.g. classified according to location. Studies have shown that the greatest motive for recording pictures and videos is to later share the experiences with others. This is emphasised when studying digital multimedia contents created with a mobile phone. However, the sharing of experiences as videos often suffers from the quality of the video contents shot when the videos are not edited. The automatic video compilation compresses the videos shot by the user or user group in a presentation in which are selected the most important segments of the video contents which are most essential in sharing the experience. Additional information for the video is obtained by utilising context data in compiling the video. Thus, the video segments are in the compilation in the chronological order or, if desired, e.g. classified according to location.
By means of the invention, the aim is at describing the experience of the user or user group as a video compilation. When creating video contents, the contents are provided with context data on the video (location and time), the aim is to also include other information on the event/situation. For instance, it is possible to save the contents of a calendar entry of the mobile phone as a part of context data. Furthermore, the service can save information on other users in the same location at the same time by means of user management and thus create a dynamic group for recording the experience. This more extensive context data is utilised in creating the video compilation describing the experience. Particularly by utilising the user data, it is possible to select in the automatic video compilation contents created in the same situation in order to be able to create of the experience a presentation as versatile as possible and describing it well.
Systems designed for recording and sharing experiences often include various mobile devices which contain sensors measuring the status and behaviour of the person. On the other hand, the experience is recorded as multimedia contents: pictures, videos and sound. Various collection and sharing services of multimedia contents exist (e.g. Flickr, YouTube), but they lack the active contents combination implemented by the service for describing the experience.
Automatic video compilation for dynamic group
In the invention by means of context data (inter alia, location and time) related to video contents, it is possible to create dynamic user groups for creating a video compilation. Based on the context data, users at a specific period are collected into one group at a specific time and a compilation of the video clips of these users is created to be shared with the group.
The libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting. Traditionally, the problem has been solved by editing a presentation of the videos by cutting and pasting them together. However, this requires special tools, time and expertise which an ordinary user does not have. Furthermore, in extensive video libraries, which include videos recorded by several users, the combining of videos into logical units involves additional challenges for the editing process. Grouping the videos without the context data of the video is extremely challenging. Furthermore, combining contents shot by others for one's own purposes is awkward as the material is difficult to obtain and edit.
The principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation. The segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction). The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values. Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering). The selection of video segments focuses on contents created by other persons and on combining them to one's own contents. The emphasis of compilation is in the contents others have created in the same location at the same time as the acquirer of the video compilation. Thus, the video compilation is provided with shots by other persons who shot the same subjects as the acquirer of the compilation.
Collaborative video compilation for group
Various social media services are very popular. Different user groups create multimedia contents and share it with other users and user groups. By means of the automatic video compilation implemented with an advantageous embodiment of the invention, it is possible to offer a group video compilations of video clips created by the group and thus to create a more extensive video of the experiences of the group members to describe the collaborative experience of the group. When creating the compilation, data on the shooting location and date of the video clips is utilised in order to be able to determine the segments selected into the compilation and to arrange them in a suitable order.
The libraries of digital mobile videos have expanded and watching them is considered wearisome and unexciting. Traditionally, the problem has been solved by editing a presentation of the videos by cutting and pasting them together. However, this requires special tools, time and expertise which an ordinary user does not have. Furthermore, in extensive video libraries, which include videos recorded by several users, the combining of videos into logical units involves additional challenges for the editing process. Grouping the videos without the context data of the video is extremely challenging. Furthermore, combining contents shot by others for one's own purposes is awkward as the material is difficult to obtain and edit. The principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation. The segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction). The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values. Video segments are compiled into scenes (video scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering). The selection of video segments focuses on the contents created by persons known by the acquirer of the video compilation. Acquaintances can be managed, inter alia, with a social network service, such as Facebook. The emphasis of compilation is in the contents the acquaintances have created in the same location at the same time as the acquirer of the video compilation. Thus, the video compilation is provided with shots by acquaintances who shot the same subjects as the acquirer of the compilation.
Gesture interface for creating video compilation
In an embodiment of the invention, the creation of the video compilation, particularly in connection with mobile phones, utilises acceleration and bearing sensors connected to the mobile station by means of which a gesture interface is implemented. By means of this, the user defines e.g. a period from the videos of which contents are selected into the video compilation. Furthermore, the user has a possibility to affect the creation of the video compilation by means of the gesture interface e.g. by changing the rhythm and tempo of the video compilation by shaking or tilting the mobile phone or equivalent.
The main principle of the invention is that the system segments the videos into smaller sections, selects the best of them and compiles them into a new video presentation. The segmenting of videos (video segmentation) is implemented by analysing the sound, picture and motion features calculated of the video (feature extraction). The selection of the video segments is implemented by giving them points by means of supplied weighting coefficients and feature values. Video segments are compiled into scenes (wcfeo scene) by clustering the segments based on the video time stamp (timestamp) and geographical data (geotagging). Finally, the scenes are combined into one video presentation (video rendering). For the creation of the compilation is used a mobile phone provided with sensors and, for combining segments, the use of sensors is utilised. The creation of video editing can be controlled by motion and bearing sensors. The sensors can e.g. limit the parameters used in the creation of the compilation: a short and fast motion creates a fast compilation including a lot of motion or limits the contents of the compilation in terms of geography or time. Equivalently, a slow and extensive motion creates a more peaceful compilation or a compilation selected more widely in terms of geography and time. During the compilation, a tap on the phone forces the segment to change and tilting can affect the effect which is created in the change of the segment.
The invention is not limited to the above exemplifying embodiment, but many variations are possible within the scope of the inventive idea defined by the claims.

Claims

Claims
1. A video editing system, characterised in that the system comprises:
- means for defining features of videos;
- means for automatically searching in a group of videos those videos or video sections i.e. segments which include the above defined features; and
- means for creating a video compilation of videos and/or segments including the above defined features.
2. A video editing system according to claim 1 , characterised in that features to be defined include the geographical region where the video has been shot and the time and/or time period when the video has been shot and recorded.
3. A video editing system according to claim 1 or 2, characterised in that the system comprises means for selecting suitable sections, segments, of the videos, which means include a video feature analysing unit which analyses sound, picture and/or motion features calculated of the video (i.e. content analysis metadata) based on which the segments are selected.
4. A video editing system according to claim 1 , 2 or 3, characterised in that the system comprises means for creating a new video i.e. video compilation of the videos or their segments most advantageously automatically.
5. A video editing system according to claim 3, characterised in that the selection of video segments is arranged to be implemented by calculating the above features and by giving each segment points based on this and by selecting those segments in the video compilation which comply with the predetermined point limits.
6. A video editing system according to claim 3 or 5, characterised in that the system comprises means for creating video scenes of selected video segments by clustering the video segments on the basis of the time stamp and/or geographical data.
7. A video editing system according to claim 3, 5 or 6, characterised in that the system comprises means for creating a new video or video presentation i.e. video compilation of the videos or their segments based on the contents of the user and those recommended to the user i.e. the selected segments such that the location, time or content analysis metadata constitute continua between the video segments.
8. A video editing system according to claim 1 , 2 or 3, characterised in that the system comprises means according to claim 1 for searching videos in a multimedia database or equivalent video database.
9. A video editing system according to claim 8, characterised in that the system comprises means for searching, by means of specific search criteria, the desired videos in a multimedia database, of which videos a video compilation is to be created.
10. A video editing system according to claim 9, characterised in that the system comprises means for combining the classified video segments into a new video i.e. video compilation.
11. A video editing system according to claim 9, characterised in that the system comprises a video database, such as a multivideo database, which is shared by several users and accessible through a data network, such as the Internet. (Kl 1.3)
12. A video editing system according to claim 11 , characterised in that the system comprises means for updating the video database with new videos.
13. A video editing system according to claim 11 or 12, characterised in that the system comprises means for automatically performing video editing according to predetermined features.
14. A video editing system according to any one of preceding claims, characterised in that the video editing system is arranged on a data network, advantageously on the Internet, where it is accessible for the users.
15. A video editing system according to claim 14, characterised in that the system comprises a music database of which the user can select music to accompany the video compilation.
16. A method for editing videos, characterised by the method comprising the following steps:
- defining features of videos; - automatically searching in a group of videos those videos or video sections i.e. segments which include the above defined features; and
- creating a video compilation of videos and/or segments including the above defined features.
PCT/FI2010/050309 2009-04-16 2010-04-16 Video editing system WO2010119181A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
FI20095415A FI20095415A0 (en) 2009-04-16 2009-04-16 Video Editing System
FI20095415 2009-04-16

Publications (1)

Publication Number Publication Date
WO2010119181A1 true WO2010119181A1 (en) 2010-10-21

Family

ID=40590296

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/FI2010/050309 WO2010119181A1 (en) 2009-04-16 2010-04-16 Video editing system

Country Status (2)

Country Link
FI (1) FI20095415A0 (en)
WO (1) WO2010119181A1 (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2012175783A1 (en) * 2011-06-21 2012-12-27 Nokia Corporation Video remixing system
WO2014001607A1 (en) 2012-06-29 2014-01-03 Nokia Corporation Video remixing system
WO2014008885A3 (en) * 2012-07-12 2014-03-06 Hochschule Mittweida (Fh) Method and device for the automatic ordering of data records into a certain data volume containing data records
WO2014037604A1 (en) * 2012-09-07 2014-03-13 Nokia Corporation Multisource media remixing
EP2887352A1 (en) * 2013-12-19 2015-06-24 Nokia Corporation Video editing
WO2015122624A1 (en) * 2014-02-12 2015-08-20 엘지전자 주식회사 Mobile terminal and method for controlling same
US9710698B2 (en) 2012-06-25 2017-07-18 Nokia Technologies Oy Method, apparatus and computer program product for human-face features extraction

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1241673A2 (en) * 2001-03-16 2002-09-18 Gateway, Inc. Automated video editing system and method
US6697564B1 (en) * 2000-03-03 2004-02-24 Siemens Corporate Research, Inc. Method and system for video browsing and editing by employing audio
US6757027B1 (en) * 2000-02-11 2004-06-29 Sony Corporation Automatic video editing
US20050152666A1 (en) * 2004-01-09 2005-07-14 Demeyer Michael F. Apparatus and method for automated video editing
WO2006065223A1 (en) * 2004-12-13 2006-06-22 Muvee Technologies Pte Ltd A method of automatically editing media recordings
US20060251382A1 (en) * 2005-05-09 2006-11-09 Microsoft Corporation System and method for automatic video editing using object recognition
WO2007082169A2 (en) * 2006-01-05 2007-07-19 Eyespot Corporation Automatic aggregation of content for use in an online video editing system
US7362946B1 (en) * 1999-04-12 2008-04-22 Canon Kabushiki Kaisha Automated visual image editing system
US20080304806A1 (en) * 2007-06-07 2008-12-11 Cyberlink Corp. System and Method for Video Editing Based on Semantic Data

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7362946B1 (en) * 1999-04-12 2008-04-22 Canon Kabushiki Kaisha Automated visual image editing system
US6757027B1 (en) * 2000-02-11 2004-06-29 Sony Corporation Automatic video editing
US6697564B1 (en) * 2000-03-03 2004-02-24 Siemens Corporate Research, Inc. Method and system for video browsing and editing by employing audio
EP1241673A2 (en) * 2001-03-16 2002-09-18 Gateway, Inc. Automated video editing system and method
US20050152666A1 (en) * 2004-01-09 2005-07-14 Demeyer Michael F. Apparatus and method for automated video editing
WO2006065223A1 (en) * 2004-12-13 2006-06-22 Muvee Technologies Pte Ltd A method of automatically editing media recordings
US20060251382A1 (en) * 2005-05-09 2006-11-09 Microsoft Corporation System and method for automatic video editing using object recognition
WO2007082169A2 (en) * 2006-01-05 2007-07-19 Eyespot Corporation Automatic aggregation of content for use in an online video editing system
US20080304806A1 (en) * 2007-06-07 2008-12-11 Cyberlink Corp. System and Method for Video Editing Based on Semantic Data

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9396757B2 (en) 2011-06-21 2016-07-19 Nokia Technologies Oy Video remixing system
CN103635967A (en) * 2011-06-21 2014-03-12 诺基亚公司 Video remixing system
EP2724343A4 (en) * 2011-06-21 2016-05-11 Nokia Technologies Oy Video remixing system
CN103635967B (en) * 2011-06-21 2016-11-02 诺基亚技术有限公司 Video remixes system
WO2012175783A1 (en) * 2011-06-21 2012-12-27 Nokia Corporation Video remixing system
US9710698B2 (en) 2012-06-25 2017-07-18 Nokia Technologies Oy Method, apparatus and computer program product for human-face features extraction
EP2868112A4 (en) * 2012-06-29 2016-06-29 Nokia Technologies Oy Video remixing system
US9940970B2 (en) 2012-06-29 2018-04-10 Provenance Asset Group Llc Video remixing system
WO2014001607A1 (en) 2012-06-29 2014-01-03 Nokia Corporation Video remixing system
WO2014008885A3 (en) * 2012-07-12 2014-03-06 Hochschule Mittweida (Fh) Method and device for the automatic ordering of data records into a certain data volume containing data records
WO2014037604A1 (en) * 2012-09-07 2014-03-13 Nokia Corporation Multisource media remixing
US9607654B2 (en) 2013-12-19 2017-03-28 Nokia Technologies Oy Video editing
EP2887352A1 (en) * 2013-12-19 2015-06-24 Nokia Corporation Video editing
WO2015122624A1 (en) * 2014-02-12 2015-08-20 엘지전자 주식회사 Mobile terminal and method for controlling same

Also Published As

Publication number Publication date
FI20095415A0 (en) 2009-04-16

Similar Documents

Publication Publication Date Title
US11328013B2 (en) Generating theme-based videos
CN102640149B (en) Melody commending system, signal conditioning package and information processing method
US9940970B2 (en) Video remixing system
US10192583B2 (en) Video editing using contextual data and content discovery using clusters
US10546010B2 (en) Method and system for storytelling on a computing device
US8566880B2 (en) Device and method for providing a television sequence using database and user inputs
WO2010119181A1 (en) Video editing system
US20140108932A1 (en) Online search, storage, manipulation, and delivery of video content
US20100042926A1 (en) Theme-based slideshows
TW201545120A (en) Automatic generation of compilation videos
WO2003088665A1 (en) Meta data edition device, meta data reproduction device, meta data distribution device, meta data search device, meta data reproduction condition setting device, and meta data distribution method
EP3322192A1 (en) Method for intuitive video content reproduction through data structuring and user interface device therefor
EP1839434A1 (en) A method of automatically editing media recordings
CN101300567A (en) Media sharing and authoring on the web
JP2000250944A (en) Information providing method and device, information receiving device and information describing method
US11775580B2 (en) Playlist preview
CN108780654A (en) Generate the mobile thumbnail for video
US20180130499A9 (en) Method for intuitively reproducing video contents through data structuring and the apparatus thereof
US20130262458A1 (en) Information processing device and program
US20220147558A1 (en) Methods and systems for automatically matching audio content with visual input
JPH08255171A (en) Information processor
EP2315167A1 (en) Artistic social trailer based on semantic analysis
Ojutkangas et al. Location based abstraction of user generated mobile videos
Sawada Recast: an interactive platform for personal media curation and distribution
AU2021250903A1 (en) Methods and systems for automatically matching audio content with visual input

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 10764151

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 10764151

Country of ref document: EP

Kind code of ref document: A1