IES20030840A2 - Multimedia management - Google Patents
Multimedia managementInfo
- Publication number
- IES20030840A2 IES20030840A2 IE20030840A IES20030840A IES20030840A2 IE S20030840 A2 IES20030840 A2 IE S20030840A2 IE 20030840 A IE20030840 A IE 20030840A IE S20030840 A IES20030840 A IE S20030840A IE S20030840 A2 IES20030840 A2 IE S20030840A2
- Authority
- IE
- Ireland
- Prior art keywords
- content
- media
- meta data
- analysis component
- component
- Prior art date
Links
- 238000004458 analytical method Methods 0.000 claims abstract description 40
- 238000003860 storage Methods 0.000 claims abstract description 18
- 238000001514 detection method Methods 0.000 claims abstract description 8
- 238000007726 management method Methods 0.000 claims description 11
- 230000009466 transformation Effects 0.000 claims description 5
- 239000000284 extract Substances 0.000 claims description 4
- 238000009826 distribution Methods 0.000 abstract description 6
- 230000003068 static effect Effects 0.000 abstract description 6
- 238000000034 method Methods 0.000 description 14
- 238000012545 processing Methods 0.000 description 7
- 230000003993 interaction Effects 0.000 description 6
- 230000033001 locomotion Effects 0.000 description 5
- 230000008569 process Effects 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 3
- 230000000750 progressive effect Effects 0.000 description 3
- 230000001360 synchronised effect Effects 0.000 description 3
- 230000002123 temporal effect Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 230000000007 visual effect Effects 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000003139 buffering effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 230000010485 coping Effects 0.000 description 1
- 238000012217 deletion Methods 0.000 description 1
- 230000037430 deletion Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000008846 dynamic interplay Effects 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000008520 organization Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 230000001960 triggered effect Effects 0.000 description 1
- 238000012800 visualization Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/74—Browsing; Visualisation therefor
- G06F16/745—Browsing; Visualisation therefor the internal structure of a single video sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7834—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using audio features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/785—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using colour or luminescence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/7854—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using shape
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/786—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using motion, e.g. object motion or camera motion
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/32—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
- G11B27/327—Table of contents
- G11B27/329—Table of contents on a disc [VTOC]
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/56—Provisioning of proxy services
- H04L67/561—Adding application-functional data or data for application control, e.g. adding metadata
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/50—Network services
- H04L67/56—Provisioning of proxy services
- H04L67/565—Conversion or adaptation of application format or content
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Information Transfer Between Computers (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
A multimedia management system (1) receives live stream content at an interface (5) or static content (6) and captures (7) it to a storage device (10). An analysis component (11) in parallel or in series generates meta data for a database (12), the meta data acting as an index for the storage device (10). Automatic detection of events by the analysis component (11) allows provision of real time distribution of content such as news clips to fixed and mobile networks (MS). Near real time media services are also provided with some annotations being performed by an annotation component (14) under control of a browser (13). <Figure 2>
Description
The invention relates to management of multimedia such as digitised audio-visual content.
Prior Art Discussion
In recent years advances have been made in the field of distribution of multimedia in mobile networks. For example the Multimedia Messaging Service (MMS) standard for mobile networks specifies peer-to-peer multimedia messaging. United States Patent Application No. US200205/442 describes a method of transmitting/receiving a broadcast message in a mobile network.
Also, advances have been made in processing of multimedia in various aspects, such as indexing and browsing. For example, the paper “Fischlar: an on-line system for indexing and browsing broadcast television”, O'Conndr N, Marlow S, Murphy N, Smeaton A, Browne P, Deasy S, Lee H and Me Donald K. ICASSP 2001 International Conference on Acoustics, Speech, and Signal Processing. Salt Lake City, UT, 7-11 May 2001., describes indexing and browsing. Television programmes are captured in MPEG-1 format and analysed using video indexing tools. Browsing interfaces allow a user to browse a visual index to locate content of interest, and selected content can then be streamed in real time.
While these developments are of benefit, there is still a need to provide more comprehensive multimedia asset management to support content-based operations for a broader range of devices and access mechanisms.
Open to pubuc inspection
UNDER SECTION 28 ANL RULE 23
IE 0 3 0 R 4 Q
-2SUMMARY OF THE INVENTION
According to the invention, there is provided a multimedia management system comprising:
a multimedia content capture component for receiving content and for writing it to a storage device;
an analysis component for analysising received content to generate meta data, and a database for storing the meta data;
a server component for distributing content or meta data to a network for delivery to subscriber devices; and a controller for coordinating operation of the components of the system to provide configured services for delivery of content to subscriber devices.
In one embodiment, the analysis component generates the meta data as an index to the content in the storage device and the server uses the meta data to access the content.
In another embodiment, the analysis component extracts key frames from received video content and segments audio and video streams.
In a further embodiment, said key frames are stored to provide a storyboard of video content with images forming part of an index.
-3£0308 40
In one embodiment, the analysis component operates in parallel with capture of content by the capture component for analysis in real time as a video content stream is being received.
In another embodiment, the analysis component uses a shot boundary detection technique to generate the meta data.
In a further embodiment, the analysis component automatically detects events in incoming content and generates notifications in real time.
In one embodiment, the controller receives the notifications and automatically controls components of the system to generate a content output triggered by the detected event.
In another embodiment, the notification is an alert message transmitted to a mobile station.
In a further embodiment, the system further comprises an annotation component for annotating content and directing storage of annotated content in the storage device.
In one embodiment, the annotation component operates in near real time.
In another embodiment, the controller directs output via the server of content in near real time to provide a subscriber service.
In a further embodiment, the system further comprises a browser connected to the storage device and to the database, the browser allowing supervisor access to content to control processing of the content.
|E 0 S 0 fi 4 0
-4In one embodiment, the browser is connected to the annotation component and directs operation of the annotation component to provide processed content in the near real-time service.
In another embodiment, the system further comprises a transcoding component connected to the browser and being for transcoding of content for delivery to the analysis component and to the storage device.
In a further embodiment, the controller directs routing of meta data from the database to a plurality of different devices, including local devices directly connected to the controller and remote subscriber devices.
In one embodiment, the analysis component writes some of the meta data to the storage devices to provide an addressing index.
In another embodiment, the controller uses meta data solely from the database for directing content to some subscriber devices.
In a further embodiment, the controller dynamically generates media interfaces with style sheet transformation, in which style sheets dynamically transform the media and code into user-viewable media and display code.
In one embodiment, the transformation performs independent processing of screen windows in a manner analogous to application or operating system display windows.
In another embodiment, the style sheet transformation code is separate from underlying functionality of the controller and other components involved in media output.
-5In a further embodiment, the system further comprises a live streaming interface connected to the content capture component.
In one embodiment, the system further comprises a scheduling component connected to the live stream circuit for activating the live stream circuit and setting recording parameters for both the live stream circuit and the capture component.
DETAILED DESCRIPTION OF THE INVENTION
Brief Description of the Drawings
The invention will be more clearly understood from the following description of some embodiments thereof, given by way of example only with reference to the accompanying drawings in which :Fig. 1 is a block diagram of a multimedia management system of the invention.
Fig. 2 is a diagram showing interfacing modes; and
Fig. 3 is a diagram illustrating media analysis, generation of meta-data, and presentation for content-based operations such as access, retrieval, sharing consumption.
Description of the Embodiments
Referring to Fig. 1 a multimedia management system 1 of the invention comprises a controller 2 which controls in a comprehensive manner various multimedia processing components to provide a variety of services. These components are as follows.
IE 0 3 0 Λ 4 0
-65: A live stream interface for receiving broadcast digital TV channels.
6: A reader device for reading multimedia content from a static, non-live source such as DVDs.
7: A capture circuit for encoding and compressing content to a format such as MPEG.
: A content storage device for storing the encoded and compressed content files together with related meta data. The latter includes images, hyperlinks, log files for data such as start/end time of capture, programme meta-data, user account information, user preferences, copyright information, number of key frames extracted, file size, and channel meta data. Some of the meta data is generated by the capture circuit 7, and the remainder by an analysis component 11.
11: An analysis component 11 which extracts representative meta-data such as key frames, audio, text, from compressed audio-visual content. Extracted meta-data is encoded in formats and resolutions (for example JPEG, GIF, MPEG) of any size that can be displayed, selected, and interacted with. The analysis component 11 both extracts existing meta data, and generates meta data which is stored in the storage device 10. However, it also writes meta data to a database 12, set out below. Additional meta data related to the precision and accuracy of analysis techniques used against audio-visual content is generated by the analysis component 11 using the shot boundary technique to generate colour histogram, edge detection, and motion analysis data.
12: A content database for storing all meta data generated by the analysis component 11. Some of the meta data is also stored with the actual content in
ΙΕ ο 3 0 8 4 ο
-Ί the storage device 10, as outlined above. The database 12 stores the meta data in a manner to provide an index for the storage device 10.
13: A browser for use by local/remote staff of the organization hosting the system 1. This is used for accessing the content using indexes of the database 12. It activates interfaces for content-based operations (non-linear playback, annotation) accomplished over end-user devices. The browser 13 can also route content to a transcoding component 15.
14: An annotation component for annotation of stored content under instruction from the browser 13.
16: A scheduling component for scheduling operation of the interface 5 for receiving and encoding live streams.
17: A Web server for routing of content to infrastructure for fixed and mobile networks for onward routing to central/edge servers and end user playback devices.
The multimedia management processes performed by the system are now described. One of the problems facing operators of communication systems such as mobile network operators is that while there is a vast amount of content available, it has not been possible to manage it so that content owners and aggregators, network operators and end users get optimum use of it. It is not just a matter of coping with the limited communication bandwidths and device capabilities of many user devices, but also syndicating content which end users wish to access.
The system 1 addresses this problem by acting as a platform which takes in the multimedia content and manages it in an optimum manner so that end users can access/receive what they subscribed to. For example, security camera systems
-8IE030840 generate a vast quantity of content, however, an end user is only interested in seeing suspicious activity. Thus, the analysis component 11 may automatically monitor live streams from such cameras and generate an alert for the controller 2 if a suspicious image and/or sound pattern is detected. The controller in turn notifies in real time the Web server 17 to retrieve the content meta data and push it onto either the fixed or mobile network to the end user subscriber playback devices. One suitable bearer for such alerts is GSM SMS to mobile devices.
In other situations some human input may be required such as downloading of a video sequence of a breaking news event or a score in a football match. In this case, staff monitor certain content channels and access audio-visual segments in non-linear fashion to achieve content-based operations for seamless syndication, transmission and retrieval over fixed and mobile networks. The content is cropped and rendered to be suitable for transmission to the subscriber’s mobile station MS. Such operations can typically be carried out in a few seconds, thus providing a near realtime service to the subscriber.
Fig. 2 illustrates the versatility of the system 1, providing content to a wide variety of devices to present a variety of interfaces. These include a portal interface 20, a search interface 21, a browsing interface 22, a playback interface 23, an annotation interface 24, remote user device browsing/swapping interfaces 25, and a remote user device playback interface 26. This diagram illustrates versatility of the architecture of the system 1. The controller 20 is at the centre allowing interfacing with both staff of the hosting organisation and remote end users in a versatile manner.
The controller 2 includes a database of access control data for interfacing with staff. It allows supervisor access to some of the components.
The controller 2 controls automatic deletion of content according to configured parameters.
-9IE 0 3 0 Β 4 0
The analysis component 11 generates the meta data is a manner whereby it forms a content index that is represented using images of video content and meta data referenced to additional related media content. Capture and analysis of content can take place in parallel, thus enabling text, image and video distribution without the need to wait for a programme to end before managing distribution to a mass audience.
The analysis component 2 generates the meta data for some content in the form of storyboards. Live and stored media files are analysed using algorithms processing specific data streams/layers, multiplexed/encapsulated in analogue and standard media compression formats (for example MPEG formats). Each algorithm can be controlled by a dynamic threshold/parameter set to analyse different types of media information (for example video with slow-movement and short cuts compared to video with fast motion and fades). The analysis techniques are used to extract representational content from media files for indexing based on natural breaks or events in the media content. Analysis results derived from files composed of different type of media information (for example news programme with newscaster and news agencies footage) are based on a combination of algorithms in order to generate the most logical representation of programme’s content for specific playback devices.
The analysis component 11 analyses whole data streams/formats and media layers by segmenting audio, detecting faces, detecting edges, detecting black frames, and generating colour histograms, The combination of these algorithms detects video frames and video shots that are sequenced in a logical scene if media programmes follow a structure at the time of production.
Referring to Fig. 3, a media file 40 is analysed in step 45 to generate content-based meta-data 46, which is then automatically stored in the storage device 10. The metadata 46 is retrieved using a style sheet 41 that can be marked-up in representation
-10ΙΕΟ 3 Ο * 4 Ο languages (for example HTML, CHTML, WML, SMIL and XML languages) useful for static/dynamic representation of media information to media devices (for example PC, PDA, and mobile phone). Dynamic representation of media content provides the basis for progressive download of media information to media devices dependent on progressive and fast retrieval of media information over narrowband networks. Thus, media data is presented or exchanged between users by communicating representational information, meta-data information, and location information rather that communicating the media data itself.
The controller 2 acts as an interface between the storage device 10 and the web server 17, the Web browser 13, and a streaming server reading commands from hyperlinks and animated media formats. Such commands are sent over the fixed and mobile network to playback devices. The Web browser 13 is automatically refreshed and mapped against media files, metafiles and meta-data found in the database 12.
Requests may be made from subscriber devices to achieve content-based operations. Each content-based operation is completed by connection (circuit switched and/or packet based) to a local media server and/or remote server depending on the availability of bandwidth that the mobile device can connect to and the location of subscriber requesting the content. Media programmes and/or media clips are subsequently and automatically sent over packet-based networks to subscriber devices in synchronised, downloadable or streamable formats that can be played using messaging software, browsers or media decoders developed specifically for the media device.
Access to interfaces and media information in the system 1 is monitored at different levels. Content-based operations requested by a subscriber can only be fulfilled after inputting a user ID. Monitored access generates information on usage of the system 1 and the media data content. Such information can be used to improve media distribution on a personalised basis.
-11ΙΕΟ 3 Oft 4 θ
Recording functions integrated in the scheduling component 16 provides system users and end users with recording options. Recording requests could be made to the system by clicking on hyperlinked media information such as programme title, and programme meta-data. End users can also rate media programmes in order to receive programme recommendations from the system.
Search interfaces with media highlights in an overview/detail view interface are also generated based upon the meta data. Media sequences are highlighted based on a search query and they can be browsed using a timeline bar in Detail View Media interface.
The system 1 also generates a search interface with a detail view based on text results and slide show media interface enabling system users to narrow down search results using text information (for example closed caption, annotation) tagged to indexed audio-visual content. A search interface for audio-visual information enables system users and end users to search audio-visual information indexed using different analysis techniques. Searches can be applied against different parameters populated by analysis techniques such a face detection, closed caption parsing, and shot detection. Search results are ranked and are based on representative meta-data. An overview media interface implements “overview first, zoom and filter, then details on demand”. This interface provides an overview of media files using representative meta-data, thus indicating to the end user what the audio-visual information is about.
The system 1 also generates a scroll bar media interface. System users browse media content using a scroll bar that acts as a very versatile controller. Clicking on the arrows of the scroll bar moves the page by each row of representative keyframes, clicking on the empty bar jumps the page by screenful, and dragging the square object provides immediate and continuous movement of representative keyframes thus allowing finer and more pleasant experience.
ΙΕ ο 3 0 β 4 0
-12Another interface is an extension of the overview media interface. This is similar to the overview/detail browser, but allows movement into a detail view. Once in a detail view, a timeline bar is used for immediate system reaction and finer control in browsing. The timeline bar can be static or dynamic. Both types of timeline bars are useful when browsing a large amount of media information because it provides temporal orientation by showing the current point of browsing in relation to the whole media file. Not only useful as a time indicator, the bar in this browser allows random access by selecting related sections of the bar. The interface with the dynamic quick bar capabilities saves tedious and continuous clickings that the static timeline bar browser may require. One sweep of the mouse allows viewing of all of the keyframes in the video. Much finer control and continuous, immediate response from the system provides 'direct manipulation'. The browsing task is less of an 'actand-react' interaction, and is more natural and paced by the user.
A slide show media interface flips through all representative indexed meta-data in a single slot. Temporally presenting representative meta-data (rather than spatially) saves screen space considerably, and is thus useful for small end user devices. People with more dynamic interaction preference may also like this browser. It is important to provide some form of temporal orientation in this kind of temporal presentation, and in this case timeline bar is used.
A speech annotation interface is useful for typists who, in an off-line mode, have to annotate and synchronise text against audio-visual content. The speech annotation interface is based on the slide show media interface and helps annotate text at specific level of granularity into media files. Annotation against video can be searched and the system can provide audio-visual search results.
IE Ο 3 08 4 Ο
-13Another annotation interface supports portal updates on remote media servers located close to end-users. The audio-visual annotation interface also enables system users to insert rights and user access information.
A media portal interface for devices with small displays provides an overview of media content available to browse and retrieve on an end user device with small display. Based on standard mark-up languages, this Interface uses media segments as retrieval units, tracking users’ access history, thus being able to recommend the most important news stories to the user, in the form of a summarised list. This way, all the user has to do is simply select a story from the list - saving awkward text input and other interaction. The idea is putting most of the processing on the system itself, rather than providing elaborate searching/browsing work to the user. Because this interface tracks user interactions, it can generate usability information useful for billing and rating mechanisms.
The server 17 also provides detailed video interfaces for devices with small display. These two interfaces make video browsing similar to video game interactions on small devices. From these interfaces, the user can control the current point of browsing, using the controller/arrows. This allows the user to use only the one hand holding the small device. Four arrow buttons are just about all that the user uses to browse representative media (other than another arrow at the bottom left, to bring back to related media information such as listing, related programmes, etc.).
Playback interfaces represent static and synchronised multimedia messaging that plays images, audio and video information in email clients, browsers, and MMS clients on MMS handsets. Receivers and viewers of MMS can interact with the media information to access related information such as related stories in video formats.
IE 0 3 Ο » 4 0
-14The system 1 provides searching capabilities across programme meta-data, meta-data parameters filled by analysis techniques, and captured and referenced closed caption data. Search results are complemented and hyperlinked to visual representation of media programmes to allow users to access media information in a random fashion thus reducing the time to play back specific media segments. Searches can be made from any web browser capable device.
To complement text-based search of media programmes, video skimming and visualisation is achieved by the system 1 to retrieve media information in a nonlinear fashion. Information rendered by browsers can be progressively downloaded on a media device for a user to interact with locally or it can be requested on-demand by selecting media information stored remotely. Based on HTML and XML mark-up languages, media browsers are designed to optimise content-based operations for media devices having unique specifications (display, resolution, media format, and/or mark-up language rendering).
The controller 2 controls playback formats, resolutions, quality and method of transmission and monitoring of the playback of a media file. Playback of a media file is supported by streaming server and streaming software, video clients, and browsers that can render proprietary and standard media formats. The management module can control processes leading to playback of media content. Processes such as buffering, progressive download, image and ad removal/insertion can be set to suit user preferences and media device specifications.
The system 1 also enables content-based interactions from file formats and streaming clients provided that the management module of the present invention can interface with the API of a streaming server.
The controller 2 generates interfaces from which a system user can automatically access, annotate, format, share and publish media segments to any file or directory.
£ 0 3 Oft 4 Ο
-15This method allows distribution of media clips once in multiple formats, over many networks, to multiple devices. Text annotation to media clips are created to be only accessed by authorised users. Annotation of media clips can be accomplished at different levels of granularity of media content - image, advertisement, logical media segment (shots, scenes, story) - to allow for more efficient retrieval at a later stage. Users operating this interface set terms-of-usage for each media file that is made available for content-based operations on local/remote servers. Terms-of-usage can be embedded in transmission delivery formats of each media file in order to track and confirm usage of each media file retrieved.
The outputs of media editing can be standard multimedia messaging services based on text, images, and synchronised multimedia presentations. Outputs can also be text, audio and video clips that can be rendered by email clients, MMSCs, browsers, streaming players, and streaming clients.
The above interfaces are dynamically generated by XSL transformation. A set of XSL stylesheets are used to take necessary information from the underlying XML data which is generated whenever a web request is done by a client. The generated XML data comes from various sources to collect necessary information and as a result contains all information to display for the user’s device in a request of the user's action. XSL stylesheets then transform the data into user-viewable code (HTML, JavaScript, Java Applet and SVG) to render on the web browser and other devices. For the web interfaces with sophisticated features, layers of frames have been used, each frame corresponding to each XSL stylesheet, transforming underlying XML data for that particular frame into presentable format such as HTML. Framed interface allows application-like interaction in which a clicked widget stays on with the effect visible on other parts of the screen, instead of the whole screen refreshing. For simpler interfaces such as for mobile devices, a single XSL stylesheet defines the layout and elements of the screen, thus corresponding to a single underlying XML data generated when requested by the user.
-16<030«40
In this way, the XML-based architecture of the system allows easy device-depended interface implementation using XSL. This also separates the underlying functionality of the system from the interface design, ensuring more user-oriented interface consideration by the designer.
The controller 2 can “push” and/or “pull” media files that have been analysed and indexed onto servers that are closer and more accessible to authorised users of media content.
It will be appreciated that the system 1 acts as the hub around which specific applications can be supported and services provided with a short lead time. The system 1 captures, manages and distributes media information requested by authorised users across retrieval devices. The system allows easy development of services.
The following is a non-exhaustive list of services which may be provided:
Distributed analysis of media information Protected editing of media information Face Detection in live and encoded video programmes Caching of referenced media information Live broadcast & indexing of media content
Real-time, near real-time, on-demand representation of media information
Near real-time alerts of media events
Image matching in live and encoded video programmes
Face matching in structured video programmes
Video object detection in live and encoded video programmes
Video object matching in live and encoded video programmes
Motion detection in live and encoded video programmes
Ig ο 3 08 4 ο
-17File merging
Advertisement insertion based on user lists Searching across closed and open caption Searching across image and video features Personalised media caching Media management for instant messaging Multicasting
Video skimming
Broadcasters transmit a continuous flow of pre-recorded and live television programmes through licensed networks (cable/satellite/terrestrial) and media device (TV/STB). Broadcast models are currently based on physical locations where a TV signal can reach viewers without necessarily identifying them and interesting them. The system 1 provides an effective mechanism to transmit beyond this traditional model and reach audiences that are mobile. This is achieved by making timesensitive broadcast events available shortly after they have occurred and in a format suitable for retrieval across personal media devices.
Archive units of content owners, producers and aggregators like parliaments and broadcasting organisations are in charge of labour intensive operations. Restoration, digitisation, cataloguing, storing, copyright clearance, copying and shipping represent of the main processes that the archive units have to manage before media assets can be available on-demand to internal and external viewers. In order to enable access to archived content, this system enables the archival process to be accomplished over a web browser capable device. It also allows retrieval of copyrighted media information based on catalogued information.
The invention is not limited to the embodiments described but may be varied in construction and detail.
Claims (5)
1. A multimedia management system comprising: a multimedia content capture (7) component for receiving content and for writing it to a storage device (10); an analysis component (11) for analysising received content to generate meta data, and a database (12) for storing the meta data; a server component (17) for distributing content or meta data to a network (MN) for delivery to subscriber devices; and a controller (2) for coordinating operation of the components of the system to provide configured services for delivery of content to subscriber devices (MS).
2. A system as claimed in claim 1, wherein the analysis component (11) generates the meta data as an index to the content in the storage device (10) and the server (17) uses the meta data to access the content; and wherein the analysis component (11) extracts key frames from received video content and segments audio and video streams; and wherein said key frames are stored to provide a storyboard of video content with images forming part of an index.
3. A system as claimed in any preceding claim, wherein the analysis component (11) operates in parallel with capture of content by the capture component (7) for analysis in real time as a video content stream is being received; and wherein the analysis component (11) uses a shot boundary detection technique to generate the meta data; and wherein the analysis component (11) automatically detects events in incoming content and generates notifications in real time. !Ε0 3 0«
4. β -194. A system as claimed in any preceding claim, wherein the controller directs routing of meta data from the database to a plurality of different devices, including local devices directly connected to the controller and remote subscriber devices; and wherein the analysis component (11) writes some of 5. The meta data to the storage devices to provide an addressing index; and wherein the controller uses meta data solely from the database for directing content to some subscriber devices; and wherein the controller dynamically generates media interfaces with style sheet transformation, in which style sheets dynamically transform the media and code into user-viewable media 6. 10 and display code.
5. A multimedia management system substantially as described with reference to the drawings.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IE20030840A IES20030840A2 (en) | 2002-11-08 | 2003-11-10 | Multimedia management |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
IE20020870 | 2002-11-08 | ||
IE20030840A IES20030840A2 (en) | 2002-11-08 | 2003-11-10 | Multimedia management |
Publications (1)
Publication Number | Publication Date |
---|---|
IES20030840A2 true IES20030840A2 (en) | 2004-05-05 |
Family
ID=32310088
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
IE20030840A IES20030840A2 (en) | 2002-11-08 | 2003-11-10 | Multimedia management |
Country Status (3)
Country | Link |
---|---|
AU (1) | AU2003278558A1 (en) |
IE (1) | IES20030840A2 (en) |
WO (1) | WO2004043029A2 (en) |
Families Citing this family (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7055166B1 (en) | 1996-10-03 | 2006-05-30 | Gotuit Media Corp. | Apparatus and methods for broadcast monitoring |
EP1618743A1 (en) * | 2003-04-17 | 2006-01-25 | Koninklijke Philips Electronics N.V. | Content analysis of coded video data |
US8316128B2 (en) | 2004-01-26 | 2012-11-20 | Forte Internet Software, Inc. | Methods and system for creating and managing identity oriented networked communication |
US8250150B2 (en) | 2004-01-26 | 2012-08-21 | Forte Internet Software, Inc. | Methods and apparatus for identifying and facilitating a social interaction structure over a data packet network |
FR2878116B1 (en) * | 2004-11-15 | 2007-05-18 | France Telecom | METHOD AND SYSTEM FOR INDEXING A MULTIMEDIA DATA STREAM AND CORRESPONDING PROCESSING METHOD AND SYSTEM |
EP1839213A1 (en) * | 2005-01-19 | 2007-10-03 | France Telecom | Method for generating a text-based index from a voice annotation |
US20060288036A1 (en) * | 2005-06-17 | 2006-12-21 | Microsoft Corporation | Device specific content indexing for optimized device operation |
WO2007130472A2 (en) * | 2006-05-01 | 2007-11-15 | Gotuit Media Corporation | Methods and systems for providing media assets over a network |
EP2151128A4 (en) | 2007-04-25 | 2011-11-16 | Miovision Technologies Inc | Method and system for analyzing multimedia content |
DE102007030131B3 (en) * | 2007-06-29 | 2008-10-23 | Whatmore Gmbh | Information i.e. video data, synchronized provision method, for e.g. interactive application of portal, involves forming community using identifier with information related to geographical location of users |
TWI425416B (en) * | 2008-06-27 | 2014-02-01 | Chi Mei Comm Systems Inc | System and method for controlling to play audio requests |
WO2010090622A1 (en) * | 2009-02-09 | 2010-08-12 | Vitamin D, Inc. | Systems and methods for video analysis |
US9330097B2 (en) * | 2009-02-17 | 2016-05-03 | Hewlett-Packard Development Company, L.P. | Projects containing media data of different types |
KR101700365B1 (en) * | 2010-09-17 | 2017-02-14 | 삼성전자주식회사 | Method for providing media-content relation information, device, server, and storage medium thereof |
EP2851900B1 (en) | 2013-09-18 | 2017-08-23 | Nxp B.V. | Media content real time analysis and semi-automated summarization for capturing a fleeting event. |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6166735A (en) * | 1997-12-03 | 2000-12-26 | International Business Machines Corporation | Video story board user interface for selective downloading and displaying of desired portions of remote-stored video data objects |
US6917965B2 (en) * | 1998-09-15 | 2005-07-12 | Microsoft Corporation | Facilitating annotation creation and notification via electronic mail |
-
2003
- 2003-11-10 IE IE20030840A patent/IES20030840A2/en not_active IP Right Cessation
- 2003-11-10 WO PCT/IE2003/000152 patent/WO2004043029A2/en not_active Application Discontinuation
- 2003-11-10 AU AU2003278558A patent/AU2003278558A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
WO2004043029A2 (en) | 2004-05-21 |
WO2004043029A3 (en) | 2004-08-05 |
AU2003278558A8 (en) | 2004-06-07 |
AU2003278558A1 (en) | 2004-06-07 |
IE20030839A1 (en) | 2005-05-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11626141B2 (en) | Method, system and computer program product for distributed video editing | |
US8589973B2 (en) | Peer to peer media distribution system and method | |
Gao et al. | Vlogging: A survey of videoblogging technology on the web | |
US8060906B2 (en) | Method and apparatus for interactively retrieving content related to previous query results | |
Chang et al. | Video adaptation: concepts, technologies, and open issues | |
US8453189B2 (en) | Method and system for retrieving information about television programs | |
US8151298B2 (en) | Method and system for embedding information into streaming media | |
US9253519B2 (en) | Method and apparatus for automatically converting source video into electronic mail messages | |
KR101009629B1 (en) | Extended Metadata Structure and Adaptive Program Service Providing System and Method for Providing Digital Broadcast Program Service | |
US9021357B2 (en) | System and method for enabling collaborative media stream editing | |
US20100070643A1 (en) | Delivery of synchronized metadata using multiple transactions | |
WO2018011689A1 (en) | Method and system for switching to dynamically assembled video during streaming of live video | |
IES20030840A2 (en) | Multimedia management | |
US8805866B2 (en) | Augmenting metadata using user entered metadata | |
JP2010538589A (en) | In-video book marking | |
WO2018011686A1 (en) | Method and system for navigation between segments of real time, adaptive and non-sequentially assembled video | |
US10389779B2 (en) | Information processing | |
IES83424Y1 (en) | Multimedia management | |
IE20030840U1 (en) | Multimedia management | |
IE84144B1 (en) | Multimedia management | |
KR102303309B1 (en) | Method and system for sharing the time link of multimedia | |
Tseng et al. | Video personalization for usage environment | |
Tjondronegoro et al. | Delivering a fully interactive mobile TV | |
Gibbon et al. | Video Content Personalization for IPTV Services | |
Meessen et al. | Content browsing and semantic context viewing through JPEG 2000-based scalable video summary |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FD4E | Short term patents deemed void under section 64 |