WO2009074940A2 - Method of annotating a recording of at least one media signal - Google Patents
Method of annotating a recording of at least one media signal Download PDFInfo
- Publication number
- WO2009074940A2 WO2009074940A2 PCT/IB2008/055137 IB2008055137W WO2009074940A2 WO 2009074940 A2 WO2009074940 A2 WO 2009074940A2 IB 2008055137 W IB2008055137 W IB 2008055137W WO 2009074940 A2 WO2009074940 A2 WO 2009074940A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- recording
- information
- media signal
- physical
- data
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
Definitions
- the invention relates to a method of annotating a recording of at least one media signal, wherein the recording relates to at least one time interval during which corresponding physical signals have been captured, which method includes augmenting the at least one media signal with information based on data representative of values of at least one physical parameter in an environment at a physical location associated with the recording.
- the invention also relates to a system for annotating a recording of at least one media signal, which recording relates to at least one time interval during which corresponding physical signals have been captured, which system includes: a signal processing system for augmenting the at least one media signal with information; and an interface to at least one sensor for measuring at least one physical parameter in an environment at a physical location associated with the recording.
- the invention also relates to a computer programme.
- US 2006/0149781 discloses metadata text files that can be used in any application where a location in a media file or even a text file can be related to sensor information. This point is illustrated in an example in which temperature and humidity readings from sensors are employed to find locations in a video that teaches cooking.
- the chef prepares a meal using special kitchen utensils such as pitchers rigged to sense if they are full of liquid, skillets that sense their temperature, and cookie cutters that sense when they are being stamped. All of these kitchen utensils transmit their sensor values to the video camera, where the readings are recorded to a metadata text file.
- the metadata text file synchronises the sensor readings with the video. When this show is packaged commercially, the metadata text file is included with the video for the show.
- a problem of the known method is that, for all relevant sensor information to be provided with the video, the video recording itself must be very long. If only sensor data captured during the actually recorded video segments are packaged with the video, then information will be missing that could be relevant to the user for determining the conditions prevailing at the location where the video was shot.
- This object is achieved by the method of annotating a recording of at least one media signal according to the invention, which method includes augmenting the at least one media signal with information based on data representative of values of at least one physical parameter in an environment at a physical location associated with the recording and pertaining at least partly to points in time outside the at least one time interval.
- the recording is augmented with information based on data representative of values of at least one physical parameter in an environment at a physical location associated with the recording. It is possible in principle to re-create those circumstances, at least to an approximation, based on that information. This provides for a more engaging playback of the media signals. Because the information is based on parameter values pertaining at least partly to points in time outside the at least one time interval, the information is more accurate. It also covers periods not covered by the media signal, e.g. intervals edited out of the media signal or periods just prior or after the media signal was captured. Thus, the capture of the media signal and the capture of the sensor data for creating the annotating information are decoupled.
- An embodiment of the method includes interpreting the parameter values to transform the parameter values into the information with which the at least one media signal is augmented.
- An embodiment of the method includes receiving at least one stream of parameter values and transforming the at least one stream of parameter values into a data stream having a lower data rate than the at least one stream of parameter values.
- An effect is to provide a form of interpretation that results in values covering longer time intervals than those to which the parameter values pertain.
- This embodiment is suitable for characterising an atmosphere at a location at which the physical signals corresponding to the media signals have been captured or rendered, since environmental conditions generally do not vary on the same short-term time scale as media signals.
- a further variant includes transforming a plurality of sequences of parameter values into a single data sequence included in the information with which the at least one media signal is augmented.
- An effect is to make the annotating information more accurate whilst keeping the amount of annotating information to an acceptable level.
- An embodiment of the method of annotating a recording includes obtaining sensor data by measuring a physical parameter in an environment at a physical location at which the physical signals corresponding to the at least one media signal are captured, and augmenting the at least one media signal with information based at least partly on the thus obtained sensor data.
- An effect is to provide information describing the ambient conditions at a location of recording. Such information is thus in harmony with the impression of the ambient conditions conveyed by the media signal.
- the annotated recording is suited to recreating the ambient conditions, or at least reinforcing an impression of the ambient conditions at playback of the recorded media signal or media signals.
- the parameter values pertain to points in time within at least part of a time interval encompassing the at least one time interval during which the corresponding physical signals are captured.
- An effect is to ensure that the media signals are annotated with information that is relevant to the at least one time interval during which the corresponding physical signals have been captured. Nevertheless, the risk of adding redundant information is relatively low, because the information is based on parameter values pertaining at least partly to points in time outside that at least one time interval.
- An embodiment of the method includes obtaining sensor data by measuring at least one physical parameter representative of a physical quantity different from that represented by the physical signals corresponding to the at least one media signal. An effect is to augment the recording with relatively relevant data.
- the system according to the invention for annotating a recording of at least one media signal includes: a signal processing system for augmenting the at least one media signal with information; and an interface to at least one device for determining values of at least one physical parameter in an environment at a physical location associated with the recording, wherein the system is capable of obtaining data representative of parameter values from the at least one device outside the at least one time interval, and of augmenting the at least one media signal with information based at least partly on those data.
- the system includes an interface to at least one device for determining values of at least one physical parameter in an environment at a physical location associated with the recording
- the system is capable of capturing data representative of ambient conditions at the time the annotated recording was produced. At least an impression of these conditions can be given by a suitable system when the annotated recording is played back. Because the system is capable of obtaining data representative of the physical parameter values outside the at least one time interval and of augmenting the recording with information based at least partly on that data, comprehensive information is provided relatively efficiently.
- the system is configured to carry out a method of annotating a recording of at least one media signal according to the invention.
- the system is configured automatically to ensure that the at least one media signal is augmented with information based on data representative of values of at least one physical parameter in an environment at a physical location associated with the recording and pertaining at least partly to points in time outside the at least one time interval.
- a computer programme including a set of instructions capable, when incorporated in a machine-readable medium, of causing a system having information processing capabilities to perform a method according to the invention.
- Fig. 1 is a schematic diagram of a recording system
- Fig. 2 is a state diagram of a recording process carried out using the recording system of Fig. 1;
- Fig. 3 is a schematic diagram of a home entertainment system.
- the recording system 1 for capturing physical signals to create an annotated recording of one or more media signals is shown.
- the recording system 1 comprises a video camera 2, a microphone 3 and first, second and third sensors 4-6.
- the video camera 2 includes a light-sensitive sensor array 7 for converting light intensity values into a digital video data stream.
- the digital video data stream will be encoded and compressed, synchronised with a digital audio data stream and recorded to a recording medium in a recording device 8, together with the digital audio data stream.
- the media signals are augmented with annotation information based on data representative of values of at least one physical parameter in an environment at the recording location.
- a physical parameter is a value of some physical quantity, i.e. a quantity relating to forces of nature.
- the video camera 2 includes a user interface in the form of a touch screen interface 9. It includes a first user control 10 for starting and stopping the capture of video and audio signals. It further includes a second user control 11 for starting and stopping the capture of annotation information based on data representative of at least one physical parameter at the recording location.
- At least one of the sensors 4-6 is provided for measuring at least one physical parameter representative of a physical quantity different from that represented by the physical signals corresponding to the digital audio and video signals.
- the first sensor 4 can measure temperature
- the second sensor 5 can measure humidity
- the third sensor 6 can measure vibration, for example.
- fewer or no sensors 4-6 are present, and the annotating information is based on e.g. the signal from the microphone 3.
- at least one of the sensors 4-6 measures a physical parameter representative of a similar quantity to those captured by the digital audio and video signals.
- one of the sensors 4-6 can measure the ambient light intensity.
- values are obtained from a system for regulating devices arranged to adjust ambient conditions, e.g. a background lighting level.
- ambient conditions e.g. a background lighting level.
- this aspect of ambient conditions is not measured directly.
- sensor data e.g. where a sensor measures wind speed and the settings for regulating floodlighting are collected also.
- Fig. 2 Some states of the recording system 1 are shown in Fig. 2.
- an operator will use the second user control 11 to commence capture and recording of the ambience at the scene of recording (state 12).
- the video camera 2 continually captures (state 13) streams 14-16 of parameter values received through its interface to the three sensors 4-6.
- These three streams 14-16 of data values are reduced to a single set 17 of ambience data values.
- the reduction comprises interpreting the streams of parameter values (state 18) and adding timing information (state 19), prior to recording the ambience information to the recording medium (state 20).
- the latter state 20 comprises recording the ambience information in text format, e.g. in xML (extensible Markup Language) format in a file.
- xML extensible Markup Language
- the three streams 14-16 of parameter values are reduced to a stream of ambience values, each value representative of an ambience at a corresponding point in time.
- Timing information to relate each ambience value to a point in time is added.
- the timing information serves to identify the time interval over which the ambience was determined, so that the ambience information relates to the entire duration of the state 12.
- the first and second streams 14,15 are reduced to a time-stamped sequence of ambience values and the first, second and third stream 16 are interpreted to arrive at a set of data characterising a further aspect of the ambience over the duration of the state 12 of capturing the ambience.
- Fig. 2 also shows a state 21 in which only media signals are recorded.
- An audio stream 22 and a video stream 23 are captured (state 24). They are synchronised (state 25) using timing information, and recorded on a recording medium in the recording device 8 (state 26).
- the general progression from and to the state 12 of capturing ambience data serves to provide in a relatively simple way more reliable information on the ambience at a recording location.
- the normal progression is from the state 12 of capturing and recording the ambience to a state 27 of capturing and recording both the audiovisual signals and the ambience and back again to the state 12 of capturing and recording the ambience, as the user actuates the first user control 10 to record video segments.
- the ambience data is based also on parameter values pertaining to points in time within the intervening time intervals, as well as points in time within the time interval preceding the recording. In an embodiment, this is automated by appropriate programming of the video camera 2.
- the set 17 of ambience information is based on values of the signal from the microphone 3, and the sensors 4-6 are not used.
- the ambience information is based on values of the microphone signal pertaining to points in time outside the time intervals of recording the audio signal, the overall information content of the annotated recording is still enhanced.
- the microphone signal is interpreted to derive information representative of an ambience (as opposed to acoustic energy).
- the ambience information can result from a determination of the average background noise level over a time interval encompassing the time intervals during which the recorded audio and video signals were captured.
- Fig. 3 illustrates a home entertainment system 28 including a home theatre 29 a television set 30 and speakers 31,32.
- the home theatre 29 is controlled by a data processing unit 36 for manipulating data held in main memory 37.
- a video output stage 38 provides a decoded video signal to the television set 30.
- An audio output stage 39 provides analogue audio signals to the speakers 31,32.
- the home theatre 29 further includes an interface 33 to first and second peripheral devices 34,35 for adjusting physical conditions in an environment of the home entertainment system 28.
- peripheral devices 34,35 are representative of a class of devices including lights adapted to emit light of varying colour and intensity; fans adapted to provide an airflow; washer light units for providing back-lighting varying in intensity and colour; and nimbler devices allowing a user to experience movement and vibration. Other sensations such as smell may also be provided.
- the data processing unit 36 controls the output of the peripheral devices 34,35 via the interface 33 by executing instructions encoded in scripts, for example scripts in a dedicated (proprietary) mark-up language.
- the scripts include timing information and information representative of settings of the peripheral devices 34,35.
- Media signals are accessed by the home theatre 29 from an internal mass storage device 40 or from a read unit 41 for reading data from a recording medium, e.g. an optical disk.
- the home theatre 29 is also capable of receiving copies of recordings of media signals via a network interface 42.
- the home theatre 29 can obtain media signals annotated with scripts indicating the settings for the peripheral devices 34,35, in a manner known per se. However, the home theatre 29 can also obtain media signals annotated with information of the type created using the method illustrated in Fig. 2.
- the home theatre 29 obtains the script itself by interpreting the information annotating the media signal according to certain rules to determine at least one target ambience, and by then transforming the target ambience or ambiences into settings for the peripheral devices 34,35, and optionally into settings for the audio output stage 39, speakers 31,32, or other components of the system for rendering the audiovisual signal.
- the annotated recording can be one obtained at an airfield. Even if there is no footage of an aeroplane taking off or coming in to land, the annotating information will still indicate a noisy ambience. This is because the ambience data is based on values of at least one physical parameter (such as noise level) pertaining at least partly to points in time outside the time interval of recording.
- the home theatre 29 translates the information indicating a noisy ambience into a script for regulating the peripheral devices 34,35 to re-create the ambience, e.g. to create a vibrating sensation and to add the sound of aeroplanes to the audio track comprised in the media signals.
- the home theatre 29 employs a database relating particular ambiences to particular settings and/or particular parameter values in algorithms for creating settings in dependence on characteristics of the media signals.
- the home entertainment system 28 is further configured to carry out a method as illustrated in Fig. 2.
- it is able to augment media signals with information based on data representative of values of at least one physical parameter in an environment at a physical location associated with a recording of the media signals being processed by it and corresponding to the location at which the media signals are rendered.
- such parameter values pertain to points in time outside the time intervals of recording the media signals originally.
- the home theatre 29 includes an interface 43 to a sensor 44 similar to the sensors 4-6 of the recording system 1 of Fig. 1.
- a recording of media signals can be augmented with annotating information representing the ambience at the time of first rendering the media signals, so that that ambience can be re-created at a later time.
- a mobile phone may operate as a recording system, being fitted with a camera for obtaining a media signal in the form of a digital image, as well as a microphone.
- the sound information is not recorded, but the sound signal over an interval encompassing the point in time at which the image was captured may be analysed to determine an ambience. For example, where a digital image is captured at a football match, the sound signal may be analysed to determine automatically the mood of the crowd.
- a distributed recording system is used. Data representative of values in a city are obtained whilst digital images are captured using wireless communications to networked sensors distributed about the city. Data representative of music listened to in the course of a time interval during which the digital images were captured are also analysed. The totality of data are analysed to derive information representative of the mood the user was in whilst the digital images were captured and/or the ambience in the city.
- Each of these embodiments allows the media signals to be augmented with information based on parameter values that are not directly derivable from the media signals themselves.
- Each of these embodiments achieves this in an efficient manner by interpreting parameter values to infer an ambience or mood, rather than recording additional signals from sensors.
- the information representative of the ambience or mood is based at least partly on parameter values pertaining to points in time outside the recording intervals, so that the reliability of the annotating information is enhanced.
- the media signal and annotating information may be recorded temporarily in a memory device, e.g. a solid-state memory device or hard disk unit, and then communicated via a network.
- a memory device e.g. a solid-state memory device or hard disk unit
- 'Means' as will be apparent to a person skilled in the art, are meant to include any hardware (such as separate or integrated circuits or electronic elements) or software (such as programs or parts of programs) which perform in operation or are designed to perform a specified function, be it solely or in conjunction with other functions, be it in isolation or in co-operation with other elements.
- 'Computer programme' is to be understood to mean any software product stored on a computer-readable medium, such as an optical disk, downloadable via a network, such as the Internet, or marketable in any other manner.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Library & Information Science (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Television Signal Processing For Recording (AREA)
- Management Or Editing Of Information On Record Carriers (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US12/746,203 US20100257187A1 (en) | 2007-12-11 | 2007-12-11 | Method of annotating a recording of at least one media signal |
JP2010537567A JP2011507379A (en) | 2007-12-11 | 2008-12-08 | Method for annotating a recording of at least one media signal |
CN200880120374XA CN101896903A (en) | 2007-12-11 | 2008-12-08 | Mark the method for the record of at least one media signal |
EP08860238A EP2235645A2 (en) | 2007-12-11 | 2008-12-08 | Method of annotating a recording of at least one media signal |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07122832.4 | 2007-12-11 | ||
EP07122832 | 2007-12-11 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2009074940A2 true WO2009074940A2 (en) | 2009-06-18 |
WO2009074940A9 WO2009074940A9 (en) | 2009-11-05 |
Family
ID=40755946
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2008/055137 WO2009074940A2 (en) | 2007-12-11 | 2008-12-08 | Method of annotating a recording of at least one media signal |
Country Status (6)
Country | Link |
---|---|
US (1) | US20100257187A1 (en) |
EP (1) | EP2235645A2 (en) |
JP (1) | JP2011507379A (en) |
KR (1) | KR20100098434A (en) |
CN (1) | CN101896903A (en) |
WO (1) | WO2009074940A2 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9473813B2 (en) * | 2009-12-31 | 2016-10-18 | Infosys Limited | System and method for providing immersive surround environment for enhanced content experience |
CA2834217C (en) * | 2011-04-26 | 2018-06-19 | The Procter & Gamble Company | Sensing and adjusting features of an environment |
KR101328270B1 (en) * | 2012-03-26 | 2013-11-14 | 인하대학교 산학협력단 | Annotation method and augmenting video process in video stream for smart tv contents and system thereof |
US11816757B1 (en) * | 2019-12-11 | 2023-11-14 | Meta Platforms Technologies, Llc | Device-side capture of data representative of an artificial reality environment |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060149781A1 (en) | 2004-12-30 | 2006-07-06 | Massachusetts Institute Of Technology | Techniques for relating arbitrary metadata to media files |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0470726A (en) * | 1990-07-11 | 1992-03-05 | Minolta Camera Co Ltd | Camera capable of recording humidity information |
JPH09205607A (en) * | 1996-01-25 | 1997-08-05 | Sony Corp | Video recording device and reproducing device |
US7253302B2 (en) * | 2002-12-09 | 2007-08-07 | Smith Ronald J | Mixed esters of dicarboxylic acids for use as pigment dispersants |
US20040167767A1 (en) * | 2003-02-25 | 2004-08-26 | Ziyou Xiong | Method and system for extracting sports highlights from audio signals |
US7149961B2 (en) * | 2003-04-30 | 2006-12-12 | Hewlett-Packard Development Company, L.P. | Automatic generation of presentations from “path-enhanced” multimedia |
US20060078288A1 (en) * | 2004-10-12 | 2006-04-13 | Huang Jau H | System and method for embedding multimedia editing information in a multimedia bitstream |
EP1878209A4 (en) * | 2005-04-29 | 2009-12-02 | Hingi Ltd | A method and an apparatus for provisioning content data |
JP2007094544A (en) * | 2005-09-27 | 2007-04-12 | Fuji Xerox Co Ltd | Information retrieval system |
-
2007
- 2007-12-11 US US12/746,203 patent/US20100257187A1/en not_active Abandoned
-
2008
- 2008-12-08 JP JP2010537567A patent/JP2011507379A/en active Pending
- 2008-12-08 CN CN200880120374XA patent/CN101896903A/en active Pending
- 2008-12-08 EP EP08860238A patent/EP2235645A2/en not_active Withdrawn
- 2008-12-08 WO PCT/IB2008/055137 patent/WO2009074940A2/en active Application Filing
- 2008-12-08 KR KR1020107015092A patent/KR20100098434A/en not_active Application Discontinuation
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060149781A1 (en) | 2004-12-30 | 2006-07-06 | Massachusetts Institute Of Technology | Techniques for relating arbitrary metadata to media files |
Also Published As
Publication number | Publication date |
---|---|
JP2011507379A (en) | 2011-03-03 |
US20100257187A1 (en) | 2010-10-07 |
EP2235645A2 (en) | 2010-10-06 |
CN101896903A (en) | 2010-11-24 |
KR20100098434A (en) | 2010-09-06 |
WO2009074940A9 (en) | 2009-11-05 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11477156B2 (en) | Watermarking and signal recognition for managing and sharing captured content, metadata discovery and related arrangements | |
US9183883B2 (en) | Method and system for generating data for controlling a system for rendering at least one signal | |
JP5485913B2 (en) | System and method for automatically generating atmosphere suitable for mood and social setting in environment | |
US6639649B2 (en) | Synchronization of music and images in a camera with audio capabilities | |
JP6589880B2 (en) | Information processing system, control method, and storage medium | |
CN101577129B (en) | Digital photo frame and method for automatically switching playing media | |
US20100257187A1 (en) | Method of annotating a recording of at least one media signal | |
CN114868186B (en) | System and apparatus for generating content | |
CN104508614B (en) | Display control unit, display control method and program | |
JP2011217183A (en) | Electronic device, image output method and program | |
CN105812927A (en) | Method for backing up scene atmosphere and television | |
JP5544030B2 (en) | Clip composition system, method and recording medium for moving picture scene | |
US11595720B2 (en) | Systems and methods for displaying a context image for a multimedia asset | |
EP3035208A1 (en) | Improving the selection and control of content files | |
CN115476789A (en) | Vehicle control method and device, vehicle machine and vehicle | |
JP2012120128A (en) | Playback system and playback method | |
JP2010263331A (en) | Mobile terminal | |
WO2006093184A1 (en) | Video edition device, video edition method, and computer program for performing video edition | |
JP2007087559A (en) | Recording and reproducing apparatus, control method of recording and reproducing apparatus, and control program of recording and reproducing apparatus |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200880120374.X Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 08860238 Country of ref document: EP Kind code of ref document: A2 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008860238 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010537567 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12746203 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 4199/CHENP/2010 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 20107015092 Country of ref document: KR Kind code of ref document: A |