GB2361094A - Producing audio/visual material and identifying equipment used in its production - Google Patents

Producing audio/visual material and identifying equipment used in its production Download PDF

Info

Publication number
GB2361094A
GB2361094A GB0008416A GB0008416A GB2361094A GB 2361094 A GB2361094 A GB 2361094A GB 0008416 A GB0008416 A GB 0008416A GB 0008416 A GB0008416 A GB 0008416A GB 2361094 A GB2361094 A GB 2361094A
Authority
GB
United Kingdom
Prior art keywords
identifier
item
database
data
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB0008416A
Other versions
GB0008416D0 (en
Inventor
Martin Rex Dorricott
Richard Daniel Foster
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Europe Ltd
Original Assignee
Sony United Kingdom Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony United Kingdom Ltd filed Critical Sony United Kingdom Ltd
Priority to GB0008416A priority Critical patent/GB2361094A/en
Publication of GB0008416D0 publication Critical patent/GB0008416D0/en
Publication of GB2361094A publication Critical patent/GB2361094A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/11Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/32Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
    • G11B27/326Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier used signal is a video-frame or a video-field (P.I.P.)
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements 
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/36Monitoring, i.e. supervising the progress of recording or reproducing
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B31/00Arrangements for the associated working of recording or reproducing apparatus with related apparatus
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/235Processing of additional data, e.g. scrambling of additional data or processing content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/435Processing of additional data, e.g. decrypting of additional data, reconstructing software from modules extracted from the transport stream
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/20Disc-shaped record carriers
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B2220/00Record carriers by type
    • G11B2220/90Tape-like record carriers

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

A method of producing video and/or audio material comprising the steps of:<BR> ```using video and/or audio equipment to produce video and/or audio material;<BR> ```producing a material identifier identifying the material;<BR> ```producing at least an item identifier identifying an item of equipment used in the production of the material;<BR> ```associating the item identifier with the material identifier; and<BR> ```transferring at least the associated identifiers to a database. The invention has particular use in monitoring which equipment has been used in order to produce the video and/or audio material in order to brace faulty equipment or monitor equipment usage.

Description

1 Producing Material 2361094 The present invention relates to producing
video and/or audio material.
Video and audio equipment includes cameras, camera/recorders, audio recorders, tapes, discs, editing equipment, VTRs, and many other items all of which are prone to faults and/or wear.
Professional audio and video production is expensive because it uses groups of highly paid people. Thus it is important to maintain equipment reliability to avoid delays due to faulty equipment Audio and video equipment uses and/or contains items which are prone to wear and have lifetimes after which they should be replaced. Especially in professional audio and video production, video and audio quality is of paramount importance. The quality is determined at least in part by the equipment.
For example worn tape heads and worn tapes impair quality. Some rechargeable batteries need to be replaced after a fixed number of recharge cycles after which they may fail to become fully charged.
Ideally other items of equipment should be checked for faults. However, professional video and audio equipment is often used 'on location' which may be far distant from its maintenance base. The equipment is also expensive and thus may be used intensively and not often return to its base.
Furthermore, an Editor, Director or Producer monitors the quality of the material which is produced. If material does not have the appropriate quality, then he or she needs to take action. If the impairment is due to faulty equipment it is necessary to identify the faulty equipment.
Even though the primary task of personnel on location is to produce high quality video and audio material, they may not be aware of the effect of faulty equipment. The editor may be remote from the location and if he or she notices that faulty equipment is impairing video or audio quality, he or she does not have the information available to identify the equipment at fault. Furthermore the equipment owners may wish to know the state of repair of their equipment.
According to one aspect of the present invention, there is provided a method of producing the video and/or audio material comprising the steps of.
using video and/or audio equipment to produce video and/or audio material; producing a material identifier identifying the material; 2 producing at least an item identifier identifying an item used in the production of the material; associating the item identifier with the material identifier; and transferring at least the associated identifiers to a database.
The material identifier may directly or indirectly identify the material. In one example of indirect identification, the material is recorded on an example of a recording medium and the identifier identifies the example of the recording medium on which the material is recorded. For example the identifier is a serial number. The identifier is associated with a unique identifier of the material recorded on the medium. In direct identification, the identifier uniquely identifies the material. It may be included in the material. If the material is recorded on a recording medium, the identifier be recorded on the medium with the material either in the material or separately but associated with the material.
Thus if an editor notices a problem in some material which may be due to an equipment fault, the editor can use the database which associates the material with the equipment to identify the equipment at fault. Also the database allows the owners to monitor the state of repair of their equipment.
According to another aspect of the present invention, there is provided a system for producing video and/or audio material, comprising:
video and/or audio equipment for producing video and/or audio material; first generator for producing a material identifier identifying the material; processor for associating an item identifier identifying an item used in the production of the recorded material with the material identifier; database processor; and data link for transferring the associated identifiers to the database processor.
In an embodiment of the invention, a second generator produces the item identifier.
In an embodiment of the invention, the equipment includes a material recorder for recording the material on a recording medium.
The recorder may also record the material identifier on the medium either in the material or separately therefrom. This allows a person reviewing the recorded material to determine, from the material identifier and the associated item identifier in the database, the equipment item(s) used to produce the material and may then allow the identification of items which reduce the quality of the material.
3 The entering of the item identifier(s) into the database also allows at least the usage of the items(s) to be tracked. This allows a maintenance schedule for the equipment to be implemented. In a preferred embodiment, maintenance data relating t) to usage and/or faults in the item are produced as the equipment is used, preferably automatically, and entered into the database allowing planned maintenance and analysis of faults.
4 For better understanding of the present invention and to show how the same may be carried into effect reference will now be made by way of example to the accompanying drawings in which:
Figure 1 is a schematic diagram of a system in accordance with the present invention; Figure 2 is-a schematic block diagram of a video camera which is arranged to communicate to a personal digital assistant (PDA), Figure 3 is-shows a more detailed representation of the body 102 of the video camera which is shown in figure 1; Figures 4 illustrates the form of the PDA illustrates a PDA and the part of the camera associated with generating metadata; Figure 6 is a schematic block diagram of a camera, acquisition unit and PDA; Figure 7 is a diagram of the connection of an acquisition unit to a camera; Figure 8 shows a UMID; Figure 9 shows a UMID with reduced data; and Figure 10 illustrates a data structure for a metadata base.
Overview Figure 1 Referring to Figure 1, video and audio equipment 101 is shown. The equipment comprises various items: a camera body 102, a lens 106, and a metadata acquisition unit 152. The equipment has other items associated with it including a tape or other recording medium (not shown), sound recording items (not shown), rechargeable batteries (not shown), lights (not shown), and a PDA 112 which will be described later. The equipment may comprise, or be associated with, other items as known to video/audio production crews. Such items may hired from an equipment owner. Also such items are subject to wear and require maintenance, or replacement, after prolonged use. Tapes need to be replaced after several record erase cycles. Some types of rechargeable battery fall to maintain charge after a known number of charge/recharge cycles. Cameras and lenses may develop faults. If the items are hired the owner needs to know where they are and may want to monitor usage to plan maintenance.
In embodiments of the present invention, items are allocated identification codes. The codes may be simply written or engraved on the items. Alternatively or in addition, some items may have the identification codes recorded in a recording medium such as magnetic tape, or in a ROM or RAM or in a data processor. For example it is known to provide data processors or memory devices in/on batteries for management of charging. It is also proposed to record tape identifiers on tapes. It is known to provide processors or memory on tape cassettes to identify the tapes therein.' Lenses may have, and cameras have, processors for the control thereof and which can produce identifiers, such as serial numbers.
Items of equipment which do not have such memory or processing have serial numbers allocated to them. The serial numbers are entered into memory in the PDA 112. The PDA 112 is arranged to automatically receive item identifiers from the items which can automatically produce them.
The PDA 112 also contains data uniquely identifying the video and /or audio material being produce using the items of equipment. Such data is or includes for example UMIDs which are unique material identifiers described in more detail in the section UMIDs below. The PDA associates the item identifiers with the UMIDs. The PDA may itself transfer the associated item identifiers and UMIDs to a database processor 430. Alternatively the PDA may transfer the associated item identifiers and UMIDs to the unit 152 via a data link 119. The unit may have an RF transmitter or other data link 174 to the database processor 430. The unit 152, the PDA 112, the camera and lens 0 11, 102, 016 and the data links 174 and 119 will be described on more detail in the section A-Box below.
In the example of Figure 1, a recording medium, in this case a tape, containing video and audio material is sent to a remote studio. The tape is identified by a tape ID which for example is recorded on the tape. The material on the tape is uniquely identified by UMIDs which may be recorded on the tape or may be in the database and related to the tape by the tape ID and other material identifiers on the tape such as IN and OUT time codes. The item identifiers are also in the database and related to the UMIDs. If, a person, for example an editor, in the studio notices a problem which may be caused by an item of equipment, the person accesses the database via the material identifiers he or she has to determine at least the identifiers of the items of equipment used in the production of the material. Preferably, the database also includes data identifying the types of equipment to which the identifiers refer. So the person can determine that for example a particular camera produced some video material and that there may be a fault in the camera.
6 Examples of the data structures which may be put in the database are given in the section Camera Metadata below.
An equipment owner may wish to monitor usage of his equipment. Thus the owner may have direct access to the database 430 or subset of the data in the database which subset is relevant to the owner. It is currently preferred that the equipment owner has access only to the subset especially if he is not the video/audio production company because the metadata in the database relating to the video/audio material is a valuable resource and may include confidential information. In a currently most preferred embodiment, the subset of data relevant to the equipment owner is transferred to a database 43)2 owned by the owner.
As is apparent from the section Camera Metadata below the camera metadata includes:
fi 1) Item IDs such as Lens ID, Tape ID, Camera ID, Tape ID and other details of equipment., The item IDs and other details allow the owner to identify the items of equipment.
2) A UMID. An extended UMID includes GPS data defining where the equipment is. Thus the owner has a means of tracking his equipment, 3) Usage data for example tape erasures, and video head hours. These allow the owner to plan maintenance and replacement of equipment.
4) Data identifying the production crew and contact names and addresses including e-mail addresses so the owner knows who has possession of his equipment and how to contact them; Other information may be provided as required by the owner. Such information may preferably include date, and location independent of GPS. Communication links may be provided to allow: the studio 431 to communicate with the owner to correct faulty equipment; the owner to send messages to the studio and or the production crew concerning equipment replacement or maintenance; the studio to send messages to the production crew concerning retaking shots:. and /or any other purpose.
7 The database may be arranged to automatically monitor usage and to compare usage with preset maximum usage values set by the owner. When the preset values are reached or exceeded, then a message is automatically generated and sent to the owner and to anyone such as a production crew who have possession of the equipment about arranging maintenance or replacement. The message may be an e-mail for example. The conununication links may include the link 174. Such links are discussed in the section A -Box below.
8 UMIDs Figures 8 and 9 A UMID is described in reference @1. Referring to Figure 8, an extended UMID is shown. It comprises a first set of 32 bytes of basic UMID and a second set of 3)2 bytes of signature metadata.
The first set of 32 bytes is the basic UMID. The components are:
A 12-byte Universal Label to identify this as a SMPTE UMID. It defines the type of material which the UMID identifies and also defines the methods by which the globally unique Material and locally unique Instance numbers are created.
A 1 -byte length value to define the length of the remaining part of the UMID.
A 3-byte Instance number which is used to distinguish between different 'instances' of material with the same Material number.
A 16-byte Material number which is used to identify each clip. Each Material number is the same for related instances of the same material.
The second set of 32 bytes of the signature metadata as a set of packed metadata items used to create an extended UMID. The extended UMID comprises the basic UMID followed immediately by signature metadata which comprises:
An 8-byte time/date code identifying the time and date of the Content Unit creation.
Unit creation.
codes tum.
A 12-byte value which defines the spatial co-ordinates at the time of Content 3 groups of 4-byte codes which register the country, organisation and user Each component of the basic and extended UMIDs will now be defined in The 12-byte Universal Label The first 12 bytes of the UMID provide identification of the UMID by the registered string value defined in table 1.
Byte No. Description Value (hex)
1 Object Identifier 06h 2 Label size 0Ch 9 3 Designation: ISO 2Bh 4 Designation: SMPTE -34h, Registry: Dictionaries 01h 6 Registry: Metadata Dictionaries 01h 7 Standard: Dictionary Number 01h.
8 Version number 01h 9 Class: Identification and location 01h Sub-class: Globally Unique Identifiers 01h 11 Type: UMID (Picture, Audio, Data, Group) 01,02,03,04h 12 Type: Number creation method XXh Table 1: Specification of the UMID Universal Label
The hex values in table 1 may be changed: the values given are examples. Also the bytes 1-12 may have designations other than those shown by way of example in the table. Referring to the Table 1, in the example shown byte 4 indicates that bytes 5-12 relate to a data format agreed by SMPTE. Byte 5 indicates that bytes 6 to 10 relate to "dictionary" data. Byte 6 indicates that such data is "metadata" defined by bytes 7 to 10. Byte 7 indicates the part of the dictionary containing metadata defined by bytes 9 and 10.. Byte 10 indicates the version of the dictionary. Byte 9 indicates the class of data and Byte 10 indicates a particular item in the class.
In the present embodiment bytes 1 to 10 have fixed preassigned values. Byte 11 is variable. Thus referring to Figure 9, and to Table 1 above, it will be noted that the bytes 1 to 10 of the label of the UMID are fixed. Therefore they may be replaced by a 1 byte 'Type' code T representing the bytes 1 to 10. The type code T is followed by a length code L. That is followed by 2 bytes, one of which is byte 11 of Table 1 and the other of which is byte 12 of Table 1, an instance number (3 bytes) and a material number (16 bytes). Optionally the material number may be followed by the signature metadata of the extended UMID and/or other metadata.
The UMID type (byte 11) has 4 separate values to identify each of 4 different data types as follows:
0 1 h' = UMID for Picture material 02h' = UMID for Audio material 03h' = UMID for Data material 04h' = UMID for Group material (i.e. a combination of related essence).
The last (12th) byte of the 12 byte label identifies the methods by which the material and instance numbers are created. This byte is divided into top and bottom nibbles where the top nibble defines the method of Material number creation and the bottom nibble defines the method of Instance number creation.
Length The Length is a 1-byte number with the value '13h' for basic UMIDs and )')h' for extended UMIDs.
Instance Number The Instance number is a unique 3-byte number which is created by one of several means defined by the standard. It provides the link between a particular 'instance' of a clip and externally associated metadata. Without this instance number, all material could be linked to any instance of the material and its associated metadata.
The creation of a new clip requires the creation of a new Material number together with a zero Instance number. Therefore, a non-zero Instance number indicates that the associated clip is not the source material. An Instance number is primarily used to identify associated metadata related to any particular instance of a clip.
Material Number The 16-byte Material number is a non-zero number created by one of several means identified in the standard. The number is dependent on a 6- byte registered port ID number, time and a random number generator.
Signature Metadata Any component from the signature metadata may be null-filled where no meaningful value can be entered. Any null-filled component is wholly null- filled to clearly indicate a downstream decoder that the component is not valid.
The Time-Date Format The date-time format is 8 bytes where the first 4 bytes are a UTC (Universal Time Code) based time component. The time is defined either by an AES3 32-bit audio sample clock or SMPTE 12M depending on the essence type.
The second 4 bytes define the date based on the Modified Julian Data (MJD) as defined in SMPTE 309M. This counts up to 999,999 days after midnight on the 17th November 1858 and allows dates to the year 4597.
11 The Spatial Co-ordinate Format The spatial co-ordinate value consists of three components defined as follows:
Altitude: 8 decimal numbers specifying up to 99,999,999 metres.
Longitude: 8 decimal numbers specifying East/West 180.00000 degrees (5' decimal places active).
Latitude: 8 decimal numbers specifying North/South 90.00000 degrees (5 decimal places active).
The Altitude value is expressed as a value in metres from the centre of the earth thus allowing altitudes below the sea level.
It should be noted that although spatial co-ordinates are static for most clips, this is not true for all cases. Material captured from a moving source such as a camera mounted on a vehicle may show changing spatial coordinate values.
Country Code The Country code is an abbreviated 4-byte alpha-numeric string according to the set defined in ISO 3166. Countries which are not re-istered can obtain a registered alpha-numeric string from the SMPTE Registration Authority.
Organisation Code The Organisation code is an abbreviated 4-byte alpha-numeric string registered with SMPTE. Organisation codes have meaning only in relation to their registered Country code so that Organisation codes can have the same value in different countries.
User Code The User code is a 4-byte alpha-numeric string assigned locally by each organisation and is not globally registered. User codes are defined in relation to their registered Organisation and Country codes so that User codes may have the same value in different organisations and countries.
Freelance Operators Freelance operators may use their country of domicile for the country code and use the Organisation and User codes concatenated to e.g. an 8 byte code which can be registered with SMPTE. These freelance codes may start with the '-' symbol ( ISO 8859 character number 7Eh) and followed by a registered 7 digit alphanumeric string.
12 Camera Metadata The following is provided, by way of example, to illustrate the possible types of metadata generated during the production of a programme, and one possible organisational approach to structuring that metadata.
Figure 10 illustrates an example structure for organising metadata. A number of tables each comprising a number of fields containing metadata are provided. The tables may be associated with each other by way of common fields within the respective tables, thereby providing a relational structure. Also, the structure may comprise a number of instances of the same table to represent multiple instances of the object that the table may represent. The fields may be formatted in a predetermined manner. The size of the fields may also be predetermined. Example sizes include "Int" which represents 2 bytes, "Long Int" which represents 4 bytes and "Double" which represents 8 bytes. Alternatively, the size of the fields may be defined with reference to the number of characters to be held within the field such as, for example, 8, 10, 16, 32, 128, and 255 characters.
Turning to the structure in more detail, there is provided a Programme Table. The Programme Table comprises a number of fields including Programme ID (PID), Title, Working Title, Genre ID, Synopsis, Aspect Ratio, Director ID and Picturestamp. Associated with the Programme Table is a Genre Table, a Keywords Table, a Script Table, a People Table, a Schedule Table and a plurality of Media Object Tables.
The Genre Table comprises a number of fields including Genre ID, which is associated with the Genre ID field of the Programme Table, and Genre Description.
The Keywords Table comprises a number of fields including Programme ID, which is associated with the Programme ID field of the Programme Table, Keyword ID and Keyword.
The Script Table comprises a number of fields including Script ID, Script Name, Script Type, Document Format, Path, Creation Date, Original Author, Version, Last Modified, Modified By, PID associated with Programme ID and Notes. The People Table comprises a number of fields including Image.
The People Table is associated with a number of Individual Tables and a number of Group Tables. Each Individual Table comprises a number of fields including Image. Each Group Table comprises a number of fields including Image.
13 Each Individual Table is associated with either a Production Staff Table or a Cast Table.
The Production Staff Table comprises a number of fields including Production Staff ID, Surname, Firstname, Contract ID, Agent, Agency ID, Email, Address, Phone Number, Role ID, Notes, Allergies, DOB, National Insurance Number and Bank ID and Picture Stamp.
The Cast Table comprises a number of fields including Cast ID, Surname, Firstname, Character Name, Contract ID, Agent, Agency ID, Equity Number, E-mail, Address, Phone Number, DOB and Bank ID and Picture Stamp. Associated with the Production Staff Table and Cast Table are a Bank Details Table and an Agency Table.
The Bank Details Table comprises a number of fields including Bank ID, which is associated with the Bank ID field of the Production Staff Table and the Bank ID field of the Cast Table, Sort Code, Account Number and Account Name.
The Agency Table comprises a number of fields including Agency ID, which is associated with the Agency ID field of the Production Staff Table and the Agency ID field of the Cast Table, Name, Address, Phone Number, Web Site and E-mail and a Picture Stamp. Also associated with the Production Staff Table is a Role Table.
The Role Table comprises a number of fields including Role ID, which is associated with the Role ID field of the Production Staff Table, Function and Notes and a Picture Stamp. Each Group Table is associated with an Organisation Table.
The Organisation Table comprises a number fields including Organisation ID, Name, Type, Address, Contract ID, Contact Name, Contact Phone Number and Web Site and a Picture Stamp.
Each Media Object Table comprises a number of fields including Media Object ID, Name, Description, Picturestamp, PID, Format, schedule ID, script ID and Master ID. Associated with each Media Object Table is the People Table, a Master Table, a Schedule Table, a Storyboard Table, a script table and a number of Shot Tables.
The Master Table comprises a number of fields including Master ID, which is associated with the Master ID field of the Media Object Table, Title, Basic UMID, EDL ID, Tape ID and Duration and a Picture Stamp.
The Schedule Table comprises a number of fields including Schedule ID, Schedule Name, Document Format, Path, Creation Date, Original Author, Start Date,
14 End Date, Version, Last Modified, Modified By and Notes and PID which is associated with the programme ID.
The contract table contains: a contract ID which is associated with the contract ID of the Production staff, cast, and organisation tables; commencement date, rate, job title, expiry date and details.
The Storyboard Table comprises a number of fields including Storyboard ID, which is associated with the Storyboard ID of the shot Table, Description, Author, Path and Media ID.
Each Shot Table comprises a number of fields including Shot ID, PID, Media ID, Title, Location ID, Notes, Picturestamp, script ID, schedule ID, and description. Associated with each Shot Table is the People Table, the Schedule Table, script table, a Location Table and a number of Take Tables.
The Location Table comprises a number of fields including Location ID, which is associated with the Location ID field of the Shot Table, GPS, Address, Description, Name, Cost Per Hour, Directions, Contact Name, Contact Address and Contact Phone Number and a Picture Stamp.
Each Take Table comprises a number of fields including Basic UMID, Take Number, Shot ID, Media ID, Timecode IN, Timecode OUT, Sign Metadata, Tape ID, Camera ID, Head Hours, Videographer, IN Stamp, OUT Staffip. Lens ID, AUTOID ingest ID and Notes. Associated with each Take Table is a Tape Table, a Task Table, a Camera Table, a lens table, an ingest table and a number of Take Annotation Tables.
The Ingest table contains an Ingest ID which is associated with the Ingest Id in the take table and a description.
The Tape Table comprises a number of fields including Tape ID, which associated with the Tape ID field of the Take Table, PID, Format, Max Duration, First Usage, Max Erasures, Current Erasure, ETA ( estimated time of arrival) and Last Erasure Date and a Picture Stamp.
The Task Table comprises a number of fields including Task ID, PID, Media ID, Shot ID, which are associated with the Media ID and Shot ID fields respectively of the Take Table, Title, Task Notes, Distribution List and CC List. Associated with the Task Table is a Planned Shot Table.
The Planned Shot Table comprises a number of fields including Planned Shot ID, PID, Media ID, Shot ID, which are associated with the PID, Media ID and Shot
ID respectively of the Task Table, Director, Shot Title, Location, Notes, Description, Videographer, Due date, Programme title, media title Aspect Ratio and Format.
The Camera Table comprises a number of fields including Camera ID, which is associated with the Camera ID field of the Take Table, Manufacturer, Model, Format, Serial Number, Head Hours, Lens ID, Notes, Contact Name, Contact Address and Contact Phone Number and a Picture Stamp.
The Lens Table comprises a number of fields including Lens ID, which is associated with the Lens ID field of the Take Table, Manufacturer, Model, Serial Number, Contact Name, Contact Address and Contact Phone Number and a Picture Stamp.
Each Take Annotation Table comprises a number of fields including Take Annotation ID, Basic UMID, Timecode, Shutter Speed, Iris, Zoom, Gamma, Shot Marker ID, Filter Wheel, Detail and Gain. Associated with each Take Annotation Table is a Shot Marker Table.
The Shot Marker Table comprises a number of fields including Shot Marker ID, which is associated with the Shot Marker ID of the Take Annotation Table, and Description.
A-Box Figures 2 to 7 Acquisition Unit Embodiments of the present inventionrelate to audio and/or video generation apparatus which may be for example television cameras, video cameras or camcorders. An embodiment of the present invention will now be described with reference to figure 2 which provides a schematic block diagram of a video camera which is arranged to conununicate to a personal digital assistant (PDA). A PDA is an example of a data processor which may be arranged in operation to generate metadata in accordance with a user's requirements. The tenn personal digital assistant is known to those acquainted with the technical field of consumer electronics as a portable or hand held personal organiser or data processor which include an alpha numeric key pad and a hand writing interface.
In figure 2 a video camera 10 1 is shown to comprise a camera body 102 which is arranged to receive light from an image source falling within a field of view of an
16 imaging arrangement 104 which may include one or more imaging lenses (not shown). The camera also includes a view finder 106 and an operating control unit 108 from which a user can control the recording of signals representative of the images formed within the field of view of the camera. The camera 10 1 also includes a microphone 110 which rnay be a plurality of microphones arranged to record sound in stereo. Also shown in figure 2 is a hand-held PDA 112 which has a screen 114 and an alphanumeric key pad 116 which also includes a portion to allow the user to write characters recognised by the PDA. The PDA 112 is arranged to be connected to the video camera 10 1 via an interface 118. The interface 118 is arranged in accordance with a predetermined standard format such as, for example an RS23)2 or the like. The interface 118 may also be effected using infra-red signals, whereby the interface 118 is a wireless communications link. The interface 118 provides a facility for communicating information with the video camera 10 1. The function and purpose of the PDA 112 will be explained in more detail shortly. However in general the PDA 112 provides a facility for sending and receiving metadata generated using the PDA 112 and which can be recorded with the audio and video signals detected and captured by the video camera 1. A better understanding of the operation of the video camera 101 in combination with the PDA 112 may be gathered from figure 3 which shows a more detailed representation of the body 102 of the video camera which is shown in flaure 2 and in which conunon parts have the same numerical designations.
In figure 3 the camera body 102 is shown to comprise a tape drive 122 having read/write heads 124 operatively associated with a magnetic recording tape 126. Also shown in figure 3 the camera body includes a metadata generation processor 128 coupled to the tape drive 122 via a connecting channel 130. Also connected to the metadata generation processor 128 is a data store 132, a clock 136 and three sensors 13 8, 140, 142. The interface unit 118 sends and receives data also shown in figure 3 via a wireless channel 119. Correspondingly two connecting channels for receiving and transmitting data respectively, connect the interface unit 118 to the metadata generation processor 128 via corresponding connecting channels 148 and 150. The metadata generation processor is also shown to receive via a connecting channel 151 the audio/video signals generated by the camera. The audio/video signals are also fed to the tape drive 122 to be recorded on to the tape 126.
The video camera 110 shown in figure 2 operates to record visual information falling within the field of view of the lens arrangement 104 onto a recording medium.
17 The visual information is converted by the camera into video signals. In combination, the visual images are recorded as video signals with accompanying sound which is detected by the microphone 10 1 and arranged to be recorded as audio signals on the recording medium with the video signals. As shown in figure 3, the recording medium is a magnetic tape 126 which is arranged to record the audio and video signals onto the recording tape 126 by the read/write heads 124. The arrangement by which the video signals and the audio signals are recorded by the read/write heads 124 onto the magnetic tape 126 is not shown in figure 3 and will not be further described as this does not provide any greater illustration of the example embodiment of the present invention. However once a user has captured visual images and recorded these images using the magnetic tape 126 as with the accompanying audio signals, metadata describing the content of the audio/video signals may be input using, the PDA 112. As will be explained shortly this metadata can be information that identifies the audio/video signals in association with a pre-planned event, such as a 'take'. As shown in figure 3) the interface unit 118 provides a facility whereby the metadata added by the user using the PDA 112 may be received within the camera body 102. Data signals may be received via the wireless channel 119 at the interface unit 118. The interface unit 118 serves to convert these signals into a form in which they can be processed by the acquisition processor 128 which receives these data signals via the connecting channels 148, 150.
Metadata is generated automatically by the metadata generation processor 128 in association with the audio/video signals which are received via the connecting channel 15 1. In the example embodiment illustrated in figure 3, the metadata generation processor 128 operates to generate time codes with reference to the clock 136, and to write these time codes on to the tape 126 in a linear recording track provided for this purpose. The time codes are formed by the metadata generation processor 128 from the clock 136. Furthermore, the metadata generation processor 128 forms other metadata automatically such as a UMID, which identifies uniquely the audio/video signals. The metadata generation processor may operate in combination with the tape driver 124, to write the UMID on to the tape with the audio/video signals.
In an alternative embodiment, the UMID, as well as other metadata may be stored in the data store 132 and communicated separately from the tape 126. In this 18 case, a tape ID is generated by the metadata generation processor 128 and written on to the tape 126, to identify the tape 126 from other tapes.
In order to generate the UMID, and other metadata identifying the contents of the audio/video signals, the metadata generation processor 128 is arranged in operation to receive signals from other sensor 138, 140, 142, as well as the clock 136. The metadata generation processor therefore operates to co-ordinate these signals and provides the metadata generation processor with metadata such as the aperture setting of the camera lens 104, the shutter speed and a signal received via the control unit 108 to indicate that the visual images captured are a "good shot". These signals and data are generated by the sensors 138, 140, 142 and received at the metadata generation processor 128. The metadata generation processor in the example embodiment is arranged to produce syntactic metadata which provides operating parameters which are used by the camera in generating the video signals. Furthermore the metadata generation processor 128 monitors the status of the camcorder 101, and in particular whether audio/video signals are being recorded by the tape drive 124. When RECORD START is detected the IN POINT time code is captured and a UMID is generated in correspondence with the IN POINT time code. Furthermore in some embodiments an extended UMID is generated, in which case the metadata generation processor is arranged to receive spatial co-ordinates which are representative of the location at which the audio/video signals are acquired. The spatial co-ordinates may be generated by a receiver which operates in accordance with the Global Positioning System (GPS). The receiver may be external to the camera, or may be embodied within the camera body 102.
When RECORD START is detected, the OUT POINT time code is captured by the metadata generation processor 128. As explained above, it is possible to generate a "good shof' marker. The "good shof' marker is generated during the recording process, and detected by the metadata generation processor. The "good shof' marker is then either stored on the tape, or within the data store 132, with the corresponding IN POINT and OUT POINT time codes.
As already indicated above, the PDA 112 is used to facilitate identification of the audio/video material generated by the camera. To this end, the PDA is arranged to associate this audio/video material with pre-planned events such as scenes, shots or takes. The camera and PDA shown in figures 2 and 3 form part of an integrated system for planning, acquiring, editing an audio/video production. During a planning 19 phase, the scenes which are required in order to produce an audio/video production are identified. Furthermore for each scene a number of shots are identified which are required in order to establish the scene. Within each shot, a number of takes may be generated and from these takes a selected number may be used to form the shot for the final edit. The planning information in this form is therefore identified at a planning stage. Data representing or identifying each of the planned scenes and shots is therefore loaded into the PDA 112 along with notes which will assist the director when the audio/video material is captured. An example of such data is shown in the table below.
A/V Production News story: BMW disposes of Rover Scene ID: 900015689 Outside Longbridge Shot 5000000199 Longbridge BMW Sign Shot 5000000200 Workers Leaving shift Shot 5000000201 Workers in car park Scene ID: 900015690 BMW HQ Munich Shot 5000000202 Press conference Shot 5000000203' Outside BMW building Scene ID: 900015691 Interview with minister Shot 5000000204 Interview In the first column of the table below the event which will be captured by the camera and for which audio/video material will be generated is shown. Each of the events which is defined in a hierarchy is provided with an identification number. Correspondingly, in the second column notes are provided in order to direct or remind the director of the content of the planned shot or scene. For example, in the first row the audio/video production is identified as being a news story, reporting the disposal of Rover by BMW. In the extract of the planning information shown in the table below, there are three scenes, each of which is provided with a unique identification number. Each of these scenes are "Outside Long Bridge", "BMW HQ Munich" and "Interview with Minister". Correspondingly for each scene a number of shots are identified and these are shown below each of the scenes with a unique shot identification number. Notes corresponding to the content of each of these shots are also entered in the second column. So, for example, for the first scene "Outside Long Bridge", three shots are identified which are "Long Bridge BMW", "Workers leaving shift" and "Workers in car park". With this information loaded onto the PDA, the director or indeed a single camera man may take the PDA out to the place where the new story is to be shot, so that the planned audio/video material can be gathered. An. illustration of the form of the PDA with the graphical user interface displaying this information is shown in figure 4.
As indicated in figure 2, the PDA 112 is arranged to communicate data to the camera 111. To this end the metadata generation processor 128 is arranged to communicate data with the PDA 112 via the interface 118. The interface 118 maybe for example an infra-red link 119 providing wireless communications in accordance with a known standard. The PDA and the parts of the camera associated with generating metadata which are shown in figure 3 are shown in more detail in figure 5.
In figure 5 the parts of the camera which are associated with generating metadata and communicating with the PDA 112 are shown in a separate acquisition unit 152. However it will be appreciated that the acquisition unit 152 could also be embodied within the camera 102. The acquisition unit 152 comprises the metadata generation processor 128, and the data store 132. The acquisition processor 152 also includes the clock 136 and the sensors 138, 140, 142 although for clarity these are not shown in figure 5. Alternatively, some or all of these features which are shown in figure 2 will be embodied within the camera 102 and the signals which are required to define the metadata such as the time codes and the audio/video signals themselves may be communicated via a communications link 153 which is coupled to an interface port 154. The metadata generation processor 128 is therefore provided with access to the time codes and the audio/video material as well as other parameters used in generating the audio/video material. Signals representing the time codes end parameters as well as the audio/video signals are received ftom the interface port 154 via the interface channel 156. The acquisition unit 152 is also provided with a screen (not shown) which is driven by a screen driver 158. Also shown in figure 6 the acquisition unit is provided with a communications processor 160 which is coupled to the metadata generation processor 128 via a connecting channel 162. Communications is effected by the communications processor 160 via a radio frequency communications channel using the antennae 164. A pictorial representation of the acquisition unit 152 is shown in figure 5.
21 The PDA 112 is also shown in figure 64. The PDA 112 is correspondingly provided with an infta-red communications port 165 for communicating data to and from the acquisition unit 152 via an infra-red link 119. A data processor 166 within the PDA 112 is arranged to communicate data to and from the infra-red port 165 via a connecting channel 166. The PDA 112 is also provided with a data store 167 and a screen driver 168 which are connected to the data processor 166.
The pictorial representation of the PDA 112 shown in figure 4 and the acquisition unit shown in figure 5 provide an illustration of an example embodiment of the present invention. A schematic diagram illustrating the arrangement and connection of the PDA 112 and the acquisition unit 152 is shown in figure 7. In the example shown in figure 7 the acquisition unit 152 is mounted on the back of a camera 101 and coupled to the camera via a six pin remote connector and to a connecting channel conveying the external signal representative of the time code recorded onto the recording tape. Thus, the six pin remote connector and the time code indicated as arrow lines form the communications channel 153 shown in figure 6. The interface port 154 is shown in figure 7 to be a VA to DNI conversion comprising an RM-P9/LTC to RS422 converter 154. RM-P9 is a camera remote control protocol, whereas LTC is Linear Time Code in the form of an analogue signal. This is arranged to communicate with a RS422 to RS232 converter 154" via a connecting channel which forms part of the interface port 154. The converter 154,' then communicates with the metadata generation processor 128 via the connecting channel 156 which operates in accordance with the RS 232 standard.
Returning to figure 6, the PDA 112 which has been loaded with the preplanned production information is arranged to communicate the current scene and shot for which audio/video material is to be generated by communicating the next shot ID number via the infra-red link 119. The preplanned information may also have been communicated to the acquisition unit 152 and stored in the data store 132 via a separate link or via the infra-red communication link 119. However in effect the acquisition unit 152 is directed to generate metadata in association with the scene or shot ID number which is currently being taken. After receiving the information of the current shot the camera 102 is then operated to make a "take of the shot". The audio/video material of the take is recorded onto the recording tape 126 with corresponding time codes. These time codes are received along with the audio/video 22 material via the interface port 154 at the metadata generation processor 128. The metadata generation processor 128 having been informed of the current pre-planned shot now being taken logs the time codes for each take of the shot. The metadata generation processor therefore logs the IN and OUT time codes of each take and stores these in the data store 132.
The information generated and logged by the metadata generation processor 128 is shown in the table below. In the first column the scene and shot are identified with the corresponding ID numbers, and for each shot several takes are made by the camera operator which are indicated in a hierarchical fashion. Thus, having received information from the PDA 112 of the current shot, each take made by the camera operator is logged by the metadata generation processor 128 and the IN and OUT points for this take are shown in the second and third columns and stored in the data store 132. This information may also be displayed on the screen of the acquisition unit 152 as shown in figure 5. Furthermore, the metadata generation processor 128 as already explained generates the UMID for each take for the audio/video material generated during the take. The UMID for each take forms the fourth column of the table. Additionally, in some embodiments, to provide a unique identification of the tape once which the material is recorded, a tape identification is generated and associated with the metadata. The tape identification may be written on to the tape, or stored on a random access memory chip which is embodied within the video tape cassette body. This random access memory chip is known as a TELEFILE (RTM) system which provides a facility for reading the tape ID number remotely. The tape ID is written onto the magnetic tape 126 to uniquely identify this tape. In preferred embodiments the TELEFILE (RTM) system is provided with a unique number which manufactured as part of the memory and so can be used as the tape ID number. In other embodiments the TELEFILE (RTM) system provides automatically the IN/OUT time codes of the recorded audio/video material items.
In one embodiment the information shown in the table below is arranged to be recorded onto the magnetic tape in a separate recording channel. However, in other embodiments the metadata shown in the table is communicated separately from the tape 126 using either the communications processor 160 or the infra-red link 119. The metadata maybe received by the PDA 112 for analysis and may be flirther communicated by the PDA.
23 Scene ID: 900015689 Tape ID: 00001 UMID:
Shot 5000000199 Take 1 IN: 00:03:45:29 OUT: 00:04:21:05 060C23B340 Take 2 IN: 00:04:21:20 OUT: 00:04:28:15 060C23B340 Take 3 IN: 00:04:28:20 OUT: 00:05:44:05 060C23B340 Shot 5000000200 Take 1 IN: 00:05:44:10 OUT: 00:08:22:05 060C23B340 Take 2 IN: 00:08:22:10 OUT: 00:08:23:05 060C23B340 The communications processor 160 may be arranged in operation to transmit the metadata generated by the metadata generation processor 128 via a wireless communications link. The metadata maybe received via the wireless communications link by a remotely located studio which can then acquire the metadata and process this metadata ahead of the audio/video material recorded onto the magnetic tape 126. This provides an advantage in improving the rate at which the audio/video production may be generated during the post production phase in which the material is edited.
A further advantageous feature provided by embodiments of the present invention is an arrangement in which a picture stamp is generated at certain temporal positions within the recorded audio/video signals. A picture stamp is known to those skilled in the art as being a digital representation of an image and in the present example embodiment is generated from the moving video material generated by the camera. The picture stamp may be of lower quality in order to reduce an amount of data required to represent the image from the video signals. Therefore the picture stamp may be compression encoded which may result in a reduction in quality. However a picture stamp provides a visual indication of the content of the audio/video material and therefore is a valuable item of metadata. Thus, the picture stamp may for example be generated at the IN and OUT time codes of a particular take. Thus, the picture stamps may be associated with the metadata generated by the metadata generation processor 128 and stored in the data store 132. The picture stamps are therefore associated with items of metadata such as, for example, the time codes which identify the place on the tape where the irnage represented by the picture stamp is recorded. The picture stamps may be generated with the "Good Shof' markers.
24 The picture stamps are generated by the metadata generation processor 128 ftom. the audio/video signals received via the communications link 153. The metadata generation processor therefore operates to effect a data sampling and compression encoding process in order to produce the picture stamps. Once the picture stamps have been generated they can be used for several purposes. They may be stored in a data file and communicated separately from the tape 126, or they may be stored on the tape 126 in compressed form in a separate recording channel. Alternatively in preferred embodiments picture stamps may be communicated using the communications processor 160 to the remotely located studio where a producer may analysis the picture stamps. This provides the producer with an indication as to whether the audio/video material generated by the camera operator is in accordance with what is required.
In a yet ftirther embodiment, the picture stamps are communicated to the PDA 112 and displayed on the PDA screen. This may be effected via the infra-red port 119 or the PDA may be provided with a further wireless link which can communicate with the communications processor 160. In this way a director having the hand held PDA 112 is provided with an indication of the current audio/video content generated by the camera. This provides an immediate indication of the artist and aesthetic quality of the audio/video material currently being generated. As already explained the picture stamps are compression encoded so that they may be rapidly communicated to the PDA.
A further advantage of the acquisition unit 152 shown in figure 6 is that the editing process is made more efficient by pro-,liding the editor at a remotely located studio with an indication of the content of the audio/video material in advance of receiving that material. This is because the picture stamps are communication with the metadata via a wireless link so that the editor is provided with an indication of the content of the audio/video material in advance of receiving the audio/video material itself. In this way the bandwidth of the audio/video material can remain high with a correspondingly high quality whilst the metadata and picture stamps are at a relatively low band width providing relatively low quality information. As a result of the low band width the metadata and picture stamps may be communicated via a wireless link on a considerably lower band width channel. This facilitates rapid communication of the metadata describing content of the audio/video material.
The picture stamps generated by the metadata generation processor 128 can be at any point during the recorded audio/video material. In one embodiment the picture stamps are generated at the IN and OUT points of each take. However in other embodiments of the present invention as an activity processor 170 is arranged. to detect relative activity within the video material. This is effected by performing a process in which a histogram of the colour components of the images represented by the video signal is compiled and the rate of change of the colour components determined and changes in these colour components used to indicate activity within the image. Alternatively or in addition, motion vectors within the image are used to indicate activity. The activity processor 176 then operates to generate a signal indicative of the relative activity within the video material. The metadata generation processor 128 then operates in response to the activity signal to generate picture stamps such more picture stamps are generated for greater activity within the images represented by the video signals.
In an alternative embodiment of the present invention the activity processor is arranged to receive the audio signals via the connecting channel 172 and to recognise speech within the audio signals. The activity processor 170 then generates content data representative of the content of this speech as text. The text data is then communicated to the data processor 128 which may be stored in the data store 1 -3)2 or communicated with other metadata via the communications processor 160 in a similar way to that already explained for the picture stamps.
As will be appreciated although the example embodiment of the present invention uses a video tape as the recording medium for storing the audio/video signals, it will be understood that alternative recording medium such as magnetic disks and random access memories may also be used.
26

Claims (1)

1 ' A method of producing video and/or audio material comprising the steps of:
using video and/or audio equipment to produce video and/or audio material; producing a material identifier identifying the material; producing at least an item identifier identifying an item of equipment used in C> the production of the material; associating the item identifier with the material identifier; and transferring at least the associated identifiers to a database.
2. A method according to claim 1, wherein the material identifier is associated with the material.
3. A method according to claim 1 or 2, wherein the material is recorded on a recording medium and the material identifier is also recorded on the medium.
4. A method according to claim 1 or 2, wherein the material identifier is in the material.
5. A method according to claim 1, 2, 3 or 4, wherein the material identifier indirectly identifies the material.
6. A method according to claim 5 when dependent on claim 3, wherein the material identifier identifies the example of the recording medium on which the material is recorded, and the material identifier is associated in the database with a further identifier which uniquely identifies the material.
7. A method according to claim 1, 2, 3 or 4, wherein the material identifier uniquely identifies the material.
27 8. A method according to any preceding claim, further comprising generating usage data relating to the amount of usage of the item of equipment, associating the usage data with the item identifier, and transferring the associated 0 usage data and item identifier to the database.
9. A method according to claim 8, further comprising the step of emitting a message when the amount of usage reaches a preset value.
10. A method according to any preceding claim, comprising the step of generating position data defining the location of the item of equipment, associating the position data with the item identifier and transferring the associated item identifier and position data to the data base.
11. A method according to claim 10, wherein the position data is data derived from a Global Positioning System.
12. A method according to claim 8, 9, 10 or 11, comprising the step of interrogating the database by reference to the item identifier to find data associated with the item identifier.
13. A method according to claim 1, comprising the steps of. producing other metadata relating to the material; associating the other metadata with the material identifier; and transferring the associated other metadata and the said material identifier to the database.
14. A method according to any preceding claim, wherein the material identifier is produced automatically by the equipment used to produce the material.
15. A method according to any preceding claim, wherein the said item identifier is produced automatically.
16. A method according to any one of claims 1 to 14, wherein the item identifier is produced manually and is automatically associated with the material identifier.
28 17. A method according to any preceding claim, wherein the database remote from the equipment and the associated identifiers, or associated identifiers and data, are transferred to the database via a communications channel.
18. A system for producing video and/or audio material, comprising: video and/or audio equipment for producing video and/or audio material; a generator for producing a material identifier identifying the material; a first processor for associating an item identifier identifying an item of equipment used in the production of the recorded material with the material identifi a database processor containing a database; and a data link for transferring the associated identifiers to the database.
er; 19. A system according to claim 18, wherein the material identifier is associated with the material.
20. A system according to claim 18 or 19, wherein the material is recorded on a recording medium and the material identifier is also recorded on the medium.
21. A system according to claim 18 or 19 or 20, wherein the material identifier is in the material.
22. A system according to claim 18, 19, 20 or 21, wherein the material identifier indirectly identifies the material.
23. A system according to claim 22 when dependent on claim 20 or 21, wherein the material identifier identifies the example of the recording medium on which the material is recorded, and the material identifier is associated in the database with a further identifier which uniquely identifies the material.
24. A system according to claim 18, 19, 20 or 21, wherein the material identifier uniquely identifies the material.
29 25. A system according to any one of claims 18 to 24, comprising a ftirther generator for generating usage data relating to the amount of usage of the item of t t> equipment, wherein the first processor is operable to associating the usage data with the item identifier, and the data link is operable to transfer the associated usage data and item identifier to the database processor.
26. A system according to claim 25, wherein the database is operable to generate a message when the amount of usage reaches a preset value.
27. A system according to any one of claims 18 to 26, comprising a position data generator for generating position data defining the location of the item of equipment, the first processor being operable to associate the position data with the item identifier and the data link being operable to transfer the associated item identifier and position data to the data base processor.
A system according to claim 10, wherein the position data generator comprises a Global Positioning System terminal.
29. A system according to any one of claims 18 to 29, wherein the database processor is operable to interrogate the database by reference to the item identifier to find data associated with the item identifier.
30. A system according to any one of claims 18 to 29, comprising a metadata producer for producing other metadata relating to the material; the first processor being operable to associate the other metadata with the material identifier; and the data link being operable to transfer the associated other metadata and the said material identifier to the database processor.
31. A system according to any one of claims 18 to 30, comprising an item identifier generator for generating the item identifier.
32. A system according to claim 3 1, wherein the said item identifier generates the identifier automatically.
33. A system according to any one of claims 18 to 30, wherein the item identifier is produced manually and the first processor automatically associates it with the material identifier.
GB0008416A 2000-04-05 2000-04-05 Producing audio/visual material and identifying equipment used in its production Withdrawn GB2361094A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB0008416A GB2361094A (en) 2000-04-05 2000-04-05 Producing audio/visual material and identifying equipment used in its production

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
GB0008416A GB2361094A (en) 2000-04-05 2000-04-05 Producing audio/visual material and identifying equipment used in its production

Publications (2)

Publication Number Publication Date
GB0008416D0 GB0008416D0 (en) 2000-05-24
GB2361094A true GB2361094A (en) 2001-10-10

Family

ID=9889299

Family Applications (1)

Application Number Title Priority Date Filing Date
GB0008416A Withdrawn GB2361094A (en) 2000-04-05 2000-04-05 Producing audio/visual material and identifying equipment used in its production

Country Status (1)

Country Link
GB (1) GB2361094A (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2330977A (en) * 1997-06-09 1999-05-05 Matsushita Electric Ind Co Ltd Video recorder/reproducer

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2330977A (en) * 1997-06-09 1999-05-05 Matsushita Electric Ind Co Ltd Video recorder/reproducer

Also Published As

Publication number Publication date
GB0008416D0 (en) 2000-05-24

Similar Documents

Publication Publication Date Title
US10200767B2 (en) Audio and/or video generation apparatus and method of generating audio and/or video signals
USRE41939E1 (en) Audio/video reproducing apparatus and method
EP1188164B1 (en) Identifying and processing of audio and/or video material
US9311962B2 (en) Audio and/or video generation apparatus and method of generating audio and/or video signals
US8457467B2 (en) Apparatus and associated methodology of recording and accessing metadata via unique identifying indicia
GB2356080A (en) Generation system for audio, video or a combination thereof where metadata is generated and stored or recorded with the audio/video signal
GB2361128A (en) Video and/or audio processing apparatus
GB2361097A (en) A system for generating audio/video productions
GB2361127A (en) Audio/video reproduction via a communications network
GB2361094A (en) Producing audio/visual material and identifying equipment used in its production
GB2361096A (en) Metadata generation in audio or video apparatus
GB2361098A (en) Editing system and method using metadata
GB2361090A (en) Generating sample images to assist video editing
GB2361130A (en) Identifying video and/or audio material

Legal Events

Date Code Title Description
WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)