WO2005099413A2 - Appareil, systeme et procede de filtrage de parties indesirables d'une presentation multimedia - Google Patents

Appareil, systeme et procede de filtrage de parties indesirables d'une presentation multimedia Download PDF

Info

Publication number
WO2005099413A2
WO2005099413A2 PCT/US2005/012448 US2005012448W WO2005099413A2 WO 2005099413 A2 WO2005099413 A2 WO 2005099413A2 US 2005012448 W US2005012448 W US 2005012448W WO 2005099413 A2 WO2005099413 A2 WO 2005099413A2
Authority
WO
WIPO (PCT)
Prior art keywords
filter
multimedia
content
memory
dvd
Prior art date
Application number
PCT/US2005/012448
Other languages
English (en)
Other versions
WO2005099413A3 (fr
Inventor
Matthew Thomas Jarman
Jason Seeley
Original Assignee
Clearplay, Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Clearplay, Inc. filed Critical Clearplay, Inc.
Publication of WO2005099413A2 publication Critical patent/WO2005099413A2/fr
Publication of WO2005099413A3 publication Critical patent/WO2005099413A3/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/162Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing
    • H04N7/163Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing by receiver means only
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/4508Management of client data or end-user data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/426Internal components of the client ; Characteristics thereof
    • H04N21/42646Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/432Content retrieval operation from a local storage medium, e.g. hard-disk
    • H04N21/4325Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • H04N21/4545Input to filtering algorithms, e.g. filtering a region of the image
    • H04N21/45457Input to filtering algorithms, e.g. filtering a region of the image applied to a time segment

Definitions

  • aspects of the present invention involve a system, method, apparatus and file formats related to filtering portions of a multimedia presentation.
  • JO content such as television programming, computer software, video games, and music.
  • Direct access media such as DVD
  • direct access media allows for accessing any arbitrary portion of the multimedia content in roughly the same amount of time as any other arbitrary portion of the multimedia content.
  • Direct access media allows for the creation and distribution of multiple versions of multimedia content, including versions that may be suitable to most ages, and storing the versions on a single medium.
  • the decoding process creates various continuous multimedia streams by identifying, selecting, retrieving and transmitting content segments from a number of available segments stored on the content source.
  • a high-level description of the basic components found in a system for presenting multimedia content may be useful.
  • such systems include a multimedia source, a decoder, and an output device.
  • the decoder is a translator between the format used to store or transmit the multimedia content and the format used for intermediate processing and ultimately presenting the multimedia content at the output device.
  • multimedia content may be encrypted to prevent piracy and compressed to conserve storage space or bandwidth.
  • the multimedia content Prior to presentation, the multimedia content must be decrypted and/or uncompressed, operations usually performed by the decoder.
  • the prior art teaches creation and distribution of multiple versions of a direct access multimedia work on a single storage medium by breaking the multimedia content into various segments and including alternate interchangeable segments where appropriate.
  • Each individually accessible segment is rated and labeled based on th& content it contains, considering such factors as subject matter, context, and explicitness.
  • One or more indexes of the segments are created for presenting each of the multiple versions of the multimedia content. For example, one index may reference segments that would be considered a "PG" version of the multimedia whereas another index may reference segments that would be considered an "R" version of the content.
  • the segments themselves or a single index may include a rating that is compared to a rating selected by a user.
  • segment indexing provides for multiple versions of a multimedia work on a single storage medium.
  • Use of storage space can be optimized because segments common to the multiple versions need only be stored once. Consumers may be given the option of setting their own level of tolerance for specific subject matter and the different multimedia versions may contain alternate segments with varying levels of explicitness.
  • segment indexing on the content source also enables the seamless playback of selected segments (i. e., without gaps and pauses) when used in conjunction with a buffer. Seamless playback is achieved by providing the segment index on the content source, thus governing the selection and ordering of the interchangeable segments prior to the data entering the buffer.
  • a buffer compensates for latency that may be experienced in reading from different physical areas of direct access media. While read mechanisms are moved from one disc location to another, no reading of the requested content from the direct access media occurs. This is a problem because, as a general rule, the playback rate for multimedia content exceeds the access rate by a fairly significant margin. For example, a playback rate of 30 frames per second is common for multimedia content. Therefore, a rando access must take less than 1/30th of a second (approximately 33 milliseconds) or the ra ndom access will result in a pause during playback while the reading mechanism moves to the next start point.
  • a 16x DVD drive for a personal computer has an average access rate of approximately 95 milliseconds, nearly three times the 33 milliseconds allowed for seamless playback. Moreover, according to a standard of the National Television Standards Committee ("NTSC"), only 5 to 6 milliseconds are allowed between painting the last pixel of one frame and painting the first pixel of the next frame.
  • NTSC National Television Standards Committee
  • Those of skill in the art will recognize that the above calculations are exemplary of the time constraints involved in reading multimedia content from direct access media for output to a PC or television, even though no time is allotted to decoding the multimedia content after it has been read, time that would need to be added to the access time for more precise latency calculations. Once access occurs, DVD drives are capable of reading multimedia content from a
  • the DVD specification teaches reading multimedia content into a track buffer.
  • the track buffer size and amount of multimedia content that must be read into the track buffer depend on several factors, including the factors described above, such as access time, decoding time, playback rate, etc.
  • a segment index as taught in the prior art, with corresponding navigation commands, identifies and orders the content segments to be read into the track buffer, enabling seamless playback of multiple version of the multimedia content.
  • segment indexes that are external to the content source are unable to completely control the navigation commands within the initial segment identification/selection/retrieval process since external indexes can interact with position codes only available at the end of the decoding process.
  • segments from separate versions of multimedia content may be interlaced. This allows for essentially sequential reading of the media, withi unwanted segments being read and discarded or skipped.
  • the skips represent relatively small movements of the read mechanism. Generally, small movements involve a much shorter access time than large movements and therefore introduce only minimal latency. Nevertheless, the prior art for including multiple versions of a multimedia work on a single direct access media suffers from several practical limitations that prevent it from wide-spread use.
  • One significant problem is that content producers must be willing to create and broadly distribute multiple versions of the multimedia work and accommodate any additional production efforts in organizing and labeling the content segments, including interchangeable segments, for use with the segment indexes or maps.
  • the indexes, in combination with the corresponding segments define a work and are stored directly on the source media at the time the media is produced.
  • the prior art offers a tool for authoring multiple versions of a multimedia work, that tool is not useful in and of itself to consumers.
  • a further problem in the prior art is that existing encoding technologies must be licensed in order to integrate segment indexes on a direct access storage medium and decoding technologies must be licensed to create a decoder that uses the segment indexes on a multimedia work to seamlessly playback multiple versions stored on the direct access medium.
  • the Motion Pictures Entertainment Group controls the compression technology for encoding and decoding multimedia files.
  • MPEG Motion Pictures Entertainment Group
  • producers of multimedia content generally want to prevent unauthorized copies o-f their multimedia work, they also employ copy protection technologies.
  • the most common copy protection technologies for DVD media are controlled by the DVD Copy Control Association (“DVD CCA”), which controls the licensing of their Content Scramble System technology (“CSS”).
  • DVD CCA DVD Copy Control Association
  • CSS Content Scramble System technology
  • Decoder developers license the relevant MPEG and CSS technology under fairly strict agreements that dictate how the technology may be used.
  • the time and cost associated with licensing existing compression and copy protection technologies or developing proprietary compression and copy protection technologies may be significant costs, prohibitive to the wide-spread use of the prior art's segment indexing for providing multiple versions of a multimedia work on a single direct access storage medium.
  • teachings of the prior art do not provide a solution for filtering direct access multimedia content that has already been duplicated and distributed without regard to presenting the content in a manner that is more suitable for most ages.
  • over 40,000 multimedia titles have been released on DVD without using the multiple version technology of the prior art to provide customers the ability to view and hear alternate versions of the content in a manner that is more suitable for most ages.
  • the prior art also has taught that audio portions of multimedia content may be identified and filtered during the decoding process by examining the closed caption information for the audio stream and muting the volume during segments of the stream that contain words matching with a predetermined set of words that are considered unsuitable for most ages.
  • This art is limited in its application since it cannot identify and filter video segments and since it can only function with audio streams that contain closed captioning information.
  • filtering audio content based on closed captioning information is imprecise due to poor synchronization between closed captioning information and the corresponding audio content.
  • aspects of the invention involve a method of filtering portions of a multimedia content presentation, the method comprising accessing at least one filter file defining a filter start indicator and a filter action; reading digital multimedia information from a memory media, the multimedia information including a location reference; comparing the location reference of the multimedia information with the filter start indicator; and responsive to the comparing operation, executing a filtering action if there is match between the location reference of the multimedia information and the filter start indicator of the at least one filterable portion of the multimedia content.
  • Figure 1 illustrates an exemplary system that provides a suitable operating 5 environment for the present invention
  • FIG. 2 is high-level block diagram showing the basic components of a system embodying the present invention
  • Figures 3A, 3B, and 3C are block diagrams of three systems that provide greater detail for the basic components shown in Figure 2;
  • L0 Figures 4A, 5A, and 7, are flowcharts depicting exemplary methods for filtering multimedia content according to the present invention.
  • Figures 4B and 5B illustrate navigation objects in relation to mocked-up position codes for multimedia content
  • Figure 6 is a flowchart portraying a method used in customizing the filtering of L5 multimedia content
  • Figures 8A and 8B are flowcharts illustrating a method conforming to aspects of the present invention
  • Figure 9 is a representative block diagram of a menu arrangement conforming to aspects of the present invention.
  • tO Figures 10A-10C are representative block diagrams illustrating a filter processing action conforming to aspects of the present invention.
  • Figure 1 1 is a representative block diagram of a menu arrangement conforming to aspects of the present invention
  • Figure 12 is a diagram illustrating aspects of a skip type filtering action conforming to 15 aspects of the present invention.
  • Figure 13 is a file format diagram for a skip type filtering action
  • Figure 14 is a diagram illustrating aspects of a mute type filtering action conforming to aspects of the present invention.
  • Figure 15 is a file format diagram for a skip type filtering action; and Figures 16-23 are file formats for indexing and filter table identification packets, conforming to aspects of the present invention.
  • Embodiments of the Invention extends to methods, systems, and computer program products for automatically identifying and filtering portions of multimedia content during the decoding process.
  • the embodiments of the present invention may comprise a special purpose or general purpose computer including various computer hardware, a television system, an audio system, and/or combinations of the foregoing. These embodiments are discussed in greater detail below. However, in all cases, the described embodiments should be viewed a exemplary of the present invention rather than as limiting it's scope.
  • Embodiments within the scope of the present invention also include computer-readable media for carrying or having computer-executable instructions or data structures stored thereon. Such computer-readable media may be any available media that can be accessed by a general purpose or special purpose computer.
  • such computer-readable media can comprise RAM, ROM, EEPROM, DVD, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
  • ROM read-only memory
  • EEPROM electrically erasable programmable read-only memory
  • DVD compact disc-read only memory
  • CD-ROM or other optical disk storage magnetic disk storage or other magnetic storage devices
  • any other medium which can be used to carry or store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
  • Computer-executable instructions comprise, for example, instructions and data which cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.
  • Figure 1 and the following discussion are intended to provide a brief, general description of a suitable computing environment in which the invention may be implemented.
  • program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
  • Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein.
  • an exemplary system for implementing the invention includes a general purpose computing device in the form of a conventional computer 20, including a processing unit 21, a system memory 22, and a system bus 23 that couples various system components including the system memory 22 to the processing unit 21.
  • the system bus 23 may be any of several types of bus structures including a memory bus or memory controller, a peripheral bus, and a local bus using any of a variety of bus architectures.
  • the system memory includes read only memory (ROM) 24 and random access memory (RAM) 25.
  • ROM read only memory
  • RAM random access memory
  • a basic input/output system (BIOS) 26 containing the basic routines that help transfer information between elements within the computer 20, such as during start-up, may be stored in ROM 24.
  • the computer 20 may also include a magnetic hard disk drive 27 for reading from and writing to a magnetic hard disk 39, a magnetic disk drive 28 for reading from or writing to a removable magnetic disk 29, and an optical disk drive 30 for reading from or writing to removable optical disk 31 such as a CD-ROM or other optical media.
  • the magnetic hard disk drive 27, magnetic disk drive 28, and optical disk drive 30 are connected to the system bus 23 by a hard disk drive interface 32, a magnetic disk drive-interface 33, and an optical drive interface 34, respectively.
  • the drives and their associated computer-readable media provide nonvolatile storage of computer-executable instructions, data structures, program modules and other data for the computer 20.
  • exemplary environment described herein employs a magnetic hard disk 39, a removable magnetic disk 29 and a removable optical disk 31
  • other types of computer readable media for storing data can be used, including magnetic cassettes, flash memory cards, digital video disks, Bernoulli cartridges, RAMs, ROMs, and the like.
  • Program code means comprising one or more program modules may be stored on the hard disk 39, magnetic disk 29, optical disk 31 , ROM 24 or RAM 25, including an operating system 35, one or more application programs 36, other program modules 37, and program data 38.
  • a user may enter commands and information into the computer 20 through keyboard 40, pointing device 42, or other input devices (not shown), such as a microphone, joy stick, game pad, satellite dish, scanner, or the like.
  • These and other input devices are often connected to the processing unit 21 through a serial port interface 46 coupled to system bus 23.
  • the input devices may be connected by other interfaces, such as a parallel port, a game port or a universal serial bus (USB).
  • a monitor 47 or another display device is also connected to system bus 23 via an interface, such as video adapter 48.
  • the computer 20 may operate in a networked environment using logical connections to one or more remote computers, such as remote computers 49a and 49b.
  • Remote computers 49a and 49b may each be another personal computer, a server, a router, a network PC, a peer device or other common network node, and typically include many or all of the elements described above relative to the computer 20, although only memory storage devices 50a and 50b and their associated application programs 36a and 36b have been illustrated in Figure 1.
  • the logical connections depicted in Figure 1 include a local area network (LAN) 51 and a wide area network (WAN) 52 that are presented here by way of example and not limitation.
  • LAN local area network
  • WAN wide area network
  • Such networking environments are commonplace in office-wide or enterprise-wide computer networks, intranets and the Internet.
  • the computer 20 When used in a LAN networking environment, the computer 20 is connected to the local network 51 through a network interface or adapter 53.
  • the computer 20 When used in a WAN networking environment, the computer 20 may include a modem 54, a wireless link, or other means for establishing communications over the wide area network 52, such as the Internet.
  • the modem 54 which may be internal or external, is connected to the system bus 23 via the serial port interface 46.
  • program modules depicted relative to the computer 20, or portions thereof may be stored in the remote memory storage device. It will be appreciated that the network connections shown are exemplary and other means of establishing communications over wide area network 52 may be used.
  • FIG. 2 a high-level block diagram identifying the basic components of a system for filtering multimedia content are shown.
  • the basic components include content source 230, decoders 250, navigator 210, and output device 270.
  • Content source 230 provides multimedia to decoder 250 for decoding, navigator 210 controls decoder 250 so that filtered content does not reach output device 270, and output device 270 plays the multimedia content it receives.
  • multimedia should be interpreted broadly to include audio content, video content, or both. The present invention does not require a particular content source 230.
  • Any data source that is capable of providing multimedia content such as a DVD, a CD, a memory, a hard disk, a removable disk, a tape cartridge, and virtually all other types of magnetic or optical media may operate as content source 230.
  • content source 230 includes read-only, read/write, and write-once varieties, whether stored in an analog or digital format. All necessary hardware and software for accessing these media types are also part of content source 230.
  • Content source 230 as described above provides an example of multimedia source means.
  • Multimedia source 230 generally provides encoded content.
  • Encoding represents a difference in the formats that are typically used for storing or transmitting multimedia content and the formats used for intermediate processing of the multimedia content.
  • Decoders 250 translate between the storage and intermediate formats. For example, stored MPEG content is both compressed and encrypted. Prior to being played at an output device, the stored MPEG content is decrypted and uncompressed by decoders 250. Decoders 250 may comprise hardware, software, or some combination of hardware and software. Due to the large amount of data involved in playing multimedia content, decoders 250 frequently have some mechanism for transferring data directly to output device 270. Decoders 250 are an exemplary embodiment of decoder means.
  • Output device 270 provides an example of output means for playing multimedia content and should be interpreted to include any device that is capable of playing multimedia content so that the content may be perceived.
  • output device 270 may include a video card, a video display, an audio card, and speakers.
  • output device 270 may be a television or audio system.
  • Television systems and audio systems cover a wide range of equipment.
  • a simple audio system may comprise little more than an amplifier and speakers.
  • a simple television system may be a conventional television that includes one or more speakers and a television screen. More sophisticated television and audio systems may include audio and video receivers that perform sophisticated processing of audio and video content to improve sound and picture quality.
  • Output device 270 may comprise combinations of computer, television, and audio systems.
  • home theaters represent a combination audio and television systems. These systems typically include multiple content sources, such as components for videotape, audiotape, DVD, CD, cable and satellite connections, etc. Audio and/or television systems also may be combined with computer systems. Therefore, output device 270 should be construed as including the foregoing audio, television, and computer systems operating either individually, or in some combination.
  • computer system whether for a consumer or operating as a server
  • television system, and audio system may identify a system's capabilities rather than its primary or ordinary use. These capabilities are not necessarily exclusive of one another. For example, a television playing music through its speakers is properly considered an audio system because it is capable of operating as an audio system.
  • Navigator 210 is software and/or hardware that control the decoders 250 by determining if the content being decoded needs to be filtered. Navigator 210 is one example of multimedia navigation means. It should be emphasized that content source 230, decoders 250, output device 270, and navigator 210 have been drawn separately only to aid in their description. Some embodiments may combine content source 230, decoders 250, and navigator 210 into a single set-top box for use with a television and/or audio system.
  • a computer system may combine portions of decoder 250 with output device 270 and portions of decoder 250 with content source 230.
  • the present invention imposes no requirement that these four components must exist separately from each other.
  • the corresponding multimedia source means, decoder means, output means, and multimedia navigation means also need not exist separately from each other and may be combined together as is appropriate for a given embodiment of the present invention.
  • content source 230, decoders 250, output device 270, and/or navigator 210 to be located remotely from each other and linked together with a communication link.
  • Figures 3A, 3B, and 3C are block diagrams of three exemplary systems that provide greater detail for the basic components shown in Figure 2.
  • Figure 3A includes navigator 310a, content source 330a, audio and video decoders 350a, and output device 370a, all located at consumer system 380a.
  • Content source 330a includes DVD 332a and DVD drive 334a.
  • the bi-directional arrow between content source 330a and audio and video decoders 350a indicates that content source 330 provides multimedia content to audio and video decoders 350a and that audio and video decoders 350a send commands to content source 330a when performing filtering operations.
  • Navigator 310a monitors decoders 350a by continuously updating the time code of the multimedia content being decoded.
  • Time codes are an example of positions used in identifying portions of multimedia content. In the case of time codes, positioning is based on an elapsed playing time from the start of the content. For other applications, positions may relate to physical quantities, such as the length of tape moving from one spool to another in a videotape or audiotape.
  • the present invention does not necessarily require any particular type of positioning for identifying portions of multimedia content.
  • the time code updates occur every 1/1 Oth of a second, but the present invention does not require any particular update interval.
  • the description of Figures 4B and 5B provides some insight regarding factors that should be considered in selecting an appropriate update interval.
  • Communication between Navigator 310a and audio and video decoders 350a occurs through a vendor independent interface 352a.
  • the vendor independent interface 352a allows navigator 310a to use the same commands for a number of different content sources.
  • Microsoft's ® DirectX ® is a set of application programming interfaces that provides a vendor independent interface for content sources 330a in computer systems running a variety of Microsoft operating systems.
  • Audio and video decoders 350a receive commands through vendor independent interface 352a and issue the proper commands for the specific content source 330a. Audio and video decoders 350a provide audio content and video content to output device 370a.
  • Output device 370a includes graphics adapter 374a, video display 372a, audio adaptor 376a, and speakers 378a.
  • Video display 372a may be any device capable of displaying video content, regardless of format, including a computer display device, a television screen, etc.
  • graphics adaptors and audio adaptors provide some decoding technology so that the amount of data moving between content source 330a and output device 370a is minimized.
  • Graphics adaptors and audio adaptors also provide additional processing for translating multimedia content from the intermediate processing format to a format more suitable for display and audio playback.
  • many graphics adaptors offer video acceleration technology to enhance display speeds by offloading processing tasks from other system components.
  • the actual transition between decoders 350a and output device 37Oa may be a somewhat fuzzy.
  • graphics adaptor 374a and audio adapter 376a perform decoding, portions of those adaptors may be properly construed as part of decoders 350a.
  • Navigator 310a includes navigation software 312a and object store 316a.
  • Bi-directional arrow 314a indicates the flow of data between navigation software 312a and object store 316a.
  • Object store 316a contains a plurality of navigation objects 320a. Within object store 316a, navigation objects may be stored as individual files that are specific to particular multimedia content, they may be stored in one or more common databases, or some other data management system may be used. The present invention does not impose any limitation on how navigation objects are stored in object store 316a.
  • Each navigation object 320a defines when (start 321 a and stop 323a) an filtering action (325a) should occur for a particular system (329a) and provides a description (327a) of why the navigation object was created.
  • Start and stop positions (321 a and 323a) are stored as time codes, in hours:minutes:seconds:frame format; actions may be either skip or mute (325a); the description is a text field (327a); and configuration is an identifier (329a) used to determine if navigation object 320a applies to a particular consumer system 380b.
  • the values indicate that the start position 321a is 00:30:10:15; stop position 323a is 00:30:15:00; the filtering action 325a is skip; the description 327a is "scene of bloodshed" and the configuration 329a is 2.1. More detail regarding navigation objects, such as navigation object 320a, will be provided with reference to Figures 4B and 5B.
  • navigator 310a monitors audio and video decoders 350a for the time code of the multimedia content currently being decoded, the time code is compared to the navigation objects in object store 316a. When the position code falls within the start and stop positions defined by a navigation object, navigator 310a activates the filtering action assigned to the navigation object. For navigation object 320a, a time code within the approximately four-second range of 00:30:10:15 - 00:30:15:00 result in navigator 310a issuing a command to audio and video decoders 350a to skip to the end of the range so that the multimedia content within the range is not decoded and is not given to output device 370a.
  • Figure 3B includes a content source 330b, audio and video decoders 350b, and output device 370b.
  • object store 316b is located at server system 390b, and all other components are located at consumer system 380b.
  • start 321b, stop 323b, action 325b, description 327b, and configuration 329b the contents of navigation object 320b remain unchanged.
  • Content source 330b, including DVD drive 334b and DVD 332b have been combined with audio and video decoders 350b, vendor independent interface 352b, and navigation software 312b into a single device.
  • Communication between navigation software 312b and object store 316b occurs over communication link 314b.
  • Communication link 314b is an example of communication means and should be interpreted to include any communication link for exchanging data between computerized systems.
  • the particular communication protocols for implementing communication link 314b will vary from one embodiment to another.
  • at least a portion of communication link 314b may include the Internet.
  • Output device 370b includes a television 372b with video input 374b and an audio receiver 377b with an audio input 376b. Audio receiver 377b is connected to speakers 378b. As noted earlier, the sophistication and complexity of output device 370b depends on the implementation of a particular embodiment. As shown, output device 370b is relatively simple, but a variety of components, such as video and audio receivers, amplifiers, additional speakers, etc., may be added without departing from the present invention. Furthermore, it is not necessary that output device 370b include both video and audio components. If multimedia content includes only audio content, the video components are not needed. Likewise, if the multimedia content includes only video data, the audio components of output device 370b may be eliminated.
  • Figure 3C includes a server/remote system 390c and a consumer system 380c.
  • navigator 310C is located at server/remote system 390c and content source 33Oc, audio and video decoders 350c, and output device 370c are located at the consumer system 380c.
  • Navigator 310c includes server navigation software 312c and object store 316c, with data being exchanged as bi-directional arrow 314c indicates.
  • Start 321c, stop 323c, action 325c, description 327c, and configuration 329c show that the contents of navigation object 320c remain unchanged from navigation objects 320b and 320a ( Figures 3B and 3A).
  • Content source 330c includes DVD drive 334c and DVD 332c
  • output device 370c includes graphics adaptor 374c, video display 372c, audio adapter 376c, and speakers 378c. 5 Because content source 330c and output device 370c are identical to the corresponding elements in Figure 3A, their descriptions will not be repeated here.
  • client navigator software 354c had been added to audio and video decoders 350c and vendor independent interface 352c.
  • Client navigator software 354c supports communication between navigation software 312c and vendor independent
  • no client navigator software 354c will be necessary whereas in other embodiments, some type of communication interface supporting communication link 356c may be necessary.
  • server/remote system 390c is a server computer, and at least a portion of communication link 356c includes the
  • Client navigator software 354c may be helpful in establishing communication link 356c and in passing information between consumer system 380c and server/remote system 390c.
  • Server/remote system 390c may be embodied in '.0 a remote control unit that controls the operation of the DVD player over an infrared or other communication channel.
  • client navigator software 354c nor vendor independent interface 352c may be needed for this case because server/remote system 390c is capable of direct communication with the DVD player and the DVD player assumes responsibility for controlling audio and video decoders 350c.
  • Figure 4A shows a sample method for filtering multimedia content according to the present invention.
  • Figures 4A, 5A, 6, and 7 show the method as a sequence of events, the
  • an object store may be part of a larger data storage.
  • a separate object store may exist '5 for dia content stored on individual DVD titles. Because many object stores have been created, at block 412 the multimedia content title is retrieved from the content source. Alternatively, a single object store may contain navigation objects corresponding to more than one DVD title.
  • the object store and corresponding navigation objects that are specific to a particular DVD title are selected. (Receive fee, block 416, will be described later, with reference to a server system.)
  • the first navigation object for the DVD title identified at 412 is retrieved.
  • a navigation object is shown in the context of multimedia content.
  • Content positions 480 identify various positions, labeled P41, P42, P43, P44, P45, P46, and P47, that are associated with the multimedia content.
  • the navigation object portion 490 of the content begins at start 491 (P42) and ends at stop 493 (P46).
  • Skip 495 is the filtering action assigned to the navigation object and scene of bloodshed 497 is a text description of the navigation object portion 490 of the multimedia content.
  • Configuration 499 identifies the hardware and software configuration of a consumer system to which the navigation object applies.
  • configuration 499 may include the make, model, and software revisions for the consumer's computer, DVD drive, graphics card, sound card, and may further identify the DVD decoder and the consumer computer's motherboard.
  • the motivation behind configuration 499 is that different consumer systems may introduce variations in how navigation objects are processed. As those variations are identified, navigation objects may be customized for a particular consumer system without impacting other consumer systems.
  • the configuration identifier may be generated according to any scheme for tracking versions of objects. In Figure 4B, the configuration identifier includes a major and minor revision, separated by a period.
  • Decision block 424 determines whether the configuration identifier of the navigation object matches the configuration of the consumer system. Matching does not necessarily require exact equality between the configuration identifier and the consumer system. For example, if major and minor revisions are used, a match may only require equality of the major revision. Alternatively, the configuration identifier of a navigation object may match all consumer configurations. Configuration identifiers potentially may include expressions with wildcard characters for matching one or more characters, numeric operators for determining the matching conditions, and the like. If no match occurs, returning to block 422 retrieves the next navigation object.
  • the decoders begin decoding the multimedia content (432) received from the DVD. Once decoded, the content is transferred (434) to the output device where in can be played for a consumer. While decoding the multimedia content, the position code is updated continuously (436).
  • a step for filtering multimed ia content includes the acts of comparing the updated position code to the navigation object identified in block 422 to determine if the updated position code lies within the navigation object and the act of activating an filtering action (444) when appropriate. If the updated position code is not within the navigation object, decoding continues at block 432. But if the updated position code is within the navigation object, the filtering action is activated (444).
  • the next navigation object is retrieved at block 422.
  • the navigation object is retrieved in block 422 and passes the configuration match test of block 424.
  • the position code is updated at block 436.
  • P41 corresponds to the updated position code. Because P41 is not within the start and stop positions (491 and 493), more multimedia content is decoded (432), transferred to the output device (434), and the position code is updated again (436). The updated position code is now P42.
  • P42 also marks the beginning of the navigation object portion 490 of the multimedia content defined by the start and stop positions (491 and 493) of the navigation object.
  • the video filtering action, skip 495 is activated in block 444. Activating the video filtering action sends a command to the decoder to discontinue decoding immediately and resume decoding at stop position 493. The content shown between P42 and P46 is skipped. Following the skip, the next navigation object is retrieved at block 422 and the acts describe above are repeated. Abruptly discontinuing and resuming the decoding may lead to noticeable artifacts that detract from the experience intended by the multimedia content. To diminish the potential for artifacts, filtering actions may be incrementally activated or separate incremental filtering action may be used.
  • a fade out (e.g., normal to blank display) filtering action may precede a skip filtering action and a fade in (e.g., blank to normal display) filtering action may follow a skip filtering action.
  • the fading out and fading in may be included as part of the skip filtering acting itself, with the start and stop positions being adjusted accordingly.
  • the length of fade out and fade in may be set explicitly or use an appropriately determined default value.
  • Incremental filtering actions need not be limited to a specific amount of change, such as normal to blank display, but rather should be interpreted to include any given change, such as normal to one-half intensity, over some interval.
  • incremental filtering actions may be used to adjust virtually any characteristic of multimedia content.
  • multimedia content includes visual information being presented to a viewer
  • unsuitable material may be localized to only a certain physical area of the scene as it is presented.
  • one or more navigation objects with reframe filtering actions may be appropriate.
  • the entire scene need not be skipped because the viewing frame may be positioned to avoid showing the unsuitable material and the remaining content may be enlarged to provide a full-size display.
  • Each reframe navigation object is capable of performing a number of reframe/resize actions, including the ability to reframe and resize on a frame-by-frame basis. Therefore, the number of reframe navigation objects used in cropping a particular scene depends on a variety of factors, including how the scene changes with time.
  • a single navigation object may be sufficient to filter a relatively static scene, whereas more dynamic scenes will likely require multiple navigation objects.
  • one navigation object may be adequate to reframe a scene showing an essentially static, full-body, view of a person with a severe leg wound to a scene that includes only the person's head and torso.
  • multiple reframe navigation objects may be required for improved results.
  • Positions P41 , P42, P43, P44, P45, P46, and P47 are separated by the update interval.
  • a shorter update interval will allow for more precise filtering. For example, if start 491 were shortly after position P42, multimedia decoding and output would continue until position P43, showing nearly 1/4 of the multimedia content that was to be filtered. With an update interval occurring ten times each second, only a minimal amount of multimedia content that should be filtered (e.g., less than 1/10th of a second) will be displayed at the output device.
  • a minimal amount of multimedia content that should be filtered e.g., less than 1/10th of a second
  • FIG. 3B provides an exemplary system where processing is shared between a server system and a consumer system. Nevertheless, the following will describe the processing as it would occur at a server system, similar to the one shown in Figure 3C, but with only the output device located at the consumer system.
  • the server receives the DVD title identifier so that the proper navigation objects can be selected in block 414.
  • the server receives a fee from the consumer system, in block 416, for allowing the consumer system access to the navigation objects.
  • the fee may be a subscription for a particular time period, a specific number of accesses, etc.
  • the first navigation object for the DVD title identified at 412 is retrieved in block 422 and checked for a configuration match in block 424. Because the configuration match is checked at the server, the consumer system supplies its configuration information or identifier.
  • receiving a content identifier (412), selecting navigation objects (414), receiving a fee (416), retrieving a navigation object (422), and determining whether the configuration identifier matches the consumer system configuration (424) have been enclosed within a dashed line to indicate that they are all examples of acts that may occur within a step for the server system providing an object store having navigation objects.
  • Decoding the multimedia content (432) may occur at either the consumer system or the server system. However, sending decoded multi edia from a server system to a consumer system requires substantial communication bandwidth.
  • the multimedia content is transferred to the output device.
  • the server system queries (436) the client system decoder to update the position code.
  • the position code may be updated (436) without making a request to the consumer system.
  • the acts of decoding (432), transferring (434), and continuously updating or querying for the position code (436) have been enclosed in a dashed line to indicate that they are examples of acts that are included within a step for the server system using a decoder to determine when multimedia content is within a navigation object (430).
  • the server system performing a step for filtering multimedia content (440) includes the acts of (i) comparing the updated position code to the navigation object identified in block 422 to determine if the updated position code lies within the navigation object, and (ii) activating or sending an filtering action (444) at the proper time. Decoding continues at
  • Figure 5A illustrates a sample method for filtering audio content, possibly included with video content, according to the present invention.
  • the steps for providing 510 and using 5 530, including the acts shown in processing blocks 512, 514, 516, 522, 524, 532, 534, and 536 are virtually identical to the corresponding steps and acts described with reference to Figures 4A. Therefore, the description of Figure 5A begins with a step for filtering (540) multimedia content.
  • Decision block 542 determines if an updated or q ueried position code (536) is within :0 the navigation object identified in blocks 522 and 524. If so, decision block 552 determines whether or not an filtering action is active. For portions of multimedia content within a navigation object where the filtering action is active or has been sent (in the case of server systems), decoding can continue at block 532. If the filtering action is not active or has not been sent, block 544 activates or sends the filtering action and then continues decoding at ,5 block 532.
  • decision block 542 determines whether or not an filtering action is active or has been sent. If no filtering action is active or has been sent, decoding continues at block 532. However, if an filtering action has been activated or sent and the 0 updated position code is no longer within the navigation object, block 546 activates or sends and end action and continues by identifying the next navigation object in blocks 522 and 524.
  • configuration 599 identifies the hardware and software configuration of a consumer system to which the navigation object applies.
  • the position code is updated at block 536.
  • P51 corresponds to the updated position code. Because P51 is not within (542) the start position 591 and stop position 593 and no filtering action is active or sent (556), more multimedia content is decoded (532), transferred to the output device (534), and the position code is updated again (536).
  • the updated position code is now P52.
  • P52 also marks the beginning of the navigation object portion 590 of the multimedia content defined by the start and stop positions (591 and 593) of the navigation object, as determined in decision block 542.
  • decision block 552 continues by activating or sending (544) the filtering action assigned to the navigation object to mute audio content, and once again, content is decoded (532), transferred to the output device (534), and the position code is updated or queried (536).
  • Muting in its most simple form, involves setting the volume level of the audio content to be inaudible. Therefore, a mute command may be sent to the output device without using the decoders. Alternatively, a mute command sent to the decoder ay eliminate or suppress the audio content.
  • audio content may include one or more channels and that muting may apply to one or more of those channels.
  • the updated or queried position code (536) is P53.
  • Decision block 542 determines that the updated or queried position code (536) is within the navigation object, but an filtering action is active or has been sent (552), so block 532 decodes content, block 524 transfers content to the output device, and block 536 up ates or queries the position code.
  • the audio content continues to be decoded and the muting action continues to be activated.
  • the updated or queried position code (536) is P54.
  • decision block 542 determines that the updated or queried position code (536) is no longer within the navigation object, but decision block 556 indicates that the muting action is active or has been sent.
  • Block 546 activates or sends and end action to end the muting of the audio content and the decoding continues at block 532.
  • the result would be that the video content is played at the output device, but the portion of the audio content containing an obscenity, as defined by the navigation object, is filtered out and not played at the output device.
  • filtering actions may be incrementally activated or separate incremental filtering action may be used.
  • a fade out (e.g., normal to no volume) filtering action may precede a mute filtering action and a fade in (e.g., no volume to normal) filtering action may follow a mute filtering action.
  • the fading out and fading in may be included as part of the mute filtering acting itself, with the start and stop positions being adjusted accordingly.
  • the length of fade out and fade in may be set explicitly or use an appropriately determi ned default value.
  • Incremental filtering actions are not limited to any particular amount of change, such as normal to no volume, but rather should be interpreted to include any change, such as normal to one-half volume, over some interval.
  • incremental filtering actions may adjust virtually any characteristic of multimedia content.
  • the method shown in Figure 5A may be practiced at both client systems and server system.
  • the methods will not be described in a server system because the distinctions between a consumer system and a server system have been adequately identified in the description of Figures 4A and 4B.
  • Figure 6 is a flowchart illustrating a method used in customizing the filtering of multimedia content.
  • a password is received to authorize disabling the navigation objects.
  • a representation of the navigation objects is displayed on or sent to (for server systems) the consumer system in block 620.
  • a response is received that identifies any navigation objects to disable and, in block 640, the identified navigation objects are disabled.
  • Navigation objects may be disabled by including an indication within the navigation objects that they should not be part of the filtering process.
  • the act of retrieving navigation objects may ignore navigation objects that have been marked as disabled so they are not retrieved. Alternatively, a separate act could be performed to eliminate disabled navigation objects from being used in filtering multimedia content.
  • the acts of receiving a password (610), displaying or sending a representation of the navigation objects (620), receiving a response identifying navigation objects to disable (630), and disabling navigation objects (640), have been enclosed in a dashed line to indicate that they are examples of acts that are included within a step for deactivating navigation objects 5 (660).
  • deactivating navigation objects may be practiced in either a consumer system or a server system.
  • Figure 7 illustrates an exemplary method for assisting a consumer system in automatically identifying and filtering portions of multimedia content.
  • a step for providing an object store (710) includes the acts of creating navigation objects (712), creating an object L0 store (714), and placing the navigation objects in the object store 716.
  • a step for providing navigation objects (720) follows.
  • the step for providing navigation objects (720) includes the acts of receiving a content identifier (722), such as a title, and receiving a request for the corresponding navigation objects (726).
  • block 732 [5 identifies the act of determining if a user has an established account. For example, if a user is a current subscriber then no charge occurs. Alternatively, the charge could be taken from a prepaid account without prompting the user (not shown). If no established account exists, the user is prompted for the fee, such as entering a credit card number or some other form of electronic currency, at block 734 and the fee is received at block 736.
  • a step for providing tO navigation objects (740) follows that includes the act of retrieving the navigation objects (742) and sending the navigation objects to the consumer system (744). The act of downloading free navigation software that makes use of the navigation objects also may be included an inducement for the fee-based service of accessing navigation objects.
  • FIG. 1 Further aspects of the present invention also involve a system, apparatus, and t5 method for a user to play a multimedia presentation, such as a movie provided on a DVD, with objectionable types of scenes and language filtered.
  • a filtering format defining event filters that may be applied to any multimedia presentation.
  • Another aspect of the invention involves a series of operations that monitor the playback of a multimedia presentation in comparison with one or more filter files, and 50 filter the playback as a function of the filter files.
  • a broad aspect of the invention involves filtering one or more portions of a mu Itimedia presentation.
  • Filtering may involve either muting objectionable language in a multimedia presentation, skipping past objectionable portions of a multimedia presentation as a function of the time of the objectionable language or video, modifying the presentation of a vid eo 15 image such as through cropping, or fading, or otherwise modifying playback to eliminate, reduce, or modify the objectionable language, images, or other content.
  • Filtering may further extend to other content that may be provided in a multimedia presentation, including close captioning text, data links, program guide information, etc.
  • a DVD can hold a full-length film with up to 133 minutes of high quality 5 audio and video compressed in accordance with a Moving Picture Experts Group ("MPEG") coding formats.
  • MPEG Moving Picture Experts Group
  • One aspect of the invention involves the lack of any modification or formatting of the multimedia presentation in order for filtering to occur.
  • the multimedia presentation need not be preformatted and stored on the DVD with any particular information related to the language or type of images being delivered at any point 0 in the multimedia presentation. Rather, filtering involves monitoring existing time codes of multimedia data read from the DVD.
  • a filter file includes a time code corresponding to a portion of the multimedia data that is intended to be skipped or muted.
  • decoding may broadly refer to any stage of processing between when multimedia information is read from a memory media to when it is presented.
  • the term “decoding” may more particularly refer to MPEG decoding.
  • the comparison between a filter file and multimedia data occurs before MPEG decoding. It is possible to perform the comparison operation after MPEG decoding; however, with current decode processing platforms, such a comparison arrangement is less efficient from a time perspective and may result in some artifacts or presentation jitter.
  • the DVD player reads the multimedia information from the DVD during conventional sequential play of the multimedia presentation.
  • the play command causes the read-write head to sequentially read portions of the video from the DVD.
  • the multimedia data may be physically located on a memory media in a non-sequential manner.
  • the multimedia information read from the DVD is stored in a buffer. At this point in the processing, all multimedia information is read from the DVD and stored to the buffer regardless of whether the audio data will be muted, or portions of the video data skipped. From the buffer, the MPEG coded multimedia information is decoded prior to display on a monitor, television, or the like.
  • a typical DVD may have several separate portions referred to as "titles.” One of the titles is the movie, and the other titles may be behind the scenes clips, copyright notices, logos, and the like.
  • filter files are applied to time sequences of the primary movie title, e.g., the sequence of frames that is associated with a particular movie, e.g., "Gladiator" provided on DVD.
  • the DVD specification defines three types of titles (not to be confused with the name of a movie): a monolithic title meant to be played straight through (one_sequential_PGC_title), a title with multiple PGCs (program chains) for varying program flow (multiple_PGC_title), and a title with multiple PGCs that are automatically selected according to the parental restrictions setting of a DVD player (parental_block_title).
  • One sequential PGC titles are the only type at the present time that have integrated timing data for time code display and searching.
  • a one_sequential_PGC_title the multimedia information being read from the DVD includes a time code.
  • the time code for the multimedia information read from a memory media and stored in a memory buffer is compared to filter files in a filter table.
  • a filter table is a collection of one or more filter files for a particular multimedia presentation.
  • a filter file is an identification of a portion of a multimedia presentation and a corresponding filtering action. The portion of the multimedia presentation may be identified by a start and end time code, by start and end physical locations on a memory media, by a time or location and an offset value (time, distance, physical location, or a combination thereof, etc.).
  • a user may activate any combination of filter files or no filter files. Table 1 below provides two examples of filter files for the movie "Gladiator".
  • a filter table for a particular multimedia presentation may be provided as a separate file on a removable memory media, in the same memory media as the multimedia presentation, on separate memory media, or otherwise loaded into the memory of a multimedia player configured to operate in accordance with aspects of the invention.
  • the first filter file (1) has a start time of 00:04:15:19 (hour:minute:second:frame) and an end time of 00:04:48:26.
  • the first filter file further has a duration of 997 frames and is a "skip" type filtering action (as opposed to a mute).
  • the first filter file is associated with two filter types.
  • the first filter type is identified as "V-D- D", which is a filter code for a violent (V) scene in which a dead (D) or decomposed (D) body is shown.
  • the second filter type is identified as "V-D-G", which is a filter code for a violent (V) scene associated with disturbing (D) and/or gruesome (G) imagery and/or dialogue.
  • Implementations of the present invention may include numerous other filter types.
  • V-D-G a filter code for a violent (V) scene associated with disturbing (D) and/or gruesome (G) imagery and/or dialogue.
  • Implementations of the present invention may include numerous other filter types.
  • the V-D-D a filter code for a violent (V) scene associated with disturbing (D) and/or gruesome (G) imagery and/or dialogue.
  • Implementations of the present invention may include numerous other filter types.
  • the 997 frames falling between 00:04:15:19 and 00:04:48:26 are skipped (not shown).
  • the V-D-G filter file is activated, the 210 frames falling between 00:04:51 :26 and 00:04:58:26 are skipped.
  • Tables 2 and 3 below provide examples of various possible filter types conforming to the present invention. Other filter types may be implemented in various embodiments of the present invention.
  • Table 2 provides a list of examples of filter types that may be provided individually or in combination in an embodiment conforming to the invention.
  • the filter types are grouped into five broad classifications, including: Sex/Nudity, Violence/Gore, Language and Crude Humor, and Mature Topics.
  • Within each of the four broad classifications are a listing of particular filter types associated with each broad classification.
  • various time sequences (between a start time and an end time) of a multimedia presentation may be identified as containing subject matter falling within one or more of the filter types.
  • multimedia time sequences may be skipped or muted when particular filter files are applied to a multimedia presentation.
  • multimedia time sequences may be skipped or muted as a function of a broad classification, e.g., Violence/Gore, in which case all portions of a multimedia presentation falling within a broad filter classification will be skipped or muted.
  • Table 3 provides a list of examples of filter types that may be provided individually or in combination in an embodiment conforming to the invention.
  • the filter types are grouped into five broad classifications, including: Violence, Sex/Nudity, Language, and Other. Within each of the four broad classifications, are a listing of particular filter types associated with each broad classification.
  • various time sequences (between a start time and an end time) of a multimedia presentation may be identified as containing subject matter falling within one or more of the filter types.
  • multimedia time sequences may be skipped or muted as a function of a particular filter type, e.g., V-S-A.
  • multimedia time sequences may be skipped or muted as a function of a broad classification, e.g., V, in which case all portions of a multimedia presentation falling within a broad filter classification will be skipped or muted.
  • FIGS 8A and 8B illustrate a flowchart of the operations involved with application of a filter file to a DVD-based multimedia presentation, such as a movie, being played on a DVD player.
  • filtration monitoring begins upon play of a multimedia presentation (operation 10).
  • operation 10 a multimedia presentation
  • play is started.
  • “Play” in the context of a movie involves the coordinated video and audio presentation of the movie on a display.
  • the user before depressing "play” the user first activates one or more filter types for the movie.
  • the user must first load the filter table in memory, or the multimedia player must first obtain the filter table, such as through some form of automatic downloading operation.
  • the multimedia information is read from the DVD and stored in a buffer (operation 15).
  • the multimedia information stored on the DVD is arranged in a generally hierarchical manner according to the DVD specifications.
  • Some implementations of the present invention operate on a portion of the multimedia data referred to as a video object unit ("VOBU").
  • the VOBU is the smallest unit of playback in accordance with the DVD specifications. However, in some implementations of the present invention, the smallest unit of playback is at the frame level.
  • a VOBU is an integer number of video fields typically ranging from 0.4 to 1 second in length, typically about 12-15 frames. Thus, playback of a VOBU may be accompanied by between 0.4 to 1 second of video, audio, or both.
  • a VOBU is a subset of a cell.
  • a cell is comprised of one or more VOBUs and is generally characterized as a group of pictures or audio blocks and is the smallest addressable portion of a program chain. Playback may be arranged through orderly designation of cells.
  • some implementations of the present invention monitor the time code of the next multimedia information to be read out of the buffer for decoding and presentation.
  • a VOBU presentation time stamp time code
  • the time code may integral with the multimedia data stored on the memory media, such as in the case of the presentation time stamp of a VOBU.
  • the buffer is a memory configured for first-in-first-out (FIFO) operation.
  • the term buffer may refer to any memory medium including RAM, Flash Memory, et.
  • the time code of the VOBU about to be transmitted from the buffer for decoding is compared with the start times of the filters identified in the filter table for the multimedia presentation (operation 20). If there is not a match (operation 25), then sequential decoding and presentation of information in the buffer continues normally
  • the type of filter event is determined (e.g., mute or skip) (operation 35).
  • mute video image playback is continued normally, but some or all of the audio portion is muted until the event end time code (operation 40).
  • Muting of the audio accounts for an analog audio output, a digital audio output, or both. For 5 audio muting, the amplitude of the audio signal is reduced to zero for the duration of the mute. For digital muting, the digital output is converted to digital 0s for the duration of the mute.
  • Figure 3B is a flowchart illustrating the operations involved with a skip.
  • playback is interpreted (operation 50).
  • the buffer is reset ,0 (operation 55).
  • a reset of the buffer may be characterized as deleting all information in the buffer or "emptying" the buffer. After a reset, all new information read into the buffer starts at the first memory address. Resetting the buffer may be accomplished in various ways, such as resetting a buffer address pointer (where the next information read from the DVD will be stored) to the first address of the buffer (i.e., allowing existing buffer data to be overwritten).
  • the DVD read unit is commanded to begin reading the frame associated with the filter end time code (operation 60).
  • the start and end of a filter file may also be designated with other values or combinations of values, besides a time code.
  • the frame associated with the filter end time code is sent to the first memory location in the buffer and playback starts again with the frame following the end
  • Figure 9 is a block diagram illustrating one possible example of an organization of on-screen menus for activating one or more filters.
  • the menus are shown in one drawing, but may be presented in separate screens in implementations conforming to aspects of the invention.
  • a first menu displays one or more filter classifications.
  • the example of Fig. 9 5 corresponds with Table 3, there are four filter classifications, including: violence, sex and nudity, language, and other.
  • filters files may not be activated based-on selecting a classification, rather the classifications are used to access a set of filters that correspond with the classification.
  • a second filter menu is displayed with a set of filters corresponding with the selected classification.
  • Fig. 9 illustrates the "violence” filters of Table 3, including: strong action violence, brutal/gory violence, and disturbing images.
  • the user selected the "strong action violence” filter, which activates the "strong action violence filter.”
  • Figs. 10A-10C are block diagrams/flow charts illustrating playback of twelve 12 portions of a multimedia presentation with the "strong action violence” filter activated, and with three portions of the multimedia (portions 5, 6, and 7) having been identified as having strong action violence (“SAV").
  • the multimedia presentation need not be modified to associate particular portions with particular filter types, or modified to associate particular portions with some form of subject matter identifier. Rather, a filter table is provided separately from the multimedia presentation.
  • the filter table has one or more filter entries, and each filter file is arranged with start and end identifiers for portions of the multimedia presentation.
  • Certain broad aspects of the invention such as reading multimedia presentation information from a memory media before filter processing, deleting all buffer contents to achieve a skip, etc., may be implemented regardless of whether the multimedia is coded with filter identifiers or otherwise modified with some form of subject matter identifier.
  • the first four portions of the multimedia presentation are read from a memory media, such as a DVD, and stored in a buffer.
  • the portions are read out of the buffer in the order they arrived, i.e., portions 1-4 are read from the buffer beginning with portion 1 and ending with portion 4.
  • the time code of each portion is compared with a filter table, and if there is no match, the portion is read from the buffer, decoded, and displayed.
  • portions 1-4 are each compared with a filter table, and because the time codes of the portions do not match a filter time code (or other start and end identifiers), the four portions are read out of the buffer, decoded, and displayed.
  • portions 5 5-7 will be skipped (not shown).
  • portions 5-7 all of the information in the buffer is deleted.
  • portions 5-7 and portions 8-10 have been read into the buffer.
  • the buffer portion may be reset to portion 8.
  • DVD read head control may be reduced or
  • Portions 8-10 do not contain strong action violence. Nonetheless, portions 8-10 are deleted from the buffer.
  • a time seek command to the filter end time code is executed. The time seek command causes the memory media to begin reading information from the media and into the buffer beginning with portion 8.
  • multimedia portions 8-12 are read from the media and stored in .5 the buffer. Because the time codes of multimedia portions 8-12 are not associated with a strong action violence filter, multimedia portions 8-12 are read from the buffer, decoded, and displayed.
  • the filtering is applied against a conventional DVD-based multimedia presentation, i.e., the DVD title does not require any tO special formatting beyond that provided in accordance with conventional DVD specifications.
  • a person plays and views the video and identifies objectionable content by way of the start and end identifiers of the objectionable content.
  • a particular range of multimedia (bounded by start and end identifiers) of a DVD title may be classified as any one or combination of filter files.
  • a filter table is loaded into a memory of the DVD player.
  • a DVD player may be configured to access a filter table by way of a network connection with a server providing filter files, by way of a removable memory media, (e.g., DVD, CD, magnetic disc, memory card, etc.) either separate from the movie title or on the 50 same memory media as the movie title, or in other ways.
  • a removable memory media e.g., DVD, CD, magnetic disc, memory card, etc.
  • Figure 11 is a block diagram illustrating one possible multimedia player on-screen menu organization. Access to the filtering menus is provided in a parental control menu.
  • the parental control menu is a conduit to various parent control functions, including conventional parental control features and parent control functionality conforming to aspects of the present invention.
  • the multimedia player is configured with a conventional "lock” parent control feature, a conventional "password” parental control feature, the filtering functionality conforming to aspects of the invention, a conventional "rating limits” parental control feature, and a conventional "unrated titles” parental control feature.
  • the multimedia player accesses a particular menu or collections of menus associated with each selection.
  • the "lock” feature allows a user to lock the DVD player, which prohibits functionality unless a correct user identification and password are entered.
  • the password menus provide the user with a means for setting up or changing a password.
  • the "rating limits” feature allows a user to prohibit viewing of titles that exceed certain ratings.
  • the rating limits feature may be aligned with MPAA (G/PG/PG-13/R/NC-17) ratings. So, for example, viewing of R-rated and above titles is prohibited.
  • the rating limits feature may be activated on a user by user basis, with particular rating limits applied to different users. Rating limits functionality may be implemented by way of V-chip technology.
  • the "unrated titles” feature allows a user to either prohibit or allow play of unrated titles. Some titles are not rated; thus, the rating limits feature would not function to prohibit or allow unrated title viewing.
  • Selection of the "Filtered Play” button causes the multimedia player to load a "Filtered Play” menu.
  • the user may navigate through the on-screen menus by way of the arrow keys on a remote, and may navigate between menus by selecting "enter” on the remote when a particular menu button is highlighted.
  • the Filtered Play menu has a "Filter Settings” button and a "Filters Available” button.
  • the Filter Settings button provides access to the filter selection menus, one example of which is illustrated in Fig. 9.
  • the Filters Available button provides access to the Filter Library menu.
  • the Filter Library menu provides a list of all filters currently in the multimedia player memory, the list is organized in alphabetical order by movie title.
  • the Filter Library menu also provides a list of filters available to download.
  • a file is included that lists all possible movie titles for which filter files are available.
  • the list of available filter files is only current as of the date that the filters were downloaded. With a network connection, it is possible to update the filer list on a regular basis so that the list is always current. If the multimedia player already includes a filter table in memory, then the user need only activate filtering, and then proceed to filtered playback. If a filter table is not already in memory, then the user uploads the filter table to memory before filtered playback. Alternatively, the user may proceed to activate certain filter types, and proceed to filtered playback without first determining whether filters for a particular multimedia title are available.
  • the DVD typically has title information accessible by a DVD player. Before filtered playback, the DVD player compares the movie title to a list of filter tables loaded in memory. If there is not a match, then the user may be prompted to load the filter table for the movie title in memory. Once a filter table is identified for a particular movie title intended for playback, the user is prompted to activate or deactivate the filter types for the movie. The user will be presented with a filter selection menu, such as shown in Fig. 9, unless filters have already been activated.
  • portions of a movie are identified in a filter table.
  • a portion of a multimedia presentation is identified as a range of time falling between the start and end time of a particular filter file. For example, if strong action violence occurs in a movie between the times of 1 :10:1 0:1 (HH:MM:SS:FF) and 1 :10:50:10, then a filter file for the movie will have a filter with a start time of 1 : 10: 10: 1 and an end time of 1 : 10:50: 10.
  • the filter file will include also include an identifier associated with "strong action violence" such as "S-A-V.”
  • an identifier associated with "strong action violence” such as "S-A-V.”
  • S-A-V strong action violence
  • the buffer may also have portions of the multimedia presentation that will be shown. Reading of the multimedia content from the memory media then restarts with the next portion of multimedia following the filter end of time. The portions of multimedia following the filter end time are read into the buffer, decoded, and presented. Due to the speed at which the DVD read head may move to the new media location and read information into the buffer, and also be decoded, it is possible to take such operations without noticeable on-screen artifacts (i.e., the skipping operation may be visibly seamless).
  • Figure 12 is a graphical illustration of one example of the format of a skip type filtering action.
  • Figure 13 is a table identifying one example of the file format for a skip type filtering action.
  • the file format represents one filter file in a filter table.
  • a skip type filter file includes a start time code and an end time code.
  • the start time code of a skip filter file occurs within VOBU N+1 , which follows VOBU N.
  • the actual frame associated with the start time code is X frames from the beginning of VOBU N+-1.
  • the end time code of the skip is occurs within VOBU N+P, which is followed by VOBU N+P+1.
  • the actual frame associated with the end time code is Y frames from the beginning of VOBU N+P.
  • the start and end times may be identified by time code (e.g., HH:MM:SS:FF) or by more particular hierarchical DVD information, discussed in greater detail below, or combination thereof.
  • VOBU N and VOBU N+P+1 are played (both audio and video) in their entirety.
  • the first X frames of VOBU N+1 are played, and the remainder of VOBU N+1 is skipped.
  • the first Y frames of VOBU N+P are skipped, and the remaining frames of VOBU N+P are played. All frames associated with any VOBU(s) falling between VOBU N+1 and VOBU N+P are skipped.
  • the table illustrates the file format for a skip type filter file, in accordance with one example of the present invention.
  • the table is organized by file format byte allocation in the left column, followed by an indication of a number of bytes for each allocation, followed by a description of the byte designations.
  • the file format is one example of a filter file format conforming to aspects of the invention.
  • a file format conforming to aspects of the invention may include some or all of the identified bytes designation, may include different byte arrangements, numbers of bytes for each designation, and other combinations and arrangements.
  • Bytes 0-7 involve packet identifiers.
  • Byte 8 is a filter action code, with 0x1 indicating a skip action, and 0x2 indicating a mute action.
  • Bytes 9-14 are reserved for filter classifications and particular filter types, such as the various classification and types discussed herein. Referring first to byte 8, it is one byte in length and identifies the event action code (e.g., skip or mute). Bytes 9-14 are coded to identify the event classification for each possible combination of event classifications, such as is shown in Table 2. When the a filtering method as discussed herein operates, a comparison is made between the filter types activated by a particular user and the filter classifications identified in bytes 9-14.
  • the event action code e.g., skip or mute
  • Bytes 15-34 are identifiers for a filter start location. The designations in bytes 15-34 may be used alone or in combination to identify the start of a filtering action.
  • Bytes 35-38 are identifiers for a filter end location. The designations in bytes 35-38 may be used alone or in combination to identify the end of a filtering action.
  • Bytes 15-18 identify the start time code of a particular filter.
  • Bytes 19-34 are also related to the start time of a filter, but provide more particular information concerning the exact location of the VOBU, which may be associated with the start time code or separate/independent.
  • Bytes 35-38 identify the end time code of a filter.
  • Bytes 39-54 are also related to the end time of the filter, but provide more particular information concerning the exact location of the VOBU associated with the end time code.
  • Bytes 55-63 involve buffering and padding.
  • Bytes 15-18 are reserved for the filter start time code (HH:MM:SS:FF), byte 15 has hour information, byte 16 has minute information, byte 17 has second information, and byte 18 has frame information. Filtering may proceed, in some implementations of the present invention, with only the start and end time code information.
  • the time code may be converted to the same format as a VOBU presentation time stamp.
  • a VOBU is made up of a sequence of frames, typically 12 to 15 frames.
  • the hour, minute, and second information may be used to identify a VOBU, and the frame information used to designate a particular frame in the VOBU.
  • the DVD player is commanded to momentarily stop playback when the start time code is encountered in the multimedia information read from a memory media, and restart playback beginning with the frame identified with the end time code.
  • VOBUs include time code information and also pointers to other VOBUs at various granularity. So, artifacts may depend on VOBU pointer granularity.
  • the DVD player may need to read some information from the DVD player to determine whether the VOBU being read includes the frame associated with the end time code. It is possible to read a number of VOBUs and assess time code information until the VOBU with the end time frame is identified, without noticeable artifacts.
  • the skip is long, then many VOBUs may need to be read before the end time frame is located. In such instances, due to the lengthy searching process, a short screen freeze may be visible. To avoid or substantially reduce artifacts or the freezing of the image on the screen, it is possible to identify the exact location on the memory of the target VOBU (the VOBU having the frame associated with a filter end time). Such precise definition allows the DVD player to avoid searching for the target VOBU.
  • the skip file format may include bytes 19-34 that identify the start chapter number, start program chain number, start program unit number, start cell number, start address of VOBU N, start address of VOBU N+1 , and frame number associated with the X frames offset from the beginning of VOBU N+1 associated with the start time for the filter event.
  • Bytes 19-34 refer to various hierarchical information as defined in various DVD specifications.
  • a VOBU includes both a time code and a logical block number. As discussed above, the time code represents the time at which the compressed multimedia information within the VOBU is intended for playback.
  • a filter file may identify a portion of a multimedia presentation based on time, and identify portions of the multimedia presentation by monitoring the time codes of VOBUs read from a DVD.
  • the logical block number is an identifier of a particular physical memory location on a DVD where the information for the VOBU is stored.
  • the physical location on the DVD may also be used in a filter file to identify the start and end of a portion of a multimedia presentation.
  • the physical location identifier of a filter file is compared with the physical location information of a VOBU.
  • filter start and end identifiers may comprise the information of the start address of VOBU N+1 , bytes 30-33 (the VOBU having the frame associated with the start of a filtering action). Filtering based on physical location as opposed to time code, has the benefit of completely or substantially avoiding translating the end time code information to a physical location on the DVD.
  • filtering based on physical location is advantageous for filtering a multimedia presentation on a memory that has multiple multimedia presentations.
  • the physical location is associated with a particular multimedia presentation, whereas a time value may require additional processing to ensure it is properly applied against the appropriate multimedia presentation.
  • Filtering based on only the VOBU information will have a granularity of the number of frames within the VOBU, typically 12-15 frames as mentioned above.
  • a frame offset value may be used.
  • the frame offset value designates a particular frame within a VOBU at which filtering begins, and also allows for frame-based playback control. Filtering based on VOBU and offset uses both the VOBU start address (bytes 30-33) and the offset value (byte 34). Alternatively, the offset value may be extracted from the frame field of the time code.
  • the VOBU (VOBU N) preceding the VOBU where a skip begins (VOBU N+1) or other preceding VOBUs may be helpful in identifying the target VOBU (where the skip begins) in fast forwarding or other operations.
  • some fast forwarding not all VOBUs are retrieved from the DVD.
  • filtering is applied in normal play as well as fast forward, the presence of one or more preceding VOBUs allows the system to identify the target VOBU in the case where the target VOBU might otherwise not be retrieved, and thus not available for comparison to the filter files.
  • the start cell number filter identifiers may be used to identify a particular cell in the DVD at which a target VOBU occurs.
  • a cell includes a number of VOBUs. It is possible to identify the start of a skip operation by a cell number and a VOBU within the cell. Referring first to byte 8, it is one byte in length and identifies the event action code (e.g., skip or mute). Bytes 9-14 are coded to identify the event classification for each possible combination of event classifications, such as is shown in Table 2. When the a filtering method as discussed herein operates, a comparison is made between the filter files
  • Multimedia information stored on a DVD is arranged hierarchically.
  • the hierarchy includes chapter information, which is divided into program chains, which is divided into program units, which is divided into cells. Cells are made up of a number of VOBUs. Thus, by identifying one or more or a combination of chapter, program chain, program unit, and
  • any particular VOBU may be precisely located without querying preceding VOBUs.
  • an offset to the VOBU may be used with the DVD hierarchical information. Additional details on the hierarchical arrangement of information on a DVD as well as other general information about DVD technology and DVD file format specifications may be found in "DVD Demystified, second addition" by Jim Taylor, copyright 2001 , 1998 by tO the McGraw-Hill Companies, Inc., the entirety of which is hereby incorporated by reference.
  • the end time code and related time coding information is identified in bytes 35-54.
  • Bytes 35-38 are reserved for the actual event end time code (HH:MM:SS: FF), while bytes 39-54 are reserved for identifying the end chapter number, end program chain number, end program unit number, end cell number, end address of VOBU N+P, and frame number t5 associated with the Y frames offset from the beginning of VOBU N+P associated with the end time for the filter event, and the start address of VOBU N+P+1.
  • Bytes 55-61 are reserved for a buffer, to make the skip event filter descriptor of the same size as an audio mute filter descriptor, and bytes 62-63 are used for padding.
  • a DVD player or other device, memory, storage media, or processing configuration, 10 configured to provide, play, display or otherwise work with a DVD or other audio/visual recording device, incorporating some or all features of the skip and mute file formats may fall within the scope of some or all aspects of the present invention.
  • MPEG 5 encoding provides I frames, B frames, and P frames.
  • An I frame includes all of the information necessary to decode and present the frame.
  • B and P frames on the other hand, rely on i nformation present in another frame for proper presentation. As such, in a skip, it is sometimes preferable to skip to an I frame, when possible. It is possible to skip to B and P frames, however, in some instances, decoding of other frames, such as an I frame, may be necessary in order to present the B or P frame.
  • Figure 13 is a graphical illustration of one example of the format of a mute type filtering action.
  • Figure 14 is a table identifying the file format for one example of a mute event.
  • a mute type filter like a skip, includes a start time code and an end time code.
  • the start time code of the mute is shown as occurring within VOBU N+1 , which follows VOBU N.
  • the actual frame associated with the start time code is X frames from the beginning of VOBU N+1.
  • the end time code of the mute is shown as occurring within VOBU N+P, which is followed by VOBU N+P+1.
  • the actual frame associated with the end time code is Y frames from the beginning of VOBU N+P.
  • the start and end times may be identified by time code (e.g., HH:MM :SS:FF) or by more particular hierarchical DVD information, discussed in greater detail below.
  • VOBU N and VOBU N+P+1 are played (both audio and video) in their entirety.
  • the first X frames of VOBU N+1 are played, and the audio of the remainder of VOBU N+1 is muted, but the video is played.
  • the audio of the first Y frames of VOBU N+P are muted (with the video played), and the remaining frames of VOBU N+P are played. All audio of the frames associated with any VOBU(s) falling between VOBU N+1 and VOBU N+P is muted, and the video is played.
  • the table of Figure 14 is organized by file format byte allocation in the left column, followed by an indication of a number of bytes for each allocation, followed by a description of the byte designations. Much of the byte allocations for a mute type filter are the same as a skip type filter. Only the differences are discussed herein.
  • Byte 15 identifies the audio channels to mute. In this implementation, seven channels of audio are provided for, and muting of any combination of channels may be specified in any particular filter. Each byte is eight bit, a digital 1 indicates a mute and a 0 indicates no mute.
  • bit 0 front center channel
  • bit 1 front right channel
  • bit 2 front left channel
  • bit 3 rear right channel
  • bit 4 rear left channel
  • bit 5 rear center channel
  • bit 6 sub woofer
  • bit 7 bit map between bits and the audio channel
  • Bytes 16-38 are related to the start time of the event, bytes 39-61 are related to the 5 end time of the event, and the remaining bytes 62-63 involve padding. Referring first to byte 8, it is one byte in length and identifies the event action code (e.g., skip or mute). Bytes 9-15 are coded to identify the event classification for each possible combination of event classifications, such as is shown in Table 2. When the event filtering method, as discussed below, operates, a comparison is made between the filters activated by a particular user and 0 the event classifications identified in bytes 9-14.
  • Byte 15 is specified for audio channel mutes, which allows muting of one particular channel of an A/V presentation provided with multiple channels of audio, such as in a 5:1 format where only the center channel may be muted, where most discussion in a movie is presented, whereas other channels may not be muted.
  • the start time code and related time coding information is identified in bytes 16-38.
  • Bytes 16-19 are reserved for the actual event start time code (HH:MM:SS:FF), byte 16 has hour information, byte 17 has minute information, byte 18 has second information, and byte 19 has frame information.
  • Bytes 20-38 are reserved for identifying the start chapter number, start program chain number, start program unit number, start cell number, start address of
  • Bytes 20-38 refer to various hierarchical information as defined in various DVD specifications.
  • Bytes 39-61 are related to the end time code of a mute type filter, with bytes 39-42 allocated to the end time code designation (HH:MM:SS:FF), and bytes 43-61 allocated t5 to hierarchical information for a particular VOBU assocatied with a particular frame where muting will be turned off.
  • aspects of the present invention further involve an indexing apparatus and method for identifying the multimedia presentations available on a particular memory media >0 containing a plurality of filter tables.
  • a particular memory media may contain hundreds or thousands of filter tables.
  • a unique identifier is generated for each multimedia presentation in which filter files have been developed, or in which there is information S5 concerning whether a filter file (table) will or will not be developed.
  • the unique identifier is generated as a function of the file size of the multimedia presentation.
  • Unique identifiers may be generated based on each DVD, or each side of each DVD, when a DVD has multiple sides.
  • Each memory media having a plurality of filter tables includes a master index with a listing of the total number of unique identifiers available on the filter disc. For each unique identifier there is a separate table providing a pointer within the multimedia to the specific filter table for that identifier (if its present) along with additional information concerning the filter table, including whether or not the filter table is actually on the memory media, whether a filter table will be generated, 10 and the MPAA rating value for the title.
  • Figure 16 is the file format for an individual unique identifier record for a particular filter disc.
  • a filter disc comprises of a collection of filter tables.
  • Byte set A are packet identification and error checking bytes.
  • Byte set B contains the unique identifier for the particular table.
  • Byte set C provides the pointer, within the disc, to the specific filter L5 information for the unique identifier, including the formats of Figure 13 and 15.
  • file format of Fig. 13 access to any particular filter file may be provided.
  • Access to any particular filter table may also be provided as a function of the title of 20 the multimedia presentation of the filter, e.g., by searching for Gladiator, access to one or more Gladiator filter tables may be achieved.
  • Filter tables are stored alphabetically (A to Z) and in ascending numerical order (1-9) based on the title of the multimedia presentation associated with a particular filter table.
  • the table 25 includes a character identifier, such as alpha characters (e.g., A-Z), numeric characters (e.g., 0-9), and other characters (e.g., !, @, #, etc.).
  • alpha characters e.g., A-Z
  • numeric characters e.g., 0-9
  • other characters e.g., !, @, #, etc.
  • each character table includes an identification of the number of filters for the character and a map to the first entry in the character table.
  • the system may generate a character-based listing, such 50 as an alphabetical listing of the filter available on the disc. Further, the listing may be accessible based on character entry. So, for example, a screen may be generated that includes an alphabetical listing, and by selecting any letter in the alphabet, the user may access a list of all filters available where the title of the multimedia presentation associated with that filter begins with the selected character.
  • Figure 17 is the file format for a character based look-up table.
  • Byte set B includes the character identifier for a particular table.
  • Byte set B provides ASCII information for each character.
  • the table for character "A” will have the ASCII value for A provided in byte set B.
  • Byte set C provides an identification of the total number of filter tables associated with 5 the particular character.
  • byte set D provide a pointer to the first filter table for the particular character. For example, for "A" the pointer will point to the first filter table for the first multimedia presentation title beginning with A., which may be arranged within the A set of filter tables in alphabetical order.
  • the filter tables on a particular memory media may further be indexed or identified L0 based upon the time of release of the filter table. For example, all filter tables released within 90 days may be highlighted.
  • new filter table releases closely track new multimedia presentation releases (new movies released on DVD, for example)
  • a user may be able to quickly determine whether a filter table for the new DVD release has been generated by searching only new releases.
  • Each new release table provides a pointer to the filter table information for the new release. Thus, a user may obtain a list of all filter tables for new releases only.
  • a particular filter table may be identified by one or more indexing tables, in various possible implementation conforming to aspects of the present invention.
  • Figures 18 - 23 represent indexing tables, that used collectively provide a map into one or a set of filter 20 tables for a particular multimedia presentation. The map provides flexibility to account for versions of filter tables, versions of a movie title, formatting variations for a multimedia presentation, filtering modes (e.g. time-based filtering and location based filtering), and other mapping efficiencies.
  • the studio release table 25 provides one or more bytes (byte set B), to identify the multimedia title (e.g., "Gladiator") for the a particular filter table or set of filter tables.
  • Byte set C includes the release number of the particular filter table. It is possible to have multiple releases of filter tales for a particular multimedia presentation.
  • Byte set D provides and identifier of the studio catalog number for a particular version of a multimedia title. Some movies, for example, may have an unrated 50 version, directors cut, extended play versions, etc. Each of which may have a unique catalogue number.
  • Bytes set E provides similar release edition information, but in the form of an alphanumeric descriptor (e.g., "Director's Cut") as opposed to a catalogue number.
  • Byte set F provides the release date for the filter table.
  • Byte set G provides a map to tables established for multi-sided releases (see discussion of Fig. 20 below).
  • Byte set H provides aspect ratio information for the particular multimedia presentation associate with a particular filter file.
  • Some multimedia titles may be associated with a plurality of physical disc sides. For example, some DVD movies, may be provided on both sides of a DVD, or a plurality of sides of a DVD. If byte set G of Fig. 19 is 1 , then the values for this table are not defined and the movie is on a single disc side. If Byte set G of Fig. 19 is 2t or more, then there are 2 or more disc side table, respectively. Referring to Fig. 20. byte set B is discussed in detail below with regard to Fig. 21. Byte set C indicates the number of DVD title packets for the disc side represented by the table. In most instances, this value will be 1 representing the main movie title.
  • filter tables for other titles that may be on the same side of a disc.
  • the main movie title e.g., Gladiator
  • another DVD title such as an interview with a director may also have a filter file.
  • Byte set D identifies the type of filter identifier applied in the filter file.
  • time code based filtering an location based filtering (as a function of VOBU) may be defined in a particular filter, in various implementations of the present invention.
  • bytes set D defines one or the filtering identifier types.
  • Byte set D also provides the MPAA rating for the particular DVD title. MPAA ratings are typically applied on a movie basis.
  • MPAA ratings may be identified on a DVD title basis.
  • Byte set F provides the filter creation date.
  • Byte set G provides information concerning the total byte length for all filter specific mapping files for the particular filter table.
  • Byte set H provides the aspect ratio for the particular DVD side.
  • the table shown in Fig. 21 provides a second unique identifier for the particular side of the DVD. This unique identifier also accounts for any changes in the unique identifier that may occur if a different length version of a multimedia presentation is released.
  • the table shown in Fig. 22 is provided when sep rate titles on a particular side of a
  • DVD have unique filters. There is a separate table for each filtered title.
  • Byte set B identifies the title.
  • Byte set C identifies the program chain number of the title.
  • Byte set D indicates a unique identifier for the particular title. With such a unique identifier, it is possible to search globally for various possible filters (e.g., search for a filter for "Gladiator") or to search for filters for various titles within a DVD disc side.
  • Bytes set E identifiers the number of different language versions that filters are available. For example, objectionable language may be different based on a particular language; thus, filtering based on objectionable language may also be different based upon the language available.
  • Byte set E provides a map to the number of language table, for which there is a separate table for each supported language.
  • the table of Fig. 23 provides the actual pointer to the specific filter file information for the multimedia presentation. Depending on the particular multimedia presentation, the pointer may address the filter files as a function of the fil title, the disc side, the DVD title, language, and other factors addressed above.
  • Byte set G indicates the number of filter files 5 in a particular filter table.
  • Byte set H is the pointer to the first filter file for the multimedia presentation.
  • the table of Fig. 23 also provides other information.
  • First, bytes set B provides a language identifier for the filter file.
  • Byte set C provides title information as shown in the diagram.
  • Byte set D is pointer into theme descriptors for the multimedia presentation.
  • LO theme descriptors do not provide filtering, but rather provide a textual description of various thematic topics presented in a particular multimedia presentation. For example, where a suicide occurs in a particular movie, the theme "suicide" may be presented to the user as a function of the thematic descriptor. As such, if the user has activated filtering, before playback begins, the thematic descriptor or descriptors will be presented to the user on the
  • Byte set E provides an identification of the particular filter types available for the multimedia presentation
  • byte set F provides an indication of the filter types not tO available.
  • Byte set G identifies the total number of activatable filter files for the multimedia presentation.
  • aspects of the present invention extend to methods, systems, and computer program products for automatically identifying and filtering portions of multimedia content (such as a multimedia presentation provided in a DVD format).
  • the embodiments of the present t5 invention may comprise a DVD player, a special purpose or general purpose computer including various computer hardware, a television system, an audio system, and/or combinations of the foregoing. These embodiments are discussed in detail above. However, in all cases, the described embodiments should be viewed a exemplary of the present invention rather than as limiting it's scope.
  • Embodiments within the scope of the present invention also include computer- readable media for carrying or having computer-executable instructions or data structures stored thereon.
  • Such computer-readable media may be any available media that can be accessed by a general purpose or special purpose computer.
  • Implementations of the present invention may be stored as computer readable instructions on a DVD along with a multimedia presentation intended to be filtered and played back with various time sequences muted or skipped.
  • a network or another communications link or connection either hardwired, wireless, or a combination of hardwired or wireless
  • the computer properly views the connection as a computer- readable medium.
  • any such a connection is properly termed a computer-readable medium.
  • Computer executable instructions comprise, for example, instructions and data which cause a DVD player, a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.
  • aspects of the invention may be deployed as computer- executable instructions, such as program modules, being executed by a DVD player.
  • program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular abstract data types.
  • Computer-executable instructions, associated data structures, and program modules represent examples of the program code means for executing steps of the methods disclosed herein. The particular sequence of such executable instructions or associated data structures represent examples of corresponding acts for implementing the functions described in such steps.
  • program code means being executed by a processing unit provides one example of a processor means.

Abstract

L'invention concerne un procédé permettant de filtrer des parties d'une présentation multimédia. Un flux de données multimédia lues à partir d'un support de mémoire est comparé à un fichier filtrant associé aux données multimédia. Ledit fichier filtrant comporte une position de départ, une position d'arrêt et une action de filtrage à réaliser sur la partie du contenu multimédia qui commence à la position de départ et se termine à la position d'arrêt. Lorsque les données multimédia lues à partir du support correspondent au fichier filtrant, l'action de filtrage désignée est réalisée. Des aspects de cette invention ont aussi trait au format du fichier filtrant, au format d'accès des fichiers filtrants sur un support de mémoire.
PCT/US2005/012448 2004-04-12 2005-04-12 Appareil, systeme et procede de filtrage de parties indesirables d'une presentation multimedia WO2005099413A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US56185104P 2004-04-12 2004-04-12
US60/561,851 2004-04-12

Publications (2)

Publication Number Publication Date
WO2005099413A2 true WO2005099413A2 (fr) 2005-10-27
WO2005099413A3 WO2005099413A3 (fr) 2006-05-18

Family

ID=35150454

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2005/012448 WO2005099413A2 (fr) 2004-04-12 2005-04-12 Appareil, systeme et procede de filtrage de parties indesirables d'une presentation multimedia

Country Status (2)

Country Link
CN (1) CN1981281A (fr)
WO (1) WO2005099413A2 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1919215A1 (fr) 2006-11-03 2008-05-07 Samsung Electronics Co., Ltd. Procédé de verrouillage de lecture et dispositif d'image correspondant
EP1949689A2 (fr) * 2005-11-15 2008-07-30 Nds Limited Systeme de zoomage video numerique
EP2016769A2 (fr) * 2006-01-30 2009-01-21 ClearPlay, Inc. Synchronisation de metadonnees de filtre avec une presentation multimedia
US9432730B2 (en) 2012-12-26 2016-08-30 Huawei Technologies Co., Ltd. Multimedia file playback method and apparatus
US9451324B2 (en) 2000-10-23 2016-09-20 Clearplay, Inc. Method and user interface for downloading audio and video content filters to a media player
US11432043B2 (en) 2004-10-20 2022-08-30 Clearplay, Inc. Media player configured to receive playback filters from alternative storage mediums

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9270748B2 (en) * 2008-12-18 2016-02-23 Telefonaktiebolaget L M Ericsson (Publ) Method for content delivery involving a policy database
CN103905904B (zh) * 2012-12-26 2018-04-10 华为技术有限公司 播放多媒体文件的方法及装置
CN106973317A (zh) * 2016-12-30 2017-07-21 华为软件技术有限公司 多媒体数据处理方法、多媒体数据提供方法、装置及系统
CN110427499B (zh) * 2018-04-26 2023-08-29 腾讯科技(深圳)有限公司 多媒体资源的处理方法、装置及存储介质和电子装置
CN109005423A (zh) * 2018-08-10 2018-12-14 北京奇艺世纪科技有限公司 一种视频播放方法及装置
JP6499364B1 (ja) * 2018-09-26 2019-04-10 株式会社Cygames 情報処理プログラム、端末装置、及び情報処理方法
CN112804584A (zh) * 2019-11-13 2021-05-14 阿里巴巴集团控股有限公司 视频、图片、音频的设置方法及装置

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6009433A (en) * 1995-04-14 1999-12-28 Kabushiki Kaisha Toshiba Information storage and information transmission media with parental control
US6091886A (en) * 1992-02-07 2000-07-18 Abecassis; Max Video viewing responsive to content and time restrictions
US6181364B1 (en) * 1997-05-16 2001-01-30 United Video Properties, Inc. System for filtering content from videos
US20040006767A1 (en) * 2002-07-02 2004-01-08 Robson Gary D. System, method, and computer program product for selective filtering of objectionable content from a program

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6091886A (en) * 1992-02-07 2000-07-18 Abecassis; Max Video viewing responsive to content and time restrictions
US6009433A (en) * 1995-04-14 1999-12-28 Kabushiki Kaisha Toshiba Information storage and information transmission media with parental control
US6181364B1 (en) * 1997-05-16 2001-01-30 United Video Properties, Inc. System for filtering content from videos
US20040006767A1 (en) * 2002-07-02 2004-01-08 Robson Gary D. System, method, and computer program product for selective filtering of objectionable content from a program

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9451324B2 (en) 2000-10-23 2016-09-20 Clearplay, Inc. Method and user interface for downloading audio and video content filters to a media player
US11432043B2 (en) 2004-10-20 2022-08-30 Clearplay, Inc. Media player configured to receive playback filters from alternative storage mediums
EP1949689A2 (fr) * 2005-11-15 2008-07-30 Nds Limited Systeme de zoomage video numerique
EP1949689A4 (fr) * 2005-11-15 2009-12-02 Nds Ltd Systeme de zoomage video numerique
US7782344B2 (en) 2005-11-15 2010-08-24 Nds Limited Digital video zooming system
EP2016769A2 (fr) * 2006-01-30 2009-01-21 ClearPlay, Inc. Synchronisation de metadonnees de filtre avec une presentation multimedia
EP2016769A4 (fr) * 2006-01-30 2010-01-06 Clearplay Inc Synchronisation de metadonnees de filtre avec une presentation multimedia
EP1919215A1 (fr) 2006-11-03 2008-05-07 Samsung Electronics Co., Ltd. Procédé de verrouillage de lecture et dispositif d'image correspondant
US9432730B2 (en) 2012-12-26 2016-08-30 Huawei Technologies Co., Ltd. Multimedia file playback method and apparatus

Also Published As

Publication number Publication date
CN1981281A (zh) 2007-06-13
WO2005099413A3 (fr) 2006-05-18

Similar Documents

Publication Publication Date Title
US20060031870A1 (en) Apparatus, system, and method for filtering objectionable portions of a multimedia presentation
US20200162787A1 (en) Multimedia content navigation and playback
US9628852B2 (en) Delivery of navigation data for playback of audio and video content
WO2005099413A2 (fr) Appareil, systeme et procede de filtrage de parties indesirables d'une presentation multimedia
JP4824846B2 (ja) ディスプレイ・モニタを備えたシステム
AU2002211296A1 (en) Filtering objectionable multimedia content
US8385719B2 (en) Video reproducing apparatus and method
KR101027466B1 (ko) 디지털 저장 장치를 구비하는 전자 프로그램 가이드
RU2304313C2 (ru) Способ и устройство воспроизведения носителя записи
US20030049014A1 (en) Method and apparatus for playing digital media and digital media for use therein
US20050152683A1 (en) Method and apparatus for repetitive playback of a video section based on subtitles
US20060051064A1 (en) Video control system for displaying user-selected scenarios
JPH07236099A (ja) 情報再生装置を内蔵したテレビジョン装置
US20090136218A1 (en) Multimedia presentation format
JP2007267259A (ja) 画像処理装置およびファイル再生方法
KR100757422B1 (ko) 방송 프로그램의 시청 이력 관리 방법
US20080159710A1 (en) Data Recording Method
KR20040034132A (ko) 타이틀 연상 이미지 표시장치

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

WWE Wipo information: entry into national phase

Ref document number: 200580019220.8

Country of ref document: CN

122 Ep: pct application non-entry in european phase