GB2505777A - Online media policy platform - Google Patents

Online media policy platform Download PDF

Info

Publication number
GB2505777A
GB2505777A GB1314833.3A GB201314833A GB2505777A GB 2505777 A GB2505777 A GB 2505777A GB 201314833 A GB201314833 A GB 201314833A GB 2505777 A GB2505777 A GB 2505777A
Authority
GB
United Kingdom
Prior art keywords
content
media
policy
platform
access
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
GB1314833.3A
Other versions
GB201314833D0 (en
Inventor
David Jacobs
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of GB201314833D0 publication Critical patent/GB201314833D0/en
Publication of GB2505777A publication Critical patent/GB2505777A/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/10Protecting distributed programs or content, e.g. vending or licensing of copyrighted material ; Digital rights management [DRM]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/60Protecting data
    • G06F21/62Protecting access to data via a platform, e.g. using keys or access control rules
    • G06F21/6218Protecting access to data via a platform, e.g. using keys or access control rules to a system of files or objects, e.g. local or distributed file system or database

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Physics & Mathematics (AREA)
  • Computer Hardware Design (AREA)
  • Computer Security & Cryptography (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Technology Law (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Bioethics (AREA)
  • General Health & Medical Sciences (AREA)
  • Storage Device Security (AREA)

Abstract

The present application seeks to provide a media policy platform to trigger actions based upon identification and/or the usage context of how the media is consumed. An online media policy platform comprises a policy engine, a sampler to sample content of digital media and to request policies from the policy engine that are applicable to the requested content. The platform further comprises a discriminator adapted to determine a condition that the content is to be consumed under, and a decision tree, which uses inputs from the discriminator to determine which policy to apply to the content. The platform retrieves enforcement instructions from the policy engine and enforces the policy at a device or network enforcement point so that access to the content is controlled. A method of controlling access to online media is also disclosed which comprises controlling access through the sampling point in dependence on content identified for a sampled media and usage conditions. A further method provides a feed of source addresses of malware sites by supplying addresses of non-permitted sources to a network provider.

Description

Online media policy platform The invention relates to a context based online media policy platform.
Copyright infringement among consumers of digital media represents a significant economic challenge to content creators and providers. As the technology to distribute digital media online has developed, piracy techniques have developed alongside, enabling the consumption of media without payment to the creator.
* The known content protection technologies control access to and use of the content and limit its unauthorized copying and redistribution. Examples of such technology can be found in W02009/1 26829, US2009/1 96465 and US2012101 1560. Parties seeking to engage in unauthorized distribution and copying of protected commercial music or video content can circumvent the content protection to obtain a decrypted copy of the content. Once a decrypted copy of the content is obtained, the content protection technology is no longer effective at governing access to it and the decrypted content may be subject to unlimited use, copying, and redistribution in particular via online techniques.
Recently content providers have begun to release media with digital watermarks.
Digital watermarks comprise identifying data, which is woven into media content such as images, movies, music or programming such as games or TV episodes, thereby giving those objects a unique, digital identity that can be used for a variety of valuable applications. An example of the possible applications can be seen in the non prior published application US2013/0205317. Digital waterrnarks are imperceptible to the human senses yet easily recognized by special software detectors, a digital watermark remains constant even through recording, manipulation and editing, compression and decompression, encryption, decryption and broadcast.
Digital watermarks are useful in establishing copyright ownership and tracing the source of an infringing work. Some media players can be adapted to recognize digital watermarks so that they can be used to recognize that a work has been copied from a different format, eg if a film has been recorded in a theatre and converted to a DVD format, it would be possible to prevent the DVD from being replayed.
Consumers of digital media can be categorised into three groups. The first group consists S consumers who will only access digital media using established services where there is no problem with piracy. The second group consists of those consumers who will never pay for online content as a mailer of principle. However, the largest group consists of people who are simply interested in. accessing a particular piece of content irrespective of whether the source is legitimate or pirated.
For this group of consumers the more limited selection of media on the established paid for services results in them looking for media through other outlets. A large amount of media is available for free via Torrents and other peer to peer (P2P) systems. Access to this media is technically difficult to control.
the present invention seeks to provide a media policy platfomi adapted to trigger actions based on identification of and/or the usage-context of how the media is consumed.
According to the invention there is provided an online media policy platform comprising a policy engine, a sampler adapted to sample content of digital media, the sampler being further adapted to request policies from the policy engine that are applicable to the requested content, the platform further comprising a discriminator adapted to determine a condition that the content is to be consumed under, the platform further comprising a decision tree, which decision tree uses inputs from the discriminator to determine which policy to apply to the content, wherein the platform retrieves enforcement instructions from the policy engine and enforces the policy.
Preferably, the platform is adapted to determine the identity of sampled content and access rules applicable to the identified content. Preferably, the discriminator determines an access status for the content. Preferably, the access status, the identity of the content and usage context conditions are fed to a decision tree to determine the policy to be applied. Preferably, the policy is returned to the sampler for enforcement. Preferably, the sampler comprises a device or network sampling and enforcement point. Preferably, the platform is adapted to control access to the media. Preferably, access to the media is blocked. Alternatively, the user is diverted from the source of the sampled content to an alternative source of the content.
According to a second aspect of the invention there is provided a method of controlling access to digital media comprising the steps of sampling The media at a sampling point, identifying the content of the media, determining any usage conditions on the content of the media and controlling further access to the media through the sampling point in dependence on the content of the media and usage conditions.
According to a third aspect of the invention there is provided a method of providing a feed of source addresses of malware sites comprising the steps of sampling the media at a sampling point, identifying the content of the media, determining any usage conditions on the content of the media, wherein the sampling point identifies the IP address or addresses of the source of the sampled media and if the usage conditions indicate that the source is not a permitted source for the content, the IP address or addresses of the source of the sampled media are supplied to a network provider.
The system of the invention can therefore identify content and determine the context of the use of the content. This in turn enables the delivery of a policy. The policy, can control access to the content and can also control behaviour of a content player.
It can further trigger notifications to a consumer such as where content can be legally accessed or alternatively, product tie-ins such as they can get a drink in the same coffee bar as the person in the content they are viewing.
Exemplary embodiment of the invention will now be described in greater detail with reference to the drawings in which: Fig. I shows schematically the system of the invention; Fig. 1a shows schematically the policy management interfaces; Fig. 2 shows schematically the usage context of streamed content; Fig. 3 shows Context Based Policy Selection; Fig. 4 shows a pirate content decision tree; Fig. 5 shows a permissible content decision tree.
Fig. 6 shows the sequence diagram for anonymising samples & notifications Fig. 7 shows the sequence diagram for collecting community decoy samples Fig. S shows the sequence diagram for donating community decoy samples Figure 1 shows schematically the context-based online media policy platform. The Context-based Online Media Policy (COMP) platform can be used by both device and network sampling and enforcement points. The platform incorporates the databases and orchestration functions for content identification, determining usage context and servicing the designated policies to be enforced. The platform further comprises a digital watermark federator function, which is adapted to communicate with a plurality of content watermark vendors so as to be able to identify watermarks in an analysed piece of digital content.
The Device-based Sampler and Enforcement Point (DSEP) is a software application that runs on consumer network devices such as residential gateways such as routers and consumption devices such as computers, set top boxes, game consoles, smart N's or tablets that are used to browse and consume digital assets.
The Network-based Sampler and Enforcement Point (NSEP) is an application that consumes information flows from a Deep Packet Inspection (DPI) probe located in a Broadband providers access network.
Rights holders, Distributors, Broadband service providers and account holders can define, configure and manage policies on an individual basis via web based interfaces. Web Services APIs are provided for integration with Rights holder compliance systems.
Streaming Content and its Usage Context Figure 2 shows schematically the usage context of streamed content. The identity of the content being streamed and the context under which it is being consumed are the critical items of information used to determine the policies to be applied. Content Sampling and the federated use of commercially available automatic content recognition services are used to identify the content being streamed. Such services are provided by companies such as Gracenote®, Civolution® and Digimarc® Simultaneous sampling of usage conditions is used to establish the full usage context of the content being consumed.
When inspecting each packet the sampler looks for a Media type PDU (Protocol Data Unit). A PDU is analogous to an envelope or package that contains an item.
Sometimes the item is a piece of paper with a message or a physical item such as a glass and at other times the item is another envelope. In the same way that an envelope has an address that tells the mail service where to deliver the envelope and the type of content that is contained inside (glass, paper, etc) a PDU has a header called a packet header that tells the device where to send the content who sent the content and the kind of content inside (called the payload). The packet header is how * 5 it can be determined that the payload is a media payload as opposed to another payload such as a text message payload or a voice call payload.
In use, the sampler first looks for media PDUs and inspects their payload. If it cannot determine the PDU type it looks inside the PDU to try and do further pattern matching to see if it is a media PDU that has somehow been hidden. It then takes the payload of several media PDUs and assembles them into a contiguous stream that becomes the media sample.
Automatic Content Recognition services use samples of the content being streamed to detect any embedded watermarks and fingerprints. Detected watermarks and fingerprints are used to reference a database, which in turn provides the streaming media's unique identity. Even in circumstances where the streaming content is user generated or not registered as copyrighted material, the absence of any detectable watermarks is still of significance as it indicates that the content is of unknown origin.
A captured series of data packets, that has either been copied or intercepted, is.
processed to recover the embedded media stream into its pre-packetized and pre-encoded form. Discrete fragments of the media stream are temporarily stored and tagged with metadata that identifies them as being samples from the stream of data packets that they were taken from.
Sampling can be performed at several locations: -On the play-out device -through interception of data packets before they reach the media player or in the network interface via the Device-based Sampling and Enforcement Point (DSEP) -In the access network via Deep Packet Inspection as the content is being streamed to an end user via the Network-based Sampling and Enforcement Point (WSEP) -In a consumer's network device such as a residential gateway or media server via a DSEP -At the Internet peering point via Deep Packet Inspection as the content is being streamed into the ISP's network via an NSEP.
Recovery of the media stream is dependent on: -The correct temporal reassembly of the data packets -Extraction of the combined data packet payload into a contiguous data stream -Decoding of the data stream into a media stream that can be used for comparison and matching purposes.
As there is no guarantee that data packets will be received in temporal order or, in the case of P2P delivery, received from the same source, the captured data packets are first processed to recover their payload to reform the original data stream.
Popular P2P algorithms are used to identify the required data packets from the set of data packets that have been captured. The sampling of the media stream can be performed using established patterns or alternatively by leveraging or adapting a third party sampler such as an Anti-virus checker. The media stream that is sampled by the sampler can thus be either a stream of requested content or content that is played out from the device's storage such as disc, non volatile memory or an external storage device such as a memory stick. Some samplers, in particular anti-virus checkers, look at memory and data storage and so the sampler in this particular context could also effectively sample content that is held on the storage of the device.
The data stream is then further processed using known media encoding algorithms (HLS, Adobe Flash, Silverlight, DASH, etc.) to recover the transmitted or stored media stream.
The usage-context is a complete or partial time-stamped array of sampled usage conditions that have been determined through combining artefacts from the sampler and enforcement point and the context-based online media policy platform. Usage conditions are captured at the same time as the content stream is being sampled.
The usage-context array is a plurality of tuples with each tuple comprising a condition class, a condition attribute and an attribute value. The array is extensible to cater for new types of condition class, condition attribute and attribute value.
Usage contexts are recalculated at regular intervals or when any of the monitored conditions change. Typical condition classes, condition attributes and attribute value examples are shown but not limited to those listed in Table 1.
Condaion Class Condition Attribute Attnbute Value Example Sampler Point Network, Gateway, Sampllng point ID Consumption device, etc. Access Method Distributor Distributor identifier such as Amazon, Corncast, etc. Delivery network Network type -waed garden, Internet Media stream source IPv4 Address address List or class of content IDs that are prohibited from being streamed from this address, Media stream destination List or class of content IDa address that are prohibited from being streamed to this address.
cv) Media Consumption device P STB, PC, Game Console, Tablet, Smariphone, etc. Location of the consumption Country, City, GPS co-device ordinates, etc. C) Consumption device network Wired, Wi-Fi, Cellular connection type Account Consumer Globally Unique D (GUID), Enterprise email etc. Government Hospitality Consumption behaviour Consumed content ID, Consumed content categories, Content I purchased, etc. Configured consumption Content categories preferences Registered devices Game consoles, Tablets, I PCs Social connections Social network IDs Device Type Model Operating System Version Player Silverlight, Flash, etc. Audio and Sound resolution Display resolution Sound configuration Table I Usage Context Attributes A Device or Network Sampling and Enforcement Point use the assembled media stream fragments together with the sampled usage-context to form the arguments for a policy request that is sent to the COMP platform.
To protect the privacy of the end-user a DSEP is configured to anonymise the arguments of the request that are sent to the COMP platform. It does so by sending multiple decoy samples together with the locally taken media and context sample for processing, see Fig 6.
DSEPs use popular P2P protocols to donate and receive decoy samples as required, see Fig 7 and Fig 8. On continuous and random intervals each DSEP inserts locally taken content and media samples into its Decoy Donation Cache. At regular intervals a DSEP's Cache Manager notifies the COMP Decoy P2P Manager with the status of its Decoy Donation Cache. Status includes cache depth and whether there are any priority decoys that need sending.
A DSEP's Cache Manager continuously monitors its Local Decoy Cache status.
When it falls below a configured threshold it gets a list of donor DSEPs from the Decoy P2P Manager. The Decoy P2P Manager will provide a random list of donors and ensure donors with urgent samples are prioritised over other donors. The DSEP Cache Manager replenishes its Local Decoy Cache with a random numbers of decoy samples from the list of supplied donors.
When a DSEP is ready to send a media and context sample to the COMP for analysis, it takes one of several recently taken samples together with a random number of decoys from its Local Decoy Cache and inserts them into a sample container in a random order. This sample container forms the argument for the policy request that is sent to the COMP. The remaining local samples are put into the Decoy Donation Cache and are marked as high priority so they will be donated as soon as possible. Prior to sending the sample container to the COMP, the DSEP stores the checksum of the container and a hash of the locally taken media and context sample as a linked pair of attributes.
Having processed all of the media and context samples in the received container, the COMP returns a response container that comprises notifications and policy responses for each of the received samples arranged in a random order. A hash for each of the original media and context samples is also inserted into the response container preceding each of their corresponding notification and policy responses.
The header of the response container contains the checksum of the original media and context sample container. Having received the response container the DSEP uses the checksum in the response containers header to retrieve the locally stored hash of the original sample. Using the hash, the DSEP extracts the notification and policy that corresponds with the original sample it sent for analysis. The other responses are stored temporarily should there be a need to request additional policy information whilst maintaining privacy.
The COMP platform uses the received media fragment to make a Content ID request to one or more content recognition services via the COMP digital watermark federation service. If the content can be identified, its unique identity is returned by the digital watermark federation service. If the content cannot be identified the returned value is UNKNOWN_CONTENTJD.
Before declaring the content to be unknown, the COMP platform submits alternative versions of the sample for identification. These versions include but are not limited to separating the audio and video streams, applying noise reduction methods and inverting the image and reattaching it to the audio.
Figure 3 shows a context based policy detemiination, in which a two-stage process is used to determine the policy to be that is to be returned to the requesting Sampling & Enforcement Point.
In the first step the system of the invention takes the Content's Identity and Access conditions and uses a discriminator function to determine the Content Access Status of the streaming content. In a second step the system then feeds the Access Status, together with the Content Identity, Full Usage Context and Content Class into a Decision tree to determine the policy that is most relevant to the sampled media stream and its usage context.
The purpose of the discriminator (Function 1) is to determine a media stream sample's Content Access Status. The Content Access Status is the real-time condition that a content item is determined as being consumed under.
The discriminator uses the Content ID to retrieve the content's access rules and their test precedence from the content access rules database. It then uses the sampled media stream and the access conditions extracted from the sampled usage conditions to test each access condition against the access rules in precedence order, see Table 2 for an example of an access rule definition where the content has been identified.
I Test* Test Focus Check Test Result Subject * Content Audh and Correct * Vt' V V V V V ntegnty Video sound track, V or or or or or or or or no video ? ? ? ? ? 2? inversion Media P Address Backhsthg
V V V V V V
stream & V or -or or or or or or source whitebsbng ? 2 ? ? ? 2 ? address Distributor IP Address Blacklisting V V V V V Vor & or or or or or 2 whitelisting ? ? ? ? ? Media Distributor Blacklisting V V V V stream name & V or rf\ or or or or -destination whitelistinq ? 2 ? ? I address Consumpti Country via Biacksflng V V V * on device location & V or C) location service / whitelisting or or or Delivery Private Blackhsting V V or network Network or & or or Internet whitelisting ? ? Consumpti Tablet, PC, Blacklisting V v or on device Internet TV & or whitelishng ? Network Wi-H, B!ackhsting V or connection Ethernet, & type Wireless whitebsting * Content Access I
-
Status $ -s < -2 P Table 2 Discriminator Access Rule Tests When a test definitively passes or if a pass or fail cannot be determined, the discriminator moves on to the next test. When all tests pass or when all tests are inconclusive or when all tests are a mixture of passed or inconclusive then the Content Access Status is set to PERMISSIBLE. When a test definitively fails the corresponding Content Access Status is set. The value returned by a definitively failed test can be, but is not limited to, PIRATED, GEO_RESTRICTED, NETWORK_RESTRICTED or ACCESS_RESTRICTED.
IP address that repeatedly show up as being a source for PIRATED content are used to both update the blacklist database and provide a real-time feed of IP address which a broadband ISP can choose to block from feeding their network or perform further investigation and take corrective action. It should be noted in this context that Pirated P2P content distribution is often associated with the distribution of maiware, in particular bats that can be turned on or off remotely. The users computer can therefore unwillingly become part of a botnet and they may be unaware that their IP address is used to distribute pirated content.
If the Content ID input to the Discriminator is set to UNKNOWN_CONTENT_ID, the discriminator sets the Content Access Condition to UNKNOWN_CONTENT.
The Content Access Status, determined by the discriminator, together with the content ID, content class and full usage-context conditions, are used by the Hyper Heuristic Decision Tree (Function 2) to determine the policy to be returned to the requesting Sampling and Enforcement point. The user profile database is referenced by function 2 when the identity of the customer forms part of the full usage context and decisions are based on user preferences or past consumption behaviour. It would be possible to use to decision trees other than hyper heuristic decision trees Figure 4 shows an example of a decision free for pirated content where decisions need to be made as to whether to prosecute an offender or give them another chance, and if giving another chance whether to incentivize with a discount.
Attributes from the sampled usage context are combined with historic activity data to form decision tree events that are used to determine the appropriate decision tree path. In this example the historic frequency of the subscribers IP address having been observed consuming illegal content is used to enable decisions to be made on the basis of the offender being a first time offender, low volume offender or high volume offender. Local databases containing availability status of the desired content and if known the subscription status of the consumer are used to form additional events to enable decisions to be made as to whether to educate, redirect or incentivise the consumer to change their behaviour. For example, a user illegally downloading a film or movie from a peer-to-peer system could be quickly pushed a pop-up message with links to purchase or rent the same content, or advising whether the title in question exists on the video on demand library of a participating distributor's own broadband network or on a third-party seller like Amazon.
Figure 5 shows a decision free for permissible content.
The decision tree uses inputs known as events to make decisions. Nodes shown as circles are events and nodes shown as squares are decisions. Events can be whether an identified address has been seen multiple times or whether content is available from another source. Decisions are selecting from one or more alternatives based on a desired outcome that is usually expressed as a probability Configuration of policies and decision trees that are applicable to specific content titles is descri bed below and shown in Figure Ia. A precedence value is assigned to all policies to resolve conflicts when more than one relevant policy is identified as being suitable for the content and circumstances under which it is being consumed.
Even when the content sample is unable to be identified and the Content Access Status is set to UNKNOWN_CONTENT a decision tree is still applied to select an appropriate policy based on the full usage context conditions.
Having determined the policy to be applied, the COMP platform retrieves the enforcement instructions from the content policy database and any relevant data needed for instruction execution from the application and user databases and passes them to the requesting sampling and enforcement point as policy enforcement metadata.
Multiple classes of enforcement instruction are combined to create the policy that is to be enforced. Table 3shows the enforcement instruction classes and their usage.
Enforcement Class Usage User is Example aware i Do Nothing e No action No Not appilcable required Stealth * Redirection No Modifying the CON Usage metric pointer to a less collection congested node.
Recording usage context and content consumed Blocking * Stopping Yes Disrupting the flow consumption of lP packets C') Information * Awareness Yes * Notification of r * Promotion promotions and * Education events interactive * Changing Yes * Gamification behaviour * Perform activity * Reduce
Table 3
An example of how a combination of enforcement classes can be used to reduce online piracy is illustrated as follows: When a known illegal content address or pirated online content is identified, access is prohibited and the user is directed to locations where legal versions of their desired content can be found.
It would also be possible to use gaming techniques to enable consumers to earn reward-points for continued use of the platform to access permissible content. Users can earn additional reward-points through interaction with platform services that are provided or sponsored by content owners and third parties such as advertisers, etc. Reward-points can be exchanged for either contributed items or additional platform services, see below: -Contributed rewards such as o Exclusive access to online events and services o Opportunities to: -Purchase tickets to exclusive artist events * Get priority access to public artist events o Free or discounted merchandising -Service rewards such as o Unified Catalogue for devices, subscriptions and digital lockers o Federated recommendations from the top curators and social media o Searches across multiple OTT vendors The COMP platform collects and stores usage metadata at ingress of sample data; egress of policies to be applied and at each processing stage leading to the determination of the context-based online media policy to be applied.
To enable correlation of collected events, the COMP platform attaches a Globally Unique ID (GUID) to all samples that it receives. This GUID is maintained at all COMP processing stages and is also embedded as an attribute of any policies returned. This mechanism facilitates correlation of all processing activities associated with a given sample for the purpose of providing: * An audit trail of actEvities * Feeding heuristic algorithms used by the Decision Tree functions * Building databases for use in subsequent processing of content and usage-context samples.
Examples of stored metadata include but are not limited to: -Time stamped metadata for received media stream samples and their full usage context -Content ID and Content Access Condition -Policy returned -Aggregates of collected data, such as o History of Content Access Conditions by IP address -used by decision trees to determine repeated scenarios such as absence or presence of repeated piracy o History of location by User ID -used to offer incentives for visiting local attractions, stores, cafés, etc. Usage metadata is also collected and stored during application of policies that require interaction with the COMP platform. Again the GUID, first applied to the sample received by the COMP platform and embedded in the returned policy, is maintained during these interactions to enable correlation of metrics over the full sample to policy enforcement.
Metadata is collected as opposed to capturing sampled data to protect privacy of individuals and to reduce the amount of data stored.
Operation using a Device-based Sampler and Enforcement Point Device Installation & User Configuration To install the system on a device, a user downloads the applicable version of the Device-based Sampler and Enforcement Point (DSEP) for the device they are using.
It would alternatively be possible for the DSEP to be pre-instalted.
If no account. exists, the user is required to create an account via the device user interface or a web browser connected to the Internet. Once signed in, the User is able to register their devices and digital lockers and also configure system preferences, which include but are not limited to: a. Interests for awards b. Parental controls In use, a device with the DSEP installed requests content from a site serving online digital media. The DSEP samples the requested URL and IP address and makes a request to the COMP platform for any registered policies to be enforced when these destinations are used.
If a known URL or IP address is detected the relevant policy is returned. Unless a policy redirecting or preventing access to the requested location is returned, the requested site's web page or storefront is presented. Having browsed the available digital assets the user selects a digital asset to consume and the site begins streaming media to the device.
The OSEP samples and captures the received data packets. The DSEP also collects as many usage context attributes as it is able to identify and builds a usage context array. The captured data packet payloads are reassembled into media stream fragments, compressed and sent to the COMP Platform together with the usage context array.
The COMP platform attempts to ascertain the sampled content's unique ID via its watermark and fingerprint federation service. The response is either the content's globally unique identifier as defined by the relevant industry organization such as MovieLabs or Unknown Content ID. Unknown Content ID may be returned because the sampled item is user generated content or not registered with any of the watermark ard fingerprint databases used by the COMP platform.
The response from the watermark federation service is passed to the Discriminator (Function 1) together with the available access conditions that have been selected from the usage context sent with the sample from the DSEP. The Discriminator uses the Content ID to retrieve the access rules and the test precedence for the access conditions associated with the identified content from the content access rules database.
Using the sampled media stream and the access conditions extracted from the usage context array, the Discriminator tests each access condition against the access rules in the order specified by the identified content's precedence selling to determine the Access Status of the streamed content. The value of the Access Status can be, but is not limited to, PIRATED, GEO_RESTRICTED, NETWORK ESTRICTED, ACCESS_RESTRICTED, PERMISSIBLE or UNKNOWN_CONTENT The newly derived Content Access Status, the Content ID, Full Usage Context and the Content Class are fed into the Hyper Heuristic Decision tree. If a user ID forms part of the full usage context, the user's profile is retrieved from the User Profile database.
The Content Access Status, Content ID and full usage context are used to select the appropriate decision tree for policy selection. The selected decision tree uses all of the available input data and the policy database to determine the policy or policies that best match the circumstances under which the content is being consumed.
When more than one policy is available, the policy with the highest precedence is selected. The selected policy is sent to the requesting DSEP. The policy returned to the DSEP can be a sequence of instructions to perform or a URL to a location to retrieve the instructions. Having received the instructions or instruction URL for the content being consumed the DSEP executes them.
PIRATED instructions can be, but are not limited to, displaying a message advising the user that the content they are consuming is illegitimate and optionally: o Advising the reason for the content's illegitimacy o Providing a list of locations where the content can be obtained legally and where applicable the costs for doing so.
o Stopping the content playing RESTRICTED instructions can be, but are not limited to, displaying a message advising the user that the content they are consuming is unauthorized and optionally: o Advising the reason for the lack of authorization o Providing a list of locations where legal and authorized versions of the content can be obtained and where applicable the costs for doing so.
o Stopping the content playing PERMISSIBLE instructions can be, but are not limited to: o Notifying the user of * Mechanisms to interact with the content they are consuming * A promotion or event related to the content they are consuming * Related content that they may like to consume * The number of reward points they will earn for various percentages of content consumption o Asking the user if they wish to share their activity via a social network o Updating a local history of the items the user has consumed and the location it was consumed at.
Using one of the OSEP menus, Consumers access COMP platform services that are native to the platform or that have been provided by third parties such as Content Owners, Advertisers and third party developers. Native services can include recommendations and library unification services. Third party services are on-boarded to the COMP platform from internal and external contributors. Usage of platform services can earn consumers additional reward-points Each user earns rewards points every time they consume permissible content via the DSEP interface. Additional rewards points may be earned through interaction with a COMP platform service or application that itself may have been sponsored or provided by a content owner, advertiser or third party application. Service rewards are made available via the DSEP user interface.
Access to each service reward requires the use of rewards points that have been earned for using the COMP platform and for consuming legitimate content.
The user is able to browse a library of available service rewards and assign the points that they have earned to enable the service rewards they wish to make use of.
Third party developers are able to contribute new services for inclusion in the service reward library. Contributed rewards are items that can be claimed via the DSEP user interface.
Claiming of a contributed reward requires the use of rewards points that have been earned for using the service and for consuming legitimate content.
The user is able to browse a library of contributed rewards and spend the points that they have earned to claim the contributed reward that they wish to have.
Operation using a Network-based Sampler and Enforcement Point Deep Packet Inspection Engine Attachment The Network-based Sampling and Enforcement Point (NSEP) is an application layer service running on or attached to a Deep Packet Inspection (DPI) engine located in the path of a Broadband Access Network.
The DPI is configured to pass all media traffic and traffic that is unable to be determined to the NSEP.
A user's device requests content from a site serving online digital media. The NSEP samples the requested URL and IP address and makes a request to the COMP platform for any registered policies to be enforced when these destinations are used.
If a known URL or IP address is detected the relevant policy is returned. Unless a policy redirecting or preventing access to the requested location is returned, the requested site's web page or storefront is returned to the user device. Having browsed the available digital assets the user selects a digital asset to consume and the site begins streaming media to the device via the DPI system The NSEP samples and captures the received data packets. The NSEP also collects from the DPI engine as many usage context attributes as it is able to identify and builds a usage context array. The captured data packet payloads are reassembled into media stream fragments, compressed and sent to the COMP Platform together with the usage context array.
The COMP platform attempts to ascertain the sampled content's unique ID via its watermark and fingerprint federation service. The response is either the content's globally unique identifier as defined by the relevant industry organization such as MovieLabs or Unknown Content ID. Unknown Content ID may be returned because the sampled item is user generated content or not registered with any of the watermark and fingerprint databases used by the COMP platform.
The response from the watermark federation service is passed to the Discriminator (Function 1) together with the available access conditions that have been selected from the usage context sent with the sample from the NSEP. The Discriminator uses the Content ID to retrieve the access rules and the test precedence for the access conditions associated with the identified content from the content access rules database.
Using the sampled media stream and the access conditions extracted from the usage context arra9, the Discriminator tests each access condition against the access rules in the order specified by the identified content's precedence setting to determine the Access Status of the streamed content. The value of the Access Status can be. but is not limited to, PIRATED, GEO_RESTRICTED, NETWORK_RESTRICTED, ACCESS_RESTRICTED, PERMISSI BLE or UNKNOWN_CONTENT The newly derived Content Access Status, the Content ID, Full Usage Context and the Content Class are fed into the Hyper Heuristic Decision tree. The Content Access Status, Content ID and full usage context are used to select the appropriate decision tree for policy selection.
The selected decision tree uses all of the available input data and the policy database to determine the policy or policies that best match the circumstances under which the content is being consumed. When more than one policy is available, the policy with the highest precedence is selected.
The selected policy is sent to the requesting NSEP. The policy returned to the NSEP pan be a sequence of instructions to perform br a URL to a location to retrieve the instructions.
Having receiyed the instructions or instruction URL for the content being consumed the NSEP executes them via the DPI API.
PIRATED and RESTRICTED instructions can be, but are not limited to: o Blocking the content o Degrading the content's audio or video quality o Replacing the content with a message advising the user that: * The content they are consuming is illegitimate * Advising the reason for the content's illegitimacy -Providing a list of locations where the content can be obtained legally and where applicable the costs for doing so.
Stopping the content playing PERMISSIBLE instructions can be, but are not limited to: a Overlaying a message on the content notifying the user of: Mechanisms to interact with the content they are consuming * A promotion or event related to the content they are consuming The media policy platform provides a user interlace for authorized rights holders, distributors, broadband service providers and consumers alike to define policies (the conditional actions) to be enforced following detection of a media item's identifier and the usage-context, i.e. an event trigger. This is shown in Figure Ia below.
Event triggers and their policies are used to make content more immersive, reduce online piracy and protect broadband providers and consumers from rights holder litigation.

Claims (16)

  1. Claims 1. An online media policy platform comprising a policy engine, a sampler adapted to sample content of digital media, the sampler being further adapted to request policies from the policy engine that are applicable to the requested content, the platform further comprising a discriminator adapted to determine a condition that the content is to be consumed under, the platform further comprising a decision tree, which decision tree uses inputs from the discriminator to determine which policy to apply to the content, wherein the platform retrieves enforcement instructions from the policy engine and enforces the policy.
  2. 2. An online media policy platform according to Claim 1, wherein the platform is adapted to determine the identity of sampled content and access rules applicable to the identified content.
  3. 3. An online media policy platform according to Claim 1 or Claim 2, wherein the discriminator determines an access status for the content.
  4. 4. An online media policy platform according to Claim 3, wherein the access status, the identity of the content and usage context conditions are fed to a decision tree to determine the policy to be applied.
  5. 5. An online media policy platform according to Claim 4, wherein the policy is returned to the sampler for enforcement.
  6. 6. An online media policy platform according to any one of Claims 1 to 5, wherein the sampler comprises a device or network sampling and enforcement point.
  7. 7. An online media policy platform according to any one of Claims 1 to 6, wherein the platform is adapted to control access to the media.
  8. 8. An online media policy platform according to Claim 7, wherein access to the media is blocked.
  9. 9. An online media policy platform according to Claim 8, wherein the user is diverted from the source of the sampled content to an alternative source of the content.
  10. 10. An online media policy platform according to any one of Claims 1 to 9, wherein the platform is adapted to trigger events related to the media.
  11. 11. An online media policy platform according to any one of Claims 1 to 10, wherein the platform is adapted to enforce a policy at an enforcement point in a network or on a device.
  12. 12. An online media policy platform according to any one of Claims 1 to 11, wherein the sampler is adapted to sample a stream of requested content.
  13. 13. An online media policy platform according to any one of Claims 1 to 12, wherein the sampler is adapted to sample content being played out from a device.
  14. 14. An online media policy platform according to any one of Claims I to 12, wherein the sampler is adapted to sample content stored on a device.
  15. 15. A method of controlling access to online media comprising the steps of sampling the media at a sampling point, identifying the content of the media, determining any usage conditions on the content of the media and controlling further access to the media through the sampling point in dependence on the content of the media and usage conditions.
  16. 16. A method of providing a feed of source addresses of malware sites comprising the steps of sampling the media at a sampling point, identifying the content of the media, determining any usage conditions on the content of the media, wherein the sampling point identifies the lP address or addresses of the source of the sampled media and if the usage conditions indicate that the source is not a permitted source for the content, the IP address or addresses of the source of the sampled media are supplied to a network provider.
GB1314833.3A 2012-08-20 2013-08-19 Online media policy platform Withdrawn GB2505777A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GBGB1214808.6A GB201214808D0 (en) 2012-08-20 2012-08-20 Online media policy platform
GBGB1217967.7A GB201217967D0 (en) 2012-08-20 2012-10-08 Online media policy platform

Publications (2)

Publication Number Publication Date
GB201314833D0 GB201314833D0 (en) 2013-10-02
GB2505777A true GB2505777A (en) 2014-03-12

Family

ID=47017026

Family Applications (3)

Application Number Title Priority Date Filing Date
GBGB1214808.6A Ceased GB201214808D0 (en) 2012-08-20 2012-08-20 Online media policy platform
GBGB1217967.7A Ceased GB201217967D0 (en) 2012-08-20 2012-10-08 Online media policy platform
GB1314833.3A Withdrawn GB2505777A (en) 2012-08-20 2013-08-19 Online media policy platform

Family Applications Before (2)

Application Number Title Priority Date Filing Date
GBGB1214808.6A Ceased GB201214808D0 (en) 2012-08-20 2012-08-20 Online media policy platform
GBGB1217967.7A Ceased GB201217967D0 (en) 2012-08-20 2012-10-08 Online media policy platform

Country Status (1)

Country Link
GB (3) GB201214808D0 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090196465A1 (en) * 2008-02-01 2009-08-06 Satish Menon System and method for detecting the source of media content with application to business rules
WO2009126829A2 (en) * 2008-04-09 2009-10-15 Level 3 Communications, Llc Rule-based content request handling
US20120011560A1 (en) * 2010-07-07 2012-01-12 Computer Associates Think, Inc. Dynamic Policy Trees for Matching Policies

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090196465A1 (en) * 2008-02-01 2009-08-06 Satish Menon System and method for detecting the source of media content with application to business rules
WO2009126829A2 (en) * 2008-04-09 2009-10-15 Level 3 Communications, Llc Rule-based content request handling
US20120011560A1 (en) * 2010-07-07 2012-01-12 Computer Associates Think, Inc. Dynamic Policy Trees for Matching Policies

Also Published As

Publication number Publication date
GB201217967D0 (en) 2012-11-21
GB201314833D0 (en) 2013-10-02
GB201214808D0 (en) 2012-10-03

Similar Documents

Publication Publication Date Title
US10523986B2 (en) Methods for identifying, disrupting and monetizing the illegal sharing and viewing of digital and analog streaming content
US9646140B2 (en) Method and apparatus for protecting online content by detecting noncompliant access patterns
US9124650B2 (en) Digital rights management in a mobile environment
US11824946B2 (en) Systems and methods for distributing content
CN105144725B (en) System and method for telescopic content delivery network request processing mechanism
JP5555271B2 (en) Rule-driven pan ID metadata routing system and network
JP4806816B2 (en) Techniques for watermark embedding and content delivery
US8464066B1 (en) Method and system for sharing segments of multimedia data
RU2463717C2 (en) Remote data accessing methods for portable devices
US20100250704A1 (en) Peer-to-peer content distribution with digital rights management
US20110197237A1 (en) Controlled Delivery of Content Data Streams to Remote Users
US10693839B2 (en) Digital media content distribution blocking
US20110126018A1 (en) Methods and systems for transaction digital watermarking in content delivery network
US11032625B2 (en) Method and apparatus for feedback-based piracy detection
WO2011041916A1 (en) Digital rights management in a mobile environment
CN103229186A (en) DRM service providing method and device
Price Sizing the piracy universe
US20060167813A1 (en) Managing digital media rights through missing masters lists
WO2007139277A1 (en) Method for executing digital right management and tracking using characteristic of virus and system for executing the method
WO2006069394A2 (en) Managing digital media rights through missing masters lists
KR20110058880A (en) Method enabling a user to keep permanently their favourite media files
US20140053233A1 (en) Online media policy platform
JP2016524732A (en) System and method for managing data assets associated with a peer-to-peer network
KR102213373B1 (en) Apparatus and method for blocking harmful contents using metadata
JP2008305371A (en) Apparatus and method for checking huge amount of content by distributed processing, and content delivery system for controlling autonomous content distribution and content use among users according to content check result

Legal Events

Date Code Title Description
WAP Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1)