WO2021033853A1 - Method and system for processing content having sensitive portion - Google Patents

Method and system for processing content having sensitive portion Download PDF

Info

Publication number
WO2021033853A1
WO2021033853A1 PCT/KR2020/000652 KR2020000652W WO2021033853A1 WO 2021033853 A1 WO2021033853 A1 WO 2021033853A1 KR 2020000652 W KR2020000652 W KR 2020000652W WO 2021033853 A1 WO2021033853 A1 WO 2021033853A1
Authority
WO
WIPO (PCT)
Prior art keywords
content
sensitive
tile
processor
sensitive portion
Prior art date
Application number
PCT/KR2020/000652
Other languages
French (fr)
Inventor
Rahul Aggarwal
Manoj Kumar
Original Assignee
Samsung Electronics Co., Ltd.
Delhi Technological University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co., Ltd., Delhi Technological University filed Critical Samsung Electronics Co., Ltd.
Publication of WO2021033853A1 publication Critical patent/WO2021033853A1/en

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/23418Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/234345Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements the reformatting operation being performed only on part of the stream, e.g. a region of the image or a time segment
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/262Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists
    • H04N21/26258Content or additional data distribution scheduling, e.g. sending additional data at off-peak times, updating software modules, calculating the carousel transmission frequency, delaying a video stream transmission, generating play-lists for generating a list of items to be played back in a given order, e.g. playlist, or scheduling item distribution according to such list
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4318Generation of visual interfaces for content selection or interaction; Content or additional data rendering by altering the content in the rendering process, e.g. blanking, blurring or masking an image region
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/4508Management of client data or end-user data
    • H04N21/4532Management of client data or end-user data involving end-user characteristics, e.g. viewer profile, preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • H04N21/4542Blocking scenes or portions of the received content, e.g. censoring scenes

Definitions

  • the present disclosure relates to process of content having sensitive portion and more particularly, relates to systems and methods for tagging and rendering content having sensitive portion.
  • Internet connectivity has allowed viewers across the globe to connect with content providers from different parts of the world to view all sorts of content. This flexibility of course offers a wider exposure to a viewer to understand and experience different cultures around the world. On the other hand, from the perspective of content providers, they also get to cater to a large pool of audience belonging to different geographic locations, which is at least a financial advantage for them.
  • the content may include sensitive portion, such as violence and disturbing imagery, nudity and sexually suggestive content, and portrayal of dangerous activities. Accordingly, there are various factors, such as age restrictions, territorial laws, and content provider policies, which govern the type of content that can be rendered to a viewer. Therefore, it is not possible to render the same content to different viewers across the globe.
  • the content providers either completely remove the content or selectively render the content to a predefined section of viewers meeting the mandatory requirements, such as age.
  • the content providers may censor the content, for example, by blurring a specific scene or by removing few sequences from the content.
  • an embodiment herein discloses a content tagging system for tagging a content having sensitive portion.
  • the content tagging system comprises a memory, and at least one processor coupled to the memory.
  • the at least one processor configured to receive a video content having at least one sensitive portion, wherein the video content includes a plurality of frames; encode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame; categorize at least one tile including the sensitive portion into at least one category among a plurality of categories of sensitive content; and tag the at least one tile with an indication related to the at least one category of the sensitive content.
  • an embodiment herein discloses a content rendering system for rendering content having sensitive portion.
  • the content rendering system comprises a memory and at least one processor coupled to the memory.
  • the at least one processor configured to receive a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content; decode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content; modify the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and render the video content with the modified at least one tile.
  • an embodiment herein discloses a method for tagging a content having sensitive portion.
  • the method comprises receiving, by at least one processor of a content tagging system, a video content having at least one sensitive portion, wherein the video content is formed of a plurality of frames; encoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame; categorizing, by the at least one processor, at least one tile including the sensitive portion into at least one category a plurality of categories of sensitive content; and tagging, by the at least one processor, at least one tile with an indication related to the at least one category of the sensitive content.
  • an embodiment herein discloses a method for rendering content having sensitive portion.
  • the method comprises receiving, by at least one processor of a content rendering system, a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content; decoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content; modifying, by the at least one processor, the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and rendering, by the at least one processor, the video content with the modified at least one tile.
  • Figure 1 illustrates an environment implementing a content tagging system and a content rendering system, according to an embodiment of the present disclosure
  • Figure 2 illustrates a block diagram of the content tagging system, according to an embodiment of the present disclosure
  • Figure 3 illustrates a block diagram depicting tagging of sensitive portion of a content by a tagging module of the content tagging system, according to an embodiment of the present disclosure
  • Figure 4 illustrates a block diagram of the content rendering system, according to an embodiment of the present disclosure
  • Figure 5 illustrates a block diagram depicting an example of notifying a user of the presence of sensitive portion in a subsequent frame of the content, according to an embodiment of the present disclosure
  • Figure 6 illustrates an example of tagging of sensitive portion of the content, based on a user instruction, according to an embodiment of the present disclosure
  • Figure 7 illustrates a block diagram depicting operations of the content tagging system and the content rendering system, according to an embodiment of the present disclosure
  • Figure 8 illustrates a method for tagging a content having sensitive portion, according to an embodiment of the present disclosure.
  • Figure 9 illustrates a method for rendering content having sensitive portion, according to an embodiment of the present disclosure.
  • Figure 1 illustrates an environment 100 implementing a content tagging system 102 and a content rendering system 104, according to an embodiment of the present disclosure.
  • the environment 100 may include a server 106 of a content provider in communication with a client device 108.
  • the server 106 and the client device 108 may be in communication with each other over a network 110.
  • the network 110 may be a wired network or a wireless network.
  • the server 106 is shown to be in communication with only one client device 108 in the illustrated embodiment. In other embodiments, the server 106 may be in communication with a plurality of client devices 108, without departing from the scope of the present disclosure.
  • the content tagging system 102 may be installed at the server 106 whereas the content rendering system 104 may be installed at the client device 108.
  • the content tagging system 102 may be configured to tag a content having at least one sensitive portion. Constructional and operational details of the content tagging system 102 are explained in the description of at least Figure 2.
  • the content rendering system 104 may be configured to render a content having tagged sensitive portion. Constructional and operational details of the content rendering system 104 are explained in the description of at least Figure 3.
  • FIG. 2 illustrates a block diagram of the content tagging system 102, according to an embodiment of the present disclosure.
  • the content tagging system 102 may include a processor 202, a memory 204, modules 206, and data 208.
  • the modules 206 and the memory 204 are coupled to the processor 202.
  • the processor 202 can be a single processing unit or a number of units, all of which could include multiple computing units.
  • the processor 202 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions.
  • the processor 202 is configured to fetch and execute computer-readable instructions and data stored in the memory 204.
  • the memory 204 may include any non-transitory computer-readable medium known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read-only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
  • volatile memory such as static random access memory (SRAM) and dynamic random access memory (DRAM)
  • DRAM dynamic random access memory
  • non-volatile memory such as read-only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
  • the modules 206 include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement data types.
  • the modules 206 may also be implemented as, signal processor(s), state machine(s), logic circuitries, and/or any other device or component that manipulate signals based on operational instructions.
  • the modules 206 can be implemented in hardware, instructions executed by a processing unit, or by a combination thereof.
  • the processing unit can comprise a computer, a processor, such as the processor 202, a state machine, a logic array or any other suitable devices capable of processing instructions.
  • the processing unit can be a general-purpose processor which executes instructions to cause the general-purpose processor to perform the required tasks or, the processing unit can be dedicated to perform the required functions.
  • the modules 206 may be machine-readable instructions (software) which, when executed by a processor/processing unit, perform any of the described functionalities.
  • the modules 206 may include a receiving module 210, an encoding module 212, a categorizing module 214, a tagging module 216, and a generating module 218.
  • the receiving module 210, the encoding module 212, the categorizing module 214, the tagging module 216, and the generating module 218 may be in communication with each other.
  • the data 208 serves, amongst other things, as a repository for storing data processed, received, and generated by one or more of the modules 206.
  • the receiving module 210 may be configured to receive a video content having at least one sensitive portion.
  • the video content may be formed of a plurality of frames.
  • the encoding module 212 may be configured to encode each frame of the video content in form of a plurality of tiles. Each tile may be indicative of a sub-frame.
  • the encoding module 212 may encode each frame in the form of the plurality of tile by High-Efficiency Video Coding (HEVC) standard technique.
  • HEVC High-Efficiency Video Coding
  • the HEVC coding is a Moving Picture Experts Group (MPEG)/ITU-T video coding standard capable of achieving more than 50% bitrate reduction compared to an Advanced Video Coding (AVC) standard at the same subjective quality.
  • MPEG Moving Picture Experts Group
  • AVC Advanced Video Coding
  • HEVC standard technique is developed particularly for Ultra-High Definition (UHD) videos. At that resolution, using the main profile and with today's HEVC encoders, a video stream may typically be delivered at a bit rate of 15-20 Mbps.
  • UHD videos or videos with a higher resolution may not be viewable by all devices, for example, owing to processing requirements, or stream-able through all networks, for example, owing to high bitrates.
  • the encoding module 212 may use the HEVC standard technique for tiling . The encoding of a frame into tiles makes it possible to encode different regions independently, within a single frame, while maintaining the use of a single decoder and efficient coding.
  • the encoding module 212 may perform motion-constrained tiling of the frames of the content.
  • the encoding module 212 may encode consecutive pictures in a video sequence such that a tile of a given picture only requires, if any, the coded information of the tiles at the same spatial position in the reference pictures, effectively creating tile tunnels in the video.
  • the tile tunnels may then be stored in ISO Base Media File Format (ISOBMFF) files and may be delivered separately.
  • ISOBMFF ISO Base Media File Format
  • a player may then decide whether to present one or more tile tunnels at a given time, based on user interaction.
  • Adaptive HTTP streaming technologies such as MPEG DASH, enable the live or on-demand delivery of media streams over HTTP.
  • the quality of the media delivered to the client may widely be adapted to user's preferences, such as language and point of view), to client's capabilities, for example, support for a given codec, and to fluctuating network conditions.
  • the categorizing module 214 may be configured to identify whether the each tile includes sensitive portion, and categorize at least one tile including the sensitive portion into at least one of a plurality of predefined categories of sensitive content.
  • the plurality of categories may be defined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions.
  • the tagging module 216 may be configured to tag at least one tile with at least one associated category of the sensitive content.
  • Figure 3 illustrates a block diagram 300 depicting tagging of the sensitive portion of a frame of the content by the tagging module 216, according to an embodiment of the present disclosure. As would be appreciated by a person skilled in the art, Figure 3 is included for providing better understanding of the present disclosure and therefore, should not be construed as limiting in any way.
  • the frame 302 is encoded in form of a plurality of tiles. As illustrated, the frame 302 is depicting a male smoking a cigarette. Further, the tagging module 216 may tag the at least one tile including the sensitive portion in the frame 302 with an indication related to the at least one associated category of the sensitive content. For example, the tagging module 216 may tag the tile having the cigarette with a category of the sensitive content, namely, "smoking". Therefore, the frame 302 is now tagged with the category "Smoking" indicative of the frame 302 having a smoking sequence.
  • the generating module 218 may be configured to generate a playlist file indicative of the sensitive content for the video content including the at least one tile tagged with the indication.
  • FIG. 4 illustrates a block diagram of the content rendering system 104, according to an embodiment of the present disclosure.
  • the content rendering system 104 may include a processor 402, a memory 404, modules 406, and data 408.
  • the modules 406 and the memory 404 are coupled to the processor 402.
  • the constructional and operational characteristics of the processor 402, the memory 404, the modules 406, and the data 408 may be similar to the processor 202, the memory 204, the modules 206, and the data 208 of the content tagging system 102. Therefore, for the sake of brevity, those details are not explained again in the description of Figure 4.
  • the modules 406 include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement data types.
  • the modules 406 may also be implemented as, signal processor(s), state machine(s), logic circuitries, and/or any other device or component that manipulate signals based on operational instructions.
  • the modules 406 can be implemented in hardware, instructions executed by a processing unit, or by a combination thereof.
  • the processing unit can comprise a computer, a processor, such as the processor 402, a state machine, a logic array or any other suitable devices capable of processing instructions.
  • the processing unit can be a general-purpose processor which executes instructions to cause the general-purpose processor to perform the required tasks or, the processing unit can be dedicated to perform the required functions.
  • the modules 406 may be machine-readable instructions (software) which, when executed by a processor/processing unit, perform any of the described functionalities.
  • the modules 406 may include a receiving module 410, a decoding module 412, a modifying module 414, a rendering module 416, an updating module 418, a detecting module 420, and a notification module 422.
  • the receiving module 410, the decoding module 412, the modifying module 414, the rendering module 416, the updating module 418, the detecting module 420, and the notification module 422 may be in communication with each other.
  • the data 408 serves, amongst other things, as a repository for storing data processed, received, and generated by one or more of the modules 406.
  • the receiving module 410 may be configured to receive a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of predefined categories of sensitive content.
  • the plurality of categories of the sensitive content may be determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions. Further, the receiving module 410 may receive the playlist file from the content tagging system 102.
  • the decoding module 412 may be configured to decode each frame of the video content in form of the plurality of tiles. Further, the modifying module 414 may be configured to modify the at least one tile having sensitive content, based on a predefined set of viewing policies for a user. In an embodiment, the rendering module 416 may be configured to render the video content with the at least one tile modified in-line with the predefined set of viewing policies.
  • the updating module 418 may be configured to update the predefined set of viewing policies for the user. Based on the updated set of viewing policies, the modifying module 414 may be configured to re-modify the at least one tile having the sensitive content.
  • the detecting module 420 may be configured to detect, while the video content is being rendered to the user, presence of a sensitive portion in one of subsequent frames of the video content. Further, the notification module 422 may be configured to notify the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user.
  • Figure 5 illustrates a block diagram 500 depicting an example of notifying a user of the presence of sensitive portion in a subsequent frame of the content, according to an embodiment of the present disclosure. As illustrated, the block diagram 500 includes 3 frames of the content, for example, at the 1st second, the 2nd second, and the 3rd second of the content. Further, the third frame of the content includes sensitive portion.
  • the detecting module 420 may detect the presence of the sensitive portion in the 3rd frame of the content. Accordingly, the notification module 422 may notify the user of the presence of the sensitive portion in the upcoming frames, for example, by prompting a message "Sensitive content in next 2 seconds".
  • the content rendering system 104 may receive a user instruction, while playing of the content.
  • the user instruction may be indicative of a tile of a frame having a sensitive content.
  • the content rendering system 104 may prompt the user to indicate a category of the sensitive content for associating the corresponding tile.
  • the content rendering system 104 may transmit such information to the content tagging system 102.
  • the content tagging system 102 may tag such tiles of the frames of the content with the user-selected category of sensitive portion. Accordingly, the content rendering system 104 may render the subsequent portions of the content, based on the tagging done based on the user instruction.
  • Figure 6 illustrates a block diagram 600 depicting an example of the tagging of sensitive portion of the content, based on user instruction, according to an embodiment of the present disclosure.
  • the user may select a tile of the frame for being tagged with a category of "Smoking".
  • any subsequent portion of the content showing smoking is automatically tagged by the content tagging system 102, and then censored by the content rendering system 104.
  • Figure 7 illustrates another environment 700 depicting operations of the content tagging system 102 and the content rendering system 104, according to another embodiment of the present disclosure.
  • the server 106 is shown to be in communication with the client device 108.
  • the client device 108 may include, but is not limited to, an adaptive streaming engine 702, the decoding module 412, and the rendering module 416.
  • the adaptive streaming engine 702 may further include, but is not limited to, a playlist parser module 704 and a tile stitching module 706.
  • the adaptive streaming engine 702 may request the server 106 to share a playlist.
  • the server 106 may provide the client device 108 with the playlist.
  • the playlist parser module 704 may retrieve information related to tiling of the frames of the content. For example, the information may relate to categories of sensitive content associated with each tile of each frame.
  • the adaptive streaming engine 702 may request the server 106 for sharing specific tiles of a frame. Further, the tile stitching module 706 may stitch the received tiles and share the stitched tiles with the decoding module 412. The decoding module 412 may decode the tiles and share the decoded tiles with the rendering module 416. The rendering module 416 may then render the content based on at least one sensitive content category. Further, a user equipment 708 may stream the content, based on at least one of content provider policies, OEM policies, and user policies.
  • the adaptive streaming engine 702 may download and parse the playlist of the content from the server 106, informing the client device 108 about the sensitive portion of each tile. Such information can then be matched with viewing policies and only the tiles that do not fall within the purview of any category of sensitive content may be played normally. Further, the tiles with sensitive portion may be censored based on the viewing policies before being shown to the user.
  • Figure 8 illustrates a method 800 for tagging a content having sensitive portion, according to an embodiment of the present disclosure.
  • the method 800 may be a computer-implemented method 800.
  • the method 800 may be executed by the processor 202 of the content tagging system 102. Further, for the sake of brevity, details of the present disclosure that are explained in details in the description of Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, and Figure 8 are not explained in detail in the description of Figure 8.
  • the method 800 includes receiving the video content having the at least one sensitive portion.
  • the video content may be formed of the plurality of frames.
  • the receiving module 210 of the content tagging system 102 may receive the video content. The receiving module 210 may be controlled by the processor 202.
  • the method 800 includes encoding each frame of the video content in form of the plurality of tiles. Each tile may be indicative of a sub-frame.
  • the encoding module 212 may encode each frame in form of the plurality of tiles. The encoding module 212 may be controlled by the processor 202.
  • the method 800 includes categorizing at least one tile including the sensitive portion into at least one of the plurality of predefined categories of sensitive content.
  • the categorizing module 214 may identify whether the each tile includes sensitive portion, and categorize each tile into the at least one of the plurality of predefined categories.
  • the categorizing module 214 may be controlled by the processor 202.
  • the method 800 includes tagging the at least one tile with an indication related to the at least one associated category of the sensitive content.
  • the tagging module 216 may tag the at least one tile with the indication, and be controlled by the processor 202.
  • the method 800 may include generating the playlist file indicative of sensitive content for the video content including the at least one tile tagged with the indication.
  • the generating module 218 may generate the playlist file, and be controlled by the processor 202.
  • Figure 9 illustrates a method 900 for rendering content having sensitive portion, according to an embodiment of the present disclosure.
  • the method 900 may be a computer-implemented method 900.
  • the method 800 may be executed by the processor 402 of the content rendering system 104. Further, for the sake of brevity, details of the present disclosure that are explained in details in the description of Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, and Figure 8 are not explained in detail in the description of Figure 9.
  • the method 900 includes receiving a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of predefined categories of sensitive content.
  • the method 900 may include receiving the playlist file generated by the content tagging system 102.
  • the plurality of categories may be defined based on at least one of the age-based restrictions, the demography-based restrictions, and the language-based restrictions.
  • the receiving module 410 of the content rendering system 104 may receive the playlist file. The receiving module 410 may be controlled by the processor 402.
  • the method 900 includes decoding each frame of the video content in form of the plurality of tiles.
  • Each tile may be indicative of a sub-frame. Further, at least one tile may be tagged with an indication related to the at least one associated category of the sensitive content.
  • the decoding module 412 may decode each frame of the video content, and be controlled by the processor 402.
  • the method 900 includes modifying the at least one tile having the sensitive content, based on a predefined set of viewing policies for the user.
  • the modifying module 414 may modify the at least one tile, and be controller by the processor 402.
  • the method 900 includes rendering the video content with the at least one tile modified in-line with the predefined set of viewing policies.
  • the rendering module 416 may render the video content with the modified at least one tile, and be controller by the processor 402.
  • the method 900 may include detecting, while the video content is being rendered to the user, presence of a sensitive portion in one of the subsequent frames of the video content. In an embodiment, the method 900 may include notifying the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user.
  • the present disclosure offers a comprehensive approach of tagging and rendering the content based on the presence of sensitive portion.
  • the proposed approach eliminates the need of manual intervention and high server storage cost in media content management.
  • the tiling of the frames of the content for being tagged with a respective category of the sensitive portion allows for saving of storage space and offer efficient following of the viewing policies.
  • the adaptability of the content tagging system 102 and the content rendering system 104 is client-driven, as the client seamlessly downloads and plays small continuous media files of a given quality and dynamically switches between alternative media files of different qualities offered by HTTP servers.
  • the client may only download the necessary quality given the situation based on adaptation configuration. Therefore, combining the concepts of tiling with adaptive HTTP streaming allows for new streaming possibilities.
  • a client can indeed select and adaptively stream only one or several tiles of interest to reduce the decoding requirements. Even when all tiles are transmitted, new adaptation configurations can be designed allowing the client to select different qualities per tile. This would reduce an overall bitrate while maintaining the best quality on the tiles of interest.
  • the proposed approach saves a lot of server storage space. Further, only MPD files need to be changed in order to accommodate any change in the viewing policies. Accordingly, the content will remain unaffected by the change in the viewing policies. Moreover, delivery of media content will be more flexible as the present disclosure provides multiple options for applying different policies. Therefore, the present disclosure offers the content tagging system 102, the content rendering system 104, the method 800, and the method 900 that are comprehensive, flexible, accurate, and intelligent.

Abstract

A method for tagging a content having sensitive portion is disclosed. The method includes receiving, by at least one processor of a content tagging system, a video content having at least one sensitive portion. The video content is formed of a plurality of frames. The method includes encoding, by the at least one processor, each frame of the video content in form of a plurality of tiles. Each tile is indicative of a sub-frame. The method includes categorizing, by the at least one processor, at least one tile including the sensitive portion into at least one category among a plurality of categories of sensitive content. The method includes tagging, by the at least one processor, the at least one tile with an indication related to the at least one category of the sensitive content.

Description

METHOD AND SYSTEM FOR PROCESSING CONTENT HAVING SENSITIVE PORTION
The present disclosure relates to process of content having sensitive portion and more particularly, relates to systems and methods for tagging and rendering content having sensitive portion.
Internet connectivity has allowed viewers across the globe to connect with content providers from different parts of the world to view all sorts of content. This flexibility of course offers a wider exposure to a viewer to understand and experience different cultures around the world. On the other hand, from the perspective of content providers, they also get to cater to a large pool of audience belonging to different geographic locations, which is at least a financial advantage for them.
It is relevant to ensure that such content is being responsibly rendered to the viewers so that it doesn't leave any undesired impression on them. For example, the content may include sensitive portion, such as violence and disturbing imagery, nudity and sexually suggestive content, and portrayal of dangerous activities. Accordingly, there are various factors, such as age restrictions, territorial laws, and content provider policies, which govern the type of content that can be rendered to a viewer. Therefore, it is not possible to render the same content to different viewers across the globe.
Now, in case the restrictions are to be imposed on such sensitive content based on the abovementioned factors, the content providers either completely remove the content or selectively render the content to a predefined section of viewers meeting the mandatory requirements, such as age. In order to restrict the viewing of banned content, the content providers may censor the content, for example, by blurring a specific scene or by removing few sequences from the content.
As would be gathered, different viewing policies are enforced in different regions and for different age groups. In fact, even within the same territory, multiple viewing policies exist for different communities or age groups. Therefore, it is not possible to generate a common version of the content that can be rendered to different groups of viewers. Therefore, for different set of viewing policies, the content providers will have to render different versions of the same content, which is a cumbersome and complicated task. Further, in case of any change in any viewing policy, the content providers will also have to remodel their entire rendering policies in order to accommodate the change. For example, the entire video content has to be regenerated again. This would of course demand a lot of media server storage for regeneration of each version of content based on a particular viewing policy. Such regeneration of new versions of the content would also require a lot of manual intervention, even in case of a minor policy change, which poses inconvenience as well. Moreover, owing to significant manual intervention, possibility of error would be higher too.
This disclosure is provided to introduce a selection of concepts, in a simplified format, that are further described in the detailed description of the invention. This summary is neither intended to identify key or essential inventive concepts of the invention and nor is it intended for determining the scope of the invention.
In accordance with an aspect of the present disclosure, an embodiment herein discloses a content tagging system for tagging a content having sensitive portion. The content tagging system comprises a memory, and at least one processor coupled to the memory. The at least one processor configured to receive a video content having at least one sensitive portion, wherein the video content includes a plurality of frames; encode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame; categorize at least one tile including the sensitive portion into at least one category among a plurality of categories of sensitive content; and tag the at least one tile with an indication related to the at least one category of the sensitive content.
In accordance with another aspect of the present disclosure, an embodiment herein discloses a content rendering system for rendering content having sensitive portion. The content rendering system comprises a memory and at least one processor coupled to the memory. The at least one processor configured to receive a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content; decode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content; modify the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and render the video content with the modified at least one tile.
In accordance with another aspect of the present disclosure, an embodiment herein discloses a method for tagging a content having sensitive portion. The method comprises receiving, by at least one processor of a content tagging system, a video content having at least one sensitive portion, wherein the video content is formed of a plurality of frames; encoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame; categorizing, by the at least one processor, at least one tile including the sensitive portion into at least one category a plurality of categories of sensitive content; and tagging, by the at least one processor, at least one tile with an indication related to the at least one category of the sensitive content.
In accordance with another aspect of the present disclosure, an embodiment herein discloses a method for rendering content having sensitive portion. The method comprises receiving, by at least one processor of a content rendering system, a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content; decoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content; modifying, by the at least one processor, the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and rendering, by the at least one processor, the video content with the modified at least one tile.
To further clarify advantages and features of the present invention, a more particular description of the invention will be rendered by reference to specific embodiments thereof, which is illustrated in the appended drawings. It is appreciated that these drawings depict only typical embodiments of the invention and are therefore not to be considered limiting of its scope. The invention will be described and explained with additional specificity and detail with the accompanying drawings.
These and other features, aspects, and advantages of the present invention will become better understood when the following detailed description is read with reference to the accompanying drawings in which like characters represent like parts throughout the drawings, wherein:
Figure 1 illustrates an environment implementing a content tagging system and a content rendering system, according to an embodiment of the present disclosure;
Figure 2 illustrates a block diagram of the content tagging system, according to an embodiment of the present disclosure;
Figure 3 illustrates a block diagram depicting tagging of sensitive portion of a content by a tagging module of the content tagging system, according to an embodiment of the present disclosure;
Figure 4 illustrates a block diagram of the content rendering system, according to an embodiment of the present disclosure;
Figure 5 illustrates a block diagram depicting an example of notifying a user of the presence of sensitive portion in a subsequent frame of the content, according to an embodiment of the present disclosure;
Figure 6 illustrates an example of tagging of sensitive portion of the content, based on a user instruction, according to an embodiment of the present disclosure;
Figure 7 illustrates a block diagram depicting operations of the content tagging system and the content rendering system, according to an embodiment of the present disclosure;
Figure 8 illustrates a method for tagging a content having sensitive portion, according to an embodiment of the present disclosure; and
Figure 9 illustrates a method for rendering content having sensitive portion, according to an embodiment of the present disclosure.
Further, skilled artisans will appreciate that elements in the drawings are illustrated for simplicity and may not have been necessarily been drawn to scale. For example, the flow charts illustrate the method in terms of the most prominent steps involved to help to improve understanding of aspects of the present invention. Furthermore, in terms of the construction of the device, one or more components of the device may have been represented in the drawings by conventional symbols, and the drawings may show only those specific details that are pertinent to understanding the embodiments of the present invention so as not to obscure the drawings with details that will be readily apparent to those of ordinary skill in the art having benefit of the description herein.
For the purpose of promoting an understanding of the principles of the invention, reference will now be made to the embodiment illustrated in the drawings and specific language will be used to describe the same. It will nevertheless be understood that no limitation of the scope of the invention is thereby intended, such alterations and further modifications in the illustrated system, and such further applications of the principles of the invention as illustrated therein being contemplated as would normally occur to one skilled in the art to which the invention relates. Unless otherwise defined, all technical and scientific terms used herein have the same meaning as commonly understood by one of ordinary skilled in the art to which this invention belongs. The system, methods, and examples provided herein are illustrative only and not intended to be limiting.
Embodiments of the present invention will be described below in detail with reference to the accompanying drawings.
For the sake of clarity, the first digit of a reference numeral of each component of the present disclosure is indicative of the Figure number, in which the corresponding component is shown. For example, reference numerals starting with digit "1" are shown at least in Figure 1. Similarly, reference numerals starting with digit "2" are shown at least in Figure 2.
Figure 1 illustrates an environment 100 implementing a content tagging system 102 and a content rendering system 104, according to an embodiment of the present disclosure. In an embodiment, the environment 100 may include a server 106 of a content provider in communication with a client device 108. In an embodiment, the server 106 and the client device 108 may be in communication with each other over a network 110. The network 110 may be a wired network or a wireless network. As would be appreciated by a person skilled in the art, for the sake of clarity, the server 106 is shown to be in communication with only one client device 108 in the illustrated embodiment. In other embodiments, the server 106 may be in communication with a plurality of client devices 108, without departing from the scope of the present disclosure.
Further, the content tagging system 102 may be installed at the server 106 whereas the content rendering system 104 may be installed at the client device 108. In an embodiment, the content tagging system 102 may be configured to tag a content having at least one sensitive portion. Constructional and operational details of the content tagging system 102 are explained in the description of at least Figure 2. Further, the content rendering system 104 may be configured to render a content having tagged sensitive portion. Constructional and operational details of the content rendering system 104 are explained in the description of at least Figure 3.
Figure 2 illustrates a block diagram of the content tagging system 102, according to an embodiment of the present disclosure. The content tagging system 102 may include a processor 202, a memory 204, modules 206, and data 208. The modules 206 and the memory 204 are coupled to the processor 202. The processor 202 can be a single processing unit or a number of units, all of which could include multiple computing units. The processor 202 may be implemented as one or more microprocessors, microcomputers, microcontrollers, digital signal processors, central processing units, state machines, logic circuitries, and/or any devices that manipulate signals based on operational instructions. Among other capabilities, the processor 202 is configured to fetch and execute computer-readable instructions and data stored in the memory 204.
The memory 204 may include any non-transitory computer-readable medium known in the art including, for example, volatile memory, such as static random access memory (SRAM) and dynamic random access memory (DRAM), and/or non-volatile memory, such as read-only memory (ROM), erasable programmable ROM, flash memories, hard disks, optical disks, and magnetic tapes.
The modules 206, amongst other things, include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement data types. The modules 206 may also be implemented as, signal processor(s), state machine(s), logic circuitries, and/or any other device or component that manipulate signals based on operational instructions.
Further, the modules 206 can be implemented in hardware, instructions executed by a processing unit, or by a combination thereof. The processing unit can comprise a computer, a processor, such as the processor 202, a state machine, a logic array or any other suitable devices capable of processing instructions. The processing unit can be a general-purpose processor which executes instructions to cause the general-purpose processor to perform the required tasks or, the processing unit can be dedicated to perform the required functions. In another embodiment of the present disclosure, the modules 206 may be machine-readable instructions (software) which, when executed by a processor/processing unit, perform any of the described functionalities.
In an implementation, the modules 206 may include a receiving module 210, an encoding module 212, a categorizing module 214, a tagging module 216, and a generating module 218. The receiving module 210, the encoding module 212, the categorizing module 214, the tagging module 216, and the generating module 218 may be in communication with each other. Further, the data 208 serves, amongst other things, as a repository for storing data processed, received, and generated by one or more of the modules 206.
In an embodiment, the receiving module 210 may be configured to receive a video content having at least one sensitive portion. In an embodiment, the video content may be formed of a plurality of frames. Further, the encoding module 212 may be configured to encode each frame of the video content in form of a plurality of tiles. Each tile may be indicative of a sub-frame.
In an embodiment, the encoding module 212 may encode each frame in the form of the plurality of tile by High-Efficiency Video Coding (HEVC) standard technique. The HEVC coding is a Moving Picture Experts Group (MPEG)/ITU-T video coding standard capable of achieving more than 50% bitrate reduction compared to an Advanced Video Coding (AVC) standard at the same subjective quality. HEVC standard technique is developed particularly for Ultra-High Definition (UHD) videos. At that resolution, using the main profile and with today's HEVC encoders, a video stream may typically be delivered at a bit rate of 15-20 Mbps. UHD videos or videos with a higher resolution may not be viewable by all devices, for example, owing to processing requirements, or stream-able through all networks, for example, owing to high bitrates. In order to offer parallelism features for encoders and decoders, the encoding module 212 may use the HEVC standard technique for tiling. The encoding of a frame into tiles makes it possible to encode different regions independently, within a single frame, while maintaining the use of a single decoder and efficient coding.
In an embodiment, the encoding module 212 may perform motion-constrained tiling of the frames of the content. The encoding module 212 may encode consecutive pictures in a video sequence such that a tile of a given picture only requires, if any, the coded information of the tiles at the same spatial position in the reference pictures, effectively creating tile tunnels in the video. The tile tunnels may then be stored in ISO Base Media File Format (ISOBMFF) files and may be delivered separately. A player may then decide whether to present one or more tile tunnels at a given time, based on user interaction. Adaptive HTTP streaming technologies, such as MPEG DASH, enable the live or on-demand delivery of media streams over HTTP. In an embodiment, the quality of the media delivered to the client may widely be adapted to user's preferences, such as language and point of view), to client's capabilities, for example, support for a given codec, and to fluctuating network conditions.
Further, in an embodiment, the categorizing module 214 may be configured to identify whether the each tile includes sensitive portion, and categorize at least one tile including the sensitive portion into at least one of a plurality of predefined categories of sensitive content. In an embodiment, the plurality of categories may be defined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions. The tagging module 216 may be configured to tag at least one tile with at least one associated category of the sensitive content. Figure 3 illustrates a block diagram 300 depicting tagging of the sensitive portion of a frame of the content by the tagging module 216, according to an embodiment of the present disclosure. As would be appreciated by a person skilled in the art, Figure 3 is included for providing better understanding of the present disclosure and therefore, should not be construed as limiting in any way.
Referring to Figure 2 and Figure 3, the frame 302 is encoded in form of a plurality of tiles. As illustrated, the frame 302 is depicting a male smoking a cigarette. Further, the tagging module 216 may tag the at least one tile including the sensitive portion in the frame 302 with an indication related to the at least one associated category of the sensitive content. For example, the tagging module 216 may tag the tile having the cigarette with a category of the sensitive content, namely, "smoking". Therefore, the frame 302 is now tagged with the category "Smoking" indicative of the frame 302 having a smoking sequence.
Referring to Figure 2, once the content is tagged by the tagging module 216, the generating module 218 may be configured to generate a playlist file indicative of the sensitive content for the video content including the at least one tile tagged with the indication.
Figure 4 illustrates a block diagram of the content rendering system 104, according to an embodiment of the present disclosure. The content rendering system 104 may include a processor 402, a memory 404, modules 406, and data 408. The modules 406 and the memory 404 are coupled to the processor 402. The constructional and operational characteristics of the processor 402, the memory 404, the modules 406, and the data 408 may be similar to the processor 202, the memory 204, the modules 206, and the data 208 of the content tagging system 102. Therefore, for the sake of brevity, those details are not explained again in the description of Figure 4.
The modules 406, amongst other things, include routines, programs, objects, components, data structures, etc., which perform particular tasks or implement data types. The modules 406 may also be implemented as, signal processor(s), state machine(s), logic circuitries, and/or any other device or component that manipulate signals based on operational instructions.
Further, the modules 406 can be implemented in hardware, instructions executed by a processing unit, or by a combination thereof. The processing unit can comprise a computer, a processor, such as the processor 402, a state machine, a logic array or any other suitable devices capable of processing instructions. The processing unit can be a general-purpose processor which executes instructions to cause the general-purpose processor to perform the required tasks or, the processing unit can be dedicated to perform the required functions. In another embodiment of the present disclosure, the modules 406 may be machine-readable instructions (software) which, when executed by a processor/processing unit, perform any of the described functionalities.
In an embodiment, the modules 406 may include a receiving module 410, a decoding module 412, a modifying module 414, a rendering module 416, an updating module 418, a detecting module 420, and a notification module 422. In an embodiment, the receiving module 410, the decoding module 412, the modifying module 414, the rendering module 416, the updating module 418, the detecting module 420, and the notification module 422 may be in communication with each other. Further, the data 408 serves, amongst other things, as a repository for storing data processed, received, and generated by one or more of the modules 406.
In an embodiment, the receiving module 410 may be configured to receive a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of predefined categories of sensitive content. In an embodiment, the plurality of categories of the sensitive content may be determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions. Further, the receiving module 410 may receive the playlist file from the content tagging system 102.
Upon receiving the playlist file, the decoding module 412 may be configured to decode each frame of the video content in form of the plurality of tiles. Further, the modifying module 414 may be configured to modify the at least one tile having sensitive content, based on a predefined set of viewing policies for a user. In an embodiment, the rendering module 416 may be configured to render the video content with the at least one tile modified in-line with the predefined set of viewing policies.
In an embodiment, the updating module 418 may be configured to update the predefined set of viewing policies for the user. Based on the updated set of viewing policies, the modifying module 414 may be configured to re-modify the at least one tile having the sensitive content.
In an embodiment, the detecting module 420 may be configured to detect, while the video content is being rendered to the user, presence of a sensitive portion in one of subsequent frames of the video content. Further, the notification module 422 may be configured to notify the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user. Figure 5 illustrates a block diagram 500 depicting an example of notifying a user of the presence of sensitive portion in a subsequent frame of the content, according to an embodiment of the present disclosure. As illustrated, the block diagram 500 includes 3 frames of the content, for example, at the 1st second, the 2nd second, and the 3rd second of the content. Further, the third frame of the content includes sensitive portion. Therefore, when the user is viewing the content, the detecting module 420 may detect the presence of the sensitive portion in the 3rd frame of the content. Accordingly, the notification module 422 may notify the user of the presence of the sensitive portion in the upcoming frames, for example, by prompting a message "Sensitive content in next 2 seconds".
Referring to Figure 4, in an embodiment, the content rendering system 104 may receive a user instruction, while playing of the content. The user instruction may be indicative of a tile of a frame having a sensitive content. In such an embodiment, the content rendering system 104 may prompt the user to indicate a category of the sensitive content for associating the corresponding tile. Upon receiving the user's indication, the content rendering system 104 may transmit such information to the content tagging system 102. Thereafter, the content tagging system 102 may tag such tiles of the frames of the content with the user-selected category of sensitive portion. Accordingly, the content rendering system 104 may render the subsequent portions of the content, based on the tagging done based on the user instruction.
Figure 6 illustrates a block diagram 600 depicting an example of the tagging of sensitive portion of the content, based on user instruction, according to an embodiment of the present disclosure. As illustrated, while watching the content, the user may select a tile of the frame for being tagged with a category of "Smoking". Subsequently, any subsequent portion of the content showing smoking is automatically tagged by the content tagging system 102, and then censored by the content rendering system 104.
Figure 7 illustrates another environment 700 depicting operations of the content tagging system 102 and the content rendering system 104, according to another embodiment of the present disclosure. As illustrated, the server 106 is shown to be in communication with the client device 108. In an embodiment, the client device 108 may include, but is not limited to, an adaptive streaming engine 702, the decoding module 412, and the rendering module 416. The adaptive streaming engine 702 may further include, but is not limited to, a playlist parser module 704 and a tile stitching module 706.
In an embodiment, the adaptive streaming engine 702 may request the server 106 to share a playlist. In response, the server 106 may provide the client device 108 with the playlist. The playlist parser module 704 may retrieve information related to tiling of the frames of the content. For example, the information may relate to categories of sensitive content associated with each tile of each frame.
In an embodiment, the adaptive streaming engine 702 may request the server 106 for sharing specific tiles of a frame. Further, the tile stitching module 706 may stitch the received tiles and share the stitched tiles with the decoding module 412. The decoding module 412 may decode the tiles and share the decoded tiles with the rendering module 416. The rendering module 416 may then render the content based on at least one sensitive content category. Further, a user equipment 708 may stream the content, based on at least one of content provider policies, OEM policies, and user policies.
Therefore, the adaptive streaming engine 702 may download and parse the playlist of the content from the server 106, informing the client device 108 about the sensitive portion of each tile. Such information can then be matched with viewing policies and only the tiles that do not fall within the purview of any category of sensitive content may be played normally. Further, the tiles with sensitive portion may be censored based on the viewing policies before being shown to the user.
Figure 8 illustrates a method 800 for tagging a content having sensitive portion, according to an embodiment of the present disclosure. In an embodiment, the method 800 may be a computer-implemented method 800. In an embodiment, the method 800 may be executed by the processor 202 of the content tagging system 102. Further, for the sake of brevity, details of the present disclosure that are explained in details in the description of Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, and Figure 8 are not explained in detail in the description of Figure 8.
At a block 802, the method 800 includes receiving the video content having the at least one sensitive portion. In an embodiment, the video content may be formed of the plurality of frames. In an embodiment, the receiving module 210 of the content tagging system 102 may receive the video content. The receiving module 210 may be controlled by the processor 202.
At a block 804, the method 800 includes encoding each frame of the video content in form of the plurality of tiles. Each tile may be indicative of a sub-frame. In an embodiment, the encoding module 212 may encode each frame in form of the plurality of tiles. The encoding module 212 may be controlled by the processor 202.
At a block 806, the method 800 includes categorizing at least one tile including the sensitive portion into at least one of the plurality of predefined categories of sensitive content. In an embodiment, the categorizing module 214 may identify whether the each tile includes sensitive portion, and categorize each tile into the at least one of the plurality of predefined categories. The categorizing module 214 may be controlled by the processor 202.
At a block 808, the method 800 includes tagging the at least one tile with an indication related to the at least one associated category of the sensitive content. In an embodiment, the tagging module 216 may tag the at least one tile with the indication, and be controlled by the processor 202.
In an embodiment, the method 800 may include generating the playlist file indicative of sensitive content for the video content including the at least one tile tagged with the indication. In an embodiment, the generating module 218 may generate the playlist file, and be controlled by the processor 202.
Figure 9 illustrates a method 900 for rendering content having sensitive portion, according to an embodiment of the present disclosure. In an embodiment, the method 900 may be a computer-implemented method 900. In an embodiment, the method 800 may be executed by the processor 402 of the content rendering system 104. Further, for the sake of brevity, details of the present disclosure that are explained in details in the description of Figure 1, Figure 2, Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, and Figure 8 are not explained in detail in the description of Figure 9.
At a block 902, the method 900 includes receiving a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of predefined categories of sensitive content. In an embodiment, the method 900 may include receiving the playlist file generated by the content tagging system 102. Further, the plurality of categories may be defined based on at least one of the age-based restrictions, the demography-based restrictions, and the language-based restrictions. In an embodiment, the receiving module 410 of the content rendering system 104 may receive the playlist file. The receiving module 410 may be controlled by the processor 402.
At a block 904, the method 900 includes decoding each frame of the video content in form of the plurality of tiles. Each tile may be indicative of a sub-frame. Further, at least one tile may be tagged with an indication related to the at least one associated category of the sensitive content. In an embodiment, the decoding module 412 may decode each frame of the video content, and be controlled by the processor 402.
At a block 906, the method 900 includes modifying the at least one tile having the sensitive content, based on a predefined set of viewing policies for the user. In an embodiment, the modifying module 414 may modify the at least one tile, and be controller by the processor 402.
At a block 908, the method 900 includes rendering the video content with the at least one tile modified in-line with the predefined set of viewing policies. In an embodiment, the rendering module 416 may render the video content with the modified at least one tile, and be controller by the processor 402.
In an embodiment, the method 900 may include detecting, while the video content is being rendered to the user, presence of a sensitive portion in one of the subsequent frames of the video content. In an embodiment, the method 900 may include notifying the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user.
As would be gathered, the present disclosure offers a comprehensive approach of tagging and rendering the content based on the presence of sensitive portion. The proposed approach eliminates the need of manual intervention and high server storage cost in media content management. The tiling of the frames of the content for being tagged with a respective category of the sensitive portion allows for saving of storage space and offer efficient following of the viewing policies.
The adaptability of the content tagging system 102 and the content rendering system 104 is client-driven, as the client seamlessly downloads and plays small continuous media files of a given quality and dynamically switches between alternative media files of different qualities offered by HTTP servers. The client may only download the necessary quality given the situation based on adaptation configuration. Therefore, combining the concepts of tiling with adaptive HTTP streaming allows for new streaming possibilities. A client can indeed select and adaptively stream only one or several tiles of interest to reduce the decoding requirements. Even when all tiles are transmitted, new adaptation configurations can be designed allowing the client to select different qualities per tile. This would reduce an overall bitrate while maintaining the best quality on the tiles of interest.
Therefore, the proposed approach saves a lot of server storage space. Further, only MPD files need to be changed in order to accommodate any change in the viewing policies. Accordingly, the content will remain unaffected by the change in the viewing policies. Moreover, delivery of media content will be more flexible as the present disclosure provides multiple options for applying different policies. Therefore, the present disclosure offers the content tagging system 102, the content rendering system 104, the method 800, and the method 900 that are comprehensive, flexible, accurate, and intelligent.
While specific language has been used to describe the present disclosure, any limitations arising on account thereto, are not intended. As would be apparent to a person in the art, various working modifications may be made to the method in order to implement the inventive concept as taught herein. The drawings and the foregoing description give examples of embodiments. Those skilled in the art will appreciate that one or more of the described elements may well be combined into a single functional element. Alternatively, certain elements may be split into multiple functional elements. Elements from one embodiment may be added to another embodiment.

Claims (14)

  1. A system for tagging a content having sensitive portion, the system comprising:
    a memory; and
    at least one processor coupled to the memory and configured to:
    receive a video content having at least one sensitive portion, wherein the video content includes a plurality of frames;
    encode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame;
    categorize at least one tile including the sensitive portion into at least one category among a plurality of categories of sensitive content; and
    tag the at least one tile with an indication related to the at least one category of the sensitive content.
  2. The system as claimed in claim 1, wherein the at least one processor further configured to generate a playlist file indicative of the sensitive content for the video content including the at least one tile tagged with the indication.
  3. The system as claimed in claim 1, wherein the plurality of categories of sensitive content are determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions.
  4. A system for rendering content having sensitive portion, the system comprising:
    a memory; and
    at least one processor coupled to the memory and configured to:
    receive a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content;
    decode each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content;
    modify the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and
    render the video content with the modified at least one tile.
  5. The system as claimed in claim 4, wherein the plurality of categories of sensitive content are determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions.
  6. The system as claimed in claim 4, wherein the at least one processor further configured to:
    update the predefined set of viewing policies for the user; and
    re-modify the at least one tile having the sensitive content, based on the updated set of viewing policies.
  7. The system as claimed in claim 4, wherein the at least one processor further configured to:
    detect, while the video content is being rendered to the user, presence of a sensitive portion in one of a subsequent frames of the video content; and
    notify the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user.
  8. A method for tagging a content having sensitive portion by a content tagging system, the method comprising:
    receiving, by at least one processor of the content tagging system, a video content having at least one sensitive portion, wherein the video content includes a plurality of frames;
    encoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame;
    categorizing, by the at least one processor, at least one tile including the sensitive portion into at least one category among a plurality of categories of sensitive content; and
    tagging, by the at least one processor, the at least one tile with an indication related to the at least one category of the sensitive content.
  9. The method as claimed in claim 8, further comprising:
    generating, by the at least one processor, a playlist file indicative of the sensitive content for the video content including the at least one tile tagged with the indication.
  10. The method as claimed in claim 8, wherein the plurality of categories of sensitive content are determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions.
  11. A method for rendering content having sensitive portion by a content rendering system, the method comprising:
    receiving, by at least one processor of the content rendering system, a playlist file indicative of a video content having sensitive portion tagged with at least one of a plurality of categories of sensitive content;
    decoding, by the at least one processor, each frame of the video content in form of a plurality of tiles, wherein each tile is indicative of a sub-frame and at least one tile is tagged with an indication related to the at least one category of the sensitive content;
    modifying, by the at least one processor, the at least one tile having sensitive content, based on a predefined set of viewing policies for a user; and
    rendering, by the at least one processor, the video content with the modified at least one tile.
  12. The method as claimed in claim 11, wherein the plurality of categories of sensitive content are determined based on at least one of age-based restrictions, demography-based restrictions, and language-based restrictions.
  13. The method as claimed in claim 11, further comprising:
    updating, by the at least one processor, the predefined set of viewing policies for the user; and
    re-modifying, by the at least one processor, the at least tile having the sensitive content, based on the updated set of viewing policies.
  14. The method as claimed in claim 11, further comprising:
    detecting, by the at least one processor, while the video content is being rendered to the user, presence of a sensitive portion in one of a subsequent frames of the video content; and
    notifying, by the at least one processor, the user of the presence of the sensitive portion in the subsequent frame, before the sensitive portion is rendered to the user.
PCT/KR2020/000652 2019-08-21 2020-01-14 Method and system for processing content having sensitive portion WO2021033853A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
IN201911033711 2019-08-21
IN201911033711 2019-08-21

Publications (1)

Publication Number Publication Date
WO2021033853A1 true WO2021033853A1 (en) 2021-02-25

Family

ID=74660241

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2020/000652 WO2021033853A1 (en) 2019-08-21 2020-01-14 Method and system for processing content having sensitive portion

Country Status (1)

Country Link
WO (1) WO2021033853A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022256539A1 (en) * 2021-06-02 2022-12-08 Google Llc Selective content masking for collaborative computing

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100628132B1 (en) * 2004-11-02 2006-09-26 엘지전자 주식회사 Broadcasting service method and apparatus for the same
KR101010523B1 (en) * 2008-08-05 2011-01-25 구진국 A System and a Method for Alternating Bad Contents during TV Broadcast
US20170094269A1 (en) * 2012-09-18 2017-03-30 Vid Scale, Inc. Method and apparatus for region of interest video coding using tiles and tile groups
US20170132495A1 (en) * 2015-11-11 2017-05-11 International Business Machines Corporation User-configurable settings for content obfuscation
KR20180071156A (en) * 2016-12-19 2018-06-27 삼성전자주식회사 Method and apparatus for filtering video

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100628132B1 (en) * 2004-11-02 2006-09-26 엘지전자 주식회사 Broadcasting service method and apparatus for the same
KR101010523B1 (en) * 2008-08-05 2011-01-25 구진국 A System and a Method for Alternating Bad Contents during TV Broadcast
US20170094269A1 (en) * 2012-09-18 2017-03-30 Vid Scale, Inc. Method and apparatus for region of interest video coding using tiles and tile groups
US20170132495A1 (en) * 2015-11-11 2017-05-11 International Business Machines Corporation User-configurable settings for content obfuscation
KR20180071156A (en) * 2016-12-19 2018-06-27 삼성전자주식회사 Method and apparatus for filtering video

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022256539A1 (en) * 2021-06-02 2022-12-08 Google Llc Selective content masking for collaborative computing

Similar Documents

Publication Publication Date Title
Concolato et al. Adaptive streaming of hevc tiled videos using mpeg-dash
US11917220B2 (en) Scene section and region of interest handling in video streaming
US10645465B2 (en) Video file universal identifier for metadata resolution
US8645990B2 (en) Dynamic advertising control
US7688889B2 (en) Methods, apparatus, and systems for insertion of overlay content into a video signal with transrating capabilities
KR102027410B1 (en) Transmission of reconstruction data in a tiered signal quality hierarchy
US10904642B2 (en) Methods and apparatus for updating media presentation data
GB2594615A (en) Method and apparatus for encoding and transmitting at least a spatial part of a video sequence
CN103918268A (en) Signaling of state information for a decoded picture buffer and reference picture lists
US8681874B2 (en) Video insertion information insertion in a compressed bitstream
US20200162731A1 (en) Motion-constrained av1 encoding method and apparatus for tiled streaming
CN102771134A (en) Method and arrangement for supporting playout of content
CN109963176B (en) Video code stream processing method and device, network equipment and readable storage medium
US20070217503A1 (en) Scalable video coding/multiplexing compatible with non-scalable decoders
WO2020178065A1 (en) Method and apparatus for encoding and decoding a video bitstream for merging regions of interest
US20200029104A1 (en) Systems and methods for securely generating live previews
WO2021033853A1 (en) Method and system for processing content having sensitive portion
CN101652931B (en) Methods, apparatus, and systems for insertion of overlay content into a video signal with transrating capabilities
KR102163440B1 (en) Picture referencing control for video decoding using a graphics processor
GB2597642A (en) Method, device, and computer program for encapsulating media data into a media file
EP4210339A1 (en) Reproduction device, transmission device, reproduction method, and transmission method
CN101605243A (en) The method that program provides, media device and user side equipment
Aggarwal Smart Sensitive Content Management MPEG-DASH based HEVC tiling technique
Zhang et al. A real-time multi-view AVS2 decoder on mobile phone
CN114339426A (en) Live video encoding and decoding forwarding system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20853838

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20853838

Country of ref document: EP

Kind code of ref document: A1