KR20140050535A - Apparatus and method for providing n screen service using group visual objects based on depth and providing contents service - Google Patents
Apparatus and method for providing n screen service using group visual objects based on depth and providing contents service Download PDFInfo
- Publication number
- KR20140050535A KR20140050535A KR1020130113380A KR20130113380A KR20140050535A KR 20140050535 A KR20140050535 A KR 20140050535A KR 1020130113380 A KR1020130113380 A KR 1020130113380A KR 20130113380 A KR20130113380 A KR 20130113380A KR 20140050535 A KR20140050535 A KR 20140050535A
- Authority
- KR
- South Korea
- Prior art keywords
- group
- unit
- independent
- visual
- screen
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/426—Internal components of the client ; Characteristics thereof
- H04N21/42607—Internal components of the client ; Characteristics thereof for processing the incoming bitstream
- H04N21/42615—Internal components of the client ; Characteristics thereof for processing the incoming bitstream involving specific demultiplexing arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/61—Network physical structure; Signal processing
- H04N21/6156—Network physical structure; Signal processing specially adapted to the upstream path of the transmission network
- H04N21/6175—Network physical structure; Signal processing specially adapted to the upstream path of the transmission network involving transmission via Internet
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8146—Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N2213/00—Details of stereoscopic systems
- H04N2213/005—Aspects relating to the "3D+depth" image format
Abstract
Description
The present invention relates to an apparatus and method for providing a multimedia content service, and more particularly, to an apparatus and method for producing a visual object for each depth-based group and providing an object for each group through an N screen service.
Today, not only 2D or 3D video and still images, but also media such as 3D video games are being serviced by real time streaming or VOD down and play. Accordingly, the field of application service technology through object unit coding based on the extraction of media objects and the MPEG-4 standard for image processing has been continuously developed.
MPEG-4 based object creation technology (Sang-Wook Kim et al., Publication No. 2003-0037614, MPEG-4 content generation method and apparatus thereof) as an application service technology field through the separation of media objects and object-based encoding based on the MPEG-4 standard. ), Image processing technology for object separation extraction (Gojongkuk 3 people, Publication No. 2012-0071226, object extraction method and apparatus) or image processing method (3, Park Ji-young and others, Publication No. 2012-0071219) to obtain depth information , 3D depth information acquisition device and method).
However, in the above-described conventional techniques, when visual objects such as a background, a person, and a car are overlapped in a 2D or 3D video and a still image scene, the viewer can see each of the objects included in the 2D or 3D video and the still picture perfectly. Can't. In other words, the visual objects hidden behind the overlapped area are not shown.
The present invention groups and authors independent visual (movie or still image) objects in group value units, extracts each grouped visual object scene in units of interest that can interact with a user, and views them on various screens. A service and an apparatus are provided.
The present invention provides a method for providing an image service using two or more different types of screens in an N screen service providing apparatus, comprising: separating and extracting independent visual objects having different depth values from the image; And grouping them according to depth values into scenes for each group, and selectively reproducing one or more scenes authored for each group to two or more screens according to a user interaction event.
The present invention provides an N screen service providing apparatus using a depth-based group visual object, an independent visual object extractor for separating and extracting independent visual objects having different depth values from an image, and the separated extracted independent visual objects according to a depth value. And a group unit visual object authoring unit for grouping and authoring scenes in groups, and an N screen unit for selectively reproducing one or more scenes authored for each group to two or more screens according to a user interaction event.
According to the present invention, in the case of a digital signage service supporting a multi-screen service such as multi-vision, if the user can separate and extract only the objects of interest among scenes being played on one screen, the user can re-view on an independent screen. Targeting advertising effect can be obtained.
1 is an MPEG-4 system reference model.
2 is a configuration diagram of an N screen service providing apparatus using a depth-based group visual object according to an exemplary embodiment.
3 is a detailed configuration diagram of a group unit visual object authoring unit according to an exemplary embodiment.
4 is a detailed configuration diagram of an N screen according to an embodiment of the present invention.
5 to 7 are flowcharts illustrating a method for providing an N screen service using a depth-based group visual object according to an embodiment of the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like reference numerals refer to the like elements throughout.
In the following description of the present invention, a detailed description of known functions and configurations incorporated herein will be omitted when it may make the subject matter of the present invention rather unclear.
The terms used throughout the specification are defined in consideration of the functions in the embodiments of the present invention and can be sufficiently modified according to the intentions and customs of the user or the operator. It should be based on the contents of.
The present invention authors independent visual (movie or still image) objects in group units, and extracts one or more scenes of each grouped visual object into units of interest object that can interact with a user. A method and apparatus for viewing services, respectively, on the screens of the present invention. MPEG-4 is mainly used as an international standard for application services for digital video synthesis, manipulation, indexing, and retrieval as well as providing high compression rate through object-based coding for such visual objects. .
1 is a structural diagram of an MPEG-4 system reference model.
Referring to FIG. 1, the MPEG-4 system reference model synthesizes a media object including an interaction function in a desired audio visual scene, and then multiplexes the media data into a bitstream to ensure quality of service (QoS). And transmit (2) the media content source (1) generated by synchronization to the receiving side. The receiving side then demultiplexes (3) the received media content source (1) to composition (5) the decoded data into BIFS, video, audio, animation, text, and the like, and outputs it (7). At this time, the receiving side has a system structure capable of interacting the
In order to solve the problem of object overlap in the system reference model such as MPEG-4, the present invention groups independent visual objects according to depth, authors each visual object in a group unit, and uses grouped visual object scenes. It interacts with and allows output through various screens. In the present invention, for convenience of explanation, the system model will be described using MPEG-4 as an example, but the present invention is not limited to MPEG-4.
2 is a configuration diagram of an N screen service providing apparatus using a depth-based group visual object according to an exemplary embodiment.
2, an N screen service providing apparatus using a depth-based group visual object according to the present invention includes an independent
The independent
According to an embodiment of the present invention, the independent
The independent visual
The group visual
The
The
3 is a detailed configuration diagram of a group unit visual object authoring unit according to an exemplary embodiment.
Referring to FIG. 3, in detail, the group visual
The independent visual
The independent visual
The playback
The interaction
The group visual
The depth-based
The playback
Through the independent visual
The scene
The media
4 is a detailed configuration diagram of the N-screen unit according to an embodiment of the present invention.
Referring to FIG. 4, when the user action is given, the
To this end, the N-screen unit includes a
The
The
The rendering and
5 is a flowchart illustrating a method of providing a depth-based group unit visual object scene through an N screen service according to an exemplary embodiment.
Referring to FIG. 5, in S510, one or more independent visual objects included in a moving image or a still image are automatically or semi-automatically extracted. In this case, according to an embodiment of the present invention, a depth value is designated for each independent visual object while extracting the independent visual object. For example, '1' may be designated as the depth value, and '2' may be designated as the depth value for 'Object 2' that is located at the deepest.
In S520, the extracted independent visual objects are grouped according to depth, and the visual object scene is authored in group units. That is, one or more independent visual objects are grouped according to a specified depth value, and scenes are authored by setting the spatiotemporal relationship information and interaction events of the visual objects in group units. This will be described in detail with reference to FIG. 6 below.
Although not shown in the figure, the authored group object is streamed to the N-screen through the network, where the N-screen receives the streamed media object to decode and compose the scene, and plays it through the N screens. In the present invention, the visual objects superimposed on the moving image or the still image are grouped according to the depth value using the N screen service, and the hidden visual objects can be clearly reproduced by unfolding them into the authored scene.
That is, as the visual object of interest is selected through interaction with the user in S530, the visual object of interest for each group may be played through the N screen in S540.
Then, S510 and S520 will be described in more detail with reference to FIG. 6.
FIG. 6 is a flowchart illustrating a group unit visual object authoring step according to an exemplary embodiment. Referring to FIG.
Referring to FIG. 6, the group unit visual
The group unit visual
The group unit visual
The group unit visual
As a result of the determination of S640, when the authoring of all the groups is not completed, the process proceeds to S650.
However, when the authoring of all the groups is completed as a result of the determination in S640, the group-based visual
In operation S670, the group unit visual
In S680, the group visual
7 is a flowchart illustrating a group-based visual object authoring step according to an embodiment of the present invention.
Referring to FIG. 7, the
In S820, the
As a result of the determination in S820, when there is a group visual object interaction, the
In S840, it is determined whether there is an independent visual object selection interaction. As a result of the determination in S840, when there is an independent visual object interaction, the
This allows only some grouping objects to interact with the user among several overlapping objects, and also allows the user to edit the grouped objects when the service provider edits objects that are intentionally hidden from the nested objects. When you take an action (by dragging with the mouse or moving it), you can make the object hidden within it appear, and then process the event to be taken by the user and the action again.
Claims (15)
Separating and extracting independent visual objects having different depth values from the image;
Authoring the separated and extracted independent visual objects into scenes by grouping the separated visual objects according to a depth value;
And selectively reproducing one or more scenes authored for each group to two or more screens according to a user interaction event.
And assigning a depth value to each of the independent visual objects.
And providing the independent visual objects by grouping the number of screens as the number of screens.
And streaming the scenes authored for each group to the N-screen via a network.
Setting space-time relationship information and user interaction event information of one or more independent visual objects;
Grouping one or more independent visual objects according to depth values;
Setting space-time relationship information and user interaction event information of visual objects included in each group;
Generating a scene composition tree hierarchically configured with the set independent visual objects and grouped visual objects;
And generating a media file by encoding the scene configuration tree and the visual objects.
Determining whether a user interaction event for a scene composed of group visual objects has occurred from a user;
When a user interaction event for a scene composed of group visual objects occurs, moving the selected visual object to any N screen, N-screen service providing method using a depth-based group visual object.
Independent Visual Object Selection When there is a user interaction event, the method further includes applying a user interaction event to the selected independent visual object.
A group unit visual object authoring unit which groups the separated and extracted independent visual objects according to a depth value and authors them as scenes for each group;
And an N screen unit comprising two or more screens for selectively playing one or more scenes authored by the group according to a user interaction event.
N-screen service providing apparatus using a depth-based group visual object, characterized in that for assigning a depth value to each of the independent visual objects.
N-screen service providing apparatus using a depth-based group visual object, characterized in that for grouping the independent visual objects as the number of the screen.
The apparatus for providing N screen service using a depth-based group visual object further comprises a streaming unit for streaming the scenes authored for each group to an N-screen through a network.
An apparatus for providing N screen service using a depth-based group visual object, characterized in that a session is established and a network channel is set by RTSP, a packet unit including sync headers is generated, and a media stream is transmitted through an IP network.
An independent visual object setting unit for setting space-time relationship information and user interaction event information of one or more independent visual objects;
A group visual object setting unit for grouping one or more independent visual objects according to a depth value, and setting space-time relationship information and user interaction event information of visual objects included in each group;
A scene composition tree manager configured to generate a scene composition tree hierarchically configured with the set independent visual objects and grouped visual objects;
And a media file generation unit for generating a media file by encoding the scene composition tree and the visual objects.
An apparatus for providing N screen service using a depth-based group visual object, characterized in that, when a user interaction event for a scene composed of group visual objects is generated from a user, the selected group visual object is moved to an arbitrary N screen.
Independent visual object selection When there is a user interaction event, the apparatus for providing N screen service using a depth-based group visual object, wherein the user interaction event is applied to the selected independent visual object.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/057,718 US20140115484A1 (en) | 2012-10-19 | 2013-10-18 | Apparatus and method for providing n-screen service using depth-based visual object groupings |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20120116919 | 2012-10-19 | ||
KR1020120116919 | 2012-10-19 |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20140050535A true KR20140050535A (en) | 2014-04-29 |
Family
ID=50655687
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020130113380A KR20140050535A (en) | 2012-10-19 | 2013-09-24 | Apparatus and method for providing n screen service using group visual objects based on depth and providing contents service |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20140050535A (en) |
-
2013
- 2013-09-24 KR KR1020130113380A patent/KR20140050535A/en not_active Application Discontinuation
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10210907B2 (en) | Systems and methods for adding content to video/multimedia based on metadata | |
US11917221B2 (en) | Encoding device and method, reproduction device and method, and program | |
EP2113112B1 (en) | Method for creating, editing, and reproducing multi-object audio contents files for object-based audio service, and method for creating audio presets | |
TWI571868B (en) | Method and apparatus for sequencing content | |
KR101571283B1 (en) | Media content transmission method and apparatus, and reception method and apparatus for providing augmenting media content using graphic object | |
US10931930B2 (en) | Methods and apparatus for immersive media content overlays | |
US20110138418A1 (en) | Apparatus and method for generating program summary information regarding broadcasting content, method of providing program summary information regarding broadcasting content, and broadcasting receiver | |
US10924779B2 (en) | Location agnostic media control room and broadcasting facility | |
WO2017219980A1 (en) | Played picture generation method, apparatus, and system | |
CN102984560A (en) | Method and device used for playing video from breaking point | |
Sreedhar et al. | Immersive media experience with MPEG OMAF multi-viewpoints and overlays | |
US10764655B2 (en) | Main and immersive video coordination system and method | |
KR20140050535A (en) | Apparatus and method for providing n screen service using group visual objects based on depth and providing contents service | |
KR102417084B1 (en) | Method And System for Transmitting and Receiving Multi-Channel Media | |
US20140115484A1 (en) | Apparatus and method for providing n-screen service using depth-based visual object groupings | |
CN115623276A (en) | Display device and advertisement insertion method | |
Srivastava | Broadcasting in the new millennium: A prediction | |
Session et al. | A framework for the development, deployment, consumption and management of 3D interactive content |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |