US20130080531A1 - Apparatus and method for clipping and sharing content at a portable terminal - Google Patents

Apparatus and method for clipping and sharing content at a portable terminal Download PDF

Info

Publication number
US20130080531A1
US20130080531A1 US13629394 US201213629394A US2013080531A1 US 20130080531 A1 US20130080531 A1 US 20130080531A1 US 13629394 US13629394 US 13629394 US 201213629394 A US201213629394 A US 201213629394A US 2013080531 A1 US2013080531 A1 US 2013080531A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
content
information
portable terminal
data
server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US13629394
Inventor
Seung-hyun Yoon
Moon-sang Lee
Min-soo Koo
Ji-in Nam
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor ; File system structures therefor
    • G06F17/30017Multimedia data retrieval; Retrieval of more than one type of audiovisual media
    • G06F17/30058Retrieval by browsing and visualisation of multimedia data

Abstract

A apparatus and method for clipping and sharing content at a portable terminal are provided. The method includes receiving a selection of specific content, analyzing related information of the selected content, and classifying the selected content, and storing the classified content according to the classification.

Description

    CROSS-REFERENCE TO RELATED APPLICATION(S) AND CLAIM OF PRIORITY
  • The present application is related to and claims the benefit under 35 U.S.C. §119(a) to a Korean Patent Application No. 10-2011-0097578, filed in the Korean Intellectual Property Office on Sep. 27, 2011, the entire disclosure of which are hereby incorporated by reference.
  • TECHNICAL FIELD OF THE INVENTION
  • The present invention relates to a clip apparatus and method based on content information of a portable terminal.
  • BACKGROUND OF THE INVENTION
  • With the popularization of a smart phone, much data are generated and consumed in a mobile environment. As the kind and amount of data increase, a need for a method for storing specific data generated by a user, so that the user or other people again consume the specific data after the lapse of arbitrary time is being requested.
  • SUMMARY OF THE INVENTION
  • To address the above-discussed deficiencies of the prior art, it is a primary object to provide an apparatus and method for clipping and sharing content at a portable terminal capable of clipping specific data, and easily sharing and consuming the clipped data.
  • Another aspect of the present disclosure is to provide an apparatus and method for designating and clipping a specific portion of a web page that a user is interested in at the portable terminal.
  • A further aspect of the present invention is to provide an apparatus and method for being capable of consuming clipped data at a portable terminal without a Uniform Resource Locator (URL) as stored data is not limited to URL data.
  • The above aspects are achieved by providing with an apparatus and method for clipping and sharing content at a portable terminal according to the present disclosure.
  • According to one aspect of the present disclosure, a method for clipping and sharing content at a portable terminal includes receiving selection of specific content, analyzing related information of the selected content, classifying the selected content, and storing the classified content according to the classification.
  • According to another aspect of the present invention, an apparatus for clipping and sharing content information at a portable terminal include a decision unit for selecting specific content, a classification unit for analyzing related information of the selected contents and for classifying the selected content, and a storage unit for storing the classified content according to the classification.
  • Before undertaking the DETAILED DESCRIPTION OF THE INVENTION below, it may be advantageous to set forth definitions of certain words and phrases used throughout this patent document: the terms “include” and “comprise,” as well as derivatives thereof, mean inclusion without limitation; the term “or,” is inclusive, meaning and/or; the phrases “associated with” and “associated therewith,” as well as derivatives thereof, may mean to include, be included within, interconnect with, contain, be contained within, connect to or with, couple to or with, be communicable with, cooperate with, interleave, juxtapose, be proximate to, be bound to or with, have, have a property of, or the like; and the term “controller” means any device, system or part thereof that controls at least one operation, such a device may be implemented in hardware, firmware or software, or some combination of at least two of the same. It should be noted that the functionality associated with any particular controller may be centralized or distributed, whether locally or remotely. Definitions for certain words and phrases are provided throughout this patent document, those of ordinary skill in the art should understand that in many, if not most instances, such definitions apply to prior, as well as future uses of such defined words and phrases.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • For a more complete understanding of the present disclosure and its advantages, reference is now made to the following description taken in conjunction with the accompanying drawings, in which like reference numerals represent like parts:
  • FIGS. 1A and 1B illustrates clipping static data according to one embodiment of the present disclosure;
  • FIGS. 2A and 2B illustrate clipping dynamic data according to one embodiment of the present disclosure;
  • FIGS. 3A and 3B illustrate exemplary binary data and eXtensible Markup Language (XML) data according to one embodiment of the present disclosure;
  • FIGS. 4A to 4D illustrate consuming a generated clip according to one embodiment of the present disclosure;
  • FIGS. 5A and 5B illustrate sharing clipped data of registered friends according to one embodiment of the present disclosure;
  • FIG. 6 is a flowchart illustrating a method for clipping content at a portable terminal according to an one embodiment of the present disclosure; and
  • FIG. 7 is a block diagram illustrating an apparatus for clipping and sharing content at a portable terminal according to an one embodiment of the present disclosure.
  • DETAILED DESCRIPTION OF THE INVENTION
  • FIGS. 1A through 7, discussed below, and the various embodiments used to describe the principles of the present disclosure in this patent document are by way of illustration only and should not be construed in any way to limit the scope of the disclosure. Those skilled in the art will understand that the principles of the present disclosure may be implemented in any suitably arranged. Preferred embodiments of the present invention will be described herein below with reference to the accompanying drawings. In the following description, well-known functions or constructions are not described in detail since they would obscure the invention in unnecessary detail. Also, parts performing like function and operations are denoted by the same symbols throughout the drawings.
  • FIGS. 1A and 18 illustrate clipping static data according to one embodiment of the present disclosure. Assume that a jacket and black sunglasses worn by a model in a picture are the latest-style items catching user's fancy while a user who is usually much interested in fashion browses a web page. The user who intends to purchase the jacket and the black sunglasses reviews information disclosed on the web page so as to obtain items information. In a case where the web page has disclosed the information on the item such as a brand name and a website address, the user can access a related website and obtain detailed information on the interested item. Also, in a case where the web page does not have disclosed the information on the interested item because the web page does not concern the fashion, or in a case where the user wants to simply store the interested item, the user can store the web page and consume stored data later.
  • In the present disclosure, the user not only can store the website address of the webpage, but also can designate and store a particular portion of the webpage that the user is interested in. As illustrated in FIG.1A, in a case where the user is interested in only a jacket and black sunglasses worn by a model in a picture, the user can designate and store a specific portion containing the jacket and black sunglasses without storing the whole web page. In certain embodiment, if the user discovers data to clip, the user can begin clipping by pressing a clip button. A pressing the clip button can be performed by selecting a User Interface (UI) icon or gesture which is recognized using a detecting sensor. After that, if the clipping process is begun, the user can designate a portion to clip and perform clipping the specific interesting portion. To designate a portion, the user can use an input tool. Desirably, the user can drag and designate a portion to clip by means of the input tool.
  • In a case where data to clip is static data such as a web page or a picture, a portion designated by a user is captured in an image file form and is stored in binary data form. Also, a file containing detailed information on clipped data can be stored in an eXtensible Markup Language (XML) data form. In certain embodiments, the file containing, detailed information on the clip can be created using not only an XML but also other languages such as a Web Ontology Language (OWL).
  • In the present disclosure, selected content are automatically classified and is stored based on related information of the selected content. As illustrated in FIG. 1B, once a specific portion is selected by a user, content is automatically classified and is stored by related information of the content. The related information of the content can be categorized into either data information or peripheral information. First, the data information of the content can include a plurality of supplementary information related to data such as a title, an original data Uniform. Resource Locator (URL), a maker, a creation date, a category, a publication authority, a user memo, an availability period, Global Positioning System (GPS) information, a data summary, information of an application in which clipping was carried out, the type of clip binary data, the size of a thumbnail image, and a display position.
  • Except for the user's memo, the related information of the content is automatically input to a portable terminal when the content is selected by the user. In detail, when the user discovers the whole or part of a web page, which can be a specific picture and moving picture, the user designates and selects a portion of the web page. Upon the selection, related information is automatically stored at the portable terminal. For example, as illustrated in FIG. 1B, in a case where the user clips a picture of a model who is wearing a jacket and black sunglasses, the category of “TREND” and the title of “FASHION LOUIS VUITTON” and the original data URL of “http://WWW.LOUISVUITTON.COM” are displayed. Besides that, related information are automatically stored such as an designer of a web page, the date of creating the web page, a publication authority, an availability period, GPS information of an area in which the clipping was carried out, a data summary, information of an application in which the clipping was carried out, the type of clipped binary data, the size of a thumbnail image, and a display position. As described above, in a case where related information does not catch user's fancy, the user can input related information. In a case where the user inputs the related information, the related information input by the user takes priority over the automatically input related information in classifying and storing of the selected content.
  • The selected content can be classified and be stored using the peripheral information. The peripheral information can include at least one of metadata of the content, text contained in content, and information input by the user. The metadata information, called attribute information is data given to content according to the certain rules for efficient searching and using target information amongst a large amount of information. Metadata can include a position and substance of content, information on a maker, the right conditions, the use conditions and a use specification. In certain embodiments, the user can easily find out his/her own desired specific data (or information) using the metadata by means of a search engine. The functions of the metadata can be extracting movie data in which an actor/actress appearing, picking out only goal-in scenes in a video of a soccer game, and also even editing data. The clipped content is automatically classified and stored using this metadata information. Also, the metadata information has a record of a size or a capacity of the file, so the stored content may be classified by a size or a capacity.
  • According to one embodiment of the present disclosure text information can be used for automatic classifying and storing of data. When a user can designate a specific portion of a web page and peripheral text information disclosed in the web page can be used for classification. That is, in the embodiment, a selected picture can be classified to the broad level category of fashion as well as be classified to narrow level category by brand name or in alphabetical order of the website address of a product. In the embodiment, the brand names of a jacket and black sunglasses, a model name of the product, and the website address of the product around the picture of a model who is wearing the jacket and the sunglasses can be used for classification. Assume that information of a jacket is disclosed as following: “maker: Louisvuitton, model name: TH-002; size: 44, 55, and 66; price: 800,000 Won; store: Samsung department store; Website address of product: http://WWW.LOUISVUITTON.COM”. And information of sunglasses is disclosed as following: “maker: Louisvuitton, model name: T-501; price: 1,000,000 Won; store: Samsung department store; and the Website address of product: http://WWW.LOUISVUITTON.COM”. If the user selects a model in a picture who is wearing the jacket and the sunglasses, the selected content is automatically classified to a category of fashion using peripheral text information. In detail, as the information on the jacket, information is automatically input such as the maker ‘Louisvuitton’, a model name ‘TH-002’, a selling size ‘44, 55, and 66’, a price ‘800,000 Won’, a current store ‘Samsung Department store’, and the website address ‘http://WWW.LOUISVUITTON.COM’ which contains detailed information of the product. Also, as the information on the sunglasses, information is automatically input such as the maker ‘Louisvuitton’, the model name ‘T-501’, the price ‘1,000,000 Won’, the current store ‘Samsung department store’, and the website address of ‘http://WWW.LOUISVUITTON.COM’. Accordingly, the user can automatically classify and store the selected content using the text information among the peripheral information.
  • Another example of the peripheral information is information that a user inputs. In a case where information automatically input using the metadata information or text information does not catch user's fancy, the user can input information. For example, when a jacket and sunglasses of a model in a picture is clipped, the clipped jacket and sunglasses can be classified to the different category than the user's intention if the peripheral information have little relationship to the clipped content or irrelevant words have been disclosed as text information around the clipped content. That is, if a tourist destination, traffic information, a tourist attraction, a suggested place have been disclosed in a text around the clipped picture taken during travel, a jacket and sunglasses can be stored in the travel category, although a user intends to store the jacket and sunglasses in the fashion category, not the travel category because he/she is interested in jacket and sunglasses in the clipped picture. Accordingly, the user can delete automatically input peripheral texts and then, input the user memo indicating fashion, classify and store to the fashion category.
  • In the present disclosure, the selected picture (image data) in which a jacket and black sunglasses worn by a model are the latest-style items catching user's fancy in a browsed web page may be classified into a ‘fashion’ folder in the portable terminal. Also, goal-in scenes (motion data) in a video of a soccer game in which uniform of a player with new style design having new colors may be classified into the ‘fashion’ folder. Further, text data with regard to recent fashion related news may be classified into the ‘fashion’ folder. In this way, there may be a plurality of folders (‘fashion’, ‘movie’, ‘health’ folders) stored in the portable terminal each of which includes a plurality of files. The files may have different file formats as described. Also, the plurality of files having different file formats may be displayed in a display of the portable terminal.
  • FIGS. 2A and 2B illustrate clipping dynamic data according to one embodiment of the present disclosure. Assuming that, while a user who is usually much interested in a movie browses web pages, the user has discovered a first-run film in which his/her most favorite actor appears. The user can store a close-up video of his/her favorite actor. In detail, the user can preset a specific run time and store data of an arbitrary time period before and after a reference time point of selection in binary data form. That is, as illustrated in FIG. 2A, in a case where the user previously designates a specific run time as 12 seconds, the user can store a video of 12 seconds before and after the reference time point of selection. Accordingly, the video is played from 12 seconds before the selected time point to 12 seconds after the selected time point.
  • Regarding detailed information on the video, as illustrated in FIG. 2B, a file having detailed information on data is stored in an eXtensible Markup Language (XML) data form. Desirably, a file describing detailed information of the clip can be created using not only an XML but also other languages such as an OWL.
  • In certain embodiments, selected content is automatically classified and stored based on related information of the selected content. As illustrated in FIG. 2B, if a specific portion is selected by a user, content is automatically classified and stored using related information of selection. The related information of the content can be classified into either data information of the content or peripheral information. The data information of the dynamic data content can include a plurality of supplementary information related to data such as a title, an original data URL, a maker, a creation date, a category, a publication authority, a user memo, an availability period, GPS information, a data summary, information of an application in which clip was carried out, the type of clip binary data, the size of a thumbnail image and a display position. Except for the user memo, the related information of the content are automatically input to a portable terminal when the content is selected by the user. In detail, when the user discovers the whole or part of a web page to clip, which can be a specific picture and moving picture, and the user designates and selects a portion of the web page, content related information is automatically stored in the portable terminal. For example, as illustrated in FIG. 28, assuming that, during movement, the user searches the previews for a movie in which a user's favorite actor appears. If the user selects a close-up video of his/her favorite actor among the previews for a first-run film, the title of “MY BLUEBERRY NIGHTS” is displayed in the category of “MOVIE”, and GPS location information of “Garosugil, Sinsa-dong, Gangnam-gu, Seoul”, where GPS signals from GPS satellites were received is stored. Besides that, related information are automatically stored such as a maker who makes a moving picture, a date of creating the moving picture, a publication authority, an availability period, a data summary, information of an application in which clipping is carried out, the type of clipped binary data, the size of a thumbnail image, and a display position. As described above, in a case where related information does not catch user's fancy, the user can input user memo for related information. In a case where the related information is input by the user, the related information input by the user takes priority over the automatically input related information for classifying and storing the clipped content.
  • As another example of automatic classifying and storing of data, an example using text information can be given. A user can designate a specific portion of web page, and use peripheral text information disclosed in the web page selection. That is if a user selects a close-up video of his/her favorite movie actor, the video can be classified and stored using a movie title, a movie company, names of a movie director, a leading actor and a supporting actor, a screen time, a playing cinema, a Website address introducing a movie, and the like around the moving picture. For example, assume that the following text has been disclosed around the web page in which a moving picture has been uploaded: “title: Bittersweet Life, director: Ji-Un KIM, genre: action, leading actor: Byeong-Heon LEE, supporting actor: Roe-Ha Ha KIM, screen time: 120 minutes, playing cinema: CGV, Website address introducing movie: http://WWW.A BITTERSWEETLIFE.com”. In a case where the user being the favor of the actor ‘Byeong-Heon LEE’ selects the moving picture, the moving picture is automatically input to the main category of ‘MOVIE’ and a subcategory of ‘Byeong-Heon LEE’. As detailed items, the title ‘bittersweet life’, the director ‘Ji-Un KIM’, the genre ‘action’, the screen time ‘120 minutes’, the playing cinema ‘CGV’, and the related website address ‘http://WWW.A BITTERSWEETLIFE.com’ are automatically input. Accordingly, the user can automatically classify and store the selected content using the text information among the peripheral information.
  • Another embodiment according to the present disclosure uses the peripheral information, which a user inputs. In a case where a moving picture of an actor taken in a restaurant has been clipped, and a restaurant shop name and related food have been disclosed, the moving picture can be stored in the food category, although a user intends to store the moving picture in the movie category, not the food category because he/she is interested in the actor in the moving picture. Accordingly, the user may delete the automatically input peripheral text, input the user memo indicating movie category and classify and store the selected moving picture to the movie category.
  • Although an example of clipping a video of a first-run film has been described in one exemplary embodiment of the present invention, a user can store a preset arbitrary time on a basis of a selected time point for clipping audio data.
  • In the present disclosure, the selected close-up video (motion data) of actor with preset run time with reference to the reference time point of selection may be classified into the ‘movie’ folder in a portable terminal. And, the selected content such as audio data in the first-run film may be classified into the ‘movie’ folder. In this way, there may be a plurality of folders (‘movie’, ‘fashion’ ‘health’ folders) stored in the portable terminal each of which includes a plurality of files. The files may have different file formats as described. Also, the plurality of files having different file formats may be displayed in a display of the portable terminal.
  • FIG. 3A illustrates clipped binary data according to one embodiment of the present disclosure. In FIG. 3A, it illustrates the various forms of the files in which clipped binary data are stored after being selected. In detail, if a specific portion of a web page is selected by a user and clipping is performed, clipped content are stored in a binary data and XML data forms. The clipped binary data do not have an extension, so the type of the data cannot be detected. Unlike this, the XML data contain information on the data, so the type of the data can be detected with the XML data. As illustrated in FIG. 3A, the plurality of selected contents such as the motion data, the audio data, and the image data may be stored as the file in the binary data in the folder in the portable terminal according to one embodiment of the present disclosure. These files may have different file formats. There may be a plurality of folders in accordance with the categories such as ‘movie’, ‘fashion’ and ‘health’.
  • FIG. 3B illustrates exemplary XML data according to one embodiment of the present disclosure. In FIG. 3B, the file containing detail information on data is stored in a XML data form. A file containing detail information of the clip may be created using not only an XML but also other languages such as an OWL. The detail information can include a plurality of supplementary information related to corresponding data such as a title, an original data URL, a maker, a creation date, a category, a publication authority, a user memo, an availability period, GPS information, a data summary, information of an application in which clip is carried out, the kind of clip binary data, the size of a thumbnail image and a display position. Desirably, when being selected, some values among the detailed information are automatically input by a terminal and, if necessary, the user can make a modification for the whole information including the value automatically input by the terminal, or input in detail.
  • FIGS. 4A to 4D illustrates an example of consuming a generated clip according to an exemplary embodiment of the present invention. Each clipped content has binary data, so the clipped content is providing sufficient information to a user only by itself. In detail, selected and previously stored content are each stored according to a specific criterion. As illustrated in FIG. 4A, each clipped content is captured and displayed in an image file form. Also, detailed information such as each title and clipped time is displayed together with the image file form, so the user can select specific data that the user wants to consume. After that, as illustrated in FIGS. 4B and 4C, if specific data is selected by the user, the user can identify detailed information on the selected data. In detail, FIG. 4B illustrates a case of consuming detailed information of selected data (i.e., a picture) and FIG. 4C illustrates a case of consuming detailed information of selected data. In detail, the detailed information of the selected data is displayed. That is, a title and classification reference (i.e., fashion) of a clipped picture, an address of a website in which a picture is uploaded, comments on a history of a brand, a clipped picture, etc. are displayed. FIG. 4D illustrates an example of accessing an original web page in a case where original URL information of a clip exists within XML data describing detailed information of clipped data. In detail, FIG. 4D illustrates that, in a case where interposed original URL information of a clipped picture exists among the detailed information of FIG. 4C, if the original URL information is selected, it is possible to access a website in which an original picture is interposed.
  • FIGS. 5A and 5B illustrate one embodiment of sharing clipped data with registered friends according to the present disclosure. As illustrated in FIG. 5A, if a user searches for clipped data about Jeju-do travel, the clipped data related to the Jeju-do travel are displayed by title or date. In FIG. 5B, if the user searches friend's clipped data and selects interested data, the user can consume the selected picture or moving picture. In detail, the user can store the clipped data in his/her own terminal, and transmit the stored data to a server by user's selection. The thus transmitted data is automatically updated at the server, and the updated data are again transmitted to other users' terminals upon the users' requests. That is, the server searches for and provides the clips corresponding to the same category among clips of users and registered friends. In certain embodiments, if the user wants to receive the data about the Jeju-do travel which are clipped by the friends, the user can select the corresponding category (i.e., Jeju-do travel), receive and consume data which are clipped during the Jeju-do travel of the friends.
  • The criterion for searching for content can include not only the same category but also the similarity of GPS information. In certain embodiment, if data have been clipped in the Jeju-do area, the user can search for the data clipped by friends using the GPS information. That is, the user can not only search a category ‘Jeju-do travel’ but also can search using GPS location information of Jeju-do and share content relating to Jeju-do travel clipped by the friends.
  • FIG. 6 is a flowchart illustrating a method for clipping content at a portable terminal according to one embodiment of the present disclosure. As illustrated in FIG. 6, in a case where the user discovers a picture, a moving picture or a sound source to clip and to store, the user executes an application SIW providing with a clip function (step 601). After that, the user designates a specific portion to clip, and performs the clipping. That is, a specific portion of content is selected by the user (step 602). In certain embodiments, if the user discovers data to clip, the user can begin clipping by pressing a clip button. The behavior of pressing the clip button can be replaced with other behaviors of forwarding a user's clipping intention such as selecting a UI icon or gesture which can be recognized by using a sensor. For detailed portion designation, the user can designate a portion using an input tool. Desirably, the user can drag and designate a portion to clip by means of the input tool.
  • If the user designates the specific portion to clip and performs clipping, clipped binary data and XML data are created (step 603). In certain embodiments, in a case where data to clip is static data such as a web page or a picture, the portion designated by the user is captured in an image file form and is stored in a binary data form. In a case where data is dynamic data such as a moving picture and a currently playing sound source, the user can preset a specific time period and store, in binary data form, data for the preset time period before and after a reference time point of selection. Also, a file containing detailed information on data can be stored in a XML data form. Desirably, a file containing detailed information of clipped data may be created using not only an XML but also other languages such as an OWL.
  • After that, the terminal inquires if the user will input data information of the selected contents (step 604). If the user will input the data information of the selected contents (step 604), the terminal receives the input of the data information from the user (step 605) and then, classifies the selected content using the user input data information, automatically input data information, and automatically input peripheral information (step 606). The content data information can include a plurality of supplementary information related to data such as a title, an original data URL, a maker, a creation date, a category, a publication authority, a user memo, an availability period, GPS information, a data summary, information of an application in which clipping was carried out, the type of clip binary data, the size of a thumbnail image, and a display position. Except for the user memo, the related information of the content is automatically input to a portable terminal when the content is selected by the user. In certain embodiments, when the user discovers the whole or part of a web page to clip, a specific picture and moving picture while the user searches content, the user designates and selects a portion, and contents related information is automatically stored in the portable terminal. Among the contents data information, the user memo is input, not automatically, by the user. Also, the user may input the contents data information.
  • If the content data information is input by the user, the contents data information input by the user takes priority over automatically stored data information. If the user does not input the data information in the inquiry process (step 604), the terminal classifies the selected content using the automatically input data information and automatically input peripheral information (step 607). The peripheral information can include at least one of metadata information of the selected content, text information of the selected content and information input by the user.
  • If the selected contents are classified, the terminal stores the classified contents and completes the clipping, (step 608). If the content is stored in the terminal, the portable terminal determines if synchronizing function to a server upon clipping will be enabled (step 609). In certain embodiment, the portable terminal can transmit to the server on every clipping events through an inquiry of the user as to whether to transmit to the server, or can automatically transmit to the server by a preset value. For example, the terminal can either transmit to the server automatically on every clippings, transmit periodically to the server, transmit to the server at a defined time, or transmit to the server when a specific condition is satisfied. That is, transmitting periodically to the server represents transmitting to the server every one hour after storing clipped data. Transmitting to the server at the defined time represents transmitting to the server at 24:00 pm every day. Transmitting to the server when the specific condition is satisfied desirably represents transmitting to the server upon connection to the network including wireless fidelity (Wifi).
  • If it is determined in the determination process (step 609) that the function of synchronizing to the server is enabled, the terminal transmits clipped content to the server (step 610) and terminates all of the processes. If it is determined in the determination process (step 609) that the function of synchronizing to the server has not been enabled, the terminal can terminate the process without transmitting the clipped content to the server.
  • The method described above in relation with FIG. 6 under of the present invention may be provided as one or more instructions in one or more software modules, or computer programs stored in an electronic device including a portable terminal. The present invention may be implemented in an electronic device including a portable terminal such as, for example, a smart phone and a mobile telecommunication terminal. Hereunder, a portable terminal is used as an example for the electronic device.
  • FIG. 7 is a block diagram illustrating a configuration of a clip apparatus based on content data information of a portable terminal according to an exemplary embodiment of the present invention. As illustrated in FIG. 7, the clip apparatus according to the present disclosure can include a controller 701, a decision unit 702, a classification unit 703, a GPS module 704, a communication module 705, an input unit 706, a display unit 707, and a storage unit 708.
  • The controller 701 controls the general operation of the portable terminal. In certain embodiments, the controller 701 can include the decision unit 702 and the classification unit 703.
  • The decision unit 702 is provided within the controller 701, and clips content and inquires if the user will input data information. For example, the decision unit 702 selects the whole or part of the specific content, inquires if the user will input data information for the selected content, and determines if it is preset to transmit the selected content to the server. Also, the decision unit 702 sends the server a request for content stored in the server, and sends to a preset different portable terminal a request for content which corresponds to the same category as that of the portable terminal and is stored in the different portable terminal.
  • The classification unit 703 automatically classifies the selected content. In the embodiment, the classification unit 703 analyzes related information of the selected content, classifies the selected content, and analyzes data information of the selected content and peripheral information. Also, if the classification unit 703 receives an input of the data information from the user, the classification unit 703 analyzes the data information received from the user, automatically input data information, and the peripheral information of the content, and classifies the selected content based on the analysis. If the user does not input the data information, the classification unit 703 analyzes the automatically input data information and the peripheral information of the content and classifies the selected content.
  • The GPS module 704 can provides with location information of a terminal upon clipping the content, and the location information can be used as supplementary information for classification. In certain embodiments, if clipped data are stored in the Jeju-do area, the GPS module 704 can inquire data which are stored at friend mobile terminals using GPS information among the supplementary information. That is, the user not only searches a category of ‘Jeju-do travel’, but also may search using GPS location information fitting to the Jeju-do area and share the clipped content.
  • The communication module 705 processes a signal transmitted/received through an antenna for voice and data communication. In certain embodiments, if it is preset to transmit stored content to a server, the communication module 705 transmits the content to the server according to the preset mode including transmitting the content to the server whenever the contents are stored, transmitting the content to the server every preset time, or transmitting the contents to the server upon meeting preset conditions. Also, the communication module 705 receives contents corresponding to the same category from other portable terminals, receives a request for content corresponding to the same category as that of a portable terminal from a preset different portable terminal, and transmits the content corresponding to the same category to the other portable terminals.
  • The input unit 706 provides with input command which is generated by user's selection to the controller 701. In certain embodiment, if an application S/W providing with a clip function is opened, the input unit 706 receives an input command of executing a clip function.
  • The display unit 707 displays status information of the portable terminal, menu screen and story information by the control of the controller 701. For instance, the display unit 707 can include a touch screen.
  • The storage unit 708 can include a program storage unit and a data storage unit. The program storage unit stores a program for controlling an operation of the portable terminal. The data storage unit stores data generated in execution of a program. In certain embodiment, the storage unit 708 stores classified content and, in a case where selected content is sound source or moving picture data, the storage unit 708 stores data during a preset arbitrary time period before and after a time point at which data is selected.
  • In the aforementioned embodiments, the controller 701 can perform the general function of the terminal. In the present disclosure, general functions are separately described in order to distinguish other functions. Thus, when a product is actually realized, the controller 701 can process all of the functions of the terminal, or the controller 701 can process only some of the functions.
  • Embodiments of the present invention according to the claims and description in the specification can be realized in the form of hardware, software or a combination of hardware and software.
  • Such software may be stored in a computer readable storage medium. The computer readable storage medium stores one or more programs (software modules), the one or more programs comprising instructions, which when executed by one or more processors in an electronic device, cause the electronic device to perform methods of the present invention.
  • Such software may be stored in the form of volatile or non-volatile storage such as, for example, a storage device like a ROM, whether erasable or rewritable or not, or in the form of memory such as, for example, RAM, memory chips, device or integrated circuits or on an optically or magnetically readable medium such as, for example, a CD, DVD, magnetic disk or magnetic tape or the like. It will be appreciated that the storage devices and storage media are embodiments of machine-readable storage that are suitable for storing a program or programs comprising instructions that, when executed, implement embodiments of the present invention. Embodiments provide a program comprising code for implementing apparatus or a method as claimed in any one of the claims of this specification and a machine-readable storage storing such a program. Still further, such programs may be conveyed electronically via any medium such as a communication signal carried over a wired or wireless connection and embodiments suitably encompass the same.
  • While the invention has been shown and described with reference to certain preferred embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention as defined by the appended claims.

Claims (40)

    What is claimed is:
  1. 1. A method for managing contents at a portable terminal, the method comprising:
    classifying selected contents based on an analysis of related information of the selected contents into a plurality of folders at the portable terminal; and
    storing the classified contents in the folders according to the classification wherein at least one folder among the plurality of folders includes contents having different file formats.
  2. 2. The method of claim 1, wherein the content comprises the whole or portion of the content.
  3. 3. The method of claim 1, wherein the related information comprises data information of the selected content and peripheral information of the selected content.
  4. 4. The method of claim 3, wherein the data information of the content comprises at least one of a title of the selected content, an original data Uniform Resource Locator (URL), a maker, a creation date, a category, a publication authority, a user memo, an availability period, Global Positioning System (GPS) information, a data summary, information of an application in which clipping was carried out, a thumbnail image size, and a display position.
  5. 5. The method of claim 4, wherein the title of the selected contents, the original data URL, the maker, the creation date, the category, the publication authority, the availability period, the GPS information, the data summary, the information of the application in which the clip is carried out, the thumbnail image size, and the display position are automatically input when the content are selected.
  6. 6. The method of claim 3, wherein the peripheral information of the content comprises at least one of metadata information of the content, text information contained in the content, and user input information.
  7. 7. The method of claim 3, further comprising:
    inquiring if a user will input data information of the selected content; and
    if receiving the input of the data information from the user, analyzing the data information received from the user, the automatically input data information, and the peripheral information of the content.
  8. 8. The method of claim 7, further comprising, if not being received the input of the data information, analyzing the automatically input data information and the peripheral information of the contents.
  9. 9. The method of claim 1, wherein storing the classified contents comprises, in a case where the selected content is sound data or moving picture data, storing selected content during a preset arbitrary time period before and after at a time point when the data is selected.
  10. 10. The method of claim 1, further comprising:
    determining if it is preset to transmit the stored content to a server; and
    if it is preset to transmit the stored content to the server, transmitting the content to the server according to a preset condition.
  11. 11. The method of claim 10, wherein transmitting the content to the server according to the preset condition is either transmitting to the server whenever the content is stored, periodically transmitting to the server, transmitting to the server every preset time, or transmitting to the server upon connection to network including wireless fidelity (Wifi).
  12. 12. The method of claim 10, further comprising:
    sending a request for the content stored at the server to the server; and
    receiving the content from the server.
  13. 13. The method of claim 1, further comprising:
    sending to the server a request for content which corresponds to the same category as that of the portable terminal and is stored at another portable terminal; and
    receiving the content which corresponds to the same category and is stored at the different portable terminal through the server.
  14. 14. The method of claim 1, further comprising:
    sending a request for content which correspond to the same category as that of the portable terminal and is stored in another portable terminal; and
    receiving the content corresponding to the same category from another portable terminal.
  15. 15. The method of claim 1, further comprising:
    receiving a request for content corresponding to the same category as that of the portable terminal from a preset second portable terminal; and
    transmitting the content corresponding to the same category to the second portable terminal.
  16. 16. The method of claim 15, wherein the content corresponding to the same category as that of the portable terminal was stored within a preset time period.
  17. 17. A method for sharing clipped content at a portable terminal, the method comprising:
    receiving, by a server, content stored at a first portable terminal from the first portable terminal;
    updating the received content; and
    transmitting the updated content to a second portable terminal by a request of the second portable terminal.
  18. 18. The method of claim 17, further comprising:
    receiving a request for content, which corresponds to the same category as that of the first portable terminal and is stored at the second portable terminal from the second portable terminal;
    searching the content, which correspond to the same category and is stored in the second portable terminal; and
    receiving the content corresponding to the same category from the second portable terminal, and transmitting the content to the first portable terminal.
  19. 19. The method of claim 1, wherein the content corresponding to the same category was stored within a specific time period.
  20. 20. A method for sharing content at a portable terminal, the method comprising:
    transmitting a request for content stored at another portable terminal to a server; and
    receiving the content from the server.
  21. 21. The method of claim 20, further comprising:
    sending the request for content which corresponds to the same category as that of the portable terminal and is stored at another portable terminal; and
    receiving the content corresponding to the same category from another portable terminal.
  22. 22. The method of claim 1, wherein the content corresponding to the same category was stored within a specific time period.
  23. 23. An apparatus for clipping and sharing content at a portable terminal, the apparatus comprising:
    a decision unit for selecting specific content;
    a classification unit for analyzing related information of the selected content, and for classifying the selected content; and
    a storage unit for storing the classified content according to the classification wherein at least one folder among a plurality of folders storing the classified content includes files having the selected content with different file formats.
  24. 24. The apparatus of claim 23, wherein the decision unit selects the whole or partial portion of the content.
  25. 25. The apparatus of claim 23, wherein the classification unit analyzes data information of the selected content and peripheral information of the selected content.
  26. 26. The apparatus of claim 25, wherein the data information of the contents comprises at least one of a title of the selected content, an original data Uniform Resource Locator (URL), a maker, a creation date, a category, a publication authority, a user memo, an availability period, Global Positioning System (GPS) information, a data summary, information of an application in which clipping was carried out, a thumbnail image size, and a display position.
  27. 27. The apparatus of claim 26, wherein the title of the selected contents, the original data URL, the maker, the creation date, the category, the publication authority, the availability period, the GPS information, the data summary, the information of the application in which the clipping was carried out, the thumbnail image size, and the display position are automatically input when the content is selected.
  28. 28. The apparatus of claim 25, wherein the peripheral information of the content comprises at least one of metadata information of the content, text information of the content and user input information.
  29. 29. The apparatus of claim 23, wherein the decision unit inquires if a user will input data information of the selected content, and
    if the data information being input, the classification unit analyzes the data information received from the user, automatically input data information, and peripheral information of the content.
  30. 30. The apparatus of claim 29, wherein, if the data information not being input, the classification unit analyzes the automatically input data information and the peripheral information of the content.
  31. 31. The apparatus of claim 23, wherein, in a case where the selected content is sound source or moving picture data, the storage unit stores data during a preset arbitrary time period before and after at a time point when the data is selected.
  32. 32. The apparatus of claim 23, further comprising a communication module for transmitting the contents to the server according to a preset condition if the decision unit determines that it is preset to transmit the stored contents to the server.
  33. 33. The apparatus of claim 32, wherein the communication module either transmits the content to the server whenever the contents are stored, transmits periodically to the server, transmits to the server every preset time, or transmits upon connection to the network including wireless fidelity (Wifi).
  34. 34. The apparatus of claim 32, wherein the decision unit sends a request for content to the server; and
    the communication module receives the content from the server.
  35. 35. The apparatus of claim 23, further comprising a communication module for sending to a server a request for content which corresponds to the same category as that of the portable terminal and is stored at another portable terminal, and for receiving the requested content from the server.
  36. 36. The apparatus of claim 23, further comprising a communication module configured to send to a preset second portable terminal a request for content which corresponds to the same category as that of a first portable terminal and is stored in the second portable terminal, the communication module being configured to receive the content corresponding to the same category from the second portable terminal.
  37. 37. The apparatus of claim 23, further comprising a communication module for receiving a request for content corresponding to the same category as that of a first portable terminal from a preset second portable terminal, and for transmitting the content corresponding to the same category, to the second portable terminal.
  38. 38. The apparatus of claim 37, wherein the content was stored within a specified time period.
  39. 39. A method for managing contents at a portable terminal, the method comprising:
    classifying selected contents based on an analysis of related information of the selected contents into a plurality of folders at the portable terminal; and
    displaying the classified contents in the folders according to the classification wherein at least one folder among the plurality of folders includes contents having different file formats.
  40. 40. A portable terminal for displaying content, the terminal comprising:
    means for classifying selected contents based on an analysis of related information of the selected contents into a plurality of folders at the portable terminal; and
    means for displaying the classified contents in the folders according to the classification wherein at least one folder among the plurality of folders includes contents having different file formats.
US13629394 2011-09-27 2012-09-27 Apparatus and method for clipping and sharing content at a portable terminal Pending US20130080531A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
KR20110097578A KR20130033721A (en) 2011-09-27 2011-09-27 Clip apparatas and method for based on contents in a portable terminal
KR10-2011-0097578 2011-09-27

Publications (1)

Publication Number Publication Date
US20130080531A1 true true US20130080531A1 (en) 2013-03-28

Family

ID=47912457

Family Applications (1)

Application Number Title Priority Date Filing Date
US13629394 Pending US20130080531A1 (en) 2011-09-27 2012-09-27 Apparatus and method for clipping and sharing content at a portable terminal

Country Status (5)

Country Link
US (1) US20130080531A1 (en)
EP (3) EP2761574A4 (en)
KR (1) KR20130033721A (en)
CN (2) CN103827913B (en)
WO (1) WO2013048091A3 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140364158A1 (en) * 2013-06-10 2014-12-11 Samsung Electronics Co., Ltd. Mobile terminal and method for controlling the same
US20140380142A1 (en) * 2013-06-20 2014-12-25 Microsoft Corporation Capturing website content through capture services

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530865A (en) * 1993-03-03 1996-06-25 Apple Computer, Inc. Method and apparatus for improved application program switching on a computer-controlled display system
US20060206495A1 (en) * 2003-04-07 2006-09-14 Johan Sebastiaan Van Gageldonk Method and apparatus for grouping content items
US20070106952A1 (en) * 2005-06-03 2007-05-10 Apple Computer, Inc. Presenting and managing clipped content
US7315848B2 (en) * 2001-12-12 2008-01-01 Aaron Pearse Web snippets capture, storage and retrieval system and method
US20120066584A1 (en) * 2010-09-15 2012-03-15 Samsung Electronics Co., Ltd. Host apparatus and method of displaying content by the same
US8539336B2 (en) * 2011-07-28 2013-09-17 Scrawl, Inc. System for linking to documents with associated annotations

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3993385B2 (en) * 2000-12-28 2007-10-17 株式会社エヌ・ティ・ティ・ドコモ Content delivery system, copyright protection system and content receiving terminal
EP1491063B1 (en) * 2002-04-03 2008-08-13 Nokia Corporation Enabling a content provider initiated content delivery via a specific radio access network
KR100573037B1 (en) * 2005-10-20 2006-04-14 (주)인트로모바일 Content extraction server on the rss and method thereof, service system for idle screen on mobile using the same
US20100070501A1 (en) 2008-01-15 2010-03-18 Walsh Paul J Enhancing and storing data for recall and use using user feedback

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5530865A (en) * 1993-03-03 1996-06-25 Apple Computer, Inc. Method and apparatus for improved application program switching on a computer-controlled display system
US7315848B2 (en) * 2001-12-12 2008-01-01 Aaron Pearse Web snippets capture, storage and retrieval system and method
US20060206495A1 (en) * 2003-04-07 2006-09-14 Johan Sebastiaan Van Gageldonk Method and apparatus for grouping content items
US20070106952A1 (en) * 2005-06-03 2007-05-10 Apple Computer, Inc. Presenting and managing clipped content
US20120066584A1 (en) * 2010-09-15 2012-03-15 Samsung Electronics Co., Ltd. Host apparatus and method of displaying content by the same
US8539336B2 (en) * 2011-07-28 2013-09-17 Scrawl, Inc. System for linking to documents with associated annotations

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140364158A1 (en) * 2013-06-10 2014-12-11 Samsung Electronics Co., Ltd. Mobile terminal and method for controlling the same
US9716989B2 (en) * 2013-06-10 2017-07-25 Samsung Electronics Co., Ltd. Mobile terminal and method for capturing a screen and extracting information to control the same
US20140380142A1 (en) * 2013-06-20 2014-12-25 Microsoft Corporation Capturing website content through capture services

Also Published As

Publication number Publication date Type
EP3370172A1 (en) 2018-09-05 application
CN103827913A (en) 2014-05-28 application
WO2013048091A2 (en) 2013-04-04 application
CN108133057A (en) 2018-06-08 application
EP2761574A4 (en) 2015-05-06 application
CN103827913B (en) 2018-03-13 grant
EP3367271A1 (en) 2018-08-29 application
WO2013048091A3 (en) 2013-05-23 application
KR20130033721A (en) 2013-04-04 application
EP2761574A2 (en) 2014-08-06 application

Similar Documents

Publication Publication Date Title
US20050234875A1 (en) Methods and systems for processing media files
US20090119614A1 (en) Method, Apparatus and Computer Program Product for Heirarchical Navigation with Respect to Content Items of a Media Collection
US20040220926A1 (en) Personalization services for entities from multiple sources
US20100030740A1 (en) System and method for context enhanced mapping
US20120173551A1 (en) Method and system for storing and retrieving tags
US20050015712A1 (en) Resolving metadata matched to media content
US20040220791A1 (en) Personalization services for entities from multiple sources
US20100049702A1 (en) System and method for context enhanced messaging
US20090100068A1 (en) Digital content Management system
US20090125571A1 (en) Arrangement for synchronizing media files with portable devices
US20110208722A1 (en) Method and apparatus for segmenting and summarizing media content
US20090259971A1 (en) Media mashing across multiple heterogeneous platforms and devices
US8108778B2 (en) System and method for context enhanced mapping within a user interface
US20120036431A1 (en) Server apparatus, electronic apparatus, electronic book providing system, electronic book providing method, electronic book displaying method, and program
US7636509B2 (en) Media data representation and management
US20100169778A1 (en) System and method for browsing, selecting and/or controlling rendering of media with a mobile device
US20120102410A1 (en) Media management for multi-user group
US20090327953A1 (en) Unified navigation model between multiple applications
US7966638B2 (en) Interactive media display across devices
US20070136247A1 (en) Computer-implemented system and method for obtaining customized information related to media content
US8566329B1 (en) Automated tag suggestions
US20060265503A1 (en) Techniques and systems for supporting podcasting
US20080288573A1 (en) Method and apparatus for filtering virtual content
US20080263103A1 (en) Digital asset management system (DAMS)
US20130160065A1 (en) Video distribution system, information providing device, and video information providing method

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD, KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YOON, SEUNG-HYUN;LEE, MOON-SANG;KOO, MIN-SOO;AND OTHERS;REEL/FRAME:029040/0571

Effective date: 20120921