US10853408B2 - Method for providing graphic effect corresponding to configuration information of object and electronic device thereof - Google Patents
Method for providing graphic effect corresponding to configuration information of object and electronic device thereof Download PDFInfo
- Publication number
- US10853408B2 US10853408B2 US15/925,585 US201815925585A US10853408B2 US 10853408 B2 US10853408 B2 US 10853408B2 US 201815925585 A US201815925585 A US 201815925585A US 10853408 B2 US10853408 B2 US 10853408B2
- Authority
- US
- United States
- Prior art keywords
- objects
- sub
- display
- electronic device
- processor
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 238000000034 method Methods 0.000 title claims abstract description 34
- 230000000694 effects Effects 0.000 title abstract description 31
- 238000001514 detection method Methods 0.000 claims abstract description 116
- 238000004891 communication Methods 0.000 claims description 47
- 230000008859 change Effects 0.000 claims description 5
- 230000006870 function Effects 0.000 description 28
- 230000001413 cellular effect Effects 0.000 description 8
- 238000012545 processing Methods 0.000 description 8
- 230000008569 process Effects 0.000 description 7
- 230000003287 optical effect Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000014509 gene expression Effects 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 230000004048 modification Effects 0.000 description 4
- 238000004364 calculation method Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 239000008280 blood Substances 0.000 description 2
- 210000004369 blood Anatomy 0.000 description 2
- 230000000295 complement effect Effects 0.000 description 2
- 238000002591 computed tomography Methods 0.000 description 2
- 238000004590 computer program Methods 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000012806 monitoring device Methods 0.000 description 2
- 238000012015 optical character recognition Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 2
- 229910052724 xenon Inorganic materials 0.000 description 2
- FHNFHKCVQCLJFQ-UHFFFAOYSA-N xenon atom Chemical compound [Xe] FHNFHKCVQCLJFQ-UHFFFAOYSA-N 0.000 description 2
- WQZGKKKJIJFFOK-GASJEMHNSA-N Glucose Natural products OC[C@H]1OC(O)[C@H](O)[C@@H](O)[C@@H]1O WQZGKKKJIJFFOK-GASJEMHNSA-N 0.000 description 1
- 230000001133 acceleration Effects 0.000 description 1
- 230000003213 activating effect Effects 0.000 description 1
- 238000002583 angiography Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000036772 blood pressure Effects 0.000 description 1
- 230000036760 body temperature Effects 0.000 description 1
- 230000010267 cellular communication Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000002567 electromyography Methods 0.000 description 1
- 239000000446 fuel Substances 0.000 description 1
- 239000008103 glucose Substances 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000006698 induction Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 238000001646 magnetic resonance method Methods 0.000 description 1
- 229910044991 metal oxide Inorganic materials 0.000 description 1
- 150000004706 metal oxides Chemical class 0.000 description 1
- 238000004377 microelectronic Methods 0.000 description 1
- 238000010295 mobile communication Methods 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 239000004984 smart glass Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000005406 washing Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/43—Querying
- G06F16/432—Query formulation
- G06F16/434—Query formulation using image data, e.g. images, photos, pictures taken by a user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04812—Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G06K9/00671—
-
- G06K9/46—
-
- G06K9/4604—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/13—Edge detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- G06K2209/03—
-
- G06K2209/25—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/02—Recognising information on displays, dials, clocks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/09—Recognition of logos
Definitions
- the present disclosure relates to a device and method for providing a graphic effect corresponding to configuration information of an object displayed on a display of an electronic device.
- electronic devices provide various functions with the development of the information & communication technology and the semiconductor technology.
- electronic devices can provide multimedia services such as a voice call service, a message service, a broadcasting service, a wireless internet service, a camera service, and a music service.
- Electronic devices can provide a searching service using a camera to improve convenience for users. For example, electronic devices can search for and output the information about a product photographed by a camera.
- An electronic device can extract an object to search for from an image acquired from a camera when a search service is provided.
- the electronic device can search for information about a corresponding object from a predetermined database, and can output the information through a display.
- a user of the electronic device may not clearly recognize an object extracted by the electronic device.
- the electronic device requires a method for clearly recognizing an object extracted from an image acquired through a camera in order to increase accuracy in searching for an object.
- Various embodiments may provide a device and method for detecting at least one object from an image displayed on a display in an electronic device.
- Various embodiments may provide a device and method for displaying detection information of an object in an image displayed on a display in an electronic device.
- an electronic device includes a display, at least one processor, and a memory coupled to the at least one processor.
- the memory can store instructions executable by the at least one processor to control the display to display an image, detect at least one object from the image, and control the display to display detection information corresponding to the at least one object based on configuration information of the at least one object.
- a method of an electronic device may include an operation of displaying an image on a display electrically connected with the electronic device, an operation of detecting at least one object from the image, and an operation of displaying detection information corresponding to the at least one object based on configuration information of the at least one object.
- various functions described below can be implemented or supported by one or more computer programs, each of which is formed from computer readable program code and embodied in a computer readable medium.
- application and “program” refer to one or more computer programs, software components, sets of instructions, procedures, functions, objects, classes, instances, related data, or a portion thereof adapted for implementation in a suitable computer readable program code.
- computer readable program code includes any type of computer code, including source code, object code, and executable code.
- computer readable medium includes any type of medium capable of being accessed by a computer, such as read only memory (ROM), random access memory (RAM), a hard disk drive, a compact disc (CD), a digital video disc (DVD), or any other type of memory.
- ROM read only memory
- RAM random access memory
- CD compact disc
- DVD digital video disc
- a “non-transitory” computer readable medium excludes wired, wireless, optical, or other communication links that transport transitory electrical or other signals.
- a non-transitory computer readable medium includes media where data can be permanently stored and media where data can be stored and later overwritten, such as a rewritable optical disc or an erasable memory device.
- FIG. 1 illustrates an electronic device in a network environment according to various embodiments
- FIG. 2 illustrates a block diagram of an electronic device according to various embodiments
- FIG. 3 illustrates a block diagram of a program module according to various embodiments
- FIG. 4 illustrates a flowchart for displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments
- FIG. 5 illustrates a screen configuration displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments
- FIG. 6 illustrates a flowchart for acquiring recognition information of an object through a server in an electronic device according to various embodiments
- FIG. 7 illustrates a screen configuration displaying a graphic effect corresponding to the kind of an object in an electronic device according to various embodiments
- FIG. 8 illustrates a flowchart for recognizing an object based on detection information of important objects in an electronic device according to various embodiments
- FIG. 9 illustrates a flowchart for detecting a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments
- FIGS. 10A-10E illustrate a screen configuration displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments
- FIG. 11 illustrates a flowchart for outputting a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments
- FIGS. 12A and 12B illustrate a screen configuration displaying a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments.
- FIG. 13 illustrates a flowchart for displaying a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments.
- FIGS. 1 through 13 discussed below, and the various embodiments used to describe the principles of the present disclosure in this patent document are by way of illustration only and should not be construed in any way to limit the scope of the disclosure. Those skilled in the art will understand that the principles of the present disclosure may be implemented in any suitably arranged system or device.
- an expression “A or B”, “A and/or B”, or the like may include all possible combinations of items enumerated together.
- expressions such as “1 st ” “2 nd ”, “first”, and “second” may be used to express corresponding constitutional elements, it is not intended to limit the corresponding constitutional elements.
- a certain (e.g., 1 st ) constitutional element is mentioned as being “operatively or communicatively coupled with/to” or “connected to” a different (e.g., 2 nd ) constitutional element, the certain constitutional element is directly coupled with/to another constitutional element or can be coupled with/to the different constitutional element via another (e.g., 3 rd ) constitutional element.
- an expression “configured to” used in the present document may be interchangeably used with, for example, “suitable for”, “having the capacity to”, “adapted to”, “made to”, “capable of”, or “designed to” in a hardware or software manner according to a situation.
- an expressed “a device configured to” may imply that the device is “capable of” together with other devices or components.
- a processor configured to perform A, B, and C may imply a dedicated processor (e.g., an embedded processor) for performing a corresponding operation or a generic-purpose processor (e.g., Central Processing Unit (CPU) or an application processor) capable of performing corresponding operations by executing one or more software programs stored in a memory device.
- a dedicated processor e.g., an embedded processor
- a generic-purpose processor e.g., Central Processing Unit (CPU) or an application processor
- An electronic device may include at least one of a smartphone, a tablet personal computer (PC), a mobile phone, a video phone, an electronic book (e-book) reader, a desktop PC, a laptop PC, a netbook computer, a workstation, a server, a personal digital assistant (PDA), a portable multimedia player (PMP), an MP3 player, a mobile medical appliance, a camera, and a wearable device (e.g., smart glasses, a head-mounted-device (HMD), electronic clothes, an electronic bracelet, an electronic necklace, an electronic appcessory, an electronic tattoo, a smart mirror, or a smart watch).
- a wearable device e.g., smart glasses, a head-mounted-device (HMD), electronic clothes, an electronic bracelet, an electronic necklace, an electronic appcessory, an electronic tattoo, a smart mirror, or a smart watch.
- the electronic device may include at least one of, for example, a television, a Digital Video Disk (DVD) player, an audio, a refrigerator, an air conditioner, a vacuum cleaner, an oven, a microwave oven, a washing machine, an air cleaner, a set-top box, a home automation control panel, a security control panel, a media box (e.g., Samsung HomeSyncTM, Apple TVTM, or Google TVTM), a game console (e.g., XboxTM and PlayStationTM, an electronic dictionary, an electronic key, a camcorder, and an electronic photo frame.
- DVD Digital Video Disk
- the electronic device may include at least one of various medical devices (e.g., various portable medical measuring devices (a blood glucose monitoring device, a heart rate monitoring device, a blood pressure measuring device, a body temperature measuring device, etc.), a Magnetic Resonance Angiography (MRA), a Magnetic Resonance Imaging (MRI), a Computed Tomography (CT) machine, and an ultrasonic machine), a navigation device, a Global Positioning System (GPS) receiver, an Event Data Recorder (EDR), a Flight Data Recorder (FDR), a Vehicle Infotainment Devices, an electronic devices for a ship (e.g., a navigation device for a ship, and a gyro-compass), avionics, security devices, an automotive head unit, a robot for home or industry, an automatic teller's machine (ATM) in banks, point of sales (POS) in a shop, or internet device of things (e.g., a light bulb, various sensors, electric or gas meter, or a light bulb
- the electronic device may include at least one of a part of furniture or a building/structure, an electronic board, an electronic signature receiving device, a projector, and various kinds of measuring instruments (e.g., a water meter, an electric meter, a gas meter, and a radio wave meter).
- the electronic device according to various embodiments of the present disclosure may be a combination of one or more of the aforementioned various devices.
- the electronic device according to some embodiments of the present disclosure may be a flexible device. Further, the electronic device according to an embodiment of the present disclosure is not limited to the aforementioned devices, and may include a new electronic device according to the development of technology
- the term “user” may indicate a person who uses an electronic device or a device (e.g., an artificial intelligence electronic device) that uses an electronic device.
- detection of an object may include an operation of extracting at least one object from an image or an operation of recognizing at least one object extracted from an image.
- FIG. 1 illustrates an electronic device 101 in a network environment according to various embodiments.
- the electronic device 101 may include a bus 110 , a processor 120 (e.g., including processing circuitry), a memory 130 , an I/O interface 150 (e.g., including input/output circuitry), a display 160 (e.g., including display circuitry), a communication interface 170 (e.g., including communication circuitry), and a camera 180 .
- the electronic device 101 may not include at least one of the components or may additionally include other components.
- the bus 110 may include a circuit that connects the components ( 120 to 180 ) and transmits signals (for example, control messages and/or data) among the components.
- the processor 120 may include one or more of a Central Processing Unit (CPU), an Application Processor (AP), a Communication Processor (CP), and an Image Signal Processor (ISP).
- the processor 120 for example, can perform calculation or data processing about control and/or communication of one or more other components of the electronic device 101 .
- the processor 120 can detect at least one object from an image displayed on the display 160 .
- the processor 120 can detect characteristic points included in an image displayed on the display 160 by analyzing the image.
- the processor 120 can detect at least one object included in an image based on characteristic points included in the image.
- the processor 120 can detect a plurality of sub-objects from one object detected from an image.
- the processor 120 can divide a sub-object corresponding to the external shape of the vehicle and a sub-object corresponding to the logo of the vehicle.
- the processor 120 when the processor 120 divides a plurality of sub-objects from one object, the processor 120 can acquire the detailed information of a sub-object by transmitting at least one sub-object to at least one server.
- the processor 120 can recognize an object including an object including the sub-object based on the detailed information of the sub-object. For example, by transmitting a plurality of sub-objects divided from an object to different servers corresponding to the sub-objects, respectively, the processor 120 can acquire the detailed information of the sub-objects. For example, by transmitting a plurality of sub-objects divided from an object to one server, the processor 120 can acquire the detailed information of the sub-objects. In this case, the processor 120 can transmit a format of detailed information to be acquired from a server to the server.
- the processor 120 when the processor 120 detects an object from an image displayed on the display 160 , the processor 120 can control the display 160 to display detection information of the object.
- the processor 120 can control the display 160 to display detection information of an object detected from an image displayed on the display 160 , using a graphic variable corresponding to the configuration information of the object.
- the configuration information of an object may include at least one of the kind (for example, a text, an image, and an outline) of an object (or a sub-object), the shape (for example, the curvature, the height, and the form of a line) of an object (or a sub-object), and the color of an object.
- the processor 120 when the processor 120 detects at least one object from an image, the processor 120 can control the display 160 to display detection information close to the object. For example, when the processor 120 divides a plurality of sub-objects from one object, the processor 120 can control the display 160 to display different detection information for the sub-objects, using graphic variables corresponding to the sub-objects, respectively. For example, when the processor 120 detects various properties (for example, a shape or a color) of one object (or a sub-object), the processor 120 can control the display 160 to display detection information corresponding to the various properties of the object (or the sub-object).
- various properties for example, a shape or a color
- the graphic variable may include at least one of the form, color, contrast, definition, and size of detection information displayed close to an object detected by the processor 120 so that a user can recognize the object.
- the form of detection information may include various shapes such as a straight line, a curved line, a polygon, a circle, and a dot.
- the processor 120 when the processor 120 detects a plurality of objects from an image displayed on the display 160 , the processor 120 can control the display 160 to display detection information corresponding to the recognition status of the object. For example, when the processor 120 can detect a plurality of objects from an image displayed on the display 160 , the processor 120 can recognize the objects. For example, the processor 120 may preferentially recognize a first object of a plurality of objects detected from an image. In this case, the processor 120 can control the display 160 to display detection information showing a recognition progression status for the first object and detection information showing recognition standby status for a second object.
- the processor 120 can control the display 160 to display detection information showing a recognition completion status for the first object and detection information showing a recognition progression status for the second object.
- the processor 120 can control the display 160 to display detection information showing a recognition completion status for the first object and the second object.
- the processor 120 can control the display 160 to display different detection information for the first object and the second object, using graphic variables corresponding to the configuration information of the first object and the second object.
- the processor 120 can acquire the detailed information of the first object and the second object detected from an image by transmitting the objects to different servers corresponding to the objects, respectively.
- the processor 120 can acquire the detailed information of the first object and the second object detected from an image by transmitting the objects to one server.
- the memory 130 may include a volatile and/or nonvolatile memory.
- the memory 130 can store, for example, commands or data related to at least one other component of the electronic device 101 .
- the memory 130 can store a graphic variable corresponding to the configuration information of an object (or a sub-object).
- the memory 130 can store software and/or a program 140 .
- the program 140 may include a kernel 141 , a middleware 143 , an Application Programming Interface (API) 145 , or an application program (or an “application”) 147 .
- At least some of the kernel 141 , middleware 143 , and API 145 may be referred to as an Operating System (OS).
- OS Operating System
- the kernel 141 can control or manage system resources (for example, the bus 110 , processor 120 , or memory 130 ) that are used to perform operations or functions that are implemented by other programs (for example, the middleware 143 , the API 145 , or the application program 147 ). Further, the kernel 141 can provide an interface that can control or manage system resources, by accessing individual components of the electronic device 101 via the middleware 143 , API 145 , or application program 147 .
- system resources for example, the bus 110 , processor 120 , or memory 130
- other programs for example, the middleware 143 , the API 145 , or the application program 147 .
- the kernel 141 can provide an interface that can control or manage system resources, by accessing individual components of the electronic device 101 via the middleware 143 , API 145 , or application program 147 .
- the middleware 143 can function as a relay so that the API 145 or application program 147 can transmit and receive data by communicating with the kernel 141 . Further, the middleware 143 can process one or more work requests received from the application program 147 in order of priority. For example, the middleware 143 can give a priority to be able to use system resources (for example, the bus 110 , the processor 120 , or the memory 130 ) of the electronic device 101 to at least one of the application programs 147 and process one or more work requests.
- system resources for example, the bus 110 , the processor 120 , or the memory 130
- the API 145 which is an interface for the application program 147 to control a function provided to the kernel 141 or the middleware 143 , for example, may include at least one interface or function (for example, instruction) for file control, window control, image control, or text control.
- the I/O interface 150 can transmit commands or data input from a user or another external device to other component(s) of the electronic device 101 or can output commands or data received from other component(s) of the electronic device 101 to a user or another external device.
- the I/O interface 150 may include at least one physical button such as a home button, a power button, and a volume control button.
- the I/O interface 150 may include a speaker for outputting audio signals and a microphone for collecting audio signals.
- the display 160 can display various contents (for example, a text, an image, a video, an icon, and/or a symbol) to a user.
- the display 160 may include a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display, an Organic Light Emitting Diode (OLED) display, or a Micro Electronic Mechanical System (MEMS) display, or an electronic paper display.
- the display 160 may include a touch screen.
- the display 160 can receive touching, gesturing, approaching, or hovering input by an electronic pen or a part of the body of a user.
- the communication interface 170 can set communication between the electronic device 101 and an external device (for example, a first external electronic device 102 , a second external electronic device 104 , or a server 106 ).
- the communication interface 170 can be connected to the network 172 and can communicate with an external device (for example, the second external electronic device 104 or the server 106 ) through wireless communication or wire communication.
- wireless communication may include cellular communication using at least one of LTE, LTE-A (LTE Advance), Code Division Multiple Access (CDMA), Wideband CDMA (WCDMA), Universal Mobile Telecommunications System (UMTS), Wireless Broadband (WiBro), and Global System for Mobile Communications (GSM).
- wireless communication 174 may include at least one of Wireless Fidelity (WiFi), Bluetooth, Bluetooth Low Energy (BLE), Zigbee, Near Field Communication (NFC), magnetic secure transmission, Radio Frequency (RF), or Body Area Network (BAN).
- the wireless communication may include GNSS.
- GNSS may be a Global Positioning System (GPS), a Global Navigation Satellite System (Glonass), a Beidou (Beidou Navigation Satellite System), or a Galileo (the European global satellite-based navigation system).
- GPS Global Positioning System
- Glonass Global Navigation Satellite System
- Beidou Beidou Navigation Satellite System
- Galileo the European global satellite-based navigation system
- wire communication may include at least one of a Universal Serial Bus (USB), a High Definition Multimedia Interface (HDMI), an RS-232 (Recommended Standard-232), power-line communication, and a Plain Old Telephone Service (POTS).
- the network 172 may include at least one of telecommunication networks, for example, a computer network (for example, LAN or WAN), the internet, and a telephone network.
- the first and second external electronic device 102 and 104 may be devices the same as or different from the electronic device 101 . According to various embodiments, all or some of operations that are performed by the electronic device 101 may be performed by another electronic device or a plurality of other electronic devices (for example, the electronic devices 102 and 104 or the server 106 ). According to an embodiment, when the electronic device 101 has to perform a function or service automatically or due to a request, the electronic device 101 can request at least partial function related to the function or service to another device (for example, the electronic devices 102 and 104 or the server 106 ) additionally or instead of performing the function or service by itself.
- Another electronic device can perform the requested function or the additional function and transmit the result to the electronic device 101 .
- the electronic device 101 can provide the requested function or service based on the received result or by additionally processing the received result.
- cloud computing, distributed computing, or client-server computing may be used.
- the camera 180 can take still images (pictures) and moving images by collecting image information about a subject.
- the camera 180 can be controlled to photograph a subject by the processor 120 .
- the camera 180 can transmit photographed data (for example, images) to the display 160 and the processor 120 .
- the camera 180 may include at least one of an image sensor, a lens, an Image Signal Processor (ISP), and a flash (for example, an LED or a xenon lamp).
- the image sensor may be a Charged Coupled Device (CCD) or a Complementary Metal-Oxide Semiconductor (CMOS).
- the camera 180 may include a plurality of cameras.
- the camera 180 may include a plurality of cameras disposed on the front or rear side of the electronic device 101 .
- the camera 180 may include a plurality of cameras disposed on the rear side of the electronic device 101 .
- FIG. 2 illustrates a block diagram of an electronic device 201 according to various embodiments.
- the electronic device 201 may include the entire or a portion of the electronic device 101 shown in FIG. 1 .
- the electronic device 201 may include one or more processors 210 (for example, AP), a communication module 220 , a subscriber identification module 224 , a memory 230 , a sensor module 240 , an input device 250 , a display 260 , an interface 270 , an audio module 280 , a camera module 291 , a power management module 295 , a battery 296 , an indicator 297 , and a motor 298 .
- processors 210 for example, AP
- a communication module 220 for example, AP
- subscriber identification module 224 for example, AP
- a memory 230 for example, a communication module 220 , a subscriber identification module 224 , a memory 230 , a sensor module 240 , an input device 250 ,
- the processor 210 can control a plurality of hardware or software components connected to the processor 210 by operating an operating system or an application and can perform processing and calculation on various data.
- the processor 210 may be a System on Chip (SoC).
- SoC System on Chip
- the processor 210 may further include a Graphic Processing Unit (GPU) and/or an Image Signal Processor (ISP).
- the processor 210 may include at least some (a cellular module 221 ) of the components shown in FIG. 2 .
- the processor 210 can load and process commands or data received from other components (for example, a nonvolatile memory) on a volatile memory and can store the resultant data on a nonvolatile memory.
- the processor 210 can control the display 260 to display detection information corresponding to the configuration information of an object detected from an image displayed on the display 260 .
- the configuration information of an object may include at least one of the kind of an object (or a sub-object), the shape of an object (or a sub-object), and the color of an object.
- the kind of an object may include at least one of a texture, an image, an outline, an Optical Character Reader (OCR), a barcode, a Quick Response (QR) code, and a logo
- the shape of an object may include at least one of the curvature, a height, and a form of an outline
- the communication module 220 may have a configuration the same as or similar to that of the communication interface 170 shown in FIG. 1 .
- the communication module 220 may include a cellular module 221 , an WiFi module 223 , a Bluetooth module 225 , an GNSS module 227 , an NFC module 228 , and an RF module 229 .
- the cellular module 221 can provide a voice call, a video call, a text service, or an internet service through a communication network.
- the cellular module 221 can identify and authenticate the electronic device 201 in a communication network, using a subscriber identification module 224 (for example, a SIM card).
- the cellular module 221 can perform at least some of the functions that the processor 210 can provide.
- the cellular module 221 may include a Communication Processor (CP).
- CP Communication Processor
- At least some (for example, two or more) of the cellular module 221 , WiFi module 223 , Bluetooth module 225 , GNSS module 227 , and NFC module 228 may be included in one Integrated Chip (IC) or IC package.
- IC Integrated Chip
- the RF Module 229 can transmit and receive communication signals (for example, RF signals).
- the RF module 229 may include a transceiver, a Power Amp Module (PAM), a frequency filter, a Low Noise Amplifier (LNA), or an antenna.
- PAM Power Amp Module
- LNA Low Noise Amplifier
- at least one of the cellular module 221 , WiFi module 223 , Bluetooth module 225 , GNSS module 227 , and NFC module 228 can transmit and receive RF signals through a separate RF module.
- the subscriber identification module 224 may include a card including a subscriber identification module or an embedded SIM and may include unique identification information (for example, Integrated Circuit Card Identifier (ICCID) or subscriber information (for example, International Mobile Subscriber Identity (IMSI).
- ICCID Integrated Circuit Card Identifier
- IMSI International Mobile Subscriber Identity
- the memory 230 may include a built-in memory 232 or an external memory 234 .
- the built-in memory 232 may include at least one of a volatile memory (for example, a DRAM, an SRAM, or an SDRAM) and a nonvolatile memory (for example, a One Time Programmable ROM (OTPROM), a PROM, an EPROM, an EEPROM, a mask ROM, a flash ROM, a flash memory, a hard drive, or a Solid State Drive (SSD).
- the external memory 234 may include a flash drive, for example, a Compact Flash (CF), a Secure Digital (SD), a Micro-SD, a Mini-SD, an Extreme Digital (xD), a Multi-Media Card (MMC), or a memory stick.
- CF Compact Flash
- SD Secure Digital
- xD Extreme Digital
- MMC Multi-Media Card
- the external memory 234 can be functionally or physically connected to the electronic device 201 through various interfaces.
- the sensor module 240 can measure physical quantities or sense operation states of the electronic device 201 and can convert the measured or sensed information into electrical signals.
- the sensor module 240 may include at least one of a gesture sensor 240 A, a gyro sensor 240 B, a barometer sensor 240 C, a magnetic sensor 240 D, an acceleration sensor 240 E, a grip sensor 240 F, a proximity sensor 240 G, a color sensor 240 H (for example, an RGB (red, green, blue) sensor), a biosensor 240 I, a temperature/moisture sensor 240 J, an illumination sensor 240 K, and an Ultraviolet (UV) sensor 240 M.
- a gesture sensor 240 A for example, a gyro sensor 240 B, a barometer sensor 240 C, a magnetic sensor 240 D, an acceleration sensor 240 E, a grip sensor 240 F, a proximity sensor 240 G, a color sensor 240 H (for example, an RGB (red, green, blue) sensor),
- the sensor module 240 may include an e-nose sensor, an Electromyography (EMG) sensor, an Electroencephalogram (EEG) sensor, an Electro-cardiogram (ECG) sensor, an Infrared (IR) sensor, an iris sensor, and/or a fingerprint sensor.
- the sensor module 240 may further include a control circuit for controlling one or more sensors therein.
- the electronic device 201 may further include a processor configured to control the sensor module 240 , separately or as a part of the processor 210 , whereby it is possible to control the sensor module 240 while the processor 210 is in a sleep state.
- the input device 250 may include a touch panel 252 , a (digital) pen sensor 254 , a key 256 , or an ultrasonic input device 258 .
- the touch panel 252 may use at least one of electrostatic, decompressing, infrared, and ultrasonic methods.
- the touch panel 252 may further include a control circuit.
- the touch panel 252 can provide a touch response (touch coordinates) to a user by further including a tactile layer.
- the (digital) pen sensor 254 for example, may include a recognition sheet that is a part of the touch panel or a separate part.
- the key 256 for example, may include a physical button, an optical button, or a keypad.
- the ultrasonic input device 258 can sense an ultrasonic wave generated from an input tool through a microphone (for example, a microphone 288 ) and find data corresponding to the sensed ultrasonic wave.
- the display 260 may include a panel 262 , a hologram device 264 , a projector 266 , and/or a control circuit for controlling these components.
- the panel 262 may be implemented to be flexible, transparent, or wearable.
- the panel 262 may be configured as one or modules together with the touch panel 252 .
- the panel 262 may include a pressure sensor (for example, a force sensor) that can measure pressure information (for example, pressure coordinates and intensity of pressure) about a touch by a user.
- the pressure sensor may be integrated with the touch panel 252 or may be composed of one or more sensors separated from the touch panel 252 .
- the hologram device 264 can show 3D images in the air, using interference of light.
- the projector 266 can show images by projecting light to a screen.
- the screen may be positioned inside or outside the electronic device 201 .
- the interface 270 may include an HDMI 272 , a USB 274 , an optical interface 276 , or a D-subminiature (D-sub) 278 .
- the interface 270 may be included in the communication interface 170 shown in FIG. 1 . Additionally or alternatively, the interface 270 , for example, may include a Mobile High-definition Link (MHL) interface, an SD card/Multi-Media Card (MMC) interface, or an interface under Infrared Data Association (IrDA).
- MHL Mobile High-definition Link
- MMC Multi-Media Card
- IrDA Infrared Data Association
- the audio module 280 can convert a sound into an electrical signal and vice versa. At least some components of the audio module 280 , for example, may be included in the I/O interface 150 shown in FIG. 1 .
- the audio module 280 can process sound information input or output through a speaker 282 , a receiver 284 , an earphone 286 , or a microphone 288 .
- the camera module 291 (for example, the camera 180 shown in FIG. 1 , for example, is a device that can take still images and moving images, and according to an embodiment, the camera module 291 may include one or more image sensors (for example, front sensors or rear sensors), lenses, ISPs, or flashes (for example, LEDs or xenon lamps).
- the power management module 295 can manage power of the electronic device 201 .
- the power management module 295 may include a Power Management Integrated Circuit (PMIC), a charging IC, or a battery or fuel gauge.
- the PMIC may have a wire and/or wireless charging method.
- the wireless charging method includes a magnetic resonance method, a magnetic induction method, or an electromagnetic wave method, and may further include an additional circuit for wireless charging, for example, a coil loop, a resonant circuit, or a rectifier.
- the battery gauge for example, can measure the remaining capacity, or a voltage, a current, or temperature in charging of a battery 296 .
- the battery 296 for example, may include a chargeable battery and/or a solar battery.
- the indicator 297 can show specific statuses such as a booting status, a message status, or a charging status of the electronic device 201 or some (for example, the processor 210 ) of the electronic device 201 .
- the motor 298 can convert electrical signals into mechanical vibration and can generate vibration or a haptic effect.
- the electronic device 201 may include a mobile TV support device (for example, a GPU) that can process media data following standards such as Digital Multimedia Broadcasting (DMB), Digital Video Broadcasting (DVB), or mediaFloTM.
- DMB Digital Multimedia Broadcasting
- DVD Digital Video Broadcasting
- mediaFloTM mediaFloTM.
- the components described herein each may be composed of one or more elements and the names of the parts may depend on the kinds of electronic devices.
- an electronic device (for example, the electronic device 201 ) may not include some of the components, may further include additional components, or may be configured as one part by combining some of the components, and can perform the functions of the components before combining.
- FIG. 3 illustrates a block diagram of a program module according to various embodiments.
- a program module 310 may include an operating system, which controls resources related to an electronic device (for example, the electronic device 101 shown in FIG. 1 ), and/or various applications (for example, the application program 147 shown in FIG. 1 ) that are executed on an operating system.
- the operating system may include AndroidTM, iOSTM, WindowsTM, SymbianTM, TizenTM, or BadaTM.
- a program module 310 may include a kernel 320 (for example, the kernel 141 shown in FIG. 1 ), a middleware 330 (for example, the middleware 143 shown in FIG. 1 ), an API 360 (for example, the API 145 shown in FIG. 1 ), and/or an application 370 (for example, the application program 147 shown in FIG. 1 ). At least a portion of the program module 310 can be pre-loaded on an electronic device or can be downloaded from an external electronic device (for example, the electronic devices 102 and 104 and the server 106 shown in FIG. 1 ).
- the kernel 320 may include a system resource manager 321 and/or a device driver 323 .
- the system resource manager 321 can control, allocate, or recover system resources.
- the system resource manager 321 may include a process manager, a memory manager, or a file system manager.
- the device driver 323 may include a display driver, a camera driver, a Bluetooth driver, a shared memory driver, a USB driver, a keypad driver, a WiFi driver, an audio driver, a touch device driver, a pressure device driver, or an Inter-Process Communication (IPC) driver.
- IPC Inter-Process Communication
- the middleware 330 can provide functions that all of the applications 370 use, or can provide various functions to the applications 370 through the API 360 so that the application 370 can use limited system resources of an electronic device.
- the middleware 330 may include at least one of a runtime library 335 , an application manager 341 , a window manager 342 , a multimedia manager 343 , a resource manager 344 , a power manager 345 , a database manager 346 , a package manager 347 , a connectivity manager 348 , a notification manager 349 , a location manager 350 , a graphic manager 351 , and a security manager 352 .
- the runtime library 335 may include a library module that is used by a compiler to add new functions, using a programming language while the application 370 is executed.
- the runtime library 335 can perform input/output management, memory management, or calculation function processing.
- the application manager 341 can manage the lifecycle of the application 370 .
- the window manager 342 can manage a GUI resource that is used for the screen.
- the multimedia manager 343 can find the formats for playing media files and encode or decode the media files, using codecs corresponding to the formats.
- the resource manager 344 can manage the source code of the application 370 or the space of a memory.
- the power manager 345 for example, can manage the capacity of a battery or power and provide power information for operating an electronic device.
- the power manager 345 can operate together with a Basic Input/Output System (BIOS).
- BIOS Basic Input/Output System
- the database manager 346 for example, can create, search for, or change a database to be used by the application 370 .
- the package manager 347 can manage installation or update of applications that are released in the type of a package file.
- the connectivity manager 348 can manage wireless connection.
- the notification manager 349 can provide events such as an arrived message, a promise, and notification of proximity to a user.
- the location manager 350 can manage the location information of an electronic device.
- the graphic manager 351 can manage a graphic effect to be provided to a user or a user interface related to the graphic effect. According to an embodiment, when an object is detected from an image displayed on the display 160 , the graphic manager 351 can manage a graphic effect displaying detection information corresponding to the configuration information of the object.
- the security manager 352 can provide system security or user authentication.
- the middleware 330 may include a telephony manager for managing a voice or video call function of an electronic device or a middleware module that can generate combinations of the functions of the components described above.
- the middleware 330 can provide modules specified for the kinds of operating systems.
- the middleware 330 can dynamically delete some of existing component or add new components.
- the API 360 may be provided to have different configurations, depending on operating systems, as a set of API programming functions. For example, for AndroidTM or iOSTM, one API set can be provided for each platform, and for TizenTM, two or more API sets can be provided for each platform.
- the application 370 may include home 371 , dialer 372 , SMS/MMS 373 , Instant Message (IM) 374 , browser 375 , camera 376 , alarm 377 , contact 378 , voice dial 379 , email 380 , calendar 381 , medial player 382 , album 383 , watch 384 , healthcare (for example, measuring the amount of exercise or blood sugar), or environment information (for example, atmospheric pressure, humidity, or temperature information) providing applications.
- the application 370 may include an information exchange application that can support information exchange between an electronic device and an external electronic device.
- the information exchange application may include a notification relay application for transmitting specific information to an external electronic device or a device management application for managing an external electronic device.
- a notification transmission application can transmit notification information generated by another application of an electronic device to an external electronic device, or can receive notification information from an external electronic device and provide the notification information to a user.
- the device management application for example, can install, delete, or update the functions of an external electronic device communicating with an electronic device (fore example, turning-on/of of the external electronic device (or some components) or adjustment of brightness (or resolution) of a display), or an application that is executed in an external electronic device.
- the application 370 may include an application designated in accordance with the property of an external electronic device (for example, a healthcare application of a mobile medical device).
- the application 370 may include an application received from an external electronic device.
- At least a portion of the program module 310 can be implemented (for example, executed) in software, firmware, hardware (for example, the processor 210 ), or a combination of at least two of them, and may include a module, a program, a routine, an instruction set, or a process for performing one or more functions.
- an electronic device includes a display, at least one processor, and a memory coupled to the processor.
- the memory can store instructions that, when being executed, make the processor control the display to display an image, detect at least one object from the image, and control the display to display detection information corresponding to the at least one object based on the configuration information of the at least one object.
- the configuration information of the at least one object may include at least one of a kind, a shape, or a color of the at least one object.
- the instructions may include instructions that make the processor divide a plurality of sub-objects from the at least one object detected from the image, determine graphic variables corresponding to the configuration information of the sub-objects, and control the display to display detection information corresponding to the sub-objects based on the graphic variables corresponding to the configuration information of the sub-objects.
- the graphic variable may include at least one of a form, a color, a contrast, a definition, or a size of detection information.
- the display is further configured to display detection information corresponding to the at least one object close to an outline of the at least one object.
- the instructions may include instructions that make the processor detect at least one graphic variable corresponding to an outline of the at least one object based on a curvature of the outline the at least one object, and control the display to display detection information showing a change in curvature of the outline of the object based on the graphic variable corresponding to the outline.
- a communication interface is further included, and the instructions may include instructions that make the processor divide a plurality of sub-objects from the object detected from the image, control the communication interface to transmit at least one sub-object of the plurality of sub-objects to at least one server corresponding to the sub-object, receive detailed information of the sub-objects through the communication interface, and recognize the object detected from the image based on the detailed information of the sub-object when the items of detailed information of the sub-object correspond to each other.
- a communication interface is further included, and the instructions may include instructions that make the processor divide a plurality of sub-objects from the object detected from the image, select important sub-object of the plurality of sub-objects, control the communication interface to transmit the important sub-object to a server, receive detailed information of the important sub-object through the communication interface, and recognize the object detected from the image based on the detailed information of the important sub-object.
- the instructions may store instructions that make the processor, when a plurality of objects is detected from the image, recognize the object, detect graphic variables corresponding to recognition statuses of the object, and control the display to display detection information corresponding to the objects based on the graphic variables corresponding to the recognition statuses of the objects.
- a camera is further included, and the instructions may store instructions that make the processor control the display to display an image acquired through the camera.
- FIG. 4 illustrates a flowchart for displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments.
- FIG. 5 illustrates a screen configuration displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- an electronic device can display at least one image on a display in operation 401 .
- the processor 120 when the processor 120 detects occurrence of an event for activating a camera application, the processor 120 can control the camera 180 to be activated.
- the processor 120 as shown in FIG. 5 , can control the display 160 to display an image 500 (for example, a preview image) acquired through the camera 180 .
- the processor 120 when the processor 120 detects occurrence of an event for displaying an image list, the processor 120 can control the display 160 to display an image list stored in the memory 130 .
- the processor 120 can control the display 160 to display at least one image corresponding to input information detected through the I/O interface 150 from the image list displayed on the display 160 .
- the electronic device can detect at least one object from an image displayed on a display in operation 403 .
- the processor 120 can detect a mobile phone-shaped object 502 included in an image by extracting characteristic points of the image 500 displayed on the display 160 .
- the processor 120 can divide a plurality of sub-objects such as an external shape 504 , a logo 506 , a projective area 508 of the external shape (for example, a camera area), and a recessed area 510 of the external shape (for example, a light area) from the mobile phone-shaped object 502 .
- the processor 120 can acquire detailed information of the sub-object.
- the processor 120 determines the object detected from the image based on the detailed information of the sub-object.
- the electronic device can output a graphic effect corresponding to the configuration information of an object detected from an image displayed on a display in operation 405 .
- the processor 120 when the processor 120 , as shown in FIG. 5 , detect one object 502 from the image 500 , the processor 120 can control the display 160 to display detection information 512 close to the object 502 .
- the processor 120 when the processor 120 divides the object 502 detected from the image 500 into a plurality of sub-objects or recognizes sub-objects using at least one server, the processor 120 can detect graphic variables corresponding to the configuration information of the sub-objects.
- the processor 120 can control the display 160 to display different detection information for the sub-objects based on the graphic variables corresponding to the configuration information of the sub-objects to be able to discriminate the sub-objects, using the detection information.
- the processor 120 can control the display 160 to display items of detection information 520 and 522 in line types on the external shape 504 of the object 502 .
- the processor 120 can control the display 160 to display different line 520 or 522 corresponding to the curvature of the external shape 504 .
- the processor 120 as shown in FIG. 5 , can control the display 160 to display detection information 524 in a circular type on the logo 506 of the object 502 .
- the processor 120 can control the display 160 to display black dot-shaped detection information 526 on the projective area 508 of the external shape of the object 502 .
- the processor 120 can control the display 160 to display transparent dot-shaped detection information 528 on the recessed area 510 of the external shape of the object 502 .
- FIG. 6 illustrates a flowchart for acquiring recognition information of an object through a server in an electronic device according to various embodiments.
- FIG. 7 illustrates a screen configuration displaying a graphic effect corresponding to the kind of an object in an electronic device according to various embodiments. An operation for detecting an object included in an image in operation 403 shown in FIG. 4 is described hereafter.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- the electronic device when an electronic device displays an image on a display (for example, operation 401 shown in FIG. 4 ), the electronic device can divide an object detected from the image displayed on the display into a plurality of sub-objects in operation 601 .
- the processor 120 can control the display 160 to display an image 700 (for example, a preview image) acquired through the camera 180 .
- the processor 120 can detect an object 710 by extracting characteristic points from the image 700 .
- the processor 120 can divide the object 710 into an external shape 712 and a logo 714 .
- the electronic device in operation 603 , can transmit information of a plurality of sub-objects divided from the object to servers corresponding to the sub-objects.
- the processor 120 can control the communication interface 170 to transmit the external shape 712 divided from the object 710 shown in FIG. 7 to a first server and transmit the logo 714 to a second server.
- the processor 120 can control the communication interface 170 to transmit the external shape 712 and the logo 714 divided from the object 710 shown in FIG. 7 to a third server.
- the processor 120 can transmit a type of detailed information to be acquired from the third server to the third server
- the processor 120 can transmit request information for separately acquiring detailed information of the external shape 712 and detailed information of the logo 714 to a server.
- the processor 120 can designate any one sub-object, of which the detailed information is acquired from the third server, of the external shape 712 and the logo 714 .
- the processor 120 can transmit request information for acquiring related detailed information of the external shape 712 and the logo 714 to a server.
- the electronic device in operation 605 , can check whether detailed information of the sub-objects is received from the servers.
- the processor 120 can check whether detailed information corresponding to the external shape 712 of the object 710 is received from the first server, through the communication interface 170 .
- the processor 120 can check whether detailed information corresponding to the logo 714 of the object 710 is received from the second server, through the communication interface 170 .
- the processor 120 can control the communication interface 170 to transmit again the information of the sub-object to a server.
- the processor 120 can check whether detailed information corresponding to the external shape 712 and the logo 714 of the object 710 is received from the third server, through the communication interface 170 . In this case, the processor 120 can receive the detailed information of the type requested to the third server.
- the electronic device When the electronic device receives detailed information of sub-objects from servers, the electronic device, in operation 607 , can check whether the items of detailed information of the sub-objects correspond to each other. For example, the processor 120 can check whether product information corresponding to the external shape 712 and company information corresponding to the logo 714 , which are acquired from the servers, correspond to each other.
- the electronic device can determine that an error occurred at the point of time of detecting the object from the image or dividing the sub-objects. Accordingly, the electronic device can detect again an object from the image in operation 601 .
- the electronic device in operation 609 , can determine that the object included in the image based on the detailed information of the sub-objects acquired from the servers. For example, the processor 120 can recognize the information of the object 710 detected from the image, based on the company information of the logo 714 and the product information of the external shape 712 acquired from the servers.
- the electronic device when the electronic device detects an object from an image displayed on the display 160 , the electronic device can display detection information of the object.
- the processor 120 as shown in FIG. 7 , can control the display 160 to display transparent dot-shaped detection information 716 close the object 710 detected from the image 700 .
- the electronic device when the electronic device divides and recognizes sub-objects, that is, the external shape 712 and the logo 714 from the vehicle-shaped object 710 , the electronic device can display different types of items of detection information on the sub-objects 712 and 714 based on graphic variables corresponding to configuration information of the sub-objects 712 and 714 .
- the processor 120 when the processor 120 , as shown in FIG. 7 , divides sub-objects, that is, the external shape 712 and the logo 714 from the vehicle-shaped object 710 , the processor 120 can control the display 160 to display detection information of the external shape 712 in a dark dot type 720 based on the kinds of the sub-objects.
- the processor 120 can control the display 160 to display detection information of the logo 714 in a light dot shape 730 based on the kinds of the sub-objects.
- the server when a server receives a plurality of objects (or sub-objects) from the electronic device, the server can search the objects (sub-objects) corresponding to the type of detailed information that the electronic device uses. For example, the server can search detailed information of the objects (or sub-objects) received from the electronic device. For example, the server can search detailed information of the objects by associating the objects (or sub-objects) received from the electronic device. For example, the server can search for vehicle information of the external shape 712 corresponding to the logo 714 received from the electronic device.
- FIG. 8 illustrates a flowchart for recognizing an object based on detection information of important objects in an electronic device according to various embodiments. An operation for detecting an object included in an image in operation 403 shown in FIG. 4 is described hereafter.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- the electronic device when an electronic device displays an image on a display (for example, operation 401 shown in FIG. 4 ), the electronic device can extract important objects from objects detected from the image displayed on the display in operation 801 .
- the processor 120 as shown in FIG. 7 , can detect the object 710 from the image 700 (for example, a preview image) acquired through the camera 180 .
- the processor 120 can detect the logo 714 as an important sub-object of the object 710 based on predetermined priorities of sub-objects, in the external shape 712 and the logo 714 included in the object 710 .
- the priorities of sub-objects may be set based on the kinds of sub-objects.
- the electronic device in operation 803 , can transmit information of the important sub-object extracted from the object to a server corresponding to the important sub-object.
- the processor 120 can control the communication interface 170 to transmit the logo 714 that is the important sub-object extracted from the object 710 shown in FIG. 7 to a logo server.
- the electronic device in operation 805 , can check whether detailed information of the important sub-objects is received from the server.
- the processor 120 can check whether detailed information corresponding to the logo 714 of the object 710 is received from the logo server, through the communication interface 170 .
- the processor 120 can control the communication interface 170 to transmit again the information of the important sub-object to a server.
- the processor 120 can determine that the processor 120 has failed to recognize the object.
- the electronic device can determine the object included in the image based on the detailed information of the important sub-object in operation 807 .
- the processor 120 can update (limit) reference external shape list to be used to determine the external shape 712 based on the company information of the logo 714 acquired from the logo server.
- the processor 120 can detect a reference external shape corresponding to the external shape 712 of the object 710 from the reference external shape list limited based on the company information of the logo 714 .
- the processor 120 can recognize information (for example, the product name) of the object 710 detected from the image 700 based on the reference external corresponding to the external shape 712 of the object 710 .
- FIG. 9 illustrates a flowchart for detecting a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments. An operation for outputting a graphic effect corresponding to the configuration information of an object in operation 405 shown in FIG. 4 is described hereafter.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- the electronic device when an electronic device detects at least one object from an image displayed on a display (for example, operation 403 shown in FIG. 4 ), the electronic device can find the configuration information of the object (or a sub-object) detected from the image displayed on the display in operation 901 .
- the processor 120 can recognize an object (or a sub-object) checked through a server (an external device) as in operations 601 to 609 shown in FIG. 6 or operations 801 to 807 shown in FIG. 8 .
- the processor 120 can detect the configuration information of an object (or a sub-object) including at least one of the kind, shape, and color of the object based on the recognition information of object (or the sub-object).
- the electronic device can check whether there is a graphic variable corresponding to the configuration information of the object (or the sub-object) in operation 903 .
- the processor 120 can check whether there is a graphic variable corresponding to the configuration information of an object (for a sub-object) in a graphic variable list stored in the memory 130 .
- the electronic device can display detection information of the object (or the sub-object) based on the graphic information corresponding to the configuration information of the object (or the sub-object) in operation 905 .
- the processor 120 detects a plurality of objects having different items of configuration information (for example, kinds) from an image displayed on the display 160 , the processor 120 can control the display 160 to display different items of detection information on the objects based on the configuration information of the objects.
- the processor 120 when the processor 120 extracts a plurality of sub-objects having different items of configuration information (for example, kinds) from one object, the processor 120 can control the display 160 to display different items of detection information on the sub-objects based on the configuration information of the sub-objects. For example, when one object (or a sub-object) has various properties, the processor 120 can control the display 160 to display various items of detection information on the object (or the sub-object) based on the various properties included in the object (or the sub-object).
- configuration information of the sub-objects for example, kinds
- the processor 120 can control the display 160 to display different items of detection information on the sub-objects based on the configuration information of the sub-objects. For example, when one object (or a sub-object) has various properties, the processor 120 can control the display 160 to display various items of detection information on the object (or the sub-object) based on the various properties included in the object (or the sub-object).
- the electronic device can find a predetermined basic graphic variable in operation 907 .
- the electronic device can display detection information of the object (or the sub-object) based on the basic graphic variable in operation 909 .
- the electronic device when there is graphic variables corresponding to the configuration information of at least some of sub-objects divided from an object, can display detection information of the some of sub-objects based on graphic variables corresponding to the some of sub-objects.
- the electronic device can display detection information of the other sub-objects of the sub-objects divided from the object based on the basic graphic variable.
- FIGS. 10A-10E illustrate a screen configuration displaying a graphic effect corresponding to the configuration information of an object in an electronic device according to various embodiments.
- an embodiment for display detection information of an object or a sub-object based on operations 901 to 909 shown in FIG. 9 may be included.
- the processor 120 can control the display 160 to display an image 1000 (for example, a preview image) acquired through the camera 180 .
- the processor 120 can detect a business card-shaped object 1002 by extracting characteristic points of the image 1000 displayed on the display 160 .
- the processor 120 can detect the object 1002 and sub-objects 1004 and 1006 included in the object 1002 based on the characteristic points of the image 1000 displayed on the display 160 .
- the processor 120 can control the display 160 to display detection information 1008 in the same type (for example, a transparent dot type) on the object 1002 and the sub-objects 1004 and 1006 .
- the processor 120 can recognize the sub-objects 1004 and 1006 divided from the object 1002 through a database stored in the electronic device 101 or through an external server.
- the processor 120 can control the display 160 to display different items of detection information 1010 and 1012 based on configuration information of the sub-objects 1004 and 1006 (for example, the kind of the object).
- the user of the electronic device 101 can discriminate the sub-objects 1004 and 1006 having different items of configuration information based on the items of detection information 1010 and 1012 displayed on the display 160 .
- the processor 120 as shown in FIG.
- the display 160 can control the display 160 to display the detection information 1010 in a light dot type on the sub-object 1004 that is a text and the detection information 1012 in a dark dot type on the sub-object 1006 that is a QR code. That is, the sub-objects 1004 and 1006 can be discriminated by the difference in contrast of the items of detection information 1010 and 1012 displayed on the display 160 .
- the processor 120 as shown in FIG. 10B , can control the display 160 to display the detection information 1020 in a light dot type on the sub-object 1004 that is a text and detection information 1022 in a rectangular type including the sub-object 1006 on the sub-object 1006 that is a QR code.
- the sub-objects 1004 and 1006 can be discriminated by the difference in shape of the items of detection information 1020 and 1022 displayed on the display 160 .
- the processor 120 can control the display 160 to make the items of detection information of sub-objects different based on accuracy in recognition depending on the kinds of sub-objects.
- the processor 120 can detect an accessory-shaped object 1032 from an image 1030 acquired through the camera 180 .
- the processor 120 can control the display 160 to display detection information 1034 in a transparent dot type on the object 1032 .
- the processor 120 when the processor 120 recognizes the kind and shape of the accessory-shaped object 1032 , the processor 120 can determine a graphic variable corresponding to the configuration information (for example, the shape) of the object 1032 .
- the processor 120 can determine the size of detection information corresponding to the form of detection information and the curvature of an outline discriminated from the shape of the object 1032 .
- the form of detection information may be determined in a shape discriminated from the pattern included in the object 1032 .
- the processor 120 as shown in FIG. 10C , can control the display 160 to display items of detection information 1036 and 1038 in a rectangular type based on the graphic variable corresponding to the configuration information of the object 1032 .
- the processor 120 can control the display 160 to differently display the sizes of the items of detection information 1036 and 1038 such that the items of detection information correspond to the curvature of the outline of the object 1032 .
- the processor 120 can detect an accessory-shaped object 1042 from an image 1040 acquired through the camera 180 .
- the processor 120 can control the display 160 to display detection information 1044 in a transparent dot type on the object 1042 .
- the processor 120 when the processor 120 determines the kind and shape of the accessory-shaped object 1042 , the processor 120 can determine a graphic variable corresponding to the configuration information (for example, the shape) of the object 1042 .
- the processor 120 can determine the size and color (or contrast) of detection information corresponding to the curvature of the outline of the object 1042 .
- the color of the detection information may be determined as a color (for example, a complementary color) discriminated from the color of the object 1042 .
- the processor 120 as shown in FIG. 10D , can control the display 160 to display items of detection information 1046 and 1048 having various shapes and colors to correspond to the curvature of the outline based on the graphic variable corresponding to the configuration information of the object 1042 .
- the processor 120 can detect a specific product-shaped object 1052 from an image 1050 acquired through the camera 180 .
- the processor 120 can control the display 160 to display detection information 1056 in a transparent dot type on the object 1052 .
- the processor 120 detects a sub-object 1054 (for example, a logo) of the object 1052 , the processor 120 can control the display 160 to display detection information 1056 in the same type on the sub-object 1054 .
- the processor 120 when the processor 120 recognizes the kind and shape of the object 1052 , the processor 120 can determine a graphic variable corresponding to the configuration information (for example, important area information) of the object 1052 .
- the processor 120 when the logo 1054 of the object 1052 is set as an important area, the processor 120 , as shown in FIG. 10E , can control the display 160 to display detection information 1060 of the logo 1054 of the object 1052 differently from another detection information 1058 . That is, the processor 120 can discriminate the important area 1054 and a non-important area 1058 from the difference in color of the items of detection information 1058 and 1060 displayed on the display 160 .
- FIG. 11 illustrates a flowchart for outputting a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments.
- FIGS. 12A and 12B illustrate a screen configuration displaying a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments.
- An operation for outputting a graphic effect corresponding to the configuration information of an object in operation 405 shown in FIG. 4 is described hereafter.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- an electronic device when an electronic device detects at least one object from an image displayed on a display (for example, operation 403 shown in FIG. 4 ), the electronic device checks whether a plurality of objects is detected from the image displayed on the display in operation 1101 .
- the processor 120 as shown in FIG. 12A , can check whether a plurality of objects such as a first object 1202 and a second object 1204 is detected from an image 1200 displayed on the display 160 .
- the electronic device When the electronic device detects one object from an image displayed on a display, the electronic device can output a graphic effect displaying detection information of the object based on a graphic variable corresponding to the configuration information of the object in operation 1107 .
- the configuration information of an object may include at least one of the kind, shape, and color of an object.
- the electronic device can check the recognition statuses of the objects in operation 1103 .
- the recognition status of an object may include at least one of a recognition standby status, a recognition progression status, and a recognition completion status.
- the electronic device can output graphic effects of objects to correspond to the recognition statuses of objects in operation 1105 .
- the processor 120 can display detection information 1212 in a dark dot type on the second object 1204 that is in the recognition progression status.
- the processor 120 can control the display 160 to display detection information 1210 in a light dot type so that the first object 1202 that is in the recognition standby status is discriminated from the second object 1204 .
- the processor 120 can display detection information 1220 in a solid line type on the second object 1204 that is in the recognition progression status.
- the processor 120 can control the display 160 to display detection information 1222 in a dotted line type so that the first object 1202 that is in the recognition standby status is discriminated from the second object 1204 .
- the electronic device when the electronic device does not recognize the kinds and shapes of a plurality of objects 1202 and 1204 detected from the image 1200 displayed on a display, as shown in FIGS. 12A and 12B , the electronic device can display detection information 1206 in the same type (for example, a transparent dot type) on the objects 1202 and 1204 .
- detection information 1206 in the same type (for example, a transparent dot type) on the objects 1202 and 1204 .
- FIG. 13 illustrates a flowchart for displaying a graphic effect corresponding to the recognition status of an object in an electronic device according to various embodiments. An operation for outputting a graphic effect corresponding to the recognition status of an object in operation 1105 shown in FIG. 11 is described hereafter.
- an electronic device may include the electronic device 101 shown in FIG. 1 or at least a portion (for example, the processor 120 ) of the electronic device 101 .
- the electronic device when an electronic device finds the recognition statuses of a plurality of objects detected from an image (for example, operation 1103 shown in FIG. 11 ), the electronic device can check whether a first object of the objects is recognized in operation 1301 .
- the processor 120 can detect a plurality of objects from an image displayed on the display 160 , the processor 120 can individually recognize the objects. Accordingly, the processor 120 can check whether an operation of recognizing the first object is performed with the other objects of the objects in the recognition standby status.
- the operation of recognizing the first object may include a series of operations of transmitting information of the first object to a first server corresponding to the first object and receiving detailed information of the first object from the first server.
- the electronic device When the electronic device recognizes the first object of the objects, the electronic device can display items of detection information corresponding to the recognition status of the first object and the standby status of the second object in operation 1303 .
- the processor 120 can control the display 160 to display detection information 1212 of the object that is in the recognition progression status in a dark dot type and detection information 1210 of the object that is in the recognition standby status in a light dot type.
- the electronic device can check whether the second object of the objects is recognized in operation 1305 .
- the processor 120 can check whether information of the second object of the objects has been transmitted to a second server corresponding to the second object.
- the electronic device can display items of detection information corresponding to the recognition status of the first object and the standby status of the second object in operation 1303 .
- the electronic device can control the display 160 to display detection information corresponding to the recognition completion status of the first object.
- the processor 120 can control the display 160 to display detection information corresponding to the recognition standby status of the second object.
- the electronic device When the electronic device recognizes the second object, the electronic device can display items of detection information corresponding to the recognition statuses of the first object and the second object in operation 1307 .
- the processor 120 when the processor 120 recognizes the first object and the second object, the processor 120 can control the display 160 to display detection information in the same type on the first object and the second object.
- the electronic device when the electronic device finishes recognizing the first object and the second object detected from the image displayed on the display, the electronic device can output graphic effects corresponding to the configuration information of the objects. For example, when the items of configuration information (for example, kinds) of the first object 1202 and the second object 1204 are different, the processor 120 , as shown in FIG. 12B , can control the display 160 to display the detection information 1222 of the first object 1202 in a dotted line type and the detection information 1220 of the second object 1204 in a solid line type.
- the processor 120 can control the display 160 to display the detection information 1222 of the first object 1202 in a dotted line type and the detection information 1220 of the second object 1204 in a solid line type.
- a method of operating an electronic device may include an operation of displaying an image on a display electrically connected with the electronic device, an operation of detecting at least one object from the image, and an operation of displaying detection information corresponding to the object based on the configuration information of the object.
- the configuration information of an object may include at least one of the kind, shape, and color of an object.
- the operation of displaying detection information may include an operation of dividing a plurality of sub-objects from an object detected from an image, an operation of determining graphic variables corresponding to the configuration information of the sub-objects, and an operation of displaying detection information corresponding to the sub-objects based on the graphic variables corresponding to the configuration information of the sub-objects.
- the graphic variable may include at least one of the form, color, contrast, definition, and size of detection information.
- the operation of displaying detection information may include an operation of displaying detection information corresponding to the object close to the outline of the object.
- the operation of displaying detection information may include an operation of detecting at least one graphic variable corresponding to the outline based on the curvature of the outline of the object and an operation of displaying detection information showing a change in curvature of the outline of the object based on the graphic variable corresponding to the outline.
- the operation of detecting an object may include an operation of dividing a plurality of sub-objects from an object detected from the image, an operation of transmitting at least one of the sub-objects to at least one server corresponding to the sub-object, an operation of receiving detailed information of the sub-objects, and an operation of recognizing the object detected from the image based on the detailed information of the sub-objects when the items of detailed information of the sub-object correspond to each other.
- the operation of detecting an object may include an operation of dividing a plurality of sub-objects from an object detected from the image, an operation of selecting important sub-objects from the sub-objects, an operation of transmitting the important sub-objects to a server, an operation of receiving detailed information of the important sub-objects, and an operation of recognizing the object detected from the image based on the detailed information of the important sub-objects.
- an operation of detecting graphic variables corresponding to the recognition statuses of objects and an operation of displaying detection information corresponding to the objects based on the graphic variables corresponding to the recognition statuses of the objects when a plurality of objects is detected from the image may be further included.
- the operation of displaying the image may include an operation of displaying an image acquired through a camera electrically connected with the electronic device.
- the electronic device and the method of operating the electronic device can increase the recognition rate on an object included in an image by acquiring information of an object displayed on a display by transmitting at least a portion of the object to at least one server.
- the electronic device and the method of operating the electronic device can enable a user of the electronic device to easily recognize an object detected by the electronic device by displaying the configuration information of an object and a graphic effect corresponding to a recognition status when an object displayed on a display is recognized.
- module as used herein may, for example, mean a unit including one of hardware, software, and firmware or a combination of two or more of them.
- the “module” may be interchangeably used with, for example, the term “unit”, “logic”, “logical block”, “component”, or “circuit”.
- the “module” may be a minimum unit of an integrated component element or a part thereof.
- the “module” may be a minimum unit for performing one or more functions or a part thereof.
- the “module” may be mechanically or electronically implemented.
- the “module” may include at least one of an Application-Specific Integrated Circuit (ASIC) chip, a Field-Programmable Gate Arrays (FPGA), and a programmable-logic device for performing operations which has been known or are to be developed hereinafter.
- ASIC Application-Specific Integrated Circuit
- FPGA Field-Programmable Gate Arrays
- programmable-logic device for performing operations which has been known or are to be developed hereinafter.
- At least some of the devices (for example, modules or functions thereof) or the method (for example, operations) according to the present disclosure may be implemented by a command stored in a computer-readable storage medium in a program module form.
- the instruction when executed by a processor (e.g., the processor 120 ), may cause the one or more processors to execute the function corresponding to the instruction.
- the computer-readable storage medium may be, for example, the memory 130 .
- the computer readable recoding medium may include a hard disk, a floppy disk, magnetic media (for example, a magnetic tape), optical media (for example, a Compact Disc Read Only Memory (CD-ROM) and a Digital Versatile Disc (DVD)), magneto-optical media (for example, a floptical disk), a hardware device (for example, a Read Only Memory (ROM), a Random Access Memory (RAM), a flash memory), and the like.
- the program instructions may include high class language codes, which can be executed in a computer by using an interpreter, as well as machine codes made by a compiler. Any of the hardware devices as described above may be configured to work as one or more software modules in order to perform the operations according to various embodiments of the present disclosure, and vice versa.
- modules or programming modules may include at least one of the above described elements, exclude some of the elements, or further include other additional elements.
- the operations performed by the modules, programming module, or other elements according to various embodiments of the present disclosure may be executed in a sequential, parallel, repetitive, or heuristic manner. Further, some operations may be executed according to another order or may be omitted, or other operations may be added.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- Mathematical Physics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims (16)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020170034018A KR102315341B1 (en) | 2017-03-17 | 2017-03-17 | Method for providing graphic effect corresponding to configuration information of object and electronic device thereof |
KR10-2017-0034018 | 2017-03-17 |
Publications (2)
Publication Number | Publication Date |
---|---|
US20180267993A1 US20180267993A1 (en) | 2018-09-20 |
US10853408B2 true US10853408B2 (en) | 2020-12-01 |
Family
ID=63520711
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/925,585 Active 2038-06-15 US10853408B2 (en) | 2017-03-17 | 2018-03-19 | Method for providing graphic effect corresponding to configuration information of object and electronic device thereof |
Country Status (2)
Country | Link |
---|---|
US (1) | US10853408B2 (en) |
KR (1) | KR102315341B1 (en) |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111679731A (en) * | 2019-03-11 | 2020-09-18 | 三星电子株式会社 | Display device and control method thereof |
WO2021025509A1 (en) * | 2019-08-06 | 2021-02-11 | Samsung Electronics Co., Ltd. | Apparatus and method for displaying graphic elements according to object |
Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5640468A (en) * | 1994-04-28 | 1997-06-17 | Hsu; Shin-Yi | Method for identifying objects and features in an image |
US6370262B1 (en) * | 1994-11-08 | 2002-04-09 | Canon Kabushiki Kaisha | Information processing apparatus and remote apparatus for object, using distance measuring apparatus |
US20040004559A1 (en) * | 2002-07-01 | 2004-01-08 | Rast Rodger H. | Keyboard device with preselect feedback |
US20050185827A1 (en) * | 2000-09-20 | 2005-08-25 | Hitachi, Ltd. | Personal identification system |
US20060078170A1 (en) * | 2004-10-08 | 2006-04-13 | Fujitsu Limited. | Biometrics authentication system registration method, biometrics authentication system, and program for same |
US20070124669A1 (en) * | 2003-09-24 | 2007-05-31 | Nokia Corporation | Presentation of large objects on small displays |
US20080183707A1 (en) * | 2006-11-20 | 2008-07-31 | Tomoyuki Asano | Verification Apparatus, Verification Method and Verification Program |
US20090147998A1 (en) * | 2007-12-05 | 2009-06-11 | Fujifilm Corporation | Image processing system, image processing method, and computer readable medium |
US20100315431A1 (en) * | 2009-06-15 | 2010-12-16 | Canon Kabushiki Kaisha | Combining overlapping objects |
US20110029635A1 (en) * | 2009-07-30 | 2011-02-03 | Shkurko Eugene I | Image capture device with artistic template design |
US20120036016A1 (en) * | 1999-02-01 | 2012-02-09 | Hoffberg Steven M | Vehicular information system and method |
US20120038671A1 (en) * | 2010-08-12 | 2012-02-16 | Pantech Co., Ltd. | User equipment and method for displaying augmented reality window |
WO2014183099A1 (en) | 2013-05-09 | 2014-11-13 | Amazon Technologies, Inc. | Recognition interfaces for computing devices |
US20140340570A1 (en) * | 2011-09-28 | 2014-11-20 | U.S. Army Research Laboratory Attn: Rdrl-Loc-I | System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles |
US8971637B1 (en) * | 2012-07-16 | 2015-03-03 | Matrox Electronic Systems Ltd. | Method and system for identifying an edge in an image |
US20150269760A1 (en) * | 2014-03-18 | 2015-09-24 | Fujitsu Limited | Display control method and system |
US20160005154A1 (en) * | 2011-09-28 | 2016-01-07 | U.S. Army Research Laboratory Attn: Rdrl-Loc-I | System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles |
KR20160009900A (en) | 2014-07-17 | 2016-01-27 | 주식회사 인프라웨어 | Method and apparatus for recommending an application based on image recognition |
US9355123B2 (en) | 2013-07-19 | 2016-05-31 | Nant Holdings Ip, Llc | Fast recognition algorithm processing, systems and methods |
KR20170005373A (en) | 2016-12-29 | 2017-01-12 | 주식회사 비즈모델라인 | Method for Providing Customized Augmented Reality |
US9626768B2 (en) * | 2014-09-30 | 2017-04-18 | Microsoft Technology Licensing, Llc | Optimizing a visual perspective of media |
US10331291B1 (en) * | 2014-12-31 | 2019-06-25 | Morpho Trust USA, LLC | Visual verification of digital identifications |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102047704B1 (en) * | 2013-08-16 | 2019-12-02 | 엘지전자 주식회사 | Mobile terminal and controlling method thereof |
-
2017
- 2017-03-17 KR KR1020170034018A patent/KR102315341B1/en active IP Right Grant
-
2018
- 2018-03-19 US US15/925,585 patent/US10853408B2/en active Active
Patent Citations (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5640468A (en) * | 1994-04-28 | 1997-06-17 | Hsu; Shin-Yi | Method for identifying objects and features in an image |
US6370262B1 (en) * | 1994-11-08 | 2002-04-09 | Canon Kabushiki Kaisha | Information processing apparatus and remote apparatus for object, using distance measuring apparatus |
US20120036016A1 (en) * | 1999-02-01 | 2012-02-09 | Hoffberg Steven M | Vehicular information system and method |
US20050185827A1 (en) * | 2000-09-20 | 2005-08-25 | Hitachi, Ltd. | Personal identification system |
US20040004559A1 (en) * | 2002-07-01 | 2004-01-08 | Rast Rodger H. | Keyboard device with preselect feedback |
US20070124669A1 (en) * | 2003-09-24 | 2007-05-31 | Nokia Corporation | Presentation of large objects on small displays |
US20060078170A1 (en) * | 2004-10-08 | 2006-04-13 | Fujitsu Limited. | Biometrics authentication system registration method, biometrics authentication system, and program for same |
US20080183707A1 (en) * | 2006-11-20 | 2008-07-31 | Tomoyuki Asano | Verification Apparatus, Verification Method and Verification Program |
US20090147998A1 (en) * | 2007-12-05 | 2009-06-11 | Fujifilm Corporation | Image processing system, image processing method, and computer readable medium |
US20100315431A1 (en) * | 2009-06-15 | 2010-12-16 | Canon Kabushiki Kaisha | Combining overlapping objects |
US20110029635A1 (en) * | 2009-07-30 | 2011-02-03 | Shkurko Eugene I | Image capture device with artistic template design |
US20120038671A1 (en) * | 2010-08-12 | 2012-02-16 | Pantech Co., Ltd. | User equipment and method for displaying augmented reality window |
US20160005154A1 (en) * | 2011-09-28 | 2016-01-07 | U.S. Army Research Laboratory Attn: Rdrl-Loc-I | System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles |
US20140340570A1 (en) * | 2011-09-28 | 2014-11-20 | U.S. Army Research Laboratory Attn: Rdrl-Loc-I | System and processor implemented method for improved image quality and generating an image of a target illuminated by quantum particles |
US8971637B1 (en) * | 2012-07-16 | 2015-03-03 | Matrox Electronic Systems Ltd. | Method and system for identifying an edge in an image |
WO2014183099A1 (en) | 2013-05-09 | 2014-11-13 | Amazon Technologies, Inc. | Recognition interfaces for computing devices |
US9355123B2 (en) | 2013-07-19 | 2016-05-31 | Nant Holdings Ip, Llc | Fast recognition algorithm processing, systems and methods |
US20150269760A1 (en) * | 2014-03-18 | 2015-09-24 | Fujitsu Limited | Display control method and system |
KR20160009900A (en) | 2014-07-17 | 2016-01-27 | 주식회사 인프라웨어 | Method and apparatus for recommending an application based on image recognition |
US9626768B2 (en) * | 2014-09-30 | 2017-04-18 | Microsoft Technology Licensing, Llc | Optimizing a visual perspective of media |
US10331291B1 (en) * | 2014-12-31 | 2019-06-25 | Morpho Trust USA, LLC | Visual verification of digital identifications |
KR20170005373A (en) | 2016-12-29 | 2017-01-12 | 주식회사 비즈모델라인 | Method for Providing Customized Augmented Reality |
Also Published As
Publication number | Publication date |
---|---|
KR102315341B1 (en) | 2021-10-20 |
KR20180106221A (en) | 2018-10-01 |
US20180267993A1 (en) | 2018-09-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10599904B2 (en) | Electronic device for measuring biometric information and method of operating same | |
US10386927B2 (en) | Method for providing notification and electronic device thereof | |
US10740604B2 (en) | Electronic device for iris recognition and operating method thereof | |
US10503390B2 (en) | Electronic device and photographing method | |
US11030288B2 (en) | Electronic device for authenticating using biometric information and method of operating electronic device | |
US10542144B2 (en) | Electronic apparatus for providing voice recognition control and operating method therefor | |
US11042240B2 (en) | Electronic device and method for determining underwater shooting | |
US10893184B2 (en) | Electronic device and method for processing image | |
US20190324279A1 (en) | Electronic device, mobile electronic device, and display device, which are controlled on basis of user gaze and input | |
US10359878B2 (en) | Method for providing events corresponding to touch attributes and electronic device thereof | |
KR102358849B1 (en) | Electronic device for providing information related to a smart watch and method for operating the same | |
EP3343876B1 (en) | Method for managing contents and electronic device thereof | |
US20170277413A1 (en) | Method for outputting screen and electronic device supporting the same | |
US20170006149A1 (en) | Method for controlling external electronic device and electronic device for processing same | |
US10033921B2 (en) | Method for setting focus and electronic device thereof | |
US10868954B2 (en) | Electronic device for providing graphic indicator for focus and method of operating electronic device | |
US10691318B2 (en) | Electronic device and method for outputting thumbnail corresponding to user input | |
US10853408B2 (en) | Method for providing graphic effect corresponding to configuration information of object and electronic device thereof | |
US10956588B2 (en) | Server, electronic device, and method for processing image by electronic device | |
US20190014240A1 (en) | Electronic device and method for providing adsorption information of foreign substance adsorbed by camera | |
US20180087890A1 (en) | Method for controlling sensor and electronic device thereof | |
KR102319983B1 (en) | Method for providing information and an electronic device thereof | |
US10395026B2 (en) | Method for performing security function and electronic device for supporting the same |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, HO YOUNG;CHOI, NARI;KI, DAEHONG;AND OTHERS;REEL/FRAME:045642/0360 Effective date: 20180319 |
|
FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |