US20190095079A1 - Method for enabling interaction using fingerprint on display and electronic device thereof - Google Patents

Method for enabling interaction using fingerprint on display and electronic device thereof Download PDF

Info

Publication number
US20190095079A1
US20190095079A1 US16/131,774 US201816131774A US2019095079A1 US 20190095079 A1 US20190095079 A1 US 20190095079A1 US 201816131774 A US201816131774 A US 201816131774A US 2019095079 A1 US2019095079 A1 US 2019095079A1
Authority
US
United States
Prior art keywords
fingerprint
electronic
fingerprints
content
display
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US16/131,774
Inventor
Gunjan Prakash DEOTALE
Samarth Shankar RAUT
Jidnya Samir SHAH
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority to IN201741032625 priority Critical
Priority to IN201741032625 priority
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Assigned to SAMSUNG ELECTRONICS CO., LTD. reassignment SAMSUNG ELECTRONICS CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: DEOTALE, GUNJAN PRAKASH, RAUT, Samarth Shankar, SHAH, JIDNYA SAMIR
Publication of US20190095079A1 publication Critical patent/US20190095079A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • G06F3/04842Selection of a displayed object
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range
    • G06F3/04845Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object or an image, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the screen or tablet into independently controllable areas, e.g. virtual keyboards, menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06KRECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K9/00Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
    • G06K9/00006Acquiring or recognising fingerprints or palmprints
    • G06K9/00013Image acquisition
    • G06K9/0002Image acquisition by non-optical methods, e.g. by ultrasonic or capacitive sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06KRECOGNITION OF DATA; PRESENTATION OF DATA; RECORD CARRIERS; HANDLING RECORD CARRIERS
    • G06K9/00Methods or arrangements for reading or recognising printed or written characters or for recognising patterns, e.g. fingerprints
    • G06K9/00006Acquiring or recognising fingerprints or palmprints
    • G06K9/00013Image acquisition
    • G06K9/0004Image acquisition by using electro-optical elements or layers, e.g. electroluminescent sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/033Indexing scheme relating to G06F3/033
    • G06F2203/0338Fingerprint track pad, i.e. fingerprint sensor used as pointing device tracking the fingertip image
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04806Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Abstract

A method for enabling interaction on an electronic device using fingerprint on display to execute an operation of the electronic device based on the detected fingerprint. The method may includes detecting at least one fingerprint from at least one touch input on a display screen, identifying an object corresponding to the at least one fingerprint, and performing at least one actions based on the object.

Description

    CROSS-REFERENCE TO RELATED APPLICATION(S)
  • This application is based on and claims priority under 35 U.S.C. § 119 from an Indian Patent Application Number 201741032625 filed on Sep. 14, 2017, and Indian Patent Application Number 201741032625 filed on Sep. 11, 2018, in the India Patent Office, the disclosures of which are incorporated by reference in their entireties.
  • BACKGROUND 1. Field
  • The disclosure relates to electronic devices, and more particularly a method and electronic device for enabling interaction using a fingerprint on display.
  • 2. Description of Related Art
  • Generally, ubiquitous input devices like fingerprint on display (FOD) provide an interactive and fast way for a user of the electronic device to interact with applications by the help of gestures (such as for e.g., touch input, long press, etc.). As the electronic devices, such as smartphone of the user, are prone to be accessed by one or more different users with or without the user's knowledge, a demand for providing security to the electronic devices and/or the applications installed therein is increased.
  • The use of a fingerprint integrated with display is an emerging trend, in which a fingerprint sensor is placed underneath a touchscreen for ensuring secure interaction with the device. Using a fingerprint integrated display, various kinds of applications, especially applications based on fingerprint information, have been gradually popularized. In this context, there remains a need to define various mechanisms for interacting with an electronic device using the fingerprint information of the user.
  • SUMMARY
  • In accordance with aspects of the disclosure, there is provided a method and electronic device for enabling interaction using a fingerprint on display.
  • In accordance with an aspect of the disclosure, one or more fingerprints may be registered with one or more graphical objects.
  • In accordance with an aspect of the disclosure, the one or more fingerprints may be received on a display screen of the electronic device for interacting with the one or more graphical objects.
  • In accordance with an aspect of the disclosure, an action may be dynamically performed in the electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, one or more fingerprints may be assigned to the one or more graphical objects.
  • In accordance with an aspect of the disclosure, the one or more fingerprints assigned to the one or more graphical objects may be stored.
  • In accordance with an aspect of the disclosure, one or more recommendations for registering the one or more fingerprints with one or more graphical objects may be provided based on content associated with the one or more graphical objects.
  • In accordance with an aspect of the disclosure, the one or more graphical objects assigned to the one or more fingerprints may be identified.
  • In accordance with an aspect of the disclosure, an action may be performed in the electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, one or more graphical objects assigned to the one or more fingerprints may be displayed in response to detecting the one or more fingerprints.
  • In accordance with an aspect of the disclosure, there is provided a method for enabling interaction on an electronic device using fingerprint on display. The method may include registering one or more fingerprints with one or more graphical objects, detecting the one or more fingerprints on a display screen of the electronic device for interacting with the one or more graphical objects and dynamically performing an action in the electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, registering the one or more fingerprints with one or more graphical objects may include receiving one or more fingerprints for the one or more graphical objects. The method may include assigning the received one or more fingerprints to the one or more graphical objects and storing the one or more fingerprints assigned to the one or more graphical objects.
  • In accordance with an aspect of the disclosure, the electronic device may provide one or more recommendations for registering the one or more fingerprints with one or more graphical objects based on content associated with the one or more graphical objects.
  • In accordance with an aspect of the disclosure, dynamically performing an action in the electronic device using the one or more fingerprints may include receiving one or more fingerprints on a display screen of the electronic device. The method may include identifying the one or more graphical objects assigned to the one or more fingerprints and performing an action in the electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, performing an action in the electronic device using the one or more fingerprints may include displaying one or more graphical objects assigned to the one or more fingerprints.
  • In accordance with an aspect of the disclosure, performing an action in the electronic device using the one or more fingerprints may include extracting content relevant to the one or more graphical objects. The method may include rendering the extracted content.
  • In accordance with an aspect of the disclosure, the method may include performing one or more actions using one or more fingerprints received on the display screen when the extracted content is being rendered.
  • In accordance with an aspect of the disclosure, performing an action in the electronic device using the one or more fingerprints may include selecting at least one content in the electronic device using the one or more fingerprints. The method may include transferring the selected content to another electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, performing an action in the electronic device using the one or more fingerprints may include selecting one or more graphical objects in the electronic device using the one or more fingerprints. The method may include accessing the selected one or more graphical objects with the one or more fingerprints.
  • In accordance with an aspect of the disclosure, performing an action in the electronic device using the one or more fingerprints may include modifying one or more the parameters of the graphical objects.
  • In accordance with an aspect of the disclosure, there is provided an electronic device for enabling interaction using fingerprint on display. The electronic device may include a registration engine and an interaction engine. The registration engine may be configured to register one or more fingerprints with one or more graphical objects. The interaction engine may be configured to receive the one or more fingerprints on a display screen of the electronic device for interacting with the one or more graphical objects. Further, the interaction engine may be configured to dynamically perform an action in the electronic device using the one or more fingerprints.
  • In accordance with an aspect of the disclosure, there is provided a method for operating an electronic device. The method may includes detecting at least one fingerprint from at least one touch input on a display screen, identifying an object corresponding to the at least one fingerprint, and performing at least one actions based on the object.
  • In accordance with an aspect of the disclosure, there is provided a method for operating an electronic device. The method may include detecting a fingerprint of a finger corresponding to a touch input on a display screen, determining an operation of the electronic device corresponding to the fingerprint, and performing the operation of the electronic device.
  • These and other aspects of the disclosure will be better appreciated and understood when considered in conjunction with the following description and the accompanying drawings. It should be understood, however, that the following descriptions, while describing embodiments and numerous specific details thereof, are given by way of illustration and not of limitation. Many changes and modifications may be made within the scope of the embodiments herein without departing from the spirit thereof, and the embodiments herein include all such modifications.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Certain aspects of the disclosure are illustrated in the accompanying drawings, throughout which like reference letters indicate corresponding parts in the various figures. The embodiments herein will be better understood from the following description with reference to the drawings, in which:
  • FIG. 1A is a block diagram of an electronic device, according to an embodiment;
  • FIG. 1B illustrates various steps performed by a fingerprint registration engine, according to an embodiment;
  • FIG. 1C illustrates various steps performed by a fingerprint recognition engine, according to an embodiment;
  • FIG. 1D illustrates installation examples of fingerprint sensors according to an embodiment;
  • FIG. 2 is a flow chart illustrating a method for enabling interaction using a fingerprint, according to an embodiment;
  • FIGS. 3A, 3B, and 3C illustrate an example for interacting with a multimedia content using a fingerprint on display, according to an embodiment;
  • FIG. 4A is a flow chart illustrating a method for interacting with a multimedia content using fingerprint on display, according to an embodiment;
  • FIG. 4B is a flow chart illustrating a method for registering the fingerprint(s) with graphical object(s), according to an embodiment;
  • FIGS. 5A, 5B, and 5C illustrate an example in which the electronic device provides interaction with graphical objects using the fingerprint on the display, according to an embodiment;
  • FIG. 6 is an example for transferring content across electronic devices using the fingerprint on display, according to an embodiment;
  • FIG. 7 is an example for assigning the fingerprints to the graphical objects, according to an embodiment;
  • FIGS. 8A, 8B, 8C, and 8D illustrate examples in which various actions are performed on the multimedia content using the fingerprint on display of the electronic device, according to an embodiment;
  • FIGS. 9A, 9B, 9C, 9D, and 9E illustrate examples of performing a zoom operation on content using various fingerprints of the user, according to an embodiment;
  • FIGS. 10A and 10B are an example of selecting characters using fingerprints, according to an embodiment;
  • FIG. 11 is an example in which the electronic device displays applications on a portion of the display screen using the fingerprints, according to an embodiment;
  • FIGS. 12A and 12B are an example of copying and pasting contents using fingerprints, according to an embodiment; and
  • FIG. 13 is an example of assigning functions using fingerprints, according to an embodiment.
  • DETAILED DESCRIPTION OF EMBODIMENTS
  • Various embodiments of the disclosure will now be described in detail with reference to the accompanying drawings. In the following description, specific details such as detailed configuration and components are provided to assist the overall understanding of these embodiments of the disclosure. Therefore, it should be apparent to those skilled in the art that various changes and modifications of the embodiments described herein can be made without departing from the scope and spirit of the disclosure. In addition, descriptions of well-known functions and constructions are omitted for clarity and conciseness.
  • Also, the various embodiments described herein are not necessarily mutually exclusive, as some embodiments can be combined with one or more other embodiments.
  • Herein, the term “or” as used herein, refers to a non-exclusive or, unless otherwise indicated. The examples used herein are intended merely to facilitate an understanding of ways in which the embodiments herein can be practiced and to further enable those skilled in the art to practice the embodiments herein. Accordingly, the examples should not be construed as limiting the scope of the embodiments herein.
  • Expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list. For example, the expression, “at least one of a, b, and c,” should be understood as including only a, only b, only c, both a and b, both a and c, both b and c, or all of a, b, and c.
  • As is traditional in the field, embodiments may be described and illustrated in terms of blocks which carry out a described function or functions. These blocks, which may be referred to herein as units, engines, manager, modules or the like, are physically implemented by analog and/or digital circuits such as logic gates, integrated circuits, microprocessors executing computer-readable instructions, microcontrollers, memory circuits, passive electronic components, active electronic components, optical components, hardwired circuits and the like, and may optionally be driven by firmware and/or software. The circuits may, for example, be embodied in one or more semiconductor chips, or on substrate supports such as printed circuit boards and the like. The circuits constituting a block may be implemented by dedicated hardware, or by a processor (e.g., one or more programmed microprocessors and associated circuitry), or by a combination of dedicated hardware to perform some functions of the block and a processor to perform other functions of the block. Each block of the embodiments may be physically separated into two or more interacting and discrete blocks without departing from the scope of the disclosure. Likewise, the blocks of the embodiments may be combined into more complex blocks without departing from the scope of the disclosure.
  • The embodiments herein generally provide for a method for enabling interaction on an electronic device using a fingerprint on a display. The method may include registering one or more fingerprints with one or more graphical objects. The method may include receiving the one or more fingerprints on a display screen of the electronic device for interacting with the one or more graphical objects. Further, the method may include dynamically performing an action in the electronic device using the one or more fingerprints.
  • The proposed method can be used for interacting with the electronic device using the fingerprint on display. With the proposed method, various fingerprints can be assigned to various graphical objects and those graphical objects can be retrieved, accessed, and/or controlled using the fingerprints on the display.
  • In some embodiments, the method may include detecting the fingerprint on the display and modifying various parameters (e.g.: appearance, zoom, scenes, audio mute, language) of the graphical objects.
  • In various embodiments, the method may include providing various recommendations for assigning the various fingerprints to the graphical objects.
  • Referring now to the drawings, and more particularly to FIGS. 1A through 13, in which similar reference characters denote corresponding features consistently throughout the figures, there are described various embodiments.
  • FIG. 1A is a block diagram of an electronic device 100, according to an embodiment. As illustrated in the FIG. 1A, the electronic device 100 includes a fingerprint registration engine 110, a fingerprint recognition engine 120, an interaction engine 130, a processor 140, a memory 150 and a display screen 160.
  • In an embodiment, fingerprint registration engine 110 can be configured to register the fingerprints of various fingers of the user, and associate the fingerprints with objects. Herein, the objects may include contents, functions, information, graphical items (icons, graphics, etc.), application components, or any combination thereof. For example, a fingerprint of the thumb is registered with a first object, a fingerprint of the index finger is registered with a second object, and a fingerprint of the middle finger is registered with a third object. More than one object may be registered with a same fingerprint.
  • The fingerprint registration engine 110 can be configured to detect the fingerprints of the various fingers in the user's hand and register these fingerprints of the various fingers in association with one or more objects. The fingerprint registration engine 110 can implement various conventional mechanisms for obtaining various fingerprints of the fingers in the user's hand. The fingerprint registration engine 110 can include one or more sensors and other components for obtaining the fingerprints of the user. The fingerprint registration engine 110 can be configured to store the one or more fingerprints detected from the user in a fingerprint database 110 a. The various steps performed by the fingerprint registration engine are described in FIG. 1B.
  • The fingerprint recognition engine 120 can be configured to recognize the detected one or more fingerprints of the user. In some embodiments, the fingerprint recognition engine 120 can be configured to identify the objects associated with the one or more fingerprints of the user.
  • In an embodiment, the interaction engine 130 can be configured to perform one or more actions in the electronic device 100 by identifying the objects assigned to the one or more fingerprints based on the fingerprints detected on the display screen 160. The various actions performed by the interaction engine 130 based on the fingerprints detected on the display screen 160 are explained in conjunction with the figures in the later parts of the description.
  • The processor 140 is coupled with the interaction engine 130 and the memory 150. The processor 140 is configured to execute instructions stored in the memory 150 and to perform various actions for providing the context based multimodal predictions. The memory 150 also stores instructions to be executed by the processor 140. The memory 150 may include non-volatile storage elements. Although the fingerprint database 110 a is illustrated as being a component of the fingerprint registration engine, the fingerprint database 110 may be stored in the memory 150 and accessed by the fingerprint registration engine 110.
  • The display screen 160 in combination with one or more other components of the electronic device 100, such as the fingerprint registration engine 110, the fingerprint recognition engine 120 and the processor 140, can be configured to detect fingerprints and to display the objects. Further, the display screen 160 can be configured to display UI changes based on the detected fingerprints of the user.
  • Although the FIG. 1A shows various hardware components of the electronic device 100, it is to be understood that other embodiments are not limited thereon. In other embodiments, the electronic device 100 may include fewer or more components. Further, the labels or names of the components are used only for illustrative purpose and do not limit the scope of the invention. One or more components can be combined together to perform same or substantially similar function to perform actions in the electronic device 100.
  • FIG. 1B illustrates various steps performed by a fingerprint registration engine 110, according to an embodiment. The fingerprint registration engine 110 captures fingerprints of the various fingers. Further, the fingerprint registration engine 110 processes the images corresponding to the fingerprints of the various fingers to uniquely identify the various fingerprints of the user. In some embodiments, fingerprint registration engine 110 can be configured to detect the fingerprints of the various fingers in the user's hand and register these fingerprints of the various fingers with one or more objects. The fingerprint registration engine 110 can be configured to associate the registration of the fingerprints with the one or more objects in the fingerprints database 110 a.
  • FIG. 1C illustrates various steps performed by a fingerprint recognition engine 120, according to an embodiment. In an embodiment, the fingerprint recognition engine 120 can be configured to identify various user interactions through the fingerprints. The fingerprint recognition engine 120 captures images of the fingerprints to identify the fingerprints of the user. Further, the fingerprint recognition engine 120 determines whether there exists any objects assigned to the detected fingerprints. Thus, the fingerprint recognition engine 120 identifies the objects assigned to the one or more fingerprints by checking the fingerprint database 110.
  • FIG. 1D illustrates installation examples of fingerprint sensors according to an embodiment. Referring to FIG. 1D, a plurality of fingerprint sensors 115 may be installed over the entire panel of the display screen 160. That is, the plurality of fingerprint sensors 115 may be installed in a form in which a plurality of sensing elements are arranged. Accordingly, fingerprint recognition may be performed through the entire region of the display screen 160. In this regard, the plurality of fingerprint sensors 115 may be regarded as a single fingerprint sensor disposed across the entirety of the panel of the display screen 160. Since the plurality of fingerprint sensors 115 is located in the entire region of the panel of the display screen 160 rather than the partial region thereof, a fingerprint can be input anywhere in a user's desired region of the panel of the display screen 160, thereby enhancing the user experience.
  • FIG. 2 is a flow chart illustrating a method for enabling interaction using fingerprint, according to an embodiment.
  • Referring to FIG. 2, at step 202, the electronic device 100 detects at least one fingerprint. In response to at least one touch input being detected on the display screen 160, the electronic device 100 detects the at least one fingerprint using at least one fingerprint sensor.
  • At step 204, the electronic device 100 identifies an object corresponding to the fingerprint. Specifically, the electronic device 100 identifies the fingerprint by matching a captured fingerprint to one of registered fingerprints, and retrieves the object in a database including mapping information related to objects and fingerprints. The mapping information may be defined for each application. For example, depending on an application, the fingerprints may correspond to graphical objects (e.g. a character, or a component of graphic model), functions (e.g., zoom in/out, or blurring), information (e.g., an identifier), interface items (e.g., virtual keys), states of the electronic device 100 (e.g., lock/unlock, or a mode) and so on.
  • At step 206, the electronic device 100 performs at least one action based on the object. The electronic device 100 performs at least one action related to the identified object. In various embodiments, the identified object may be used as at least one of a lookup parameter, a command, an address, a classifier and so on. For example, the electronic device 100 may reproduce a portion corresponding to the object in a video, store or transmit data to a destination corresponding to the object, display a user interface (UI) corresponding to the object, transition a mode corresponding to the object, apply a modification corresponding to the object to contents, or perform a function corresponding to the object.
  • FIGS. 3A-3C are examples for interacting with multimedia content using a fingerprint on display, according to embodiments of the disclosure.
  • As illustrated in the FIG. 3A, graphical objects 302 a, 302 b and 302 c are assigned to the fingerprints of thumb, index finger and little finger, respectively. The graphical objects 302 a, 302 b and 302 c are associated with the fingerprints of the thumb, the index finger and the little finger. These graphical objects 302 a, 302 b and 302 c can be retrieved or accessed using the fingerprints of the thumb, the index finger and the little finger. In an example, the graphical objects 302 a, 302 b and 302 c can be images, human faces, real world objects, buildings, application icons, labels, buttons, User Interface (UI) elements or the like.
  • When the graphical objects are considered as human faces, each human face can be assigned to each fingerprint of the thumb, the index finger and the little finger. As illustrated in the FIG. 3B, in order to access the multimedia content having the human face(s), the user performs a tap using the index finger on the multimedia content. When the user taps on the multimedia content using the index finger, the electronic device identifies the presence of the human face (i.e., the graphical object 302 b) in the multimedia content. Further, the electronic device renders the content having only the human face (i.e., the graphical object 302 b) as shown in the FIG. 3C. Further, when the user performs a tap using the little finger (which is assigned to graphical object 302 c), the electronic device renders the content having only the human face (i.e., the graphical object 302 c) as shown in the FIG. 3C.
  • FIG. 4A is a flow chart illustrating a method for interacting with a multimedia content using fingerprint on display, according to an embodiment.
  • Referring to FIG. 4A, at step 402, the electronic device 100 registers the one or more fingerprints with one or more graphical objects. The fingerprint registration engine 110 registers the one or more fingerprints with one or more graphical objects.
  • At step 404, the electronic device 100 detects the one or more fingerprints on a display screen 160 for interacting with the one or more graphical objects. The fingerprint recognition engine 120 detects the one or more fingerprints on the display screen 160 of the electronic device 100 for interacting with the one or more graphical objects.
  • At step 406, the electronic device 100 performs an action using the one or more fingerprints. The interaction engine 130 dynamically performs an action in the electronic device 100 using the one or more fingerprints.
  • The various actions, acts, blocks, steps, or the like in the flow diagram 300 may be performed in the order presented, in a different order or simultaneously. Further, in some embodiments, some of the actions, acts, blocks, steps, or the like may be omitted, added, modified, skipped, or the like without departing from the scope of the invention.
  • FIG. 4B is a flow chart illustrating a method for registering the fingerprint(s) with graphical object(s), according to an embodiment.
  • Referring to FIG. 4B, at step 412, the electronic device 100 detects the one or more fingerprints for the one or more graphical objects. The fingerprint registration engine 110 detects the one or more fingerprints for the one or more graphical objects.
  • At step 414, the electronic device 100 assigns the detected one or more fingerprints to the one or more graphical objects. The fingerprint registration engine 110 assigns the detected one or more fingerprints to the one or more graphical objects.
  • At step 416, the electronic device 100 stores the one or more fingerprints assigned to the one or more graphical objects. The fingerprint registration engine 110 stores the one or more fingerprints assigned to the one or more graphical objects in the fingerprint database 110 a.
  • The various actions, acts, blocks, steps, or the like in the flow diagram 400 may be performed in the order presented, in a different order or simultaneously. Further, in some embodiments, some of the actions, acts, blocks, steps, or the like may be omitted, added, modified, skipped, or the like without departing from the scope of the invention.
  • FIGS. 5A-5C illustrate an example in which the electronic device provides interaction with graphical objects using the fingerprint on the display, according to an embodiment.
  • As illustrated in the FIG. 5A, the user registers fingerprints with each component of a 3D CAD model. The fingerprint of the index finger (of the right hand) is registered with a trigger and the index finger (of the left hand) is registered with a spring. Thus, each fingerprint of each finger is registered with a each component of the 3D CAD model as shown in the FIG. 5A.
  • As illustrated in the FIG. 5B, the user interacts with the electronic device using the index finger of the right hand. As the index finger is registered with the trigger, the electronic device 100 displays the 3D CAD model by highlighting the trigger as shown in the FIG. 5C.
  • FIG. 6 is an example for transferring content across electronic devices using the fingerprint on display, according to an embodiment. FIG. 6 illustrates an example of copying content in a first electronic device 100 a and transferring the content to a second electronic device 100 b. As illustrated in the FIG. 6, the user can select the graphical objects, which can be application icons, labels, text, buttons, content or the like. The user selects various graphical objects by assigning the fingerprints for each of the selected object. The selected objects copied in the first electronic device 100 are extracted at the server. It should be noted that the server shown in the maintains a mapping of the fingerprints and the associated graphical objects.
  • Further, when the user taps on the second electronic device 100 b using the fingerprints, the server 600 transfers the selected graphical objects to the second electronic device as shown in the FIG. 6. Thus, with the proposed method, by assigning the fingerprints to the graphical objects, the graphical objects can be copied in the first electronic device 100 a and are transferred to the second electronic device 100 b.
  • FIG. 7 is an example for assigning the graphical objects for the fingerprints, according to an embodiment. The electronic device 100 provides various recommendations to the user for assigning fingerprints to the graphical objects. For example, the index finger is commonly used by the user for interacting with the electronic device 100. Thus, the electronic device 100 provides recommendations to the user for registering the fingerprint of the index finger to a graphical object (such as a human face) that appears most frequently in a multimedia content. In another example, the electronic device 100 provides recommendations to the user for registering the fingerprint of the index finger to a graphical object (such as a human face) that appears for a maximum duration in the multimedia content. Thus, the electronic device 100 provides various recommendation to the for registering the fingerprint(s) based on various parameters including but not limited to on-screen focus time, on-screen location of the graphical object in the multimedia content.
  • In some embodiments, the electronic device 100 provides various recommendations to the user for registering the fingerprint(s) based on usage pattern of the graphical objects.
  • As illustrated in the FIG. 7, the electronic device generates recommendations for the graphical objects 102 a-102 d for index finger, little finger and thumb finger respectively for the user to access or to interact with the graphical objects.
  • FIGS. 8A-8D are examples in which various actions are performed on the multimedia content using the fingerprint on display of the electronic device, according to an embodiment. FIGS. 8A-8D illustrate examples of content retrieval and content view management in the electronic device using the fingerprints.
  • As illustrated in the FIG. 8A, the user taps on the screen using a registered fingerprint to skip music related content or content related to a specific character associated with the registered fingerprint.
  • As illustrated in the FIG. 8B, the content of a video to be displayed is modified to show only selected content. The selected content is shown based on the fingerprint detected on the display screen of the electronic device 100.
  • As illustrated in the FIG. 8C, if the user intends to only display only photographs or video content without revealing identities of people in the video or image, the user can tap with the fingerprint on a selected photograph to enable blurring of the face of the people in the photograph.
  • As illustrated in the FIG. 8D, the user can perform zoom on a selected face in the photograph by applying the fingerprint on the selected face. Although, the FIGS. 8A-8D show limited examples of performing actions in the electronic device using the fingerprints, it should be noted that the user can perform various actions using the fingerprints assigned to the graphical objects.
  • FIGS. 9A-9E are examples of performing a zoom operation on content using various fingerprints of the user, according to an embodiment. The user assigns various fingerprints A to E on a navigational map application as shown in the FIG. 9A. Various zoom levels can be achieved using the fingerprints A to E. As illustrated in the FIG. 9B, when the fingerprints A and B are used, a first zoom level is achieved. As illustrated in the FIG. 9C, when the fingerprints A and C are used, a second zoom level is achieved. As illustrated in the FIG. 9D, when the fingerprints A and D are used, a third zoom level is achieved. Similarly, when the fingerprints A and E are used, the highest zoom level is achieved. Thus, with the proposed method, the user can perform zooming at various level by assigning the fingerprints on the navigational map application.
  • FIGS. 10A and 10B are an example of selecting characters using fingerprints, according to an embodiment. As illustrated in the FIG. 10A, various characters can be selected or inputted using the fingerprints assigned to those characters. For example, the characters A, S, D, F and G are assigned to the index figure and similarly the other characters are assigned to other fingers at Level 1 as shown in the FIG. 10A. Further, referring to FIG. 10B, at Level 2, when the user taps the index finger on the display screen, the assigned characters A, S, D, F and G are displayed. The user can select the characters using any of the fingers. Further, when the user releases the tap on the display screen, the characters disappear automatically.
  • FIG. 11 is an example in which the electronic device displays applications on a portion of the display screen using the fingerprints, according to an embodiment. As illustrated in the FIG. 11, the various applications on a portion of the display screen can be registered with fingerprint. As shown in FIG. 11, consider that the user has registered a fingerprint on the top portion of display screen with App 1 and App 2. When the user taps on the top portion of display screen using the fingerprint, App 1 and App 2 are displayed. Further, if the user registers a fingerprint on a mid portion of display screen with App 3 and App 4 and when the user taps on the mid portion of display screen, App 3 and App 4 are displayed.
  • FIGS. 12A and 12B illustrate an example of copying and pasting contents using fingerprints, according to an embodiment. Referring to FIG. 12A, a user can copy text contents 1202 a and 1202 b into a clipboard by using a copy command. Herein, the user uses an index finger when copying the first text content 1202 a, and uses a middle finger when copying the second text content 1202 b. According to an embodiment, the electronic device 100 stores text contents 1202 a and 1202 b with mapping information. The mapping information indicates at least one fingerprint mapped to each of text contents 1202 a and 1202 b. Accordingly, as illustrated in FIG. 12B, when the user pastes a text content by using a paste command, the pasted text content depends on which finger is touched on the display screen 160. For example, as in example 1220 a, if the user uses the index finger, the electronic device 100 display the first text content 1202 a based on the mapping information. Otherwise, if the user uses the middle finger, the electronic device 100 adds the second text content 1202 b based on the mapping information.
  • In an embodiment described with FIGS. 12A and 12B, copying and pasting operations are exemplified. In various embodiments, aforementioned embodiment is applied to different operation (e.g., moving operation). Further, aforementioned embodiment is applied to different type of contents (e.g., file, icon and so on).
  • FIG. 13 is an example of assigning functions using fingerprints, according to an embodiment. Referring to FIG. 13, functions related to an application are assigned to each of fingers. In an embodiment, a set of functions assigned to the fingers can changed based on which application is running in the electronic device 100. As illustrated in FIG. 13, functions 1 a-1 e are assigned to the fingers while an application #1 is running, and functions 2 a-2 e are assigned to the fingers while an application #2 is running. Hence, although the user inputs a command by using same finger, a function to be performed depends on an application being running. For example, the index finger can be used to command a function to load a file while the application #1 is running, and the index finger can be used to command to capture a screen while the application #2 is running. To do so, the electronic device 100 stores mapping information that defines mapping between a set of function and an application with each fingerprint. Further, the electronic device 100 can perform a procedure for assigning functions to the fingers to generate the mapping information.
  • The embodiments disclosed herein can be implemented through at least one software program running on at least one hardware device and performing network management functions to control the elements. The elements shown in FIGS. 1-14 include blocks which can be at least one of a hardware device, or a combination of hardware device and software module.
  • The foregoing description of the specific embodiments will so fully reveal the general nature of the embodiments herein that others can, by applying current knowledge, readily modify and/or adapt for various applications such specific embodiments without departing from the generic concept, and, therefore, such adaptations and modifications should and are intended to be comprehended within the meaning and range of equivalents of the disclosed embodiments. It is to be understood that the phraseology or terminology employed herein is for the purpose of description and not of limitation. Therefore, while the embodiments herein have been described in terms of preferred embodiments, those skilled in the art will recognize that the embodiments herein can be practiced with modification within the spirit and scope of the embodiments as described herein.

Claims (20)

1. A method for operating an electronic device, the method comprising:
detecting at least one fingerprint from at least one touch input on a display screen;
identifying an object corresponding to the at least one fingerprint; and
performing at least one actions based on the object.
2. The method of claim 1, wherein the performing at least one actions comprises,
in response to a copy command being occurred, storing a content that is selected by the copy command with mapping information indicating that the content is mapped to the fingerprint; and
in response to a paste command being occurred with detecting the fingerprint, displaying the content based on the mapping information.
3. The method of claim 1, wherein the performing at least one actions comprises,
performing one of functions in a set of functions assigned an application that is running.
4. The method of claim 3, wherein the performing at least one actions comprises
extracting a portion of a content based on the object; and
displaying the extracted portion.
5. The method of claim 1, wherein the performing at least one actions comprises,
selecting at least one content in the electronic device using the at least one fingerprint; and
transferring the selected content to another electronic device using the at least one fingerprint.
6. The method of claim 1, wherein the performing at least one actions comprises,
selecting one or more graphical objects in the electronic device using the at least one fingerprint; and
accessing the selected one or more graphical objects with the at least one fingerprint.
7. The method of claim 1, wherein the performing at least one actions comprises,
modifying one or more the parameters of a content based on the at least one fingerprint.
8. The method of claim 1, wherein the performing at least one actions comprises,
displaying a virtual keypad comprising a set of keys corresponding to the at least one fingerprint.
9. The method of claim 1, wherein the performing at least one actions comprises,
display a map in a zoom level corresponding to the at least one fingerprint.
10. The method of claim 1, further comprising:
providing one or more recommendations for assigning the at least one fingerprint with one or more objects.
11. An electronic device comprising:
a display screen;
at least one touch sensor; and
a processor coupled to the display screen and the at least one touch sensor and configured to:
detect at least one fingerprint from at least one touch input on a display screen;
identify an object corresponding to the at least one fingerprint; and
perform at least one actions based on the object.
12. The electronic device of claim 11, wherein the processor is configured to:
in response to a copy command being occurred, store a content that is selected by the copy command with mapping information indicating that the content is mapped to the fingerprint; and
in response to a paste command being occurred with detecting the fingerprint, display the content based on the mapping information.
13. The electronic device of claim 11, wherein the processor is configured to:
perform one of functions in a set of functions assigned an application that is running.
14. The electronic device of claim 13, the processor is configured to:
extract a portion of a content based on the object; and
display the extracted portion.
15. The electronic device of claim 11, the processor is configured to:
select at least one content in the electronic device using the at least one fingerprint; and
transfer the selected content to another electronic device using the at least one fingerprint.
16. The electronic device of claim 11, wherein the processor is configured to:
select one or more graphical objects in the electronic device using the at least one fingerprint; and
access the selected one or more graphical objects with the at least one fingerprint.
17. The electronic device of claim 11, wherein the processor is configured to:
modify one or more the parameters of a content based on the at least one fingerprint.
18. The electronic device of claim 11, wherein the processor is configured to:
display a virtual keypad comprising a set of keys corresponding to the at least one fingerprint.
19. The electronic device of claim 11, wherein the processor is configured to:
display a map in a zoom level corresponding to the at least one fingerprint.
20. The electronic device of claim 11, the processor is configured to:
provide one or more recommendations for assigning the at least one fingerprint with one or more objects.
US16/131,774 2017-09-14 2018-09-14 Method for enabling interaction using fingerprint on display and electronic device thereof Pending US20190095079A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
IN201741032625 2017-09-14
IN201741032625 2017-09-14

Publications (1)

Publication Number Publication Date
US20190095079A1 true US20190095079A1 (en) 2019-03-28

Family

ID=65724091

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/131,774 Pending US20190095079A1 (en) 2017-09-14 2018-09-14 Method for enabling interaction using fingerprint on display and electronic device thereof

Country Status (3)

Country Link
US (1) US20190095079A1 (en)
EP (1) EP3659025A4 (en)
WO (1) WO2019054796A1 (en)

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9271150B2 (en) * 2013-11-12 2016-02-23 Lg Electronics Inc. Mobile terminal and control method thereof
KR20150056353A (en) * 2013-11-15 2015-05-26 엘지전자 주식회사 The mobile terminal and the control method thereof
KR20150087667A (en) * 2014-01-22 2015-07-30 삼성전자주식회사 Electronic device and method for providing control function using fingerprint sensor
KR20150092479A (en) * 2014-02-05 2015-08-13 엘지전자 주식회사 Display device and method for controlling the same
CN104537366A (en) * 2015-01-07 2015-04-22 小米科技有限责任公司 Touch key and fingerprint identification implementation method and device and terminal equipment
KR20170021514A (en) * 2015-08-18 2017-02-28 엘지전자 주식회사 Display apparatus and controlling method thereof
KR20170029757A (en) * 2015-09-08 2017-03-16 엘지전자 주식회사 Mobile terminal and method for controlling the same

Also Published As

Publication number Publication date
EP3659025A1 (en) 2020-06-03
EP3659025A4 (en) 2020-08-19
WO2019054796A1 (en) 2019-03-21

Similar Documents

Publication Publication Date Title
US10599316B2 (en) Systems and methods for adjusting appearance of a control based on detected changes in underlying content
US10754603B2 (en) Systems, devices, and methods for dynamically providing user interface controls at a touch-sensitive secondary display
CN104036196B (en) For configuring password and means for unlocking and method
US8941591B2 (en) User interface elements positioned for display
US20170010848A1 (en) Multi-Device Pairing and Combined Display
US9158907B2 (en) Alternative unlocking patterns
US20170131859A1 (en) Hierarchical User Interface
US20190294327A1 (en) Glow touch feedback for virtual input devices
US10198109B2 (en) Supplementing a touch input mechanism with fingerprint detection
KR101737166B1 (en) Application for viewing images
AU2017100879B4 (en) Systems, devices, and methods for dynamically providing user interface controls at touch-sensitive secondary display
EP2701152B1 (en) Media object browsing in a collaborative window, mobile client editing, augmented reality rendering.
US9164659B2 (en) Simultaneous input across multiple applications
CN103718187B (en) Input via the safety of touch screen
US20160070432A1 (en) Manipulating Documents in Touch Screen File Management Applications
US9864504B2 (en) User Interface (UI) display method and apparatus of touch-enabled device
US9606717B2 (en) Content composer
CN102981728B (en) Semantic zoom
CN102999274B (en) Semantic zoom animation
CA2788106C (en) Multi-screen pinch and expand gestures
CA2788200C (en) Multi-screen hold and page-flip gesture
US9430145B2 (en) Dynamic text input using on and above surface sensing of hands and fingers
DE602004013116T2 (en) Haptic key-controlled data entry
EP2613238A2 (en) Method and apparatus for managing icon in portable terminal
JP2017534115A (en) Sweep and group tabs

Legal Events

Date Code Title Description
AS Assignment

Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DEOTALE, GUNJAN PRAKASH;RAUT, SAMARTH SHANKAR;SHAH, JIDNYA SAMIR;REEL/FRAME:047856/0393

Effective date: 20180917

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STCB Information on status: application discontinuation

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION