US20140176593A1 - Mobile device user interface having enhanced visual characteristics - Google Patents
Mobile device user interface having enhanced visual characteristics Download PDFInfo
- Publication number
- US20140176593A1 US20140176593A1 US14/189,893 US201414189893A US2014176593A1 US 20140176593 A1 US20140176593 A1 US 20140176593A1 US 201414189893 A US201414189893 A US 201414189893A US 2014176593 A1 US2014176593 A1 US 2014176593A1
- Authority
- US
- United States
- Prior art keywords
- computing device
- visual characteristic
- application
- age
- elements
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/001—Texturing; Colouring; Generation of texture or colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04817—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/02—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators characterised by the way in which colour is displayed
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04804—Transparency, e.g. transparent or translucent windows
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/26—Devices for calling a subscriber
- H04M1/27—Devices whereby a plurality of signals may be stored simultaneously
- H04M1/274—Devices whereby a plurality of signals may be stored simultaneously with provision for storing more than one subscriber number at a time, e.g. using toothed disc
- H04M1/2745—Devices whereby a plurality of signals may be stored simultaneously with provision for storing more than one subscriber number at a time, e.g. using toothed disc using static electronic memories, e.g. chips
- H04M1/27453—Directories allowing storage of additional subscriber data, e.g. metadata
- H04M1/2746—Sorting, e.g. according to history or frequency of use
Definitions
- a camera coupled to the computing device may capture an image, and each respective visual effect may include at least a portion of the image.
- the image may comprise an image of a user associated with the computing device.
Abstract
A computing device includes a user interface module operable by at least one processor to output, for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of application features, determine an age for a first application feature of the plurality of application features, the age being a period of time that has elapsed since the first application feature was enabled in the application, generate, based on the age, a visual characteristic that modifies at least one of a chrominance value and a luminance value associated with a first element of the plurality of elements associated with the first application feature, and output, for display, the visual characteristic.
Description
- This application is a continuation of U.S. patent application Ser. No. 13/854,806, filed Apr. 1, 2013, which claims the benefit of U.S. provisional application No. 61/714,680, filed Oct. 16, 2012, and U.S. provisional application No. 61/789,029, filed on Mar. 15, 2013, the entire content of each of which is incorporated herein by reference.
- Mobile computing devices, such as mobile phones and tablet computers, are becoming increasingly powerful and ubiquitous, and applications for these computerized devices have begun to offer extensive functionality. Many of these mobile computing devices are provided with several pre-installed applications (often referred to as “apps”) that offer at least a basic set of functions (e.g., telephony, photo capture, web browsing, email, and navigation). In addition, numerous such apps are available for downloaded and installation from online app stores or markets.
- As the functionality of mobile computing devices and applications executable thereby has expanded, so has the importance of the user interface in ensuring a lightweight and intuitive mechanism for providing input to and receiving feedback from the mobile computing device.
- In one aspect, the disclosure describes a method that includes outputting, by a computing device and for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of features, determining, by the computing device, an age for a first application feature of the plurality of application features, the age being a period of time that has elapsed since the first application feature was enabled in the application, generating, by the computing device, based on the age, a visual characteristic that modifies at least one of a chrominance or luminance value associated with a first element of the plurality of elements associated with the first application feature, and outputting, by the computing device and for display, the visual characteristic.
- In another aspect, this disclosure describes a method that includes outputting, by a computing device for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, determining, by the computing device, an age for a first element of the plurality of elements, the age being a period of time that has elapsed since the first element was last selected in the application, generating, by the computing device, based on the age of the first element, a visual characteristic that modifies at least one of chrominance or luminance value indicating the age of the first element, and outputting, by the computing device for display, the visual characteristic for the first element.
- In another aspect, the disclosure describes a computing device that includes at least one processor, and a user interface module operably by the at least one processor. In accordance with this aspect of the disclosure, the user interface module may be operable by the at least one processor to output, for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of features, determine an age for a first application feature of the plurality of application features, the age being a period of time that has elapsed since the first application feature was enabled in the application, generate, based on the age, a visual characteristic that modifies at least one of a chrominance value and a luminance value associated with a first element of the plurality of elements associated with the first application feature, and output, for display, the visual characteristic.
- In another example, A device includes at least one processor and a user interface module operable by the at least one processor. The user interface module is operable by the at least one processor to output for display at a display device, a graphical user interface of an application including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of features, determine an age for a first application feature of the plurality of application features, the age being a period of time that has elapsed since the first application feature was enabled in the application, generate, based on the age, a visual characteristic that modifies at least one of a respective chrominance or luminance value associated with a first element of the plurality of elements associated with the first application feature, and output for display, the visual characteristic.
- The details of one or more examples are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the disclosure will be apparent from the description and drawings, and from the claims.
-
FIG. 1 is a conceptual diagram illustrating an example computing device that generates and outputs enhanced visual characteristics based on various factors in accordance with one or more aspects of the present disclosure. -
FIG. 2 is a block diagram illustrating an example computing device, in accordance with one or more aspects of the present disclosure. -
FIGS. 3A-3B are conceptual diagrams that illustrate visual characteristics associated with user interface elements based on a period of time since a user has interacted with the user interface elements, in accordance with one or more aspects of the present disclosure. -
FIG. 4 is a conceptual diagram that illustrates generating and outputting enhanced visual characteristics associated with a user interface elements that represent one or more contacts, in accordance with one or more aspects of the present disclosure. -
FIG. 5 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure. -
FIG. 6 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. -
FIG. 7 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. -
FIG. 8 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. - This disclosure describes techniques for constructing and outputting, from a computing device, a graphical user interface (GUI) in which interface elements have enhanced visual characteristics based on various factors. The enhanced visual characteristics may improve the experience of a user of the computing device by drawing the attention of the user to particular interface elements and, for example, by providing an indication, based on the associated visual characteristic, of a classification of each of the user interface elements.
- As one example, a computing device may construct and output user interface elements for an application having visual characteristics indicative of an “age” (e.g., an amount of time since the installation or receipt by the device) of one or more capabilities or features of the application. As a result, a user may be able to more readily distinguish “newer” apps and/or application features from “older” apps and/or application features.
- As a second example, the computing device may output user interface elements for an application having visual characteristics based on an amount of time elapsed since a most-recent user selection of or interaction with the respective user interface element. In some instances, the respective user interface element may be associated with a particular capability or feature of application. In this manner, the visual characteristics may provide an indication of application features and associated user interface elements which the user has not recently used or with which the user has not recently interacted. Such visual characteristics may signal to a user which application features are most-frequently used, and thereby more easily identify any unused features of the application.
-
FIG. 1 is a conceptual diagram illustrating an example computing device that generates and outputs enhanced visual characteristics based on various factors in accordance with one or more aspects of the present disclosure. In the example ofFIG. 1 ,computing device 2 may be a mobile phone. However, in other examples,computing device 2 may be a personal digital assistant (PDA), a desktop computer, a laptop computer, a tablet computer, a portable gaming device, a portable media player, a camera, an e-book reader, a watch, or another type of computing device. Moreover,computing device 2 may include one or more stand-alone devices or may be part of a larger system.User 102 interacts withcomputing device 2 using various input and output devices, examples of which are shown in greater detail inFIG. 2 . - In the example shown in
FIG. 1 ,computing device 2 includes a user interface device (UID) 4, user interface device module 6 (UI device module 6), application feature module 8, andapplication modules 17A-17N.User 102 may be associated with computing device, 2, and may perform various actions, described in further detail below, usingcomputing device 2. -
UID 4 ofcomputing device 2 may function as an input device and/or an output device forcomputing device 2. For example, UID 4 ofcomputing device 2 may include a presence-sensitive display, such as a touchscreen configured to receive tactile user input from a user ofcomputing device 2. UID 4 may receive tactile user input as one or more taps and/or gestures. UID 4 may detect taps or other gestures in response to the user touching or pointing to one or more locations ofUID 4 with a finger or a stylus pen. UID 4 may be implemented using various technologies. For example, UID 4 may be implemented using a resistive touchscreen, a surface acoustic wave touchscreen, a capacitive touchscreen, a projective capacitance touchscreen, a pressure sensitive screen, an acoustic pulse recognition touchscreen, or another presence-sensitive display technology. Furthermore, UID 4 may include any one or more of a liquid crystal display (LCD) dot matrix display, light emitting diode (LED) display, organic light-emitting diode (OLED) display, e-ink, or similar monochrome or color display capable of outputting visible information touser 102 ofcomputing device 2. - UID 4 may output a graphical user interface (GUI), such as GUI 10 and GUI 18, which may be related to functionality provided by
computing device 2. For example, UID 4 may outputGUI 10 andGUI 18, which are associated with one or more ofapplication modules 17A-17N, and which are described in further detail below. -
UI device module 6 may comprise hardware or software that interacts with and/or controlsUID 4, and which manages the layout and user interaction associated with elements of a GUI. As an example,UI device module 6 may comprise one or more graphics processing units (GPUs), drivers, software, and/or firmware that may control the output and input fromUID 4.UI device module 6 may also manage the layout and user interaction with the elements of a user interface, such as GUI 10 and GUI 18, whichapplication modules 17A-17N may generate. In some examples,UI device module 6 may also generate and output enhanced visual characteristics, which may modify luminance and/or chrominance values associated with elements of the user interface based on various factors in accordance with the techniques of this disclosure. - Application modules 17-17N (“application modules 17”) represent one or more applications of
computing device 2. As some examples, application modules 17 may include a restaurant finder application, a contact management application, which may allow interaction with contacts associated withuser 102, for example using text messaging (e.g., SMS or MMS messaging), voice, and video communications. Application modules 17 may also include various applications thatuser 102 may download and install from an application store or marketplace.User 102 orcomputing device 2 may update the version or capabilities of one or more of application modules 17. In some examples, the updated applications may include new application capabilities, also referred to herein as software features, which application feature module 8 may monitor. - More specifically, application feature module 8 may monitor installation and usage of individual capabilities provided by application modules 17. In some examples, application feature module 8 may track when new application capabilities are added to each of application modules 17, e.g., by way of a software update. As part of tracking application capabilities, application feature module 8 may determine a date and/or timestamp associated with one or more capabilities of each of application modules 17. Based on the date associated with each application capability, application feature module 8 may determine a visual characteristic for a user interface element associated with that application capability, and
UID 4 may output the visual characteristics in accordance with the techniques of this disclosure. In some examples, the visual effects may modify luminance and/or chrominance values associated with elements of the user interface elements. Although shown separately from application modules 17, application feature module 8 may be a component or module integrated within one or more of application modules 17 or within an operating system or other software component ofcomputing device 2. - In the example of
FIG. 1 , one of application modules 17 (e.g.,application module 17A) may generateGUI 10 andGUI 18 at time (t1) 28 and time (t1) 30, respectively. When constructing and outputtingGUIs UID 4 may output user interface elements ofGUIs GUIs UID 4 may control the visual characteristics for each UI element based on the age - As a more detailed illustration of the first example of outputting enhanced visual characteristics,
user 102 may, at a time (t0), download and install a restaurant finder application which may include UI elements that are associated with the enhanced visual characteristics. The restaurant application may comprise one of application modules 17, which subsequently causeUID 4 tooutput GUI 10 at a time (t1) 28 to includeuser interface elements User interface elements user 102 to search for restaurants with various characteristics. Moreover, application feature module 8 determines an age of each of features to whichuser interface elements user interface elements user 102 installed the restaurant finder application, the age of the capabilities ofuser interface elements GUI 10 was constructed. - At a later time (t2) 30,
user interface device 4outputs GUI 18. In betweentime 28 andtime 30,computing device 2 may update the restaurant finder application to a different version of the application having new capabilities. As an example,GUI 18 may includeUI elements GUI 10, but further includesUI element 12C, andUI element 14. In this example, bothUI element 12C andUI element 14 are associated with new application capabilities. For example,element 12C may be associated with a capability that allowsuser 102 to search for restaurants that have Wi-Fi, andUI element 14 may be associated with a search capability that may allowuser 102 to search for restaurants, e.g., based on a textual description of the restaurant. When computingdevice 2 generatesGUI 18,UI device module 6 may constructUI elements computing device 2 hasoutput GUI 18 such thatUI element 12C associated with the new Wi-Fi property capability has enhanced visual characteristic 13 andUI element 14 associated with the text search capability has enhanced visual characteristic 15. As a result, auser 102 may be able to more readily distinguish newer application features associated withUI elements visual characteristics - As a more detailed illustration of the second example described above, application feature module 8 may determine an age associated with UI elements based on the time since
user 102 last selected each element, and based on the age of each UI element,UID 4 may output a visual characteristic indicative of the age associated with each UI element. In the example ofFIG. 1 ,user 102 may selectUI element 12A at time (t1) 28, but may not have selectedUI element 12B.Application feature module 28 may determine an age associated withUI elements UI elements UI element 12A may be older than the age ofUI element 12B. As a result,UI device module 6 may generate, andUID 4 may output,UI elements GUI 18 to have visual characteristics indicative of this age difference. In some examples, the enhanced visual characteristics, may modify luminance and/or chrominance values associated with elements, e.g., ofUI elements UID 4 may displayUI element 12A to have a first visual characteristic, and the first visual characteristic may differ from a second visual characteristic that is associated withUI element 12B, based on the different ages ofUI elements UI element 12A may be displayed with certain shading and rendering indicative of an older object, such as a texturing that visually simulates a brittle or fragile state. Alternatively, or in addition,UI element 12 having a newer age may be rendered with bright, clean lines. In this way,user 102 may more easily access and utilize more frequently used features and may more easily identify any unused features of the software application, for example. - As described above,
computing device 2 may output enhanced visual characteristics based on various factors in accordance with the techniques of this disclosure. The enhanced visual characteristics may enhance the experience of a user, such asuser 102, ofcomputing device 2, by drawing the attention of the user to the interface elements that are associated with the improved visual effects. -
FIG. 2 is a block diagram illustrating an example computing device, in accordance with one or more aspects of the present disclosure.Computing device 2 ofFIG. 2 is described below within the context ofFIG. 1 . As shown in the example ofFIG. 2 ,computing device 2 includes user interface device 4 (“UID 4”), one ormore processors 40, one ormore input devices 42, one ormore communication units 44, one ormore output devices 46, and one ormore storage devices 48. In this example,UID 4 also includes presence-sensitive display 20. -
Storage devices 48 ofcomputing device 2 may also includeuser interface module 6, application feature module 8,application modules 17A-17N, visualcharacteristic module 16,contacts 22, andoperating system 54.Communication channels 50 may interconnect each of thecomponents communication channels 50 may include a system bus, a network connection, an inter-process communication data structure, or any other method for communicating data. - One or
more input devices 42 ofcomputing device 2 may receive input. Examples of input are tactile, audio, and video input.Input devices 42 ofcomputing device 2, in one example, includes a presence-sensitive display (e.g., presence-sensitive display 20), touch-sensitive screen, mouse, keyboard, voice responsive system, video camera, microphone, accelerometer, gyroscope, light sensor, or any other type of device for detecting input from a human or machine. - One or
more output devices 46 ofcomputing device 2 may generate output. Examples of output are tactile, audio, and video output.Output devices 46 ofcomputing device 2, in one example, includes a presence-sensitive display, sound card, video graphics adapter card, speaker, cathode ray tube (CRT) monitor, liquid crystal display (LCD), or any other type of device for generating output to a human or machine. - One or
more communication units 44 ofcomputing device 2 may communicate with external devices via one or more networks by transmitting and/or receiving network signals on the one or more networks. For example,computing device 2 may usecommunication unit 44 to transmit and/or receive radio signals on a radio network such as a cellular radio network. Likewise,communication units 44 may transmit and/or receive satellite signals on a satellite network such as a GPS network. Examples ofcommunication unit 44 include a network interface card (e.g. such as an Ethernet card), an optical transceiver, a radio frequency transceiver, a GPS receiver, or any other type of device that can send and/or receive information. Other examples ofcommunication units 44 may include Bluetooth®, GPS, 3G, 4G, and Wi-Fi® radios found in mobile devices as well as Universal Serial Bus (USB) controllers. -
UID 4 ofFIG. 2 includes presence-sensitive display 20 (hereafter “display 20”).Computing device 2 may usedisplay 20 as an input device and an output device. For example, display 20 ofUID 4 may include a touchscreen configured to receive tactile user input from a user ofcomputing device 2.Display 20 ofUID 4 may also include an LED display capable of outputting visible information to the user ofcomputing device 2.UID 4 may present a user interface ondisplay 20, such asGUIs FIG. 1 , that may be related to functionality provided bycomputing device 2. In some examples,display 20 ofUID 4 may present various functions and applications, such as an electronic message client, a contacts GUI, and a notifications GUI. - One or
more storage devices 48 withincomputing device 2 may store information required for use during operation of computing device 2 (e.g., application modules 17 ofcomputing device 2 may store information related to operation of the respective one of one or more applications).Storage devices 48, in some examples, have the primary purpose of being short term and not long-term computer-readable storage mediums.Storage devices 48 oncomputing device 2 may be configured for short-term storage of information as volatile memory and therefore not retain stored contents if powered off. Examples of volatile memories include random access memories (RAM), dynamic random access memories (DRAM), static random access memories (SRAM), and other forms of volatile memories known in the art.Storage devices 48 may further be configured for long-term storage of information as non-volatile memory space and retain information after power on/off cycles. Examples of non-volatile memories include magnetic hard discs, optical discs, floppy discs, flash memories, or forms of electrically programmable memories (EPROM) or electrically erasable and programmable (EEPROM) memories.Storage devices 48 may store program instructions and/or data associated withUI interface module 16, visualcharacteristic module 16, and application modules 17. -
Storage devices 48 may also includeoperating system 54. In some examples,operating system 54 may manage various processes and components ofcomputing device 2, such as storage devices, input, output, and peripheral devices.Operating system 54 may also generate events, e.g., based on timers, and/or interrupts, that applications ofcomputing device 2 may receive, for example using an event handler. - Applications 17 may store a variety of data to
storage devices 22, such as contact data (“contacts 22”).Contacts 22 may, for example, comprise a database or other data structure that an application, such as a phone dialer application, a notifications application, or a messaging application may access. Although illustrated as being stored locally onstorage devices 48, some contacts ofcontacts 22 may also be stored remotely, for example in a cloud. - One or
more processors 40 may implement functionality and/or execute instructions withincomputing device 2. For example,processors 40 oncomputing device 2 may read and execute instructions stored bystorage devices 48 that execute the functionality application ofUI device module 6, application feature module 8,user interface elements characteristic module 16, application modules 17, andoperating system 54. These instructions executed byprocessors 40 may causecomputing device 2 to store information withinstorage devices 48 during program execution, such as information related to the age of user interface elements, notification objects, and/or information associated withcontacts 22. - As described above with respect to
FIG. 1 ,computing device 2 may be configured to generate and output user interface elements having enhanced visual characteristics based on various factors. In some examples, the visual characteristics may modify at least one of a chrominance and a luminance value associated with the user interface elements. As a first example,storage devices 48 may include application feature module 8 operable to determine an age associated with each of a plurality of features of applications modules 17. Application feature module 8 may include a database or other data that may store the age of each of the features of one of the applications of application modules 17 instorage devices 48. Based on the age of each UI element, application feature module 8 may generate a visual characteristic that is associated with each UI element of the application of application modules 17. In some examples, visualcharacteristic module 16 may modify the quality of the visual effect beforeUID 4 outputs the visual characteristic to presence-sensitive display 20. As an example, visualcharacteristic module 16 may modify the chrominance, luminance, or other properties of the visual characteristic beforeUID 4 outputs the visual characteristic at presence-sensitive display 20. Aftervisual character module 16 modifies the quality of the visual characteristics,UID 4 outputs the visual characteristics associated with each UI element of the application of application modules 17. - As a second example of generating enhanced visual characteristics, application feature module 8 may also be configured to determine a time since a user, such as
user 102 has interacted with a particular user interface element of an application, and based on that time,UID 4 may output different visual characteristics to presence-sensitive display 20. In thisexample user 102 may select one or more UI elements. Based on the time thatuser 102 selects a first one of the UI elements of the application, application feature module 8 may store a date onstorage devices 48. Based on the date, application feature module 8 may determine an age for the first one of the UI elements of the application. Application feature module 8 may generate visual characteristics, which may modify at least one of a chrominance or luminance value of the first one of the user interface elements, based on the age, which visualcharacteristic module 16 may modify. After visualcharacteristic module 16 modifies the visual characteristics,UI device 4 outputs the visual characteristics for display at presence-sensitive display 20. Based on the visual characteristics associated with each of the UI elements,user 102 ofcomputing device 2 may be able to more easily identify application capabilities and UI elements thatuser 102 has interacted with more recently and less recently. - As described above,
computing device 2 may output enhanced visual characteristics based on various factors in accordance with the techniques of this disclosure. The enhanced visual characteristics may enhance the experience of a user, such asuser 102, ofcomputing device 2, by drawing the attention of the user to the interface elements that are associated with the improved visual effects. -
FIGS. 3A-3B are conceptual diagrams that illustrate visual characteristics associated with user interface elements controlled by computingdevice 2 based on a period of time since a user has interacted with the user interface elements, in accordance with one or more aspects of the present disclosure.FIGS. 3A and 3B illustrate two GUIs,GUI 80, andGUI 88.GUIs user interface elements Graphical elements GUI 80, andgraphical elements GUI 88, that include diagonal, horizontal, and/or vertical hashing, may indicate UI elements to which visual characteristic module 16 (FIG. 2 ) has applied one or more visual characteristics. - In some examples, the visual characteristics applied to
UI elements UI elements user 102 as though they have reflective properties, such as the reflective properties of a piece of metal or a shiny piece of plastic or another material. Visualcharacteristic module 16 may also adjust various properties of the visual characteristics based on various criteria. - As one example, visual
characteristic module 16 may receive a brightness measurement of the environment ofcomputing device 2 from a brightness sensor, which may be one of input devices 44 (FIG. 2 ). The brightness measurement may indicate an amount of light detected by the brightness sensor. Based on the brightness measurement, visualcharacteristic module 16 may adjust the intensity of the visual characteristics for each of the user interface elements. As an example, if the detected brightness measurement increases, the intensity of the visual effect, e.g., the reflectivity or shininess may increase or decrease. -
Computing device 2 may also adjust the visual characteristics for each of the UI elements of a GUI based on the movement ofcomputing device 2. As an example, an accelerometer, which may comprise one ofinput devices 44 may detect an indication (e.g., a movement) thatcomputing device 2 has moved, and may generate an indication, such as a signal responsive to detecting the movement indication. Visualcharacteristic module 16 may receive the indication thatcomputing device 2 has moved, and may update the visual characteristics for each of theUI elements characteristic module 16 may update a reflection characteristic, such as a lens flare effect generated by computing device as if a light source were present. That is, visualcharacteristic module 16 may dynamically update a graphical depiction of a lens flare due to movement ofcomputing device 2 that would change the position of the computing device relative to the position of the imaginary light source. - Visual
characteristic module 16 may update the visual characteristics for each of the UI elements 81-85 based on an image captured using a camera ofcomputing device 2. In this example, the camera may capture an image, which in some cases may comprise an image of a user ofcomputing device 2, such asuser 102. Visualcharacteristic module 16 receives the captured image and updates the visual characteristics for each of UI elements 81-85 such that the visual characteristics include at least a portion of the captured image. - In the example of
FIG. 3A ,UID 4outputs GUI 80 for display at presence-sensitive display 20 at time (t1) 90. In accordance with the techniques of this disclosure of generating enhanced visual characteristics based on various factors, application feature module 8 may determine a respective age for each ofUI elements FIGS. 1 and 2 , application feature module 8 may keep a record of the various UI elements of an application, such as any of the applications illustrated inFIGS. 1 , 3, and 4. Application feature module 8 may keep a record of the last time that a user, such asuser 102 has selected each UI element, and may determine age for each UI element that is based on the amount of time that has elapsed since each ofUI elements characteristic module 16 may generate a respective visual characteristic for each of UI elements 81-85, andUID 4 may output the respective visual characteristics for each of the UI elements 81-85. - At a later time (t2) 92,
user 102 may interact withUI element 84 ofGUI 88. Based on the interaction withUI element 84, application feature module 8 may update the time sinceUI element 84 was last selected (e.g., the age of the UI element). Application feature module 8 may cause visualcharacteristic module 16 to update the visual characteristic associated withUI element 84.User interface device 4 may then output the updated visual characteristic forUI element 84. - As examples of how
visual effects module 16 generates the visual characteristics for UI elements 81-85 based on the age of UI elements 81-85,UID 4 may no longer display a visual characteristic forUI element 84 based on the updated age ofUI element 84due user 102 selectingUI element 84 at time (t1) 90. Additionally, because time has elapsed betweentime 92 andtime 90, the age ofUI elements FIG. 3B may have increased. Accordingly, visualcharacteristic module 16 updates the visual characteristics, anduser interface device 4 outputs the updated visual characteristics forUI elements FIGS. 3A and 3B illustrated with UI elements 81-83 represent changed visual characteristics forUI elements -
FIG. 4 is a conceptual diagram that illustrates generating and outputting enhanced visual characteristics associated with a user interface elements that represent one or more contacts, in accordance with one or more aspects of the present disclosure.FIG. 4 illustrates aGUI 400 ofcomputing device 2 that may includeuser interface elements user 102 ofcomputing device 2.UI elements contacts 22. -
GUI 400 may represent, in the example ofFIG. 4 , an application, such as a notifications application that may informuser 102 of various new events, such as communications thatcomputing device 2 has received sinceuser 102 last interacted withcomputing device 2. The events may include missed or received phone calls, receipt of messages, such as text and voice messages, received e-mails, and notifications from applications, such as social networking applications. - A module of
computing device 2, such as one of application modules 17 may determine a classification of one or more of contacts 22 (e.g, a “preferred” status) based on various criteria, such as frequency of communication thatuser 102 has had with a particular contact, the time that has elapsed sinceuser 102 has communicated with a particular contact, a number of messages or conversations with a particular contact, or other communication metrics or preferences ofuser 102. Forgraphical elements user 102, visualcharacteristic module 16 may generate anduser interface device 4 may output a visual characteristic indicative of the status of the contact. - One of application modules 17 may also determine that the contact represented by
UI element 404 represents does not correspond to a preferred contact. Based on that determination, visualcharacteristic module 16 may generate, anduser interface device 4 may output a second, different visual characteristic forUI element 404. In this example,UI elements FIG. 2 ). - In some examples, visual
characteristic module 16 may also generate the visual characteristic for the UI element associated with each contact based on the communication with the contact and information associated with the contact. For example, visualcharacteristic module 16 may generate different visual effects for a contact associated with a social network application, than for a contact in a phone application or a text messaging application. In another example, visualcharacteristic module 16 may generate different visual effects based on the frequency of communication with each contact. - Although
GUI 400 is illustrated as a notification application, the techniques of this disclosure and with respect toFIG. 4 may apply to other applications, as well. Some example applications that may output visual characteristics based on the determination that a UI element is associated with a preferred contact may include a phone dialer application, text or voice messaging application, social networking application, or any other application that may include UI elements that correspond tocontacts 22 ofcomputing device 2. - In general, visual
characteristic module 16 may generate, anduser interface device 4 may output any of the visual characteristics described in this disclosure for any ofUI elements contacts 22. Such visual characteristics may include shining, shimmering, highlight and/or glimmering effects, and visual characteristics that include at least a portion of an image captured by a camera ofcomputing device 2. -
FIG. 5 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure. Graphical content, generally, may include any visual information that may be output for display, such as text, images, a group of moving images, etc. The example shown inFIG. 5 includes acomputing device 500, presence-sensitive display 501,communication unit 510,projector 520,projector screen 522,tablet device 526, and visual display device 130. Although shown for purposes of example inFIGS. 1 and 2 as a stand-alone computing device 500, a computing device may, generally, be any component or system that includes a processor or other suitable computing environment for executing software instructions and, for example, need not include a presence-sensitive display. - As shown in the example of
FIG. 5 ,computing device 500 may be a processor that includes functionality as described with respect toprocessor 40 inFIG. 2 . In such examples,computing device 500 may be operatively coupled to presence-sensitive display 501 by acommunication channel 502A, which may be a system bus or other suitable connection.Computing device 500 may also be operatively coupled tocommunication unit 510, further described below, by acommunication channel 502B, which may also be a system bus or other suitable connection. Although shown separately as an example inFIG. 5 ,computing device 500 may be operatively coupled to presence-sensitive display 501 andcommunication unit 510 by any number of one or more communication channels. - In other examples, such as illustrated previously in
FIGS. 1-2 ,computing device 500 may be a portable or mobile device such as mobile phones (including smart phones), laptop computers, etc. In some examples,computing device 500 may be a desktop computers, tablet computers, smart television platforms, cameras, personal digital assistants (PDAs), servers, mainframes, etc. - Presence-
sensitive display 501, as shown inFIG. 5 , may includedisplay device 503 and presence-sensitive input device 505.Display device 503 may, for example, receive data fromcomputing device 500 and display the graphical content. In some examples, presence-sensitive input device 505 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures, etc.) at presence-sensitive display 501 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input tocomputing device 500 usingcommunication channel 502A. In some examples, presence-sensitive input device 505 may be physically positioned on top ofdisplay device 503 such that, when a user positions an input unit over a graphical element displayed bydisplay device 503, the location at which presence-sensitive input device 505 corresponds to the location ofdisplay device 503 at which the graphical element is displayed. - As shown in
FIG. 5 ,computing device 500 may also include and/or be operatively coupled withcommunication unit 510.Communication unit 510 may include functionality ofcommunication unit 44 as described inFIG. 2 . Examples ofcommunication unit 510 may include a network interface card, an Ethernet card, an optical transceiver, a radio frequency transceiver, or any other type of device that can send and receive information. Other examples of such communication units may include Bluetooth, 3G, and WiFi radios, Universal Serial Bus (USB) interfaces, etc.Computing device 500 may also include and/or be operatively coupled with one or more other devices, e.g., input devices, output devices, memory, storage devices, etc. that are not shown inFIG. 3 for purposes of brevity and illustration. -
FIG. 5 also illustrates aprojector 520 andprojector screen 522. Other such examples of projection devices may include electronic whiteboards, holographic display devices, and any other suitable devices for displaying graphical content.Projector 520 andproject screen 522 may include one or more communication units that enable the respective devices to communicate withcomputing device 500. In some examples, the one or more communication units may enable communication betweenprojector 520 andprojector screen 522.Projector 520 may receive data fromcomputing device 500 that includes graphical content.Projector 520, in response to receiving the data, may project the graphical content ontoprojector screen 522. In some examples,projector 520 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures, etc.) at projector screen using optical recognition or other suitable techniques and send indications of such user input using one or more communication units tocomputing device 500. -
Projector screen 522, in some examples, may include a presence-sensitive display 524. Presence-sensitive display 524 may include a subset of functionality or all of the functionality ofUI device 4 as described in this disclosure. In some examples, presence-sensitive display 524 may include additional functionality. Projector screen 522 (e.g., an electronic whiteboard), may receive data fromcomputing device 500 and display the graphical content. In some examples, presence-sensitive display 524 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures, etc.) atprojector screen 522 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units tocomputing device 500. -
FIG. 5 also illustratestablet device 526 and visual display device 530.Tablet device 526 and visual display device 530 may each include computing and connectivity capabilities. Examples oftablet device 526 may include e-reader devices, convertible notebook devices, hybrid slate devices, etc. Examples of visual display device 530 may include televisions, computer monitors, etc. As shown inFIG. 5 ,tablet device 526 may include a presence-sensitive display 528. Visual display device 530 may include a presence-sensitive display 532. Presence-sensitive displays 528, 532 may include a subset of functionality or all of the functionality ofUI device 4 as described in this disclosure. In some examples, presence-sensitive displays 528, 532 may include additional functionality. In any case, presence-sensitive display 532, for example, may receive data fromcomputing device 500 and display the graphical content. In some examples, presence-sensitive display 532 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures, etc.) at projector screen using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units tocomputing device 500. - As described above, in some examples,
computing device 500 may output graphical content for display at presence-sensitive display 501 that is coupled tocomputing device 500 by a system bus or other suitable communication channel.Computing device 500 may also output graphical content for display at one or more remote devices, such asprojector 520,projector screen 522,tablet device 526, and visual display device 530. For instance,computing device 500 may execute one or more instructions to generate and/or modify graphical content in accordance with techniques of the present disclosure.Computing device 500 may output the data that includes the graphical content to a communication unit ofcomputing device 500, such ascommunication unit 510.Communication unit 510 may send the data to one or more of the remote devices, such asprojector 520,projector screen 522,tablet device 526, and/or visual display device 530. In this way,computing device 500 may output the graphical content for display at one or more of the remote devices. In some examples, one or more of the remote devices may output the graphical content at a presence-sensitive display that is included in and/or operatively coupled to the respective remote devices. - In some examples,
computing device 500 may not output graphical content at presence-sensitive display 501 that is operatively coupled tocomputing device 500. In other examples,computing device 500 may output graphical content for display at both a presence-sensitive display 501 that is coupled tocomputing device 500 bycommunication channel 502A, and at one or more remote devices. In such examples, the graphical content may be displayed substantially contemporaneously at each respective device. For instance, some delay may be introduced by the communication latency to send the data that includes the graphical content to the remote device. In some examples, graphical content generated by computingdevice 500 and output for display at presence-sensitive display 501 may be different than graphical content display output for display at one or more remote devices. -
Computing device 500 may send and receive data using any suitable communication techniques. For example,computing device 500 may be operatively coupled toexternal network 514 usingnetwork link 512A. Each of the remote devices illustrated inFIG. 5 may be operatively coupled to networkexternal network 514 by one ofrespective network links External network 514 may include network hubs, network switches, network routers, etc., that are operatively inter-coupled thereby providing for the exchange of information betweencomputing device 500 and the remote devices illustrated inFIG. 5 . In some examples, network links 512A-512D may be Ethernet, ATM or other network connections. Such connections may be wireless and/or wired connections. - In some examples,
computing device 500 may be operatively coupled to one or more of the remote devices included inFIG. 5 usingdirect device communication 518.Direct device communication 518 may include communications through whichcomputing device 500 sends and receives data directly with a remote device, using wired or wireless communication. That is, in some examples ofdirect device communication 518, data sent by computingdevice 500 may not be forwarded by one or more additional devices before being received at the remote device, and vice-versa. Examples ofdirect device communication 518 may include Bluetooth, Near-Field Communication, Universal Serial Bus, WiFi, infrared, etc. One or more of the remote devices illustrated inFIG. 5 may be operatively coupled withcomputing device 500 bycommunication links 516A-516D. In some examples,communication links 512A-512D may be connections using Bluetooth, Near-Field Communication, Universal Serial Bus, infrared, etc. Such connections may be wireless and/or wired connections. - In accordance with techniques of the disclosure,
computing device 500 may be operatively coupled to visual display device 530 usingexternal network 514.Computing device 500 may output for display at visual display device 530 a graphical user interface of an application including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of application capabilities. Presence-sensitive input device 505 may determine a respective age of each one of the plurality of application capabilities, each respective age being a period of time that has elapsed since that application capability was added to the application.Computing device 500 may generate, based on the respective age of each one of the plurality of application capabilities associated with each of the plurality of elements and by the computing device, a respective visual characteristic for each of the plurality of elements. In some examples, the enhanced visual characteristics may modify at least one of a chrominance and a luminance value associated with the plurality of elements. Presence-sensitive input device 505 may output and for display at the display device, the respective visual characteristic for each one of the plurality of application capabilities associated with each of the plurality of elements in accordance with the techniques of this disclosure. -
FIG. 6 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. The operations ofFIG. 6 may be performed by one or more processors of a computing device,processors 40 ofcomputing device 2 illustrated inFIG. 1 andFIG. 2 . For purposes of illustration,FIG. 5 is described below within the context ofcomputing devices 2 ofFIG. 1 andFIG. 2 orcomputing device 500 ofFIG. 5 . -
Computing device 2 may output for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of application features (600).User interface device 4 may determine an age for a first application feature of the plurality of application features, the age being a period of time that has elapsed since that application feature was enabled in the application (601). Visualcharacteristic module 16 may generate, based on the age, a visual characteristic that modifies at least one of a chrominance or luminance value associated with the first application feature (602).UID 4 may output and for display at the display device, the visual characteristic (603). - In some examples, a camera coupled to the computing device may capture an image, and each the visual effect may include at least a portion of the image. In some examples, the image may comprise an image of a user associated with the computing device.
- The operations of the method illustrated in
FIG. 6 may further include capturing, by a light sensor of computingdevice computing device 2, a brightness measurement of an environment of the computing device, and adjusting, based on the brightness measurement of the environment, an intensity of the visual characteristic. - In some examples, an accelerometer of
computing device 2 may detect an indication that the computing device has moved.Computing device 2 may receive the indication that the computing device has moved, and responsive to receiving the indication that the computing device has moved, may update the visual characteristic based on the movement indication.UID 4 may output the updated visual characteristic. In some examples, the visual characteristic comprises one of a glimmer, shimmer, highlight, and shine. - In another example, generating the visual characteristic may comprise generating an intensity for the visual characteristic from the plurality of elements, and the intensity of each visual characteristic may decrease as the age of the application capability associated with that element increases.
-
FIG. 7 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. The operations ofFIG. 7 may be performed by one or more processors of a computing device,processors 40 ofcomputing device 2 illustrated inFIG. 1 andFIG. 2 orcomputing device 500 ofFIG. 5 . For purposes of illustration,FIG. 7 is described below within the context ofcomputing devices 2 ofFIG. 1 andFIG. 2 . -
Computing device 2 may output for display, a graphical user interface of an application, the graphical user interface including a plurality of elements (700). Application feature module 8 ofcomputing device 2 may determine an age for a first element of the plurality of elements, the age being a period of time that has elapsed since the first element was last selected in the application (701). Visualcharacteristic module 16 may generate, based on the respective age of the first element, a visual characteristic that modifies at least one of a chrominance or luminance value that indicates the age of the first element (702), andcomputing device 2 may output for display, the visual characteristic for the first element (703). - In some examples, a camera coupled to the computing device may capture an image, and the visual effect may include at least a portion of the image. In some examples, the image may comprise an image of a user associated with the computing device.
- The operations of the method illustrated in
FIG. 7 may further include capturing, by a light sensor of computingdevice computing device 2, a brightness measurement of an environment of the computing device, and adjusting, based on the brightness measurement of the environment, an intensity of the respective visual characteristic. - In some examples, an accelerometer of
computing device 2 may detect an indication that the computing device has moved.Computing device 2 may receive the indication that the computing device has moved, and responsive to receiving the indication that the computing device has moved, may update each respective visual characteristic based on the movement indication.UID 4 may output the updated visual characteristic for display at presence-sensitive display 20. In some examples, each respective visual characteristic comprises one of a glimmer, shimmer, highlight, and shine. - In an example,
computing device 2 may generate an intensity for the visual characteristic. The intensity of the respective visual characteristic may decrease as the period of time that has elapsed since that element was last selected increases. -
FIG. 8 is a flow diagram illustrating an example operation of the computing device, in accordance with one or more aspects of the present disclosure. The operations ofFIG. 8 may also be performed by one or more processors of a computing device,processors 40 ofcomputing device 2 illustrated inFIG. 1 andFIG. 2 . For purposes of illustration,FIG. 7 is described below within the context ofcomputing devices 2 ofFIG. 1 andFIG. 2 and/orcomputing device 500 ofFIG. 5 . -
Computing device 2 may receive a communication associated with a contact (800).Computing device 2 may also determine whether the contact is classified as a preferred contact based on a variety of factors, as described above (801). Responsive to determining that the contact is classified as a preferred contact, visualcharacteristic module 16 may generate based on the communication and information associated with the contact, a graphical element having a first visual characteristic, the first visual characteristic being one of a glimmer, shimmer, highlight, and shine (803). Responsive to determining that the contact is not classified as a preferred contact, visualcharacteristic module 16 may generate, based on the communication and information associated with the contact, a graphical element having a second visual characteristic, the second visual characteristic being different from the first visual characteristic (802).Computing device 2 may output for display at presence-sensitive display 20, a graphical user interface including the graphical element. - In some examples, a camera coupled to the computing device may capture an image, and each respective visual effect may include at least a portion of the image. In some examples, the image may comprise an image of a user associated with the computing device.
- The operations of the method illustrated in
FIG. 8 may further include capturing, by a light sensor of computingdevice computing device 2, a brightness measurement of an environment of the computing device, and adjusting, based on the brightness measurement of the environment, an intensity of each respective visual characteristic. - In some examples, an accelerometer of
computing device 2 may detect an indication that the computing device has moved.Computing device 2 may receive the indication that the computing device has moved, and responsive to receiving the indication that the computing device has moved, may update each respective visual characteristic based on the movement indication for output byUID 4. - In one or more examples, the functions described herein may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored on or transmitted over, as one or more instructions or code, a computer-readable medium and executed by a hardware-based processing unit. Computer-readable media may include computer-readable storage media, which corresponds to a tangible medium such as data storage media, or communication media including any medium that facilitates transfer of a computer program from one place to another, e.g., according to a communication protocol. In this manner, computer-readable media generally may correspond to (1) tangible computer-readable storage media, which is non-transitory or (2) a communication medium such as a signal or carrier wave. Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for implementation of the techniques described in this disclosure. A computer program product may include a computer-readable medium.
- By way of example, and not limitation, such computer-readable storage media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage, or other magnetic storage devices, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer. Also, any connection is properly termed a computer-readable medium. For example, if instructions are transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of medium. It should be understood, however, that computer-readable storage media and data storage media do not include connections, carrier waves, signals, or other transient media, but are instead directed to non-transient, tangible storage media. Disk and disc, as used herein, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc, where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
- Instructions may be executed by one or more processors, such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor,” as used herein may refer to any of the foregoing structure or any other structure suitable for implementation of the techniques described herein. In addition, in some aspects, the functionality described herein may be provided within dedicated hardware and/or software modules. Also, the techniques could be fully implemented in one or more circuits or logic elements.
- The techniques of this disclosure may be implemented in a wide variety of devices or apparatuses, including a wireless handset, an integrated circuit (IC) or a set of ICs (e.g., a chip set). Various components, modules, or units are described in this disclosure to emphasize functional aspects of devices configured to perform the disclosed techniques, but do not necessarily require realization by different hardware units. Rather, as described above, various units may be combined in a hardware unit or provided by a collection of interoperative hardware units, including one or more processors as described above, in conjunction with suitable software and/or firmware.
- Various examples have been described. These and other examples are within the scope of the following claims.
Claims (20)
1. A method comprising:
outputting, by a computing device and for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with a respective feature of a plurality of features of the application;
determining, by the computing device, an age for a first feature of the plurality of features, the age being a period of time that has elapsed since the first feature was enabled in the application;
generating, by the computing device, based on the age, a visual characteristic that modifies a visual appearance of the respective element of the plurality of elements associated with the first feature; and
outputting, by the computing device and for display, the visual characteristic in the graphical user interface.
2. The method of claim 1 , wherein the visual characteristic includes at least a portion of an image.
3. The method of claim 2 , wherein the image represents a user associated with the computing device.
4. The method of claim 1 , further comprising:
capturing, by a light sensor of the computing device, a brightness measurement of an environment of the computing device; and
adjusting, by the computing device and based on the brightness measurement of the environment, an aspect of the visual characteristic.
5. The method of claim 1 , further comprising:
receiving, by the computing device, an indication that the computing device has moved;
responsive to receiving the indication that the computing device has moved, updating the visual characteristic; and
outputting the updated visual characteristic.
6. The method of claim 1 , wherein generating the visual characteristic comprises:
determining an intensity for the visual characteristic, the intensity of the visual characteristic decreasing as the age of the feature associated with the element increases.
7. A method comprising:
outputting, by a computing device for display, a graphical user interface of an application, the graphical user interface including a plurality of elements;
determining, by the computing device, an age for a first element of the plurality of elements, the age being a period of time that has elapsed since the first element was last selected in the application;
generating, by the computing device, based on the age of the first element, a visual characteristic that indicates the age of the first element; and
outputting, by the computing device for display, the visual characteristic for the first element.
8. The method of claim 7 , further comprising:
generating, by the computing device, an updated visual characteristic including at least a portion of an image.
9. The method of claim 8 , wherein the image comprises an image of a user of the computing device.
10. The method of claim 7 , further comprising:
capturing, by a light sensor of the computing device, a brightness measurement of an environment of the computing device; and
adjusting an aspect of the visual characteristic based on the brightness measurement of the environment of the computing device.
11. The method of claim 7 , further comprising:
receiving, by the computing device, an indication that the computing device has moved; and
responsive to receiving the indication, updating the visual characteristic.
12. The method of claim 7 , further comprising:
determining an intensity for the visual characteristic, wherein the visual characteristic decreases as the period of time that has elapsed since the first element was last selected increases.
13. A device comprising:
at least one processor;
at least one module operable by the at least one processor to:
output, for display, a graphical user interface of an application, the graphical user interface including a plurality of elements, each respective element of the plurality of elements being associated with one of a plurality of features;
determine an age for a first feature of the plurality of application features, the age being a period of time that has elapsed since the first feature was enabled in the application;
generate, based on the age, a visual characteristic that modifies a visual appearance of the respective element of the plurality of elements associated with the first feature; and
output, for display, the visual characteristic in the graphical user interface.
14. The device of claim 13 , wherein each respective visual characteristic includes at least a portion of an image.
15. The device of claim 14 , wherein the image comprises an image of a user associated with the computing device.
16. The device of claim 13 , further comprising:
a light sensor to capture a brightness measurement of an environment of the computing device,
wherein the at least one module is further operable by the at least one processor to adjust, based on the brightness measurement of the environment, an aspect of the visual characteristic.
17. The device of claim 13 , further comprising:
an accelerometer to detect an indication that the computing device has moved,
wherein the at least one module is further operable by the at least one processor to:
receive the indication that the computing device has moved;
responsive to receiving the indication that the computing device has moved, update the visual characteristic; and
output the updated visual characteristic.
18. The device of claim 13 , wherein the at least one module is further operable by the at least one processor to determine an intensity for the visual characteristic from the plurality of elements, the intensity of each visual characteristic decreasing as the age of the feature associated with the element increases.
19. A device comprising:
at least one processor;
at least one operable by the at least one processor to:
output, for display at a display device, a graphical user interface of an application, the graphical user interface including a plurality of elements;
determine an age for a first element of the plurality of elements, the age being a period of time that has elapsed since the first element was last selected in the application;
generate, based on the age, a visual characteristic that indicates the age of the first element; and
output, for display, the visual characteristic for the first element in the graphical user interface.
20. The device of claim 19 , further comprising:
an accelerometer that detects movement of the computing device,
wherein the user interface module is further operable by the at least one processor to:
receive an indication that the computing device has moved; and
responsive to receiving the indication, update the visual characteristic.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/189,893 US20140176593A1 (en) | 2012-10-16 | 2014-02-25 | Mobile device user interface having enhanced visual characteristics |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261714680P | 2012-10-16 | 2012-10-16 | |
US201361789029P | 2013-03-15 | 2013-03-15 | |
US13/854,806 US8698835B1 (en) | 2012-10-16 | 2013-04-01 | Mobile device user interface having enhanced visual characteristics |
US14/189,893 US20140176593A1 (en) | 2012-10-16 | 2014-02-25 | Mobile device user interface having enhanced visual characteristics |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/854,806 Continuation US8698835B1 (en) | 2012-10-16 | 2013-04-01 | Mobile device user interface having enhanced visual characteristics |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140176593A1 true US20140176593A1 (en) | 2014-06-26 |
Family
ID=50441450
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/854,806 Expired - Fee Related US8698835B1 (en) | 2012-10-16 | 2013-04-01 | Mobile device user interface having enhanced visual characteristics |
US14/189,893 Abandoned US20140176593A1 (en) | 2012-10-16 | 2014-02-25 | Mobile device user interface having enhanced visual characteristics |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/854,806 Expired - Fee Related US8698835B1 (en) | 2012-10-16 | 2013-04-01 | Mobile device user interface having enhanced visual characteristics |
Country Status (2)
Country | Link |
---|---|
US (2) | US8698835B1 (en) |
WO (1) | WO2014062680A2 (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10503357B2 (en) * | 2014-04-03 | 2019-12-10 | Oath Inc. | Systems and methods for delivering task-oriented content using a desktop widget |
US20160004393A1 (en) * | 2014-07-01 | 2016-01-07 | Google Inc. | Wearable device user interface control |
US10425372B2 (en) * | 2015-08-18 | 2019-09-24 | Google Llc | Notification targeting based on downstream user engagement |
US10627988B2 (en) * | 2016-01-14 | 2020-04-21 | Keithley Instruments, Llc | Visually identifiable features for applications installed on electronic devices |
CN109145231A (en) * | 2017-06-27 | 2019-01-04 | 阿里巴巴集团控股有限公司 | A kind of generation method, device and the electronic equipment of interface element identifier |
US11474819B1 (en) * | 2021-05-21 | 2022-10-18 | Digital.Ai Software, Inc. | Unified view, analytics, and auditability for software development processes spanning distinct software development applications |
Family Cites Families (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5564004A (en) * | 1994-04-13 | 1996-10-08 | International Business Machines Corporation | Method and system for facilitating the selection of icons |
EP1431864B2 (en) * | 1995-02-13 | 2012-08-22 | Intertrust Technologies Corporation | Systems and methods for secure transaction management and electronic rights protection |
US6948070B1 (en) * | 1995-02-13 | 2005-09-20 | Intertrust Technologies Corporation | Systems and methods for secure transaction management and electronic rights protection |
US7256770B2 (en) * | 1998-09-14 | 2007-08-14 | Microsoft Corporation | Method for displaying information responsive to sensing a physical presence proximate to a computer input device |
US7215813B2 (en) * | 2001-12-03 | 2007-05-08 | Apple Computer, Inc. | Method and apparatus for color correction |
US20080177994A1 (en) * | 2003-01-12 | 2008-07-24 | Yaron Mayer | System and method for improving the efficiency, comfort, and/or reliability in Operating Systems, such as for example Windows |
US20060047704A1 (en) * | 2004-08-31 | 2006-03-02 | Kumar Chitra Gopalakrishnan | Method and system for providing information services relevant to visual imagery |
US20060218191A1 (en) * | 2004-08-31 | 2006-09-28 | Gopalakrishnan Kumar C | Method and System for Managing Multimedia Documents |
US20060097978A1 (en) * | 2004-10-22 | 2006-05-11 | Ng Kee Y | Field-sequential color display with feedback control |
US8677274B2 (en) * | 2004-11-10 | 2014-03-18 | Apple Inc. | Highlighting items for search results |
US7512904B2 (en) * | 2005-03-22 | 2009-03-31 | Microsoft Corporation | Operating system launch menu program listing |
US7440006B2 (en) * | 2005-04-05 | 2008-10-21 | Realtek Semiconductor Corp. | System for gracefully aging inactive areas of a video display |
KR100800995B1 (en) * | 2005-07-11 | 2008-02-05 | 삼성전자주식회사 | Apparatus and method for displaying icon |
JP4834729B2 (en) * | 2005-07-22 | 2011-12-14 | カンガルー メディア インコーポレイテッド | Systems and methods for promoting the spectator experience of live sporting events |
US8947297B2 (en) * | 2006-01-30 | 2015-02-03 | The Invention Science Fund I, Llc | Positional display elements |
US8139026B2 (en) * | 2006-08-02 | 2012-03-20 | Research In Motion Limited | System and method for adjusting presentation of text and images on an electronic device according to an orientation of the device |
US7844915B2 (en) * | 2007-01-07 | 2010-11-30 | Apple Inc. | Application programming interfaces for scrolling operations |
US20080263449A1 (en) * | 2007-04-20 | 2008-10-23 | Microsoft Corporation | Automated maintenance of pooled media content |
US8234564B2 (en) * | 2008-03-04 | 2012-07-31 | Apple Inc. | Transforms and animations of web-based content |
US8819172B2 (en) * | 2010-11-04 | 2014-08-26 | Digimarc Corporation | Smartphone-based methods and systems |
US8121618B2 (en) * | 2009-10-28 | 2012-02-21 | Digimarc Corporation | Intuitive computing methods and systems |
US9197736B2 (en) * | 2009-12-31 | 2015-11-24 | Digimarc Corporation | Intuitive computing methods and systems |
US20110113371A1 (en) * | 2009-11-06 | 2011-05-12 | Robert Preston Parker | Touch-Based User Interface User Error Handling |
US20120154633A1 (en) * | 2009-12-04 | 2012-06-21 | Rodriguez Tony F | Linked Data Methods and Systems |
US8839150B2 (en) * | 2010-02-10 | 2014-09-16 | Apple Inc. | Graphical objects that respond to touch or motion input |
US20120206485A1 (en) * | 2010-02-28 | 2012-08-16 | Osterhout Group, Inc. | Ar glasses with event and sensor triggered user movement control of ar eyepiece facilities |
US9129295B2 (en) * | 2010-02-28 | 2015-09-08 | Microsoft Technology Licensing, Llc | See-through near-eye display glasses with a fast response photochromic film system for quick transition from dark to clear |
JP2013521576A (en) * | 2010-02-28 | 2013-06-10 | オスターハウト グループ インコーポレイテッド | Local advertising content on interactive head-mounted eyepieces |
US8589815B2 (en) * | 2010-03-10 | 2013-11-19 | Microsoft Corporation | Control of timing for animations in dynamic icons |
US8660355B2 (en) * | 2010-03-19 | 2014-02-25 | Digimarc Corporation | Methods and systems for determining image processing operations relevant to particular imagery |
US20110242142A1 (en) * | 2010-03-30 | 2011-10-06 | Ati Technologies Ulc | Multiple display chrominance and luminance method and apparatus |
CN102835124B (en) * | 2010-04-09 | 2016-04-13 | Lg电子株式会社 | Image display and the method for operating image display |
US20120254804A1 (en) * | 2010-05-21 | 2012-10-04 | Sheha Michael A | Personal wireless navigation system |
US8913056B2 (en) * | 2010-08-04 | 2014-12-16 | Apple Inc. | Three dimensional user interface effects on a display by using properties of motion |
US20120166946A1 (en) * | 2010-12-22 | 2012-06-28 | Jens Bombolowsky | Dynamic handling of instructional feedback elements based on usage statistics |
WO2012155144A1 (en) * | 2011-05-12 | 2012-11-15 | John Devecka | An interactive mobile-optimized icon-based profile display and associated social network functionality |
US10699063B2 (en) * | 2011-08-19 | 2020-06-30 | Apple Inc. | Authoring content for digital books |
US20130111345A1 (en) * | 2011-10-31 | 2013-05-02 | Nokia Corporation | Portable electronic device, associated apparatus and methods |
WO2013075071A1 (en) * | 2011-11-18 | 2013-05-23 | Ayman Hammad | Mobile wallet store and service injection platform apparatuses, methods and systems |
-
2013
- 2013-04-01 US US13/854,806 patent/US8698835B1/en not_active Expired - Fee Related
- 2013-10-15 WO PCT/US2013/065054 patent/WO2014062680A2/en active Application Filing
-
2014
- 2014-02-25 US US14/189,893 patent/US20140176593A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
WO2014062680A3 (en) | 2015-01-29 |
US20140104296A1 (en) | 2014-04-17 |
WO2014062680A2 (en) | 2014-04-24 |
US8698835B1 (en) | 2014-04-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10127089B2 (en) | Dynamic deep links to targets | |
US9049261B2 (en) | Prioritized management and presentation of notifications | |
EP3545410B1 (en) | Application program interface for managing complication data | |
US8473871B1 (en) | Multiple seesawing panels | |
US10353661B2 (en) | Method for sharing screen between devices and device using the same | |
US8756533B2 (en) | Multiple seesawing panels | |
US9100357B2 (en) | Notification classification and display | |
US20180288739A1 (en) | Electronic device and method of providing notification by electronic device | |
US9037455B1 (en) | Limiting notification interruptions | |
US8698835B1 (en) | Mobile device user interface having enhanced visual characteristics | |
US20160147406A1 (en) | Method for providing graphical user interface and electronic device for supporting the same | |
US20180188906A1 (en) | Dynamically generating a subset of actions | |
KR20160141838A (en) | Expandable application representation | |
US8601561B1 (en) | Interactive overlay to prevent unintentional inputs | |
US10992779B2 (en) | Limiting alerts on a computing device | |
US9335452B2 (en) | System and method for capturing images | |
US20160350136A1 (en) | Assist layer with automated extraction | |
US10938767B2 (en) | Outputting reengagement alerts by a computing device | |
US9065872B1 (en) | Sharing service management | |
US20140049492A1 (en) | Electronic device and method for changing an object according to a bending state | |
US20190369827A1 (en) | Remote data input framework | |
US20170371535A1 (en) | Device, method and graphic user interface used to move application interface element | |
US20180136789A1 (en) | Sender-initiated control of information display within multiple-partition user interface | |
US20200019305A1 (en) | Method for altering display ratio of application, and electronic device that realises same | |
US10241634B2 (en) | Method and apparatus for processing email in electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: GOOGLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SHIPLACOFF, DANIEL MARC GATAN;REEL/FRAME:032296/0964 Effective date: 20130329 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |
|
AS | Assignment |
Owner name: GOOGLE LLC, CALIFORNIA Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044144/0001 Effective date: 20170929 |