US20130019204A1 - Adjusting content attributes through actions on context based menu - Google Patents
Adjusting content attributes through actions on context based menu Download PDFInfo
- Publication number
- US20130019204A1 US20130019204A1 US13/542,962 US201213542962A US2013019204A1 US 20130019204 A1 US20130019204 A1 US 20130019204A1 US 201213542962 A US201213542962 A US 201213542962A US 2013019204 A1 US2013019204 A1 US 2013019204A1
- Authority
- US
- United States
- Prior art keywords
- content
- input
- context based
- attribute
- slider
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04812—Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0489—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using dedicated keyboard keys or combinations thereof
- G06F3/04895—Guidance during keyboard input operation, e.g. prompting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
- G06F9/453—Help systems
Definitions
- font size/color/style for textual content is provided in preset values listed in a drop-down or similar menu in conventional systems, which may not be a natural way for a user to interact using touch or gestures.
- graphics such as lines may have various attributes (thickness, corner styles, color, dash style, etc.) that need to be presented to a user in an intuitive, user-friendly manner for the user to be able to interact through new input technologies.
- Embodiments are directed to adjusting content attributes through a context based menu.
- Context based menus may be deployed for variety of scenarios for content management.
- An application according to embodiments may present the context based menu upon selection of content or upon activation of a context based menu control.
- Content attributes such as font size, style, graphics line weight, style, and similar ones may be presented in an intuitive manner such that selection can be made easily through non-traditional input methods such as touch or gesture. Some attributes may be presented with visual examples.
- submenus may be activated for further detailed selection of available commands through the selection of items on a top level context based menu.
- selection of attribute values on a linear or otherwise spectrum may be enabled through a sliding action that allows continuous value selection as opposed to distinct values presented in traditional menus.
- FIGS. 1A and 1B illustrate some example devices, where context based font or attribute menus may be employed
- FIG. 2 illustrates an example context based menu with controls to adjust an attribute of content according to embodiments
- FIG. 3 illustrates example scenarios for using context based menu according to embodiments
- FIG. 4 illustrates other example scenarios for using context based menu according to embodiments
- FIG. 5 illustrates examples of context based menus for adjusting font types and styles according to embodiments
- FIG. 6 illustrates an example of a context based menu for adjusting a line thickness attribute of graphical content according to embodiments
- FIG. 7 is a networked environment, where a system according to embodiments may be implemented.
- FIG. 8 is a block diagram of an example computing operating environment, where embodiments may be implemented.
- FIG. 9 illustrates a logic flow diagram for a process of adjusting content attributes through a context based menu in touch and gesture enabled devices according to embodiments.
- a user interface of an application executing on a device may present a context based menu in relation to displayed content.
- the context based menu may provide controls to adjust the font size of the content.
- the application may detect a user action associated with the context based menu.
- the application may execute a command to adjust the font size based on the detected user action, which may include touch, gesture, keyboard entry, mouse click, and/or pen input.
- Content attributes as used herein may be any aspect of content that can be represented along a scale either discretely or continuously. Continuous representation can be interacted with by moving a finger (or equivalent gesture or similar input) across a circle.
- Discrete options may be interacted with by selecting a range of points along a line, for example, by snapping one's finger to a range of points along the line.
- a non-exhaustive, illustrative set of examples for content attributes may include an amount of space between lines, a number of columns, a size of margins, an amount of indentation, a font kerning, a zoom action, and comparable ones.
- program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types.
- embodiments may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and comparable computing devices.
- Embodiments may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network.
- program modules may be located in both local and remote memory storage devices.
- Embodiments may be implemented as a computer-implemented process (method), a computing system, or as an article of manufacture, such as a computer program product or computer readable media.
- the computer program product may be a computer storage medium readable by a computer system and encoding a computer program that comprises instructions for causing a computer or computing system to perform example process(es).
- the computer-readable storage medium is a computer-readable memory device.
- the computer-readable storage medium can for example be implemented via one or more of a volatile computer memory, a non-volatile memory, a hard drive, a flash drive, a floppy disk, or a compact disk, and comparable media.
- a user interface of a touch-enabled or gesture-enabled device may employ context based menus to adjust content attributes such as font size, font style, font boldness, graphics style, color, thickness, a line color, a line type, a corner shape, a shading attribute, a fill color, a spacing, a zoom action on displayed content, and a fill type, etc.
- a context based menu may make use of features specific to touch or gesture enabled computing devices, but may also work with a traditional mouse and keyboard.
- Context based menus in general, may be used to provide quick access to commonly used commands while viewing or editing displayed content such as documents, emails, contact lists, other communications, or any content (e.g., audio, video, etc.).
- Context based menus may appear as part of a user interface's regular menu, in a separate viewing pane (e.g., a window) outside or inside the user interface, and so on.
- context based menus present a limited set of commands for easy user access, but additional submenus may be presented upon user selection.
- Commonly used context based menus may appear over the viewed document.
- a tap or swipe action as used herein may be provided by a user through a finger, a pen, a mouse, or similar device, as well as through predefined keyboard entry combinations or a voice command.
- FIGS. 1A and 1B illustrate some example devices, where a context based menus may be employed.
- touch and gesture based technologies are proliferating and computing devices employing those technologies are becoming common, user interface arrangement becomes a challenge.
- Touch and/or gesture enabled devices specifically portable devices, tend to have smaller screen size, which means less available space for user interfaces.
- a virtual keyboard may have to be displayed further limiting the available space (“real estate”).
- real estate real estate
- Embodiments are directed to a context based font or attribute menus to adjust content attributes such as font size.
- a context based menu may be provided dynamically based on presented content and available space while providing ease of use without usurping much needed display area.
- a context based font or attribute menu may be provided according to embodiments.
- a context based menu or a context based attribute menu may be an embodiment of a context based menu.
- Embodiments may be implemented in touch and/or gesture enabled devices or others with keyboard/mouse/pen input, with varying form factors and capabilities.
- Device 104 in FIG. 1A is an example of a large size display device, where a user interface may be provided on screen 106 . Functionality of various applications may be controlled through hardware controls 108 and/or soft controls such as a context based menu displayed on screen 106 . A user may be enabled to interact with the user interface through touch actions or gestures (detected by a video capture device). A launcher indicator may be presented at a fixed location or at a dynamically adjustable location for the user to activate the context based menu. Examples of device 104 may include public information display units, large size computer monitors, and so on.
- Device 112 in FIG. 1A is an example for use of a context based menu to control functionality.
- a user interface may be displayed on a screen or projected on a surface and actions of user 110 may be detected as gestures through video capture device 114 .
- the user's gestures may activate a context based menu to adjust an attribute of a content displayed on the device 112 .
- FIG. 1B includes several example devices such as touch enabled computer monitor 116 , laptop computer 118 , handheld computer 124 , smart phone 126 , tablet computer (or slate) 128 , and mobile computing device 132 , which may be used for computing, communication, control, measurement, and a number of other purposes.
- the example devices in FIG. 1B are shown with touch activation 120 .
- any of these and other example devices may also employ gesture enabled activation of context based menus to adjust an attribute of content such as font size.
- tools such as pen 130 may be used to provide touch input.
- a context based menu may be controlled also through conventional methods such as a mouse input or input through a keyboard 122 .
- FIG. 2 illustrates an example context based menu with controls to adjust an attribute of content.
- Diagram 200 displays a context based menu 202 with control items to adjust content attributes such as font size, style, color, type, etc.
- a user interface of an application may display context based menu 202 with control item 218 to access a prior context based menu.
- the prior context based menu may be a higher level menu providing top level functionality such as copy and paste.
- Control item 218 may also be used to change the context based menu 202 to a collapsed state (i.e., visually minimized).
- the context based menu 202 may have a font size control item 204 to access another context based menu such as menu 302 in FIG. 3 to adjust the font size of textual content.
- the control item(s) may display an indicator to show the current font type, style, and size of selected textual content.
- a contemplated but not yet applied font size, style, type, etc. may also be displayed on the control items (e.g., when a user hovers on a font size but does not click or equivalent touch/gesture actions).
- Another control item 206 may enable a user to underline the selected textual content.
- a control item 208 may enable the user to access a context based menu to alter content color. Examples may include a context based menu showing a color gradient to enable a user to pick a color for the content.
- the context based menu 202 may provide a control item 210 to make the font of the selected content “bold.” Similarly, the context based menu 202 may enable a user to italicize the selected content using control item 212 . Furthermore, the context based menu 202 may enable a user to adjust the font style using control item 214 .
- the control item 214 may be assigned to a single style. Alternatively, the control item 214 may launch a context based submenu to adjust the style of the content such as the menu 516 in FIG. 5 .
- control items may be a one command control (i.e., the underlying command is executed upon selection of the control and the context based menu may be collapsed or hidden following the execution) or associated with one or more submenus.
- selection of the control item 206 for underlining textual content may result in deployment of a submenu that presents options for underline style (e.g. dashed, straight), underline thickness, underline color, etc.
- the associated submenus may be launched through a selection of the particular control (e.g., tapping on the control or swiping from the center toward the control) or through an action (e.g., tap) on an activation area along the perimeter of the context based menu next to each control (the white arrows in diagram 200 ).
- a selection of the particular control e.g., tapping on the control or swiping from the center toward the control
- an action e.g., tap
- FIG. 3 illustrates example scenarios for using context based menu according to embodiments.
- Diagram 300 displays example context based menu 302 executing a variety of user actions.
- a user action may be a touch action, a gesture action, a keyboard input, a mouse input, pen input, or similar action selecting a particular control item on the context based menu.
- the user action may be a swipe over a portion of the context based menu 302 .
- a user may swipe 304 a finger over a slider 308 to adjust the font size of the textual content in a continuous manner (i.e., even though each available font size is not listed on the menu, the swipe-based slider selection may enable the user to select unlisted font sizes easily).
- the application may display the slider 308 centered on a center command button 318 of the context based menu 302 .
- the application may position the slider 308 at a start position on a current font size value of the managed content.
- An example may include the font size of selected content that a user may wish to adjust.
- the application may move the slider clockwise subsequent to detecting the clockwise swipe action 304 (or alternatively through another clockwise swipe action 306 ) starting from a start position of the slider 308 .
- the application may increase the font size according to the clockwise swipe, up to a maximum font size.
- Maximum font size may be displayed on the context based menu 302 and may be determined according to a system setting or may be user adjustable according to application privileges.
- An example of a font size increase may include moving the slider 308 from a starting position at font size 10 to font size 20 in reaction to matching user action. Subsequent to the user action, the application may change the font size of the content from 10 to 20.
- the application may move the slider 308 counterclockwise subsequent to detecting a counterclockwise swipe from the start position.
- the application may decrease the font size according to the counterclockwise swipe, to a minimum font size.
- An example may include moving the slider 308 from a starting position at font size 20 to font size 10 in reaction to a matching user action. Subsequent to the user action, the application may change the font size of the content from 20 to 10.
- the application may display an indicator 310 on the context based menu 302 of a previous font size (e.g., font size 13).
- the indicator 310 may be presented as a bold, an italicized, an underlined, a highlighted, or an enclosed font size value or just as another number of the menu that does not appear at one of the displayed axes of the menu.
- the enclosed font size value may be a circled font size value, a font size value with a square around it, or any other shape to differentiate the previous font size value from others on the context based menu 302 .
- the application may integrate font styles into the context based menu for managing font sizes.
- the application may place a font style action 312 between displayed font size values.
- An example may include adjusting the content to a selected font style such as “heading 1” subsequent to detecting a user action 312 between displayed font size values on the menu 302 .
- the application may detect a user action moving clockwise past a maximum font size value and ending in a dead zone 314 .
- the user action may end between the maximum font size value and the minimum font size value.
- the application may position the slider 308 on the maximum font size value subsequent to the user action ending in the dead zone 314 .
- the application may detect a user action moving counterclockwise past a minimum font size value and ending in the dead zone 314 .
- the user action may end between the minimum font size value and the maximum font size value.
- the application may position the slider 308 on the minimum font size value subsequent to the user action ending in the dead zone 314 .
- Another aspect of the dead zone is that it may allow a user to continue to move their finger clockwise or counterclockwise but not have their font change immediately from the largest to the smallest font. Instead it may create a buffer within which the font does not actually change until one reaches the other edge of the buffer zone. For example, if the user moves counterclockwise past the minimum size it may stay at the minimum through the dead zone but once the user passes the maximum size it may switch to the maximum size (and then start shrinking again, etc.).
- the application may display font state 320 including present font colors, font type, and font shape on the font size values. Additionally, the application may detect a tap action 316 as the user action on a font size value. Subsequent to the user action on the font size value, the application may position the slider 308 on the font size value and adjust the font size of the content according to the font size value associated with tap action 316 .
- FIG. 4 illustrates other example scenarios for using a context based menu to adjust font attributes according to embodiments.
- Diagram 400 displays example context based menu 402 executing a variety of user actions.
- the application may display a tooltip 410 subsequent to a user action 404 such as increasing the font size.
- the application may show a tooltip 414 subsequent to user action 406 adjusting the font size.
- the user action 406 may provide more granularity compared to user action 404 due to larger arc available to a user to pick a font size according to distance of the user actions 404 and 406 from the center of the menu 402 .
- the relative position of the user action with respect to a center of the menu may be taken into consideration enabling increased granularity especially in slider style menu actions.
- the application may calculate the font size according to a polynomial equation to determine the font size for the user action ending in between font size values.
- a polynomial equation may include a user moving the slider in between font size 36 and 72.
- the polynomial curve may help determine what font size to pick if the user is 1 ⁇ 3 of the way in between, halfway in between, etc., so that it feels like a natural increase.
- the application may determine the font size according to the polynomial calculation determining where the slider 408 stops in relation to both font size increments.
- the application may select the font size value closest to an end position of the user action 412 subsequent to detecting the end position outside an action area of the font size values.
- An action area may be an oval area determined by system settings centered around a control such as a font size value.
- the application may adjust the length of the slider 408 to position the slider on the end position of the user action 412 . Adjusting the slider 408 length may enable the user to more accurately view a selected font size value to the corresponding user action.
- Tooltips 410 and 414 may be helpful when the user selection is between presented font size values displaying to the user what the selected font looks like.
- a speed of the slider may be adjusted dynamically to provide more granularity in selecting the font size (or comparable attribute).
- An example may be a slower moving slider 408 to enable the user to pick a desired font size increment.
- “snap points” may be used along the slider at each of the displayed font sizes. As the user moves by sliding/dragging closer to those snap points the line may be automatically moved to the next size, making it efficient to get those sizes.
- a back-off logic may also be employed. For example, if the user then goes in the reverse direction, the stickiness may be removed to that point and allow the user to select a point very close but not exactly on any of the axes (e.g., if the user's finger was very close to 72 , they would still be able to select font 71 ).
- FIG. 5 illustrates examples of context based menus for adjusting font types and styles according to embodiments.
- Diagram 500 displays context based menu 502 for adjusting a font type and context based menu 516 for adjusting a font style.
- a user interface of an application supporting context based menus may present a context based menu 502 to enable adjustment of a font type of textual content using a slider 506 .
- a user may move a slider 506 with a swipe action 504 moving clockwise or counterclockwise on the context based menu 502 selecting among available font types.
- the different font types may be displayed on the context based menu 502 (e.g., along the perimeter).
- a list of available font types 510 may be presented adjacent to the context based menu 502 . Alternatively, the list may be presented in the middle of the radial menu with a dial around the edge.
- font types that correspond to the slider's current position may be indicated on the list through a graphical, color, textual, shading, or similar scheme.
- the slider 506 may correspond to a position for “Arial” font. This may be emphasized on the list of available font types 510 through highlighting 512 .
- the available font types on the list may be displayed using the respective fonts to provide the user further visual assistance or in a single font (for example, in small display areas, where the size of the menu may make reading different font types relatively difficult).
- the context based menu 502 may display an indicator 508 showing the current font type of the content.
- the indicator 508 may also show other state information of the content such as font size and style.
- the context based menu 502 may provide a access button 514 to collapse the context based menu 502 and/or launch a prior context based menu such as top level menu to provide other functionality to manage the content.
- Context based menu 516 may provide a user with similar functionality as discussed above in conjunction with context based menu 502 to adjust a font style of textual content.
- the menu 516 may display indicator 518 to show the current font style of the selected content.
- the list of available styles 520 and the highlighter 522 may be used to present the styles in an easily visible manner as the user may be moving their finger over the menu (thereby, moving the slider).
- the user may be enabled to scroll through the list of available styles 520 through a swipe action, for example, and select a font style by tapping on it on the list.
- Context based menu may be a submenu under a top level menu for adjusting font attributes and it may include additional submenus depending on the number of styles, for example.
- a predefined number of styles may be presented to prevent overlapping problems.
- a slider and a separate list of available items
- higher granularity and, thereby, larger number of available items may be presented for selection through a context based menu.
- FIG. 6 illustrates an example of a context based attribute menu adjusting an attribute of graphical content according to embodiments.
- Diagram 600 displays context based attribute menu 602 to adjust line weights of a selected content such as a drawing.
- a user interface may display menu 602 to adjust an attribute of content such as line weights.
- the menu 602 may display line weight values 614 corresponding to line weights illustrated by list 604 .
- a line thickness (or the attribute in general) may also be displayed on the slider itself. For example, where the slider is at 1 ⁇ 4 it may be a 1 ⁇ 4 thickness and then the slider may become thicker as it goes around until it is at its thickest at 6. And alternatively, the thicknesses (or the attribute) may be relative and not exact. Thus, it could be just that the thickness increases but does not match exactly the numbers.
- a user action 612 may move a slider 610 through swipe and tap actions as described in context based menus selecting a line weight.
- the application may adjust the line weight of the graphical content (any shape that includes lines) according to the user action.
- the menu 602 may display the selected line weight through a highlighter 606 similar to context based menus described above.
- the application may detect a swipe action as the user action starting from a center control button 608 and moving away from the context based attribute menu 602 .
- the application may determine an attribute value 614 intersecting the swipe action. Subsequently, the application may adjust the attribute according to the attribute value 614 .
- the application may display an attribute increase control at an end of the slider 610 to increase the attribute beyond the displayed end value.
- the application may display an attribute decrease control at the other end of the slider 610 to decrease the attribute beyond the displayed end value.
- a context based attribute menu may be used to adjust any attribute of textual or graphical content. Examples may include line weights, corner styles, gradients used to pick a color, shading schemes for graphical objects, and comparable ones.
- the application may display incremental values, discrete attributes, or continuous selection of values on the attribute menu. Displayed values may be selected according to linear, logarithmic, or other scales. The displayed values may also be user selectable according to application privileges, user preferences, and/or customization.
- FIGS. 1 through 6 The example commands, links, submenus, configurations, and context based font and attribute menus depicted in FIGS. 1 through 6 are provided for illustration purposes only. Embodiments are not limited to the shapes, forms, and content shown in the example diagrams, and may be implemented using other textual, graphical, and similar schemes employing the principles described herein.
- FIG. 7 is a networked environment, where a system according to embodiments may be implemented.
- a context based font/attribute menu for touch and/or gesture enabled devices may be also be employed in conjunction with hosted applications and services that may be implemented via software executed over one or more servers 706 or individual server 708 .
- a hosted service or application may communicate with client applications on individual computing devices such as a handheld computer 701 , a desktop computer 702 , a laptop computer 703 , a smart phone 704 , a tablet computer (or slate), 705 (‘client devices’) through network(s) 710 and control a user interface presented to users.
- a context based menu may be used for adjusting an attribute (such as font size) of the content provided by the hosted service or application.
- an attribute such as font size
- a browser application, a word processing application, a spreadsheet application, a calendar application, a note taking application, a graphics application, and comparable ones may make use of a context based menu according to embodiments.
- the context based menu may be activated through a variety of user actions such as selection of content, activation of a launcher indicator, detection of a predetermined touch or gesture action, etc.
- Client devices 701 - 705 are used to access the functionality provided by the hosted service or application.
- One or more of the servers 706 or server 708 may be used to provide a variety of services as discussed above.
- Relevant data may be stored in one or more data stores (e.g. data store 714 ), which may be managed by any one of the servers 706 or by database server 712 .
- Network(s) 710 may comprise any topology of servers, clients, Internet service providers, and communication media.
- a system according to embodiments may have a static or dynamic topology.
- Network(s) 710 may include a secure network such as an enterprise network, an unsecure network such as a wireless open network, or the Internet.
- Network(s) 710 may also coordinate communication over other networks such as PSTN or cellular networks.
- Network(s) 710 provides communication between the nodes described herein.
- network(s) 710 may include wireless media such as acoustic, RF, infrared and other wireless media.
- FIG. 8 and the associated discussion are intended to provide a brief, general description of a suitable computing environment in which embodiments may be implemented.
- a block diagram of an example computing operating environment is illustrated, such as computing device 800 .
- computing device 800 may be any device in stationary, mobile, or other form such as the example devices discussed in conjunction with FIGS. 1A , 1 B, and 7 , and include at least one processing unit 802 and system memory 804 .
- Computing device 800 may also include a plurality of processing units that cooperate in executing programs.
- the system memory 804 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.) or some combination of the two.
- System memory 804 typically includes an operating system 805 suitable for controlling the operation of the platform, such as the WINDOWS ®, WINDOWS MOBILE®, or WINDOWS PHONE® operating systems from MICROSOFT CORPORATION of Redmond, Washington.
- the system memory 804 may also include one or more software applications such as program modules 806 , application 822 , context based attribute menu module 824 , and detection module 826 .
- Context based attribute menu module 824 may operate in conjunction with the operating system 805 or application 822 and provide a context based font or attribute menu as discussed previously. Context based font attribute module 824 may also provide commands, links, and submenus to adjust attribute of displayed content. Detection module 826 may detect user actions and execute a command associated with the user action to adjust an attribute such as the font size. This basic configuration is illustrated in FIG. 8 by those components within dashed line 808 .
- Computing device 800 may have additional features or functionality.
- the computing device 800 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape.
- additional storage is illustrated in FIG. 8 by removable storage 809 and non-removable storage 810 .
- Computer readable storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.
- System memory 804 , removable storage 809 and non-removable storage 810 are all examples of computer readable storage media.
- Computer readable storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computing device 800 . Any such computer readable storage media may be part of computing device 800 .
- Computing device 800 may also have input device(s) 812 such as keyboard, mouse, pen, voice input device, touch input device, an optical capture device for detecting gestures, and comparable input devices.
- Output device(s) 814 such as a display, speakers, printer, and other types of output devices may also be included. These devices are well known in the art and need not be discussed at length here.
- Computing device 800 may also contain communication connections 816 that allow the device to communicate with other devices 818 , such as over a wireless network in a distributed computing environment, a satellite link, a cellular link, and comparable mechanisms.
- Other devices 818 may include computer device(s) that execute communication applications, other directory or policy servers, and comparable devices.
- Communication connection(s) 816 is one example of communication media.
- Communication media can include therein computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media.
- modulated data signal means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal.
- communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media.
- Example embodiments also include methods. These methods can be implemented in any number of ways, including the structures described in this document. One such way is by machine operations, of devices of the type described in this document.
- Another optional way is for one or more of the individual operations of the methods to be performed in conjunction with one or more human operators performing some. These human operators need not be collocated with each other, but each can be only with a machine that performs a portion of the program.
- FIG. 9 illustrates a logic flow diagram for a process of adjusting content attributes through actions on a context based menu in touch and/or gesture enabled devices according to embodiments.
- Process 900 may be implemented as part of an application or an operating system of any computing device capable of accepting touch, gesture, keyboard, mouse, pen, or similar inputs.
- Process 900 begins with operation 910 , where a context based menu may be presented by a user interface of an application displaying the context based menu on a device screen to manage content.
- the context based menu may be a context based attribute menu to adjust an incremental attribute of the content such as font characteristics, layout features, line weights, gradient based attributes, and similar ones.
- the user interface may detect a user action on the context based menu at operation 920 .
- the user action may be a tap, a swipe, a press and hold, or similar user action including, but not limited to touch, gestures, keyboard entries, mouse clicks, pen inputs, optically captured gestures, voice commands, etc.
- the application may adjust a font attribute of the content according to the user action.
- process 900 The operations included in process 900 are for illustration purposes. Adjusting content attributes through actions on context based menus according to embodiments may be implemented by similar processes with fewer or additional steps, as well as in different order of operations using the principles described herein.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
- This application claims the benefit of U.S. Provisional Patent Applications Serial No. 61/507,983 and 61/556,945 filed on Jul. 14, 2011 and Nov. 8, 2011, respectively. The disclosures of the provisional patent applications are hereby incorporated by reference for all purposes.
- With the proliferation of computing and networking technologies, two aspects of computing devices have become prevalent: non-traditional (e.g., mouse and keyboard) input mechanisms and smaller form factors. User interfaces for all kinds of software applications have been designed taking typical screen sizes and input mechanisms into account. Thus, user interactions in conventional systems are presumed to be through keyboard and mouse type input devices and a minimum screen size that enables users to interact with the user interface at a particular precision.
- Limited display real estate burdens many portable devices from providing full featured content management functionality. Furthermore, gestural commanding is not efficient using conventional menus including support for limited displays or just taking into account where the user's finger/mouse/pen is. Additionally, display devices such as projectors, monitors, and televisions may lack controls for providing content management functionality. Modern software solutions such as on screen keyboards may be awkward to type and encompass valuable display area. Lack of adequate software solutions for managing content on non-traditional devices largely limit device use to content consumption. Carrying multiple devices for content management and consumption defeats portability and unnecessarily takes away from an enriching singular source for content consumption and management.
- One of the challenging functionality control areas is content attribute management such as font size/color/style for textual content, thickness and various styles for graphical content. For example, font sizes are provided in preset values listed in a drop-down or similar menu in conventional systems, which may not be a natural way for a user to interact using touch or gestures. Similarly, graphics such as lines may have various attributes (thickness, corner styles, color, dash style, etc.) that need to be presented to a user in an intuitive, user-friendly manner for the user to be able to interact through new input technologies.
- This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This summary is not intended to exclusively identify key features or essential features of the claimed subject matter, nor is it intended as an aid in determining the scope of the claimed subject matter.
- Embodiments are directed to adjusting content attributes through a context based menu. Context based menus may be deployed for variety of scenarios for content management. An application according to embodiments may present the context based menu upon selection of content or upon activation of a context based menu control. Content attributes such as font size, style, graphics line weight, style, and similar ones may be presented in an intuitive manner such that selection can be made easily through non-traditional input methods such as touch or gesture. Some attributes may be presented with visual examples. Through a hierarchical approach submenus may be activated for further detailed selection of available commands through the selection of items on a top level context based menu. Furthermore, selection of attribute values on a linear or otherwise spectrum may be enabled through a sliding action that allows continuous value selection as opposed to distinct values presented in traditional menus.
- These and other features and advantages will be apparent from a reading of the following detailed description and a review of the associated drawings. It is to be understood that both the foregoing general description and the following detailed description are explanatory and do not restrict aspects as claimed.
-
FIGS. 1A and 1B illustrate some example devices, where context based font or attribute menus may be employed; -
FIG. 2 illustrates an example context based menu with controls to adjust an attribute of content according to embodiments; -
FIG. 3 illustrates example scenarios for using context based menu according to embodiments; -
FIG. 4 illustrates other example scenarios for using context based menu according to embodiments; -
FIG. 5 illustrates examples of context based menus for adjusting font types and styles according to embodiments; -
FIG. 6 illustrates an example of a context based menu for adjusting a line thickness attribute of graphical content according to embodiments; -
FIG. 7 is a networked environment, where a system according to embodiments may be implemented; -
FIG. 8 is a block diagram of an example computing operating environment, where embodiments may be implemented; and -
FIG. 9 illustrates a logic flow diagram for a process of adjusting content attributes through a context based menu in touch and gesture enabled devices according to embodiments. - As briefly described above, a user interface of an application executing on a device may present a context based menu in relation to displayed content. The context based menu may provide controls to adjust the font size of the content. The application may detect a user action associated with the context based menu. The application may execute a command to adjust the font size based on the detected user action, which may include touch, gesture, keyboard entry, mouse click, and/or pen input. Content attributes as used herein may be any aspect of content that can be represented along a scale either discretely or continuously. Continuous representation can be interacted with by moving a finger (or equivalent gesture or similar input) across a circle. Discrete options may be interacted with by selecting a range of points along a line, for example, by snapping one's finger to a range of points along the line. A non-exhaustive, illustrative set of examples for content attributes may include an amount of space between lines, a number of columns, a size of margins, an amount of indentation, a font kerning, a zoom action, and comparable ones.
- In the following detailed description, references are made to the accompanying drawings that form a part hereof, and in which are shown by way of illustrations specific embodiments or examples. These aspects may be combined, other aspects may be utilized, and structural changes may be made without departing from the spirit or scope of the present disclosure. The following detailed description is therefore not to be taken in the limiting sense, and the scope of the present invention is defined by the appended claims and their equivalents. While the embodiments will be described in the general context of program modules that execute in conjunction with an application program that runs on an operating system on a personal computer, those skilled in the art will recognize that aspects may also be implemented in combination with other program modules.
- Generally, program modules include routines, programs, components, data structures, and other types of structures that perform particular tasks or implement particular abstract data types. Moreover, those skilled in the art will appreciate that embodiments may be practiced with other computer system configurations, including hand-held devices, multiprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers, and comparable computing devices. Embodiments may also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a communications network. In a distributed computing environment, program modules may be located in both local and remote memory storage devices.
- Embodiments may be implemented as a computer-implemented process (method), a computing system, or as an article of manufacture, such as a computer program product or computer readable media. The computer program product may be a computer storage medium readable by a computer system and encoding a computer program that comprises instructions for causing a computer or computing system to perform example process(es). The computer-readable storage medium is a computer-readable memory device. The computer-readable storage medium can for example be implemented via one or more of a volatile computer memory, a non-volatile memory, a hard drive, a flash drive, a floppy disk, or a compact disk, and comparable media.
- According to embodiments, a user interface of a touch-enabled or gesture-enabled device may employ context based menus to adjust content attributes such as font size, font style, font boldness, graphics style, color, thickness, a line color, a line type, a corner shape, a shading attribute, a fill color, a spacing, a zoom action on displayed content, and a fill type, etc. A context based menu may make use of features specific to touch or gesture enabled computing devices, but may also work with a traditional mouse and keyboard. Context based menus, in general, may be used to provide quick access to commonly used commands while viewing or editing displayed content such as documents, emails, contact lists, other communications, or any content (e.g., audio, video, etc.). Context based menus may appear as part of a user interface's regular menu, in a separate viewing pane (e.g., a window) outside or inside the user interface, and so on. Typically, context based menus present a limited set of commands for easy user access, but additional submenus may be presented upon user selection. Commonly used context based menus may appear over the viewed document. A tap or swipe action as used herein may be provided by a user through a finger, a pen, a mouse, or similar device, as well as through predefined keyboard entry combinations or a voice command.
-
FIGS. 1A and 1B illustrate some example devices, where a context based menus may be employed. As touch and gesture based technologies are proliferating and computing devices employing those technologies are becoming common, user interface arrangement becomes a challenge. Touch and/or gesture enabled devices, specifically portable devices, tend to have smaller screen size, which means less available space for user interfaces. For example, in a user interface that enables editing of a document (text and/or graphics), in addition to the presented portion of the document, a virtual keyboard may have to be displayed further limiting the available space (“real estate”). Thus, in such scenarios, providing a full control menu may be impractical or impossible. Embodiments are directed to a context based font or attribute menus to adjust content attributes such as font size. - As mentioned above, smaller available display space, larger content, and different aspect ratios make conventional menus impractical. Existing touch-based devices such as tablet PCs and similar ones are typically directed to data consumption (i.e., viewing). On the other hand, commonly used applications such as word processing applications, spreadsheet applications, presentation applications, and comparable ones are directed to creation (generating and editing documents with textual, graphical, and other content). Currently available context based menus are either invisible most of the time or they block the content when they are visible. A context based menu according to some embodiments may be provided dynamically based on presented content and available space while providing ease of use without usurping much needed display area.
- Referring to
FIGS. 1A and 1B , some example devices are illustrated, where a context based font or attribute menu may be provided according to embodiments. A context based menu or a context based attribute menu may be an embodiment of a context based menu. Embodiments may be implemented in touch and/or gesture enabled devices or others with keyboard/mouse/pen input, with varying form factors and capabilities. -
Device 104 inFIG. 1A is an example of a large size display device, where a user interface may be provided onscreen 106. Functionality of various applications may be controlled through hardware controls 108 and/or soft controls such as a context based menu displayed onscreen 106. A user may be enabled to interact with the user interface through touch actions or gestures (detected by a video capture device). A launcher indicator may be presented at a fixed location or at a dynamically adjustable location for the user to activate the context based menu. Examples ofdevice 104 may include public information display units, large size computer monitors, and so on. -
Device 112 inFIG. 1A is an example for use of a context based menu to control functionality. A user interface may be displayed on a screen or projected on a surface and actions ofuser 110 may be detected as gestures throughvideo capture device 114. The user's gestures may activate a context based menu to adjust an attribute of a content displayed on thedevice 112. -
FIG. 1B includes several example devices such as touch enabledcomputer monitor 116,laptop computer 118,handheld computer 124,smart phone 126, tablet computer (or slate) 128, andmobile computing device 132, which may be used for computing, communication, control, measurement, and a number of other purposes. The example devices inFIG. 1B are shown withtouch activation 120. However, any of these and other example devices may also employ gesture enabled activation of context based menus to adjust an attribute of content such as font size. In addition, tools such aspen 130 may be used to provide touch input. A context based menu may be controlled also through conventional methods such as a mouse input or input through akeyboard 122. -
FIG. 2 illustrates an example context based menu with controls to adjust an attribute of content. Diagram 200 displays a context basedmenu 202 with control items to adjust content attributes such as font size, style, color, type, etc. - According to some embodiments, a user interface of an application may display context based
menu 202 withcontrol item 218 to access a prior context based menu. The prior context based menu may be a higher level menu providing top level functionality such as copy and paste.Control item 218 may also be used to change the context basedmenu 202 to a collapsed state (i.e., visually minimized). - According to other embodiments, the context based
menu 202 may have a fontsize control item 204 to access another context based menu such asmenu 302 inFIG. 3 to adjust the font size of textual content. The control item(s) may display an indicator to show the current font type, style, and size of selected textual content. Alternatively, a contemplated but not yet applied font size, style, type, etc. may also be displayed on the control items (e.g., when a user hovers on a font size but does not click or equivalent touch/gesture actions). Anothercontrol item 206 may enable a user to underline the selected textual content. Additionally, acontrol item 208 may enable the user to access a context based menu to alter content color. Examples may include a context based menu showing a color gradient to enable a user to pick a color for the content. - According to yet other embodiments, the context based
menu 202 may provide acontrol item 210 to make the font of the selected content “bold.” Similarly, the context basedmenu 202 may enable a user to italicize the selected content usingcontrol item 212. Furthermore, the context basedmenu 202 may enable a user to adjust the font style usingcontrol item 214. Thecontrol item 214 may be assigned to a single style. Alternatively, thecontrol item 214 may launch a context based submenu to adjust the style of the content such as themenu 516 inFIG. 5 . - Some or all of the above-discussed control items may be a one command control (i.e., the underlying command is executed upon selection of the control and the context based menu may be collapsed or hidden following the execution) or associated with one or more submenus. For example, selection of the
control item 206 for underlining textual content may result in deployment of a submenu that presents options for underline style (e.g. dashed, straight), underline thickness, underline color, etc. The associated submenus may be launched through a selection of the particular control (e.g., tapping on the control or swiping from the center toward the control) or through an action (e.g., tap) on an activation area along the perimeter of the context based menu next to each control (the white arrows in diagram 200). -
FIG. 3 illustrates example scenarios for using context based menu according to embodiments. Diagram 300 displays example context basedmenu 302 executing a variety of user actions. - As previously stated, a user action may be a touch action, a gesture action, a keyboard input, a mouse input, pen input, or similar action selecting a particular control item on the context based menu. Alternatively, the user action may be a swipe over a portion of the context based
menu 302. A user may swipe 304 a finger over aslider 308 to adjust the font size of the textual content in a continuous manner (i.e., even though each available font size is not listed on the menu, the swipe-based slider selection may enable the user to select unlisted font sizes easily). - According to some embodiments, the application may display the
slider 308 centered on acenter command button 318 of the context basedmenu 302. The application may position theslider 308 at a start position on a current font size value of the managed content. An example may include the font size of selected content that a user may wish to adjust. The application may move the slider clockwise subsequent to detecting the clockwise swipe action 304 (or alternatively through another clockwise swipe action 306) starting from a start position of theslider 308. Next, the application may increase the font size according to the clockwise swipe, up to a maximum font size. Maximum font size may be displayed on the context basedmenu 302 and may be determined according to a system setting or may be user adjustable according to application privileges. An example of a font size increase may include moving theslider 308 from a starting position at font size 10 tofont size 20 in reaction to matching user action. Subsequent to the user action, the application may change the font size of the content from 10 to 20. - According to other embodiments, the application may move the
slider 308 counterclockwise subsequent to detecting a counterclockwise swipe from the start position. The application may decrease the font size according to the counterclockwise swipe, to a minimum font size. An example may include moving theslider 308 from a starting position atfont size 20 to font size 10 in reaction to a matching user action. Subsequent to the user action, the application may change the font size of the content from 20 to 10. - According to yet other embodiments, the application may display an
indicator 310 on the context basedmenu 302 of a previous font size (e.g., font size 13). Theindicator 310 may be presented as a bold, an italicized, an underlined, a highlighted, or an enclosed font size value or just as another number of the menu that does not appear at one of the displayed axes of the menu. The enclosed font size value may be a circled font size value, a font size value with a square around it, or any other shape to differentiate the previous font size value from others on the context basedmenu 302. - According to other embodiments, the application may integrate font styles into the context based menu for managing font sizes. The application may place a
font style action 312 between displayed font size values. An example may include adjusting the content to a selected font style such as “heading 1” subsequent to detecting auser action 312 between displayed font size values on themenu 302. - According to other embodiments, the application may detect a user action moving clockwise past a maximum font size value and ending in a
dead zone 314. The user action may end between the maximum font size value and the minimum font size value. In this scenario, the application may position theslider 308 on the maximum font size value subsequent to the user action ending in thedead zone 314. - According to other embodiments, the application may detect a user action moving counterclockwise past a minimum font size value and ending in the
dead zone 314. The user action may end between the minimum font size value and the maximum font size value. In this scenario, the application may position theslider 308 on the minimum font size value subsequent to the user action ending in thedead zone 314. Another aspect of the dead zone is that it may allow a user to continue to move their finger clockwise or counterclockwise but not have their font change immediately from the largest to the smallest font. Instead it may create a buffer within which the font does not actually change until one reaches the other edge of the buffer zone. For example, if the user moves counterclockwise past the minimum size it may stay at the minimum through the dead zone but once the user passes the maximum size it may switch to the maximum size (and then start shrinking again, etc.). - According to other embodiments, the application may display
font state 320 including present font colors, font type, and font shape on the font size values. Additionally, the application may detect atap action 316 as the user action on a font size value. Subsequent to the user action on the font size value, the application may position theslider 308 on the font size value and adjust the font size of the content according to the font size value associated withtap action 316. -
FIG. 4 illustrates other example scenarios for using a context based menu to adjust font attributes according to embodiments. Diagram 400 displays example context basedmenu 402 executing a variety of user actions. - According to some embodiments, the application may display a
tooltip 410 subsequent to auser action 404 such as increasing the font size. In addition, the application may show atooltip 414 subsequent touser action 406 adjusting the font size. Theuser action 406 may provide more granularity compared touser action 404 due to larger arc available to a user to pick a font size according to distance of theuser actions menu 402. Thus, the relative position of the user action with respect to a center of the menu may be taken into consideration enabling increased granularity especially in slider style menu actions. - According to other embodiments, the application may calculate the font size according to a polynomial equation to determine the font size for the user action ending in between font size values. An example may include a user moving the slider in between
font size slider 408 stops in relation to both font size increments. - According to yet other embodiments, the application may select the font size value closest to an end position of the
user action 412 subsequent to detecting the end position outside an action area of the font size values. An action area may be an oval area determined by system settings centered around a control such as a font size value. The application may adjust the length of theslider 408 to position the slider on the end position of theuser action 412. Adjusting theslider 408 length may enable the user to more accurately view a selected font size value to the corresponding user action.Tooltips slider 408 to enable the user to pick a desired font size increment. In some examples, “snap points” may be used along the slider at each of the displayed font sizes. As the user moves by sliding/dragging closer to those snap points the line may be automatically moved to the next size, making it efficient to get those sizes. A back-off logic may also be employed. For example, if the user then goes in the reverse direction, the stickiness may be removed to that point and allow the user to select a point very close but not exactly on any of the axes (e.g., if the user's finger was very close to 72, they would still be able to select font 71). -
FIG. 5 illustrates examples of context based menus for adjusting font types and styles according to embodiments. Diagram 500 displays context basedmenu 502 for adjusting a font type and context basedmenu 516 for adjusting a font style. - According to some embodiments, a user interface of an application supporting context based menus may present a context based
menu 502 to enable adjustment of a font type of textual content using aslider 506. A user may move aslider 506 with aswipe action 504 moving clockwise or counterclockwise on the context basedmenu 502 selecting among available font types. In some examples, the different font types may be displayed on the context based menu 502 (e.g., along the perimeter). In other examples, a list ofavailable font types 510 may be presented adjacent to the context basedmenu 502. Alternatively, the list may be presented in the middle of the radial menu with a dial around the edge. As the user moves theslider 506 on the menu, font types that correspond to the slider's current position may be indicated on the list through a graphical, color, textual, shading, or similar scheme. For example, in the illustrated example of context basedmenu 502, theslider 506 may correspond to a position for “Arial” font. This may be emphasized on the list ofavailable font types 510 through highlighting 512. - The available font types on the list may be displayed using the respective fonts to provide the user further visual assistance or in a single font (for example, in small display areas, where the size of the menu may make reading different font types relatively difficult).
- According to other embodiments, the context based
menu 502 may display anindicator 508 showing the current font type of the content. Theindicator 508 may also show other state information of the content such as font size and style. Additionally, the context basedmenu 502 may provide aaccess button 514 to collapse the context basedmenu 502 and/or launch a prior context based menu such as top level menu to provide other functionality to manage the content. - Context based
menu 516 may provide a user with similar functionality as discussed above in conjunction with context basedmenu 502 to adjust a font style of textual content. Themenu 516 may displayindicator 518 to show the current font style of the selected content. The list ofavailable styles 520 and thehighlighter 522 may be used to present the styles in an easily visible manner as the user may be moving their finger over the menu (thereby, moving the slider). In some embodiments, the user may be enabled to scroll through the list ofavailable styles 520 through a swipe action, for example, and select a font style by tapping on it on the list. Context based menu may be a submenu under a top level menu for adjusting font attributes and it may include additional submenus depending on the number of styles, for example. Typically, a predefined number of styles may be presented to prevent overlapping problems. However, using a slider (and a separate list of available items) and by adjusting the speed of the slider, higher granularity and, thereby, larger number of available items may be presented for selection through a context based menu. -
FIG. 6 illustrates an example of a context based attribute menu adjusting an attribute of graphical content according to embodiments. Diagram 600 displays context basedattribute menu 602 to adjust line weights of a selected content such as a drawing. - According to some embodiments, a user interface may display
menu 602 to adjust an attribute of content such as line weights. Themenu 602 may display line weight values 614 corresponding to line weights illustrated bylist 604. In some examples, a line thickness (or the attribute in general) may also be displayed on the slider itself. For example, where the slider is at ¼ it may be a ¼ thickness and then the slider may become thicker as it goes around until it is at its thickest at 6. And alternatively, the thicknesses (or the attribute) may be relative and not exact. Thus, it could be just that the thickness increases but does not match exactly the numbers. Auser action 612 may move aslider 610 through swipe and tap actions as described in context based menus selecting a line weight. The application may adjust the line weight of the graphical content (any shape that includes lines) according to the user action. Additionally, themenu 602 may display the selected line weight through ahighlighter 606 similar to context based menus described above. - According to other embodiments, the application may detect a swipe action as the user action starting from a
center control button 608 and moving away from the context basedattribute menu 602. The application may determine anattribute value 614 intersecting the swipe action. Subsequently, the application may adjust the attribute according to theattribute value 614. Additionally, the application may display an attribute increase control at an end of theslider 610 to increase the attribute beyond the displayed end value. Furthermore, the application may display an attribute decrease control at the other end of theslider 610 to decrease the attribute beyond the displayed end value. - A context based attribute menu may be used to adjust any attribute of textual or graphical content. Examples may include line weights, corner styles, gradients used to pick a color, shading schemes for graphical objects, and comparable ones. The application may display incremental values, discrete attributes, or continuous selection of values on the attribute menu. Displayed values may be selected according to linear, logarithmic, or other scales. The displayed values may also be user selectable according to application privileges, user preferences, and/or customization.
- The example commands, links, submenus, configurations, and context based font and attribute menus depicted in
FIGS. 1 through 6 are provided for illustration purposes only. Embodiments are not limited to the shapes, forms, and content shown in the example diagrams, and may be implemented using other textual, graphical, and similar schemes employing the principles described herein. -
FIG. 7 is a networked environment, where a system according to embodiments may be implemented. In addition to locally installed applications, such asapplication 822 discussed below, a context based font/attribute menu for touch and/or gesture enabled devices may be also be employed in conjunction with hosted applications and services that may be implemented via software executed over one ormore servers 706 orindividual server 708. A hosted service or application may communicate with client applications on individual computing devices such as ahandheld computer 701, adesktop computer 702, alaptop computer 703, asmart phone 704, a tablet computer (or slate), 705 (‘client devices’) through network(s) 710 and control a user interface presented to users. - As previously discussed, a context based menu may be used for adjusting an attribute (such as font size) of the content provided by the hosted service or application. For example, a browser application, a word processing application, a spreadsheet application, a calendar application, a note taking application, a graphics application, and comparable ones may make use of a context based menu according to embodiments. The context based menu may be activated through a variety of user actions such as selection of content, activation of a launcher indicator, detection of a predetermined touch or gesture action, etc.
- Client devices 701-705 are used to access the functionality provided by the hosted service or application. One or more of the
servers 706 orserver 708 may be used to provide a variety of services as discussed above. Relevant data may be stored in one or more data stores (e.g. data store 714), which may be managed by any one of theservers 706 or bydatabase server 712. - Network(s) 710 may comprise any topology of servers, clients, Internet service providers, and communication media. A system according to embodiments may have a static or dynamic topology. Network(s) 710 may include a secure network such as an enterprise network, an unsecure network such as a wireless open network, or the Internet. Network(s) 710 may also coordinate communication over other networks such as PSTN or cellular networks. Network(s) 710 provides communication between the nodes described herein. By way of example, and not limitation, network(s) 710 may include wireless media such as acoustic, RF, infrared and other wireless media.
- Many other configurations of computing devices, applications, data sources, and data distribution systems may be employed to adjust attribute of content through a context based font/attribute menu. Furthermore, the networked environments discussed in
FIG. 7 are for illustration purposes only. Embodiments are not limited to the example applications, modules, or processes. -
FIG. 8 and the associated discussion are intended to provide a brief, general description of a suitable computing environment in which embodiments may be implemented. With reference toFIG. 8 , a block diagram of an example computing operating environment according to embodiments is illustrated, such ascomputing device 800. In a basic configuration,computing device 800 may be any device in stationary, mobile, or other form such as the example devices discussed in conjunction withFIGS. 1A , 1B, and 7, and include at least oneprocessing unit 802 andsystem memory 804.Computing device 800 may also include a plurality of processing units that cooperate in executing programs. Depending on the exact configuration and type of computing device, thesystem memory 804 may be volatile (such as RAM), non-volatile (such as ROM, flash memory, etc.) or some combination of the two.System memory 804 typically includes anoperating system 805 suitable for controlling the operation of the platform, such as the WINDOWS ®, WINDOWS MOBILE®, or WINDOWS PHONE® operating systems from MICROSOFT CORPORATION of Redmond, Washington. Thesystem memory 804 may also include one or more software applications such asprogram modules 806,application 822, context basedattribute menu module 824, anddetection module 826. - Context based
attribute menu module 824 may operate in conjunction with theoperating system 805 orapplication 822 and provide a context based font or attribute menu as discussed previously. Context basedfont attribute module 824 may also provide commands, links, and submenus to adjust attribute of displayed content.Detection module 826 may detect user actions and execute a command associated with the user action to adjust an attribute such as the font size. This basic configuration is illustrated inFIG. 8 by those components within dashedline 808. -
Computing device 800 may have additional features or functionality. For example, thecomputing device 800 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Such additional storage is illustrated inFIG. 8 byremovable storage 809 andnon-removable storage 810. Computer readable storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer readable instructions, data structures, program modules, or other data.System memory 804,removable storage 809 andnon-removable storage 810 are all examples of computer readable storage media. Computer readable storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can be accessed by computingdevice 800. Any such computer readable storage media may be part ofcomputing device 800.Computing device 800 may also have input device(s) 812 such as keyboard, mouse, pen, voice input device, touch input device, an optical capture device for detecting gestures, and comparable input devices. Output device(s) 814 such as a display, speakers, printer, and other types of output devices may also be included. These devices are well known in the art and need not be discussed at length here. -
Computing device 800 may also containcommunication connections 816 that allow the device to communicate withother devices 818, such as over a wireless network in a distributed computing environment, a satellite link, a cellular link, and comparable mechanisms.Other devices 818 may include computer device(s) that execute communication applications, other directory or policy servers, and comparable devices. Communication connection(s) 816 is one example of communication media. Communication media can include therein computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. The term “modulated data signal” means a signal that has one or more of its characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media includes wired media such as a wired network or direct-wired connection, and wireless media such as acoustic, RF, infrared and other wireless media. - Example embodiments also include methods. These methods can be implemented in any number of ways, including the structures described in this document. One such way is by machine operations, of devices of the type described in this document.
- Another optional way is for one or more of the individual operations of the methods to be performed in conjunction with one or more human operators performing some. These human operators need not be collocated with each other, but each can be only with a machine that performs a portion of the program.
-
FIG. 9 illustrates a logic flow diagram for a process of adjusting content attributes through actions on a context based menu in touch and/or gesture enabled devices according to embodiments.Process 900 may be implemented as part of an application or an operating system of any computing device capable of accepting touch, gesture, keyboard, mouse, pen, or similar inputs. -
Process 900 begins withoperation 910, where a context based menu may be presented by a user interface of an application displaying the context based menu on a device screen to manage content. The context based menu may be a context based attribute menu to adjust an incremental attribute of the content such as font characteristics, layout features, line weights, gradient based attributes, and similar ones. Subsequently, the user interface may detect a user action on the context based menu atoperation 920. The user action may be a tap, a swipe, a press and hold, or similar user action including, but not limited to touch, gestures, keyboard entries, mouse clicks, pen inputs, optically captured gestures, voice commands, etc. Atoperation 930, the application may adjust a font attribute of the content according to the user action. - The operations included in
process 900 are for illustration purposes. Adjusting content attributes through actions on context based menus according to embodiments may be implemented by similar processes with fewer or additional steps, as well as in different order of operations using the principles described herein. - The above specification, examples and data provide a complete description of the manufacture and use of the composition of the embodiments. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims and embodiments.
Claims (20)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/542,962 US20130019204A1 (en) | 2011-07-14 | 2012-07-06 | Adjusting content attributes through actions on context based menu |
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161507983P | 2011-07-14 | 2011-07-14 | |
US201161556945P | 2011-11-08 | 2011-11-08 | |
US13/542,962 US20130019204A1 (en) | 2011-07-14 | 2012-07-06 | Adjusting content attributes through actions on context based menu |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130019204A1 true US20130019204A1 (en) | 2013-01-17 |
Family
ID=47519676
Family Applications (6)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/339,569 Active 2032-02-05 US9086794B2 (en) | 2011-07-14 | 2011-12-29 | Determining gestures on context based menus |
US13/341,074 Active US9250766B2 (en) | 2011-07-14 | 2011-12-30 | Labels and tooltips for context based menus |
US13/349,691 Active US9021398B2 (en) | 2011-07-14 | 2012-01-13 | Providing accessibility features on context based radial menus |
US13/542,962 Abandoned US20130019204A1 (en) | 2011-07-14 | 2012-07-06 | Adjusting content attributes through actions on context based menu |
US13/543,976 Abandoned US20130019208A1 (en) | 2011-07-14 | 2012-07-09 | Managing content color through context based color menu |
US13/549,397 Active 2033-01-13 US9116602B2 (en) | 2011-07-14 | 2012-07-13 | Providing customization of context based menus |
Family Applications Before (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/339,569 Active 2032-02-05 US9086794B2 (en) | 2011-07-14 | 2011-12-29 | Determining gestures on context based menus |
US13/341,074 Active US9250766B2 (en) | 2011-07-14 | 2011-12-30 | Labels and tooltips for context based menus |
US13/349,691 Active US9021398B2 (en) | 2011-07-14 | 2012-01-13 | Providing accessibility features on context based radial menus |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/543,976 Abandoned US20130019208A1 (en) | 2011-07-14 | 2012-07-09 | Managing content color through context based color menu |
US13/549,397 Active 2033-01-13 US9116602B2 (en) | 2011-07-14 | 2012-07-13 | Providing customization of context based menus |
Country Status (1)
Country | Link |
---|---|
US (6) | US9086794B2 (en) |
Cited By (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130076642A1 (en) * | 2011-09-22 | 2013-03-28 | Microsoft Corporation | User interface for editing a value in place |
US20140006033A1 (en) * | 2012-06-29 | 2014-01-02 | Samsung Electronics Co., Ltd. | Method and apparatus for processing multiple inputs |
US20140157200A1 (en) * | 2012-12-05 | 2014-06-05 | Samsung Electronics Co., Ltd. | User terminal apparatus and method of controlling the same |
WO2014113507A1 (en) * | 2013-01-15 | 2014-07-24 | Leap Motion, Inc. | Dynamic user interactions for display control and customized gesture interpretation |
EP2816460A1 (en) * | 2013-06-21 | 2014-12-24 | BlackBerry Limited | Keyboard and touch screen gesture system |
US20150026642A1 (en) * | 2013-07-16 | 2015-01-22 | Pinterest, Inc. | Object based contextual menu controls |
US20150370469A1 (en) * | 2013-01-31 | 2015-12-24 | Qualcomm Incorporated | Selection feature for adjusting values on a computing device |
US9285893B2 (en) | 2012-11-08 | 2016-03-15 | Leap Motion, Inc. | Object detection and tracking with variable-field illumination devices |
US9355073B2 (en) | 2013-06-18 | 2016-05-31 | Microsoft Technology Licensing, Llc | Content attribute control interface including incremental, direct entry, and scrollable controls |
US20160231914A1 (en) * | 2015-02-11 | 2016-08-11 | Adobe Systems Incorporated | Dynamic text control for mobile devices |
US9436998B2 (en) | 2012-01-17 | 2016-09-06 | Leap Motion, Inc. | Systems and methods of constructing three-dimensional (3D) model of an object using image cross-sections |
US9465461B2 (en) | 2013-01-08 | 2016-10-11 | Leap Motion, Inc. | Object detection and tracking with audio and optical signals |
US9495613B2 (en) | 2012-01-17 | 2016-11-15 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging using formed difference images |
US9613262B2 (en) | 2014-01-15 | 2017-04-04 | Leap Motion, Inc. | Object detection and tracking for providing a virtual device experience |
US9632658B2 (en) | 2013-01-15 | 2017-04-25 | Leap Motion, Inc. | Dynamic user interactions for display control and scaling responsiveness of display objects |
US9679215B2 (en) | 2012-01-17 | 2017-06-13 | Leap Motion, Inc. | Systems and methods for machine control |
US9702977B2 (en) | 2013-03-15 | 2017-07-11 | Leap Motion, Inc. | Determining positional information of an object in space |
US9747696B2 (en) | 2013-05-17 | 2017-08-29 | Leap Motion, Inc. | Systems and methods for providing normalized parameters of motions of objects in three-dimensional space |
US9952658B2 (en) | 2015-03-17 | 2018-04-24 | Wipro Limited | System and method for improving viewing experience on a digital device |
US9996638B1 (en) | 2013-10-31 | 2018-06-12 | Leap Motion, Inc. | Predictive information for free space gesture control and communication |
US10268367B2 (en) | 2010-02-19 | 2019-04-23 | Microsoft Technology Licensing, Llc | Radial menus with bezel gestures |
CN110083296A (en) * | 2019-03-15 | 2019-08-02 | 努比亚技术有限公司 | Wearable device and its exchange method, computer readable storage medium |
US10444966B2 (en) * | 2014-04-11 | 2019-10-15 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling number input in an electronic device |
US10558338B2 (en) * | 2014-05-28 | 2020-02-11 | Facebook, Inc. | Systems and methods for providing responses to and drawings for media content |
US10609285B2 (en) | 2013-01-07 | 2020-03-31 | Ultrahaptics IP Two Limited | Power consumption in motion-capture systems |
US10620709B2 (en) | 2013-04-05 | 2020-04-14 | Ultrahaptics IP Two Limited | Customized gesture interpretation |
US10691219B2 (en) | 2012-01-17 | 2020-06-23 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US10846942B1 (en) | 2013-08-29 | 2020-11-24 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US11074572B2 (en) | 2016-09-06 | 2021-07-27 | Apple Inc. | User interfaces for stored-value accounts |
US11100498B2 (en) | 2018-06-03 | 2021-08-24 | Apple Inc. | User interfaces for transfer accounts |
US11099653B2 (en) | 2013-04-26 | 2021-08-24 | Ultrahaptics IP Two Limited | Machine responsiveness to dynamic user movements and gestures |
EP3869311A1 (en) * | 2015-12-22 | 2021-08-25 | Daikin Industries, Ltd. | Setting value change device |
US20210279082A1 (en) * | 2018-11-30 | 2021-09-09 | Snap Inc. | Interface to configure media content |
US11169830B2 (en) | 2019-09-29 | 2021-11-09 | Apple Inc. | Account management user interfaces |
US11221744B2 (en) | 2017-05-16 | 2022-01-11 | Apple Inc. | User interfaces for peer-to-peer transfers |
US11222325B2 (en) | 2017-05-16 | 2022-01-11 | Apple Inc. | User interfaces for peer-to-peer transfers |
US11321731B2 (en) | 2015-06-05 | 2022-05-03 | Apple Inc. | User interface for loyalty accounts and private label accounts |
US11328352B2 (en) * | 2019-03-24 | 2022-05-10 | Apple Inc. | User interfaces for managing an account |
US11353962B2 (en) | 2013-01-15 | 2022-06-07 | Ultrahaptics IP Two Limited | Free-space user interface and control using virtual constructs |
US11373373B2 (en) * | 2019-10-22 | 2022-06-28 | International Business Machines Corporation | Method and system for translating air writing to an augmented reality device |
US11386189B2 (en) | 2017-09-09 | 2022-07-12 | Apple Inc. | Implementation of biometric authentication |
US11393258B2 (en) | 2017-09-09 | 2022-07-19 | Apple Inc. | Implementation of biometric authentication |
US11481769B2 (en) | 2016-06-11 | 2022-10-25 | Apple Inc. | User interface for transactions |
US11514430B2 (en) | 2018-06-03 | 2022-11-29 | Apple Inc. | User interfaces for transfer accounts |
US11567578B2 (en) | 2013-08-09 | 2023-01-31 | Ultrahaptics IP Two Limited | Systems and methods of free-space gestural interaction |
US11681537B2 (en) | 2019-09-29 | 2023-06-20 | Apple Inc. | Account management user interfaces |
US11720180B2 (en) | 2012-01-17 | 2023-08-08 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US11740705B2 (en) | 2013-01-15 | 2023-08-29 | Ultrahaptics IP Two Limited | Method and system for controlling a machine according to a characteristic of a control object |
US11775033B2 (en) | 2013-10-03 | 2023-10-03 | Ultrahaptics IP Two Limited | Enhanced field of view to augment three-dimensional (3D) sensory space for free-space gesture interpretation |
US11778159B2 (en) | 2014-08-08 | 2023-10-03 | Ultrahaptics IP Two Limited | Augmented reality with motion sensing |
US11783305B2 (en) | 2015-06-05 | 2023-10-10 | Apple Inc. | User interface for loyalty accounts and private label accounts for a wearable device |
US11836725B2 (en) | 2014-05-29 | 2023-12-05 | Apple Inc. | User interface for payments |
US11875012B2 (en) | 2018-05-25 | 2024-01-16 | Ultrahaptics IP Two Limited | Throwable interface for augmented reality and virtual reality environments |
US11983702B2 (en) | 2021-02-01 | 2024-05-14 | Apple Inc. | Displaying a representation of a card with a layered structure |
US11994377B2 (en) | 2012-01-17 | 2024-05-28 | Ultrahaptics IP Two Limited | Systems and methods of locating a control object appendage in three dimensional (3D) space |
US12002042B2 (en) | 2016-06-11 | 2024-06-04 | Apple, Inc | User interface for transactions |
US12032746B2 (en) | 2015-02-13 | 2024-07-09 | Ultrahaptics IP Two Limited | Systems and methods of creating a realistic displacement of a virtual object in virtual reality/augmented reality environments |
US12118134B2 (en) | 2015-02-13 | 2024-10-15 | Ultrahaptics IP Two Limited | Interaction engine for creating a realistic experience in virtual reality/augmented reality environments |
US12118562B2 (en) | 2020-05-29 | 2024-10-15 | Apple Inc. | Configuring an account for a second user identity |
US12131374B2 (en) | 2023-04-13 | 2024-10-29 | Apple Inc. | User interfaces for managing an account |
Families Citing this family (300)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8359548B2 (en) | 2005-06-10 | 2013-01-22 | T-Mobile Usa, Inc. | Managing subset of user contacts |
US8370769B2 (en) | 2005-06-10 | 2013-02-05 | T-Mobile Usa, Inc. | Variable path management of user contacts |
US8370770B2 (en) | 2005-06-10 | 2013-02-05 | T-Mobile Usa, Inc. | Variable path management of user contacts |
US7685530B2 (en) | 2005-06-10 | 2010-03-23 | T-Mobile Usa, Inc. | Preferred contact group centric interface |
US8255281B2 (en) | 2006-06-07 | 2012-08-28 | T-Mobile Usa, Inc. | Service management system that enables subscriber-driven changes to service plans |
USD609714S1 (en) * | 2007-03-22 | 2010-02-09 | Fujifilm Corporation | Electronic camera |
US8803881B2 (en) * | 2007-09-26 | 2014-08-12 | Autodesk, Inc. | Navigation system for a 3D virtual scene |
US8196042B2 (en) * | 2008-01-21 | 2012-06-05 | Microsoft Corporation | Self-revelation aids for interfaces |
US8577350B2 (en) | 2009-03-27 | 2013-11-05 | T-Mobile Usa, Inc. | Managing communications utilizing communication categories |
US9210247B2 (en) | 2009-03-27 | 2015-12-08 | T-Mobile Usa, Inc. | Managing contact groups from subset of user contacts |
US9369542B2 (en) | 2009-03-27 | 2016-06-14 | T-Mobile Usa, Inc. | Network-based processing of data requests for contact information |
US9195966B2 (en) * | 2009-03-27 | 2015-11-24 | T-Mobile Usa, Inc. | Managing contact groups from subset of user contacts |
US9355382B2 (en) | 2009-03-27 | 2016-05-31 | T-Mobile Usa, Inc. | Group based information displays |
US20120272144A1 (en) * | 2011-04-20 | 2012-10-25 | Microsoft Corporation | Compact control menu for touch-enabled command execution |
US9792017B1 (en) | 2011-07-12 | 2017-10-17 | Domo, Inc. | Automatic creation of drill paths |
US10001898B1 (en) | 2011-07-12 | 2018-06-19 | Domo, Inc. | Automated provisioning of relational information for a summary data visualization |
US9202297B1 (en) | 2011-07-12 | 2015-12-01 | Domo, Inc. | Dynamic expansion of data visualizations |
US9086794B2 (en) * | 2011-07-14 | 2015-07-21 | Microsoft Technology Licensing, Llc | Determining gestures on context based menus |
CA2746065C (en) * | 2011-07-18 | 2013-02-19 | Research In Motion Limited | Electronic device and method for selectively applying message actions |
US9417754B2 (en) | 2011-08-05 | 2016-08-16 | P4tents1, LLC | User interface system, method, and computer program product |
DE102012110278A1 (en) * | 2011-11-02 | 2013-05-02 | Beijing Lenovo Software Ltd. | Window display methods and apparatus and method and apparatus for touch operation of applications |
KR101812657B1 (en) * | 2011-11-22 | 2018-01-31 | 삼성전자주식회사 | A method and apparatus for recommending applications based on context information |
DE112011105888T5 (en) * | 2011-12-23 | 2014-09-11 | Hewlett-Packard Development Company, L.P. | Input command based on hand gesture |
JP5906097B2 (en) * | 2012-01-31 | 2016-04-20 | キヤノン株式会社 | Electronic device, its control method, program, and recording medium |
US8539375B1 (en) * | 2012-02-24 | 2013-09-17 | Blackberry Limited | Method and apparatus for providing a user interface on a device enabling selection of operations to be performed in relation to content |
EP2631747B1 (en) * | 2012-02-24 | 2016-03-30 | BlackBerry Limited | Method and apparatus for providing a user interface on a device that indicates content operators |
US9041727B2 (en) | 2012-03-06 | 2015-05-26 | Apple Inc. | User interface tools for selectively applying effects to image |
US20130238747A1 (en) | 2012-03-06 | 2013-09-12 | Apple Inc. | Image beaming for a media editing application |
US9202433B2 (en) | 2012-03-06 | 2015-12-01 | Apple Inc. | Multi operation slider |
US9131192B2 (en) | 2012-03-06 | 2015-09-08 | Apple Inc. | Unified slider control for modifying multiple image properties |
US9081833B1 (en) * | 2012-04-06 | 2015-07-14 | Google Inc. | Providing a tooltip based on search results |
EP3410287B1 (en) | 2012-05-09 | 2022-08-17 | Apple Inc. | Device, method, and graphical user interface for selecting user interface objects |
WO2013169875A2 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for displaying content associated with a corresponding affordance |
WO2013169865A2 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for moving a user interface object based on an intensity of a press input |
WO2013169854A2 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for providing feedback for changing activation states of a user interface object |
WO2013169849A2 (en) | 2012-05-09 | 2013-11-14 | Industries Llc Yknots | Device, method, and graphical user interface for displaying user interface objects corresponding to an application |
JP6082458B2 (en) | 2012-05-09 | 2017-02-15 | アップル インコーポレイテッド | Device, method, and graphical user interface for providing tactile feedback of actions performed within a user interface |
WO2013169842A2 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for selecting object within a group of objects |
WO2013169843A1 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for manipulating framed graphical objects |
WO2013169845A1 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for scrolling nested regions |
WO2013169851A2 (en) | 2012-05-09 | 2013-11-14 | Yknots Industries Llc | Device, method, and graphical user interface for facilitating user interaction with controls in a user interface |
KR101683868B1 (en) | 2012-05-09 | 2016-12-07 | 애플 인크. | Device, method, and graphical user interface for transitioning between display states in response to gesture |
CN105260049B (en) | 2012-05-09 | 2018-10-23 | 苹果公司 | For contacting the equipment for carrying out display additional information, method and graphic user interface in response to user |
EP2847661A2 (en) | 2012-05-09 | 2015-03-18 | Apple Inc. | Device, method, and graphical user interface for moving and dropping a user interface object |
US9582146B2 (en) * | 2012-05-29 | 2017-02-28 | Nokia Technologies Oy | Causing display of search results |
CN103472972A (en) * | 2012-06-06 | 2013-12-25 | 联发科技(新加坡)私人有限公司 | Text display format setting unit and text display format setting method for mobile terminals and mobile terminal |
CN102799361A (en) * | 2012-06-21 | 2012-11-28 | 华为终端有限公司 | Method for calling application object out and mobile terminal |
CN102750628B (en) * | 2012-06-25 | 2016-09-14 | 华为技术有限公司 | The method of information management and terminal |
USD732555S1 (en) * | 2012-07-19 | 2015-06-23 | D2L Corporation | Display screen with graphical user interface |
US9256351B2 (en) * | 2012-07-20 | 2016-02-09 | Blackberry Limited | Method and electronic device for facilitating user control of a menu |
USD733167S1 (en) * | 2012-07-20 | 2015-06-30 | D2L Corporation | Display screen with graphical user interface |
TW201409338A (en) * | 2012-08-16 | 2014-03-01 | Hon Hai Prec Ind Co Ltd | Electronic apparatus and displaying method of button icon |
US20140297488A1 (en) * | 2012-09-11 | 2014-10-02 | MonyDesktop, Inc. | Method for handling refunds in a budgeting system |
USD819651S1 (en) | 2012-09-11 | 2018-06-05 | Mx Technologies, Inc. | Display screen or portion thereof with a graphical user interface |
US9195368B2 (en) * | 2012-09-13 | 2015-11-24 | Google Inc. | Providing radial menus with touchscreens |
US9261989B2 (en) | 2012-09-13 | 2016-02-16 | Google Inc. | Interacting with radial menus for touchscreens |
US9104313B2 (en) * | 2012-09-14 | 2015-08-11 | Cellco Partnership | Automatic adjustment of selectable function presentation on electronic device display |
USD736232S1 (en) * | 2012-09-24 | 2015-08-11 | Robert Bosch Gmbh | Display screen with graphical user interface |
US20140115539A1 (en) * | 2012-10-18 | 2014-04-24 | Yahoo! Inc. | Customized shortcuts for resource browsing method and apparatus |
US9729695B2 (en) * | 2012-11-20 | 2017-08-08 | Dropbox Inc. | Messaging client application interface |
USD835118S1 (en) | 2012-12-05 | 2018-12-04 | Lg Electronics Inc. | Television receiver with graphical user interface |
US20140181720A1 (en) * | 2012-12-20 | 2014-06-26 | Htc Corporation | Menu management methods and systems |
CN105144057B (en) | 2012-12-29 | 2019-05-17 | 苹果公司 | For moving the equipment, method and graphic user interface of cursor according to the cosmetic variation of the control icon with simulation three-dimensional feature |
JP6097843B2 (en) | 2012-12-29 | 2017-03-15 | アップル インコーポレイテッド | Device, method and graphical user interface for determining whether to scroll or select content |
JP6093877B2 (en) | 2012-12-29 | 2017-03-08 | アップル インコーポレイテッド | Device, method, and graphical user interface for foregoing generation of tactile output for multi-touch gestures |
KR101905174B1 (en) | 2012-12-29 | 2018-10-08 | 애플 인크. | Device, method, and graphical user interface for navigating user interface hierachies |
WO2014105279A1 (en) | 2012-12-29 | 2014-07-03 | Yknots Industries Llc | Device, method, and graphical user interface for switching between user interfaces |
EP2939098B1 (en) | 2012-12-29 | 2018-10-10 | Apple Inc. | Device, method, and graphical user interface for transitioning between touch input to display output relationships |
US9652109B2 (en) * | 2013-01-11 | 2017-05-16 | Microsoft Technology Licensing, Llc | Predictive contextual toolbar for productivity applications |
US20140215373A1 (en) * | 2013-01-28 | 2014-07-31 | Samsung Electronics Co., Ltd. | Computing system with content access mechanism and method of operation thereof |
USD742389S1 (en) * | 2013-01-31 | 2015-11-03 | Directdex Inc. | Display screen portion with icon |
US9569092B2 (en) * | 2013-02-01 | 2017-02-14 | Allscripts Software, Llc | Radial control system and method |
KR20140099760A (en) * | 2013-02-04 | 2014-08-13 | 삼성전자주식회사 | Method and apparatus for display controlling in terminal |
USD739872S1 (en) * | 2013-02-22 | 2015-09-29 | Samsung Electronics Co., Ltd. | Display screen with animated graphical user interface |
USD716819S1 (en) | 2013-02-27 | 2014-11-04 | Microsoft Corporation | Display screen with graphical user interface |
USD702250S1 (en) * | 2013-02-27 | 2014-04-08 | Microsoft Corporation | Display screen with graphical user interface |
USD702253S1 (en) * | 2013-02-27 | 2014-04-08 | Microsoft Corporation | Display screen with graphical user interface |
USD702252S1 (en) * | 2013-02-27 | 2014-04-08 | Microsoft Corporation | Display screen with graphical user interface |
USD702251S1 (en) * | 2013-02-27 | 2014-04-08 | Microsoft Corporation | Display screen with graphical user interface |
US10025459B2 (en) | 2013-03-14 | 2018-07-17 | Airwatch Llc | Gesture-based workflow progression |
US9792014B2 (en) | 2013-03-15 | 2017-10-17 | Microsoft Technology Licensing, Llc | In-place contextual menu for handling actions for a listing of items |
US20140281991A1 (en) * | 2013-03-18 | 2014-09-18 | Avermedia Technologies, Inc. | User interface, control system, and operation method of control system |
US9785240B2 (en) * | 2013-03-18 | 2017-10-10 | Fuji Xerox Co., Ltd. | Systems and methods for content-aware selection |
WO2014146164A1 (en) * | 2013-03-22 | 2014-09-25 | Lifi Labs Inc | Color selection |
US20140317573A1 (en) * | 2013-04-17 | 2014-10-23 | Samsung Electronics Co., Ltd. | Display apparatus and method of displaying a context menu |
KR20140124721A (en) * | 2013-04-17 | 2014-10-27 | 삼성전자주식회사 | Display apparatus and method for displaying contextual menu |
KR102148809B1 (en) * | 2013-04-22 | 2020-08-27 | 삼성전자주식회사 | Apparatus, method and computer readable recording medium for displaying shortcut window |
WO2014178306A1 (en) * | 2013-04-30 | 2014-11-06 | グリー株式会社 | Display information provision method, display information provision program, and server device |
JP6188405B2 (en) * | 2013-05-01 | 2017-08-30 | キヤノン株式会社 | Display control apparatus, display control method, and program |
KR102169521B1 (en) * | 2013-05-14 | 2020-10-23 | 삼성전자 주식회사 | Input apparatus, display apparatus and control method thereof |
US10043172B2 (en) * | 2013-05-29 | 2018-08-07 | Ebay Inc. | Tap and hold |
USD741898S1 (en) * | 2013-05-29 | 2015-10-27 | Microsoft Corporation | Display screen with animated graphical user interface |
KR102191965B1 (en) * | 2013-07-01 | 2020-12-16 | 삼성전자주식회사 | Mobile terminal and operating method thereof |
JP6153007B2 (en) * | 2013-07-19 | 2017-06-28 | 株式会社コナミデジタルエンタテインメント | Operation system, operation control method, operation control program |
USD745533S1 (en) * | 2013-08-27 | 2015-12-15 | Tencent Technology (Shenzhen) Company Limited | Display screen or a portion thereof with graphical user interface |
USD817994S1 (en) | 2013-09-03 | 2018-05-15 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with transitional graphical user interface |
KR20150026424A (en) * | 2013-09-03 | 2015-03-11 | 삼성전자주식회사 | Method for controlling a display and an electronic device |
USD767587S1 (en) * | 2013-09-03 | 2016-09-27 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
USD857738S1 (en) | 2013-09-03 | 2019-08-27 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with transitional graphical user interface |
USD759662S1 (en) * | 2013-10-07 | 2016-06-21 | Suraj Bhagwan Panjabi | Display screen with animated graphical user interface |
US20150121314A1 (en) * | 2013-10-24 | 2015-04-30 | Jens Bombolowsky | Two-finger gestures |
CN103616992B (en) | 2013-11-13 | 2017-10-17 | 华为技术有限公司 | Application control method and device |
WO2015075717A2 (en) * | 2013-11-20 | 2015-05-28 | Hisep Technology Ltd. | Apparatus and method for displaying relative location of persons, places or objects |
US20150169531A1 (en) * | 2013-12-17 | 2015-06-18 | Microsoft Corporation | Touch/Gesture-Enabled Interaction with Electronic Spreadsheets |
USD751606S1 (en) * | 2013-12-30 | 2016-03-15 | Beijing Qihoo Technology Co., Ltd. | Display screen with animated graphical user interface |
US9558180B2 (en) | 2014-01-03 | 2017-01-31 | Yahoo! Inc. | Systems and methods for quote extraction |
US9971756B2 (en) | 2014-01-03 | 2018-05-15 | Oath Inc. | Systems and methods for delivering task-oriented content |
USD761810S1 (en) * | 2014-01-03 | 2016-07-19 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with icon |
US10503357B2 (en) | 2014-04-03 | 2019-12-10 | Oath Inc. | Systems and methods for delivering task-oriented content using a desktop widget |
USD760776S1 (en) * | 2014-01-03 | 2016-07-05 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with icon |
USD738898S1 (en) | 2014-01-09 | 2015-09-15 | Microsoft Corporation | Display screen with graphical user interface |
US10198148B2 (en) | 2014-01-17 | 2019-02-05 | Microsoft Technology Licensing, Llc | Radial menu user interface with entry point maintenance |
US9882996B2 (en) * | 2014-01-23 | 2018-01-30 | Google Llc | Determining data associated with proximate computing devices |
WO2015133657A1 (en) * | 2014-03-03 | 2015-09-11 | Lg Electronics Inc. | Terminal and method for controlling the same |
US10817159B2 (en) | 2014-03-26 | 2020-10-27 | Unanimous A. I., Inc. | Non-linear probabilistic wagering for amplified collective intelligence |
US11269502B2 (en) | 2014-03-26 | 2022-03-08 | Unanimous A. I., Inc. | Interactive behavioral polling and machine learning for amplification of group intelligence |
US20150277678A1 (en) * | 2014-03-26 | 2015-10-01 | Kobo Incorporated | Information presentation techniques for digital content |
US11151460B2 (en) | 2014-03-26 | 2021-10-19 | Unanimous A. I., Inc. | Adaptive population optimization for amplifying the intelligence of crowds and swarms |
US12079459B2 (en) | 2014-03-26 | 2024-09-03 | Unanimous A. I., Inc. | Hyper-swarm method and system for collaborative forecasting |
US10817158B2 (en) | 2014-03-26 | 2020-10-27 | Unanimous A. I., Inc. | Method and system for a parallel distributed hyper-swarm for amplifying human intelligence |
US12099936B2 (en) | 2014-03-26 | 2024-09-24 | Unanimous A. I., Inc. | Systems and methods for curating an optimized population of networked forecasting participants from a baseline population |
WO2017004475A1 (en) * | 2015-07-01 | 2017-01-05 | Unanimous A.I., Inc. | Methods and systems for enabling a credit economy in a real-time collaborative intelligence |
US12001667B2 (en) | 2014-03-26 | 2024-06-04 | Unanimous A. I., Inc. | Real-time collaborative slider-swarm with deadbands for amplified collective intelligence |
US11941239B2 (en) | 2014-03-26 | 2024-03-26 | Unanimous A.I., Inc. | System and method for enhanced collaborative forecasting |
US20150286361A1 (en) * | 2014-04-04 | 2015-10-08 | Monkey Inferno, Inc. | Single gesture video capture and share |
US10025461B2 (en) * | 2014-04-08 | 2018-07-17 | Oath Inc. | Gesture input for item selection |
US10120557B2 (en) * | 2014-04-14 | 2018-11-06 | Ebay, Inc. | Displaying a plurality of selectable actions |
US10560975B2 (en) * | 2014-04-16 | 2020-02-11 | Belkin International, Inc. | Discovery of connected devices to determine control capabilities and meta-information |
US10466876B2 (en) * | 2014-04-17 | 2019-11-05 | Facebook, Inc. | Assisting a user of a software application |
US20150324100A1 (en) * | 2014-05-08 | 2015-11-12 | Tictoc Planet, Inc. | Preview Reticule To Manipulate Coloration In A User Interface |
JP1535035S (en) * | 2014-05-25 | 2015-10-13 | ||
USD768191S1 (en) * | 2014-06-05 | 2016-10-04 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with icon |
USD765669S1 (en) * | 2014-06-10 | 2016-09-06 | Microsoft Corporation | Display screen with graphical user interface |
US9873205B2 (en) * | 2014-06-24 | 2018-01-23 | Spectrum Brands, Inc. | Electric grooming appliance |
KR102037481B1 (en) | 2014-07-31 | 2019-10-28 | 삼성전자주식회사 | Display apparatus, method of controlling the display apparatus and recordable medium storing for program for performing the method |
KR20160018269A (en) * | 2014-08-08 | 2016-02-17 | 삼성전자주식회사 | Device and method for controlling the same |
USD795916S1 (en) * | 2014-08-19 | 2017-08-29 | Google Inc. | Display screen with animated graphical user interface |
US11494056B1 (en) | 2014-08-29 | 2022-11-08 | Open Invention Network Llc | Dynamic document updating application interface and corresponding control functions |
US10025462B1 (en) | 2014-08-29 | 2018-07-17 | Open Invention Network, Llc | Color based search application interface and corresponding control functions |
US10656788B1 (en) * | 2014-08-29 | 2020-05-19 | Open Invention Network Llc | Dynamic document updating application interface and corresponding control functions |
US10534500B1 (en) | 2014-08-29 | 2020-01-14 | Open Invention Network Llc | Color based search application interface and corresponding control functions |
KR20160029509A (en) * | 2014-09-05 | 2016-03-15 | 삼성전자주식회사 | Electronic apparatus and application executing method thereof |
US10347018B2 (en) | 2014-09-08 | 2019-07-09 | Tableau Software, Inc. | Interactive data visualization user interface with hierarchical filtering based on gesture location on a chart |
US11017569B2 (en) * | 2014-09-08 | 2021-05-25 | Tableau Software, Inc. | Methods and devices for displaying data mark information |
US10380770B2 (en) | 2014-09-08 | 2019-08-13 | Tableau Software, Inc. | Interactive data visualization user interface with multiple interaction profiles |
US10347027B2 (en) | 2014-09-08 | 2019-07-09 | Tableau Software, Inc. | Animated transition between data visualization versions at different levels of detail |
US10635262B2 (en) | 2014-09-08 | 2020-04-28 | Tableau Software, Inc. | Interactive data visualization user interface with gesture-based data field selection |
USD800758S1 (en) | 2014-09-23 | 2017-10-24 | Seasonal Specialties, Llc | Computer display screen with graphical user interface for lighting |
WO2016049907A1 (en) * | 2014-09-30 | 2016-04-07 | 深圳市大疆创新科技有限公司 | Operation interface processing method and display device |
US10108320B2 (en) * | 2014-10-08 | 2018-10-23 | Microsoft Technology Licensing, Llc | Multiple stage shy user interface |
EP3007050A1 (en) * | 2014-10-08 | 2016-04-13 | Volkswagen Aktiengesellschaft | User interface and method for adapting a menu bar on a user interface |
CN105573574A (en) * | 2014-10-09 | 2016-05-11 | 阿里巴巴集团控股有限公司 | Application interface navigation method and apparatus |
USD766261S1 (en) * | 2014-10-10 | 2016-09-13 | Salesforce.Com, Inc. | Display screen or portion thereof with animated graphical user interface |
USD768673S1 (en) | 2014-10-10 | 2016-10-11 | Salesforce.Com, Inc. | Display screen or portion thereof with animated graphical user interface |
US10949075B2 (en) | 2014-11-06 | 2021-03-16 | Microsoft Technology Licensing, Llc | Application command control for small screen display |
US9922098B2 (en) | 2014-11-06 | 2018-03-20 | Microsoft Technology Licensing, Llc | Context-based search and relevancy generation |
US10235130B2 (en) | 2014-11-06 | 2019-03-19 | Microsoft Technology Licensing, Llc | Intent driven command processing |
US20160132992A1 (en) | 2014-11-06 | 2016-05-12 | Microsoft Technology Licensing, Llc | User interface scaling for devices based on display size |
USD762722S1 (en) * | 2014-11-10 | 2016-08-02 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with transitional graphical user interface |
CN104407854A (en) * | 2014-11-13 | 2015-03-11 | 浪潮电子信息产业股份有限公司 | Method for displaying menu in color under DOS system |
USD788788S1 (en) | 2014-11-18 | 2017-06-06 | Google Inc. | Display screen with animated graphical user interface |
KR102390647B1 (en) * | 2014-11-25 | 2022-04-26 | 삼성전자주식회사 | Electronic device and method for controlling object in electronic device |
US20160148396A1 (en) * | 2014-11-26 | 2016-05-26 | Blackberry Limited | Method and Apparatus for Controlling Display of Mobile Communication Device |
US9619043B2 (en) * | 2014-11-26 | 2017-04-11 | At&T Intellectual Property I, L.P. | Gesture multi-function on a physical keyboard |
US20160147433A1 (en) * | 2014-11-26 | 2016-05-26 | General Electric Company | Reference command storage and pattern recognition for user interface improvement |
US20160162131A1 (en) * | 2014-12-08 | 2016-06-09 | Etzer Zamor | Social network |
USD759081S1 (en) * | 2014-12-11 | 2016-06-14 | Microsoft Corporation | Display screen with animated graphical user interface |
USD768702S1 (en) | 2014-12-19 | 2016-10-11 | Amazon Technologies, Inc. | Display screen or portion thereof with a graphical user interface |
EP3040831A1 (en) * | 2014-12-29 | 2016-07-06 | Dassault Systèmes | Setting a parameter |
EP3040838B1 (en) * | 2014-12-29 | 2021-04-21 | Dassault Systèmes | Setting a parameter |
EP3040832A1 (en) * | 2014-12-29 | 2016-07-06 | Dassault Systèmes | Setting a parameter |
US20160188171A1 (en) * | 2014-12-31 | 2016-06-30 | Microsoft Technology Licensing, Llc. | Split button with access to previously used options |
US10048839B2 (en) * | 2015-01-22 | 2018-08-14 | Flow Labs, Inc. | Hierarchy navigation in a user interface |
US9632664B2 (en) | 2015-03-08 | 2017-04-25 | Apple Inc. | Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback |
US10048757B2 (en) | 2015-03-08 | 2018-08-14 | Apple Inc. | Devices and methods for controlling media presentation |
US9990107B2 (en) * | 2015-03-08 | 2018-06-05 | Apple Inc. | Devices, methods, and graphical user interfaces for displaying and using menus |
US9645732B2 (en) | 2015-03-08 | 2017-05-09 | Apple Inc. | Devices, methods, and graphical user interfaces for displaying and using menus |
US10095396B2 (en) | 2015-03-08 | 2018-10-09 | Apple Inc. | Devices, methods, and graphical user interfaces for interacting with a control object while dragging another object |
DK201500581A1 (en) * | 2015-03-08 | 2017-01-16 | Apple Inc | Devices, Methods, and Graphical User Interfaces for Displaying and Using Menus |
US9639184B2 (en) | 2015-03-19 | 2017-05-02 | Apple Inc. | Touch input cursor manipulation |
US9785305B2 (en) | 2015-03-19 | 2017-10-10 | Apple Inc. | Touch input cursor manipulation |
FR3034218A1 (en) * | 2015-03-27 | 2016-09-30 | Orange | METHOD OF RAPID ACCESS TO APPLICATION FUNCTIONALITIES |
US20170045981A1 (en) | 2015-08-10 | 2017-02-16 | Apple Inc. | Devices and Methods for Processing Touch Inputs Based on Their Intensities |
US10152208B2 (en) | 2015-04-01 | 2018-12-11 | Apple Inc. | Devices and methods for processing touch inputs based on their intensities |
US9980304B2 (en) | 2015-04-03 | 2018-05-22 | Google Llc | Adaptive on-demand tethering |
JP1566076S (en) * | 2015-04-21 | 2016-12-26 | ||
JP1566075S (en) * | 2015-04-21 | 2016-12-26 | ||
USD783654S1 (en) * | 2015-04-21 | 2017-04-11 | Jingtao HU | Display screen with graphic user interface |
CN108353126B (en) | 2015-04-23 | 2019-08-23 | 苹果公司 | Handle method, electronic equipment and the computer readable storage medium of the content of camera |
US10200598B2 (en) | 2015-06-07 | 2019-02-05 | Apple Inc. | Devices and methods for capturing and interacting with enhanced digital images |
US9674426B2 (en) | 2015-06-07 | 2017-06-06 | Apple Inc. | Devices and methods for capturing and interacting with enhanced digital images |
US9891811B2 (en) | 2015-06-07 | 2018-02-13 | Apple Inc. | Devices and methods for navigating between user interfaces |
US9860451B2 (en) | 2015-06-07 | 2018-01-02 | Apple Inc. | Devices and methods for capturing and interacting with enhanced digital images |
US9830048B2 (en) | 2015-06-07 | 2017-11-28 | Apple Inc. | Devices and methods for processing touch inputs with instructions in a web page |
US10346030B2 (en) | 2015-06-07 | 2019-07-09 | Apple Inc. | Devices and methods for navigating between user interfaces |
US10082931B2 (en) * | 2015-06-12 | 2018-09-25 | Microsoft Technology Licensing, Llc | Transitioning command user interface between toolbar user interface and full menu user interface based on use context |
US9939923B2 (en) * | 2015-06-19 | 2018-04-10 | Microsoft Technology Licensing, Llc | Selecting events based on user input and current context |
US20160370974A1 (en) * | 2015-06-22 | 2016-12-22 | Here Global B.V. | Causation of Expansion of a Supplemental Content Overlay |
US10296168B2 (en) * | 2015-06-25 | 2019-05-21 | Northrop Grumman Systems Corporation | Apparatus and method for a multi-step selection interface |
KR20170011583A (en) * | 2015-07-23 | 2017-02-02 | 삼성전자주식회사 | Operating Method For Contents Searching Function and electronic device supporting the same |
US9880735B2 (en) | 2015-08-10 | 2018-01-30 | Apple Inc. | Devices, methods, and graphical user interfaces for manipulating user interface objects with visual and/or haptic feedback |
US10248308B2 (en) | 2015-08-10 | 2019-04-02 | Apple Inc. | Devices, methods, and graphical user interfaces for manipulating user interfaces with physical gestures |
US10416800B2 (en) | 2015-08-10 | 2019-09-17 | Apple Inc. | Devices, methods, and graphical user interfaces for adjusting user interface objects |
US10235035B2 (en) | 2015-08-10 | 2019-03-19 | Apple Inc. | Devices, methods, and graphical user interfaces for content navigation and manipulation |
US10896532B2 (en) | 2015-09-08 | 2021-01-19 | Tableau Software, Inc. | Interactive data visualization user interface with multiple interaction profiles |
US20170068414A1 (en) * | 2015-09-09 | 2017-03-09 | Microsoft Technology Licensing, Llc | Controlling a device |
US9671244B2 (en) | 2015-09-18 | 2017-06-06 | Les Solutions Cyclelabs Inc. | Electronic device and method for providing travel information |
KR102354329B1 (en) * | 2015-10-08 | 2022-01-21 | 삼성전자주식회사 | ELECTRONIC DEVICE AND METHOD FOR DISPLAYING a plurality of items |
WO2017061702A1 (en) * | 2015-10-08 | 2017-04-13 | Samsung Electronics Co., Ltd. | Electronic device and method of displaying plurality of items |
US20170109026A1 (en) * | 2015-10-16 | 2017-04-20 | Hewlett Packard Enterprise Development Lp | Dial control for touch screen navigation |
US10386997B2 (en) * | 2015-10-23 | 2019-08-20 | Sap Se | Integrating functions for a user input device |
USD801986S1 (en) * | 2015-12-04 | 2017-11-07 | Airbus Operations Gmbh | Display screen or portion thereof with graphical user interface |
US10310618B2 (en) * | 2015-12-31 | 2019-06-04 | Microsoft Technology Licensing, Llc | Gestures visual builder tool |
US10599324B2 (en) | 2015-12-31 | 2020-03-24 | Microsoft Technology Licensing, Llc | Hand gesture API using finite state machine and gesture language discrete values |
US10831337B2 (en) * | 2016-01-05 | 2020-11-10 | Apple Inc. | Device, method, and graphical user interface for a radial menu system |
US10901573B2 (en) * | 2016-02-05 | 2021-01-26 | Airwatch Llc | Generating predictive action buttons within a graphical user interface |
US10514826B2 (en) | 2016-02-08 | 2019-12-24 | Microsoft Technology Licensing, Llc | Contextual command bar |
US20170255455A1 (en) * | 2016-03-03 | 2017-09-07 | International Business Machines Corporation | Automated customization of software feature availability based on usage patterns and history |
USD788166S1 (en) * | 2016-03-07 | 2017-05-30 | Facebook, Inc. | Display screen with animated graphical user interface |
KR20170108340A (en) * | 2016-03-17 | 2017-09-27 | 삼성전자주식회사 | A display apparatus and a method for operating in a display apparatus |
EP3223130A1 (en) * | 2016-03-22 | 2017-09-27 | Continental Automotive GmbH | Method of controlling an input device for navigating a hierarchical menu |
USD811420S1 (en) * | 2016-04-01 | 2018-02-27 | Google Llc | Display screen portion with a transitional graphical user interface component |
FR3050293A1 (en) * | 2016-04-18 | 2017-10-20 | Orange | METHOD FOR AUDIO ASSISTANCE OF TERMINAL CONTROL INTERFACE, PROGRAM AND TERMINAL |
KR102485448B1 (en) * | 2016-04-20 | 2023-01-06 | 삼성전자주식회사 | Electronic device and method for processing gesture input |
US20170315704A1 (en) * | 2016-05-02 | 2017-11-02 | Microsoft Technology Licensing, Llc | Application user interfaces with scrollable color palettes |
USD810755S1 (en) * | 2016-05-20 | 2018-02-20 | Quantum Interface, Llc | Display screen or portion thereof with graphical user interface |
USD832289S1 (en) * | 2016-05-30 | 2018-10-30 | Compal Electronics, Inc. | Portion of a display screen with icon |
USD814499S1 (en) * | 2016-06-01 | 2018-04-03 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
US9912860B2 (en) | 2016-06-12 | 2018-03-06 | Apple Inc. | User interface for camera effects |
USD794065S1 (en) * | 2016-06-17 | 2017-08-08 | Google Inc. | Display screen with an animated graphical user interface |
US10365822B2 (en) * | 2016-06-20 | 2019-07-30 | Dell Products L.P. | Information handling system multi-handed hybrid interface devices |
US10049087B2 (en) | 2016-07-19 | 2018-08-14 | International Business Machines Corporation | User-defined context-aware text selection for touchscreen devices |
US10409487B2 (en) | 2016-08-23 | 2019-09-10 | Microsoft Technology Licensing, Llc | Application processing based on gesture input |
USD816688S1 (en) * | 2016-08-26 | 2018-05-01 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
USD817340S1 (en) | 2016-08-26 | 2018-05-08 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
USD835143S1 (en) * | 2016-08-26 | 2018-12-04 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
USD817341S1 (en) * | 2016-08-26 | 2018-05-08 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with transitional graphical user interface |
USD870738S1 (en) * | 2016-08-30 | 2019-12-24 | Verizon Patent And Licensing Inc. | Display panel or screen with graphical user interface |
USD802624S1 (en) * | 2016-10-03 | 2017-11-14 | Salesforce.Com, Inc. | Display screen or portion thereof with animated graphical user interface |
CN107977138A (en) * | 2016-10-24 | 2018-05-01 | 北京东软医疗设备有限公司 | A kind of display methods and device |
US10440794B2 (en) | 2016-11-02 | 2019-10-08 | LIFI Labs, Inc. | Lighting system and method |
USD812093S1 (en) * | 2016-12-02 | 2018-03-06 | Salesforce.Com, Inc. | Display screen or portion thereof with graphical user interface |
US10775985B2 (en) * | 2016-12-29 | 2020-09-15 | Konica Minolta Laboratory U.S.A., Inc. | Dialog transparency adjustability |
US11226734B1 (en) | 2016-12-30 | 2022-01-18 | Dassault Systemes Solidworks Corporation | Triggering multiple actions from a single gesture |
USD824405S1 (en) * | 2017-01-13 | 2018-07-31 | Adp, Llc | Display screen or portion thereof with a graphical user interface |
USD840428S1 (en) * | 2017-01-13 | 2019-02-12 | Adp, Llc | Display screen with a graphical user interface |
CN108536273B (en) * | 2017-03-01 | 2024-10-18 | 深圳巧牛科技有限公司 | Man-machine menu interaction method and system based on gestures |
CN107066173B (en) * | 2017-03-28 | 2018-06-05 | 腾讯科技(深圳)有限公司 | Method of controlling operation thereof and device |
USD916712S1 (en) * | 2017-04-21 | 2021-04-20 | Scott Bickford | Display screen with an animated graphical user interface having a transitional flower design icon |
US10168879B1 (en) | 2017-05-12 | 2019-01-01 | Snap Inc. | Interactive image recoloring |
JP6914728B2 (en) * | 2017-05-26 | 2021-08-04 | キヤノン株式会社 | Communication equipment, communication methods, and programs |
US10949222B2 (en) | 2017-05-30 | 2021-03-16 | Citrix Systems, Inc. | System and method for displaying customized user guides in a virtual client application |
CN107422938A (en) * | 2017-06-21 | 2017-12-01 | 网易(杭州)网络有限公司 | Information processing method, device, electronic equipment and storage medium |
GB2567130B (en) * | 2017-07-25 | 2022-11-30 | Tesla Engineering Ltd | Cryostat arrangements and mounting arrangements for cryostats |
US11237699B2 (en) * | 2017-08-18 | 2022-02-01 | Microsoft Technology Licensing, Llc | Proximal menu generation |
US11301124B2 (en) | 2017-08-18 | 2022-04-12 | Microsoft Technology Licensing, Llc | User interface modification using preview panel |
WO2019038774A1 (en) * | 2017-08-20 | 2019-02-28 | Rolllo Ltd | Systems and methods for providing single touch graphical user interface in computerized devices |
USD846585S1 (en) * | 2017-08-22 | 2019-04-23 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with graphical user interface |
US20190095052A1 (en) * | 2017-09-27 | 2019-03-28 | Fomtech Limited | User Interface Elements for Compact Menu |
USD901522S1 (en) * | 2017-09-27 | 2020-11-10 | Toyota Research Institute, Inc. | Vehicle heads-up display screen or portion thereof with a graphical user interface |
CN107807823A (en) * | 2017-10-30 | 2018-03-16 | 江西博瑞彤芸科技有限公司 | A kind of generation method of shortcut menu |
AU2017439358B2 (en) | 2017-11-10 | 2023-11-16 | Razer (Asia-Pacific) Pte. Ltd. | Machine learning zero latency input device |
USD875742S1 (en) * | 2017-11-22 | 2020-02-18 | Samsung Electronics Co., Ltd. | Display screen or portion thereof with transitional graphical user interface |
USD857749S1 (en) * | 2017-12-01 | 2019-08-27 | Agco Corporation | Display screen or portion thereof with graphical user interface |
USD884714S1 (en) * | 2018-01-12 | 2020-05-19 | Delta Electronics, Inc. | Display screen with graphical user interface |
USD845332S1 (en) * | 2018-02-06 | 2019-04-09 | Krikey, Inc. | Display panel of a programmed computer system with a graphical user interface |
US11084408B2 (en) * | 2018-03-30 | 2021-08-10 | Honda Motor Co., Ltd. | Dedicated massage function button for vehicle |
USD916888S1 (en) * | 2018-04-20 | 2021-04-20 | Samsung Display Co., Ltd. | Smartphone display screen with graphical user interface |
KR101972264B1 (en) * | 2018-05-14 | 2019-04-24 | 이성만 | Method and apparatus for providing reward using icon |
US11770601B2 (en) | 2019-05-06 | 2023-09-26 | Apple Inc. | User interfaces for capturing and managing visual media |
WO2020080650A1 (en) | 2018-10-16 | 2020-04-23 | Samsung Electronics Co., Ltd. | Apparatus and method of operating wearable device |
US20200133645A1 (en) * | 2018-10-30 | 2020-04-30 | Jpmorgan Chase Bank, N.A. | User interface and front end application automatic generation |
CN111752444A (en) * | 2019-03-29 | 2020-10-09 | 杭州海康威视数字技术股份有限公司 | Knocking event detection method and device |
USD896245S1 (en) * | 2019-04-01 | 2020-09-15 | Sg Gaming, Inc. | Display screen with animated graphical user interface |
USD896244S1 (en) * | 2019-04-01 | 2020-09-15 | Sg Gaming, Inc. | Display screen with transitional grapical user interface |
USD896246S1 (en) * | 2019-04-01 | 2020-09-15 | Sg Gaming, Inc. | Display screen with animated graphical user interface |
CN111857897A (en) * | 2019-04-25 | 2020-10-30 | 北京小米移动软件有限公司 | Information display method and device and storage medium |
USD921506S1 (en) | 2019-04-26 | 2021-06-08 | SmartHalo Technologies Inc. | Electronic device for providing travel information |
US11706521B2 (en) | 2019-05-06 | 2023-07-18 | Apple Inc. | User interfaces for capturing and managing visual media |
US11379104B2 (en) * | 2019-06-07 | 2022-07-05 | Microsoft Technology Licensing, Llc | Sharing user interface customization across applications |
US11550540B2 (en) * | 2019-08-15 | 2023-01-10 | Lenovo (Singapore) Pte. Ltd. | Content input selection and switching |
CN112492365B (en) | 2019-09-11 | 2024-06-14 | 新加坡商欧之遥控有限公司 | Remote controller navigation interface assembly |
USD923021S1 (en) * | 2019-09-13 | 2021-06-22 | The Marsden Group | Display screen or a portion thereof with an animated graphical user interface |
US10866721B1 (en) * | 2019-09-20 | 2020-12-15 | Valve Corporation | Selecting properties using handheld controllers |
US11429687B2 (en) | 2019-10-10 | 2022-08-30 | Kyndryl, Inc. | Context based URL resource prediction and delivery |
USD914710S1 (en) * | 2019-10-31 | 2021-03-30 | Eli Lilly And Company | Display screen with a graphical user interface |
WO2021104919A1 (en) | 2019-11-26 | 2021-06-03 | Signify Holding B.V. | Method and system for filtering information in a remotely managed lighting system |
US10908811B1 (en) * | 2019-12-17 | 2021-02-02 | Dell Products, L.P. | System and method for improving a graphical menu |
US20210240339A1 (en) * | 2020-01-31 | 2021-08-05 | Salesforce.Com, Inc. | Unified hover implementation for touch screen interfaces |
TWD210778S (en) * | 2020-05-06 | 2021-04-01 | 宏碁股份有限公司 | Graphical user interface for a display screen |
CN111752438B (en) * | 2020-06-29 | 2022-01-25 | 高新兴科技集团股份有限公司 | Method for displaying mobile terminal multi-trigger update dynamic label |
USD1009070S1 (en) | 2020-07-10 | 2023-12-26 | Schlumberger Technology Corporation | Electronic device with display screen and graphical user interface |
USD1006820S1 (en) | 2020-07-10 | 2023-12-05 | Schlumberger Technology Corporation | Electronic device with display screen and graphical user interface |
USD983810S1 (en) | 2020-07-10 | 2023-04-18 | Schlumberger Technology Corporation | Electronic device with display screen and graphical user interface |
USD984464S1 (en) * | 2020-12-22 | 2023-04-25 | Google Llc | Display screen or portion thereof with graphical user interface |
US11409432B2 (en) | 2020-12-23 | 2022-08-09 | Microsoft Technology Licensing, Llc | Pen command for ink editing |
US11995297B2 (en) * | 2021-03-08 | 2024-05-28 | Samsung Electronics Co., Ltd. | Enhanced user interface (UI) button control for mobile applications |
USD971938S1 (en) * | 2021-03-11 | 2022-12-06 | Salesforce, Inc. | Display screen or portion thereof with graphical user interface |
US11949638B1 (en) | 2023-03-04 | 2024-04-02 | Unanimous A. I., Inc. | Methods and systems for hyperchat conversations among large networked populations with collective intelligence amplification |
USD1046896S1 (en) * | 2023-03-12 | 2024-10-15 | Beijing Zitiao Network Technology Co., Ltd. | Display screen or portion thereof with a graphical user interface |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5392388A (en) * | 1992-12-04 | 1995-02-21 | International Business Machines Corporation | Method and system for viewing graphic images in a data processing system |
US6448987B1 (en) * | 1998-04-03 | 2002-09-10 | Intertainer, Inc. | Graphic user interface for a digital content delivery system using circular menus |
US20050285965A1 (en) * | 2004-06-24 | 2005-12-29 | Apple Computer, Inc. | User-interface design |
US20060092177A1 (en) * | 2004-10-30 | 2006-05-04 | Gabor Blasko | Input method and apparatus using tactile guidance and bi-directional segmented stroke |
US20080143681A1 (en) * | 2006-12-18 | 2008-06-19 | Xiaoping Jiang | Circular slider with center button |
US7735020B2 (en) * | 2002-03-15 | 2010-06-08 | Apple Inc. | Method and apparatus for determining font attributes |
US20100185983A1 (en) * | 2009-01-16 | 2010-07-22 | Corel Corporation | Curved Slider Control |
US7853877B2 (en) * | 2001-06-29 | 2010-12-14 | Microsoft Corporation | Gallery user interface controls |
US20110013203A1 (en) * | 2009-07-20 | 2011-01-20 | Aryk Erwin Grosz | Method for Determining Effective Core Aspect Ratio for Display of Content Created in an Online Collage-Based Editor |
US20110157046A1 (en) * | 2009-12-30 | 2011-06-30 | Seonmi Lee | Display device for a mobile terminal and method of controlling the same |
US20110246943A1 (en) * | 2008-12-18 | 2011-10-06 | Toshihiko Fujibayashi | Slide bar display control device and slide bar display control method |
US8040142B1 (en) * | 2006-03-31 | 2011-10-18 | Cypress Semiconductor Corporation | Touch detection techniques for capacitive touch sense systems |
US20120036480A1 (en) * | 2010-08-09 | 2012-02-09 | Peter Warner | Two-dimensional slider control |
US8239784B2 (en) * | 2004-07-30 | 2012-08-07 | Apple Inc. | Mode-based graphical user interfaces for touch sensitive input devices |
US8578294B2 (en) * | 2008-01-11 | 2013-11-05 | Sungkyunkwan University Foundation For Corporate Collaboration | Menu user interface providing device and method thereof |
Family Cites Families (101)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
USD341848S (en) * | 1991-12-09 | 1993-11-30 | Microsoft Corporation | Typeface |
US6259446B1 (en) | 1992-12-23 | 2001-07-10 | Object Technology Licensing Corporation | Menu state system |
US5615320A (en) * | 1994-04-25 | 1997-03-25 | Canon Information Systems, Inc. | Computer-aided color selection and colorizing system using objective-based coloring criteria |
US6281879B1 (en) | 1994-06-16 | 2001-08-28 | Microsoft Corporation | Timing and velocity control for displaying graphical information |
US5798760A (en) | 1995-06-07 | 1998-08-25 | Vayda; Mark | Radial graphical menuing system with concentric region menuing |
AUPN360195A0 (en) * | 1995-06-16 | 1995-07-13 | Canon Information Systems Research Australia Pty Ltd | Colour selection tool |
US6073036A (en) * | 1997-04-28 | 2000-06-06 | Nokia Mobile Phones Limited | Mobile station with touch input having automatic symbol magnification function |
US7256770B2 (en) | 1998-09-14 | 2007-08-14 | Microsoft Corporation | Method for displaying information responsive to sensing a physical presence proximate to a computer input device |
US6501491B1 (en) | 1998-09-21 | 2002-12-31 | Microsoft Corporation | Extensible user interface for viewing objects over a network |
GB2342196A (en) * | 1998-09-30 | 2000-04-05 | Xerox Corp | System for generating context-sensitive hierarchically-ordered document service menus |
US6259436B1 (en) * | 1998-12-22 | 2001-07-10 | Ericsson Inc. | Apparatus and method for determining selection of touchable items on a computer touchscreen by an imprecise touch |
JP2001075712A (en) * | 1999-08-31 | 2001-03-23 | Sony Corp | Information processor, its method and program storage medium |
US7152207B1 (en) * | 1999-11-05 | 2006-12-19 | Decentrix Inc. | Method and apparatus for providing conditional customization for generating a web site |
US20020156870A1 (en) * | 2000-11-08 | 2002-10-24 | Equate Systems, Inc. | Method and apparatus for dynamically directing an application to a pre-defined target multimedia resource |
US6918091B2 (en) * | 2000-11-09 | 2005-07-12 | Change Tools, Inc. | User definable interface system, method and computer program product |
US6925611B2 (en) | 2001-01-31 | 2005-08-02 | Microsoft Corporation | Navigational interface for mobile and wearable computers |
US6828988B2 (en) | 2001-02-27 | 2004-12-07 | Microsoft Corporation | Interactive tooltip |
US7415671B2 (en) * | 2001-06-08 | 2008-08-19 | Computer Associates Think, Inc. | Interactive hierarchical status display |
CA2357969A1 (en) * | 2001-09-28 | 2003-03-28 | Dirk Alexander Seelemann | Customazation of object property layout for a user interface |
US7345671B2 (en) | 2001-10-22 | 2008-03-18 | Apple Inc. | Method and apparatus for use of rotational user inputs |
US6941521B2 (en) * | 2002-03-29 | 2005-09-06 | Intel Corporation | Method for dynamically generating a user interface from XML-based documents |
US7570943B2 (en) | 2002-08-29 | 2009-08-04 | Nokia Corporation | System and method for providing context sensitive recommendations to digital services |
US7180524B1 (en) * | 2002-09-30 | 2007-02-20 | Dale Axelrod | Artists' color display system |
US20040113941A1 (en) * | 2002-12-12 | 2004-06-17 | Xerox Corporation | User interface customization |
US7895536B2 (en) * | 2003-01-08 | 2011-02-22 | Autodesk, Inc. | Layer editor system for a pen-based computer |
US7158123B2 (en) * | 2003-01-31 | 2007-01-02 | Xerox Corporation | Secondary touch contextual sub-menu navigation for touch screen interface |
US7103852B2 (en) * | 2003-03-10 | 2006-09-05 | International Business Machines Corporation | Dynamic resizing of clickable areas of touch screen applications |
US20040212601A1 (en) * | 2003-04-24 | 2004-10-28 | Anthony Cake | Method and apparatus for improving accuracy of touch screen input devices |
US7827495B2 (en) * | 2003-09-02 | 2010-11-02 | Research In Motion Limited | Method and data structure for user interface customization |
KR100739682B1 (en) * | 2003-10-04 | 2007-07-13 | 삼성전자주식회사 | Information storage medium storing text based sub-title, processing apparatus and method thereof |
US7480863B2 (en) * | 2003-11-26 | 2009-01-20 | International Business Machines Corporation | Dynamic and intelligent hover assistance |
US20050216834A1 (en) | 2004-03-29 | 2005-09-29 | Microsoft Corporation | Method, apparatus, and computer-readable medium for dynamically rendering a user interface menu |
US20070192711A1 (en) * | 2006-02-13 | 2007-08-16 | Research In Motion Limited | Method and arrangement for providing a primary actions menu on a handheld communication device |
US7653883B2 (en) * | 2004-07-30 | 2010-01-26 | Apple Inc. | Proximity detector in handheld device |
US7895531B2 (en) | 2004-08-16 | 2011-02-22 | Microsoft Corporation | Floating command object |
US7712049B2 (en) | 2004-09-30 | 2010-05-04 | Microsoft Corporation | Two-dimensional radial user interface for computer software applications |
US20060073814A1 (en) | 2004-10-05 | 2006-04-06 | International Business Machines Corporation | Embedded specification of menu navigation for mobile devices |
RU2004133946A (en) | 2004-11-23 | 2006-05-10 | Самсунг Электроникс Ко., Лтд. (KR) | METHOD FOR ADAPTIVE CONTEXT HELP FORMATION |
US8019843B2 (en) * | 2005-05-24 | 2011-09-13 | CRIF Corporation | System and method for defining attributes, decision rules, or both, for remote execution, claim set II |
US7661074B2 (en) * | 2005-07-01 | 2010-02-09 | Microsoft Corporation | Keyboard accelerator |
US8239882B2 (en) | 2005-08-30 | 2012-08-07 | Microsoft Corporation | Markup based extensibility for user interfaces |
US7739612B2 (en) * | 2005-09-12 | 2010-06-15 | Microsoft Corporation | Blended editing of literal and non-literal values |
US10983695B2 (en) * | 2005-10-24 | 2021-04-20 | Kinoma, Inc. | Focus management system |
US20080059504A1 (en) * | 2005-11-30 | 2008-03-06 | Jackie Barbetta | Method and system for rendering graphical user interface |
US7730425B2 (en) * | 2005-11-30 | 2010-06-01 | De Los Reyes Isabelo | Function-oriented user interface |
US7603633B2 (en) | 2006-01-13 | 2009-10-13 | Microsoft Corporation | Position-based multi-stroke marking menus |
US7644372B2 (en) | 2006-01-27 | 2010-01-05 | Microsoft Corporation | Area frequency radial menus |
TWI328185B (en) * | 2006-04-19 | 2010-08-01 | Lg Electronics Inc | Touch screen device for potable terminal and method of displaying and selecting menus thereon |
JP2007293132A (en) * | 2006-04-26 | 2007-11-08 | Pioneer Electronic Corp | Mobile information input/output device and general purpose braille output device |
US20070256029A1 (en) | 2006-05-01 | 2007-11-01 | Rpo Pty Llimited | Systems And Methods For Interfacing A User With A Touch-Screen |
EP1860534A1 (en) | 2006-05-22 | 2007-11-28 | LG Electronics Inc. | Mobile terminal and menu display method thereof |
US7509348B2 (en) * | 2006-08-31 | 2009-03-24 | Microsoft Corporation | Radially expanding and context-dependent navigation dial |
USD563972S1 (en) | 2006-10-25 | 2008-03-11 | Microsoft Corporation | User interface for a portion of a display screen |
US7873710B2 (en) | 2007-02-06 | 2011-01-18 | 5O9, Inc. | Contextual data communication platform |
US8302006B2 (en) * | 2007-02-28 | 2012-10-30 | Rockwell Automation Technologies, Inc. | Interactive tooltip to display and navigate to different resources of a data point |
US8650505B2 (en) | 2007-02-28 | 2014-02-11 | Rpx Corporation | Multi-state unified pie user interface |
US8352881B2 (en) * | 2007-03-08 | 2013-01-08 | International Business Machines Corporation | Method, apparatus and program storage device for providing customizable, immediate and radiating menus for accessing applications and actions |
US20080228853A1 (en) | 2007-03-15 | 2008-09-18 | Kayxo Dk A/S | Software system |
US8059101B2 (en) * | 2007-06-22 | 2011-11-15 | Apple Inc. | Swipe gestures for touch screen keyboards |
US8645863B2 (en) * | 2007-06-29 | 2014-02-04 | Microsoft Corporation | Menus with translucency and live preview |
US8869065B2 (en) * | 2007-06-29 | 2014-10-21 | Microsoft Corporation | Segment ring menu |
US20090037813A1 (en) | 2007-07-31 | 2009-02-05 | Palo Alto Research Center Incorporated | Space-constrained marking menus for mobile devices |
US7941758B2 (en) * | 2007-09-04 | 2011-05-10 | Apple Inc. | Animation of graphical objects |
US8060818B2 (en) * | 2007-12-14 | 2011-11-15 | Sap Ag | Method and apparatus for form adaptation |
US20090160768A1 (en) * | 2007-12-21 | 2009-06-25 | Nvidia Corporation | Enhanced Presentation Capabilities Using a Pointer Implement |
US7941765B2 (en) * | 2008-01-23 | 2011-05-10 | Wacom Co., Ltd | System and method of controlling variables using a radial control menu |
US8120616B2 (en) * | 2008-02-27 | 2012-02-21 | Autodesk, Inc. | Color sampler |
US20090234716A1 (en) * | 2008-03-17 | 2009-09-17 | Photometria, Inc. | Method of monetizing online personal beauty product selections |
US8826181B2 (en) | 2008-06-28 | 2014-09-02 | Apple Inc. | Moving radial menus |
US8245156B2 (en) * | 2008-06-28 | 2012-08-14 | Apple Inc. | Radial menu selection |
WO2010041092A1 (en) * | 2008-10-07 | 2010-04-15 | Tiki'labs | Method and device for controlling an inputting data |
US20100185985A1 (en) * | 2009-01-19 | 2010-07-22 | International Business Machines Corporation | Managing radial menus in a computer system |
US20100192102A1 (en) | 2009-01-29 | 2010-07-29 | International Business Machines Corporation | Displaying radial menus near edges of a display area |
JP2010224658A (en) * | 2009-03-19 | 2010-10-07 | Smk Corp | Operation input device |
JP5553673B2 (en) * | 2009-05-11 | 2014-07-16 | キヤノン株式会社 | Imaging apparatus and display control method |
US9436380B2 (en) | 2009-05-19 | 2016-09-06 | International Business Machines Corporation | Radial menus with variable selectable item areas |
US8375295B2 (en) * | 2009-05-21 | 2013-02-12 | Sony Computer Entertainment Inc. | Customization of GUI layout based on history of use |
US8418165B2 (en) * | 2009-05-27 | 2013-04-09 | Microsoft Corporation | Package design and generation |
US8549432B2 (en) | 2009-05-29 | 2013-10-01 | Apple Inc. | Radial menus |
US8219930B2 (en) | 2009-06-26 | 2012-07-10 | Verizon Patent And Licensing Inc. | Radial menu display systems and methods |
US9213466B2 (en) * | 2009-07-20 | 2015-12-15 | Apple Inc. | Displaying recently used functions in context sensitive menu |
US8375329B2 (en) * | 2009-09-01 | 2013-02-12 | Maxon Computer Gmbh | Method of providing a graphical user interface using a concentric menu |
US8578295B2 (en) | 2009-09-16 | 2013-11-05 | International Business Machines Corporation | Placement of items in cascading radial menus |
US20130047123A1 (en) * | 2009-09-24 | 2013-02-21 | Ringguides Inc. | Method for presenting user-defined menu of digital content choices, organized as ring of icons surrounding preview pane |
CA2680602C (en) | 2009-10-19 | 2011-07-26 | Ibm Canada Limited - Ibm Canada Limitee | System and method for generating and displaying hybrid context menus |
US20110191711A1 (en) | 2010-02-04 | 2011-08-04 | Gill George M | Customer and vehicle dynamic grouping |
EP2360570A3 (en) | 2010-02-15 | 2012-05-16 | Research In Motion Limited | Graphical context short menu |
US9367205B2 (en) | 2010-02-19 | 2016-06-14 | Microsoft Technolgoy Licensing, Llc | Radial menus with bezel gestures |
US8818027B2 (en) * | 2010-04-01 | 2014-08-26 | Qualcomm Incorporated | Computing device interface |
US9542038B2 (en) * | 2010-04-07 | 2017-01-10 | Apple Inc. | Personalizing colors of user interfaces |
US8689110B2 (en) * | 2010-04-08 | 2014-04-01 | Oracle International Corporation | Multi-channel user interface architecture |
US20110248928A1 (en) | 2010-04-08 | 2011-10-13 | Motorola, Inc. | Device and method for gestural operation of context menus on a touch-sensitive display |
US8591334B2 (en) * | 2010-06-03 | 2013-11-26 | Ol2, Inc. | Graphical user interface, system and method for implementing a game controller on a touch-screen device |
US20120030624A1 (en) * | 2010-07-30 | 2012-02-02 | Migos Charles J | Device, Method, and Graphical User Interface for Displaying Menus |
DE102010036906A1 (en) * | 2010-08-06 | 2012-02-09 | Tavendo Gmbh | Configurable pie menu |
US9639265B2 (en) * | 2010-09-03 | 2017-05-02 | Microsoft Technology Licensing, Llc | Distance-time based hit-testing for displayed target graphical elements |
US20120124472A1 (en) * | 2010-11-15 | 2012-05-17 | Opera Software Asa | System and method for providing interactive feedback for mouse gestures |
US9645986B2 (en) * | 2011-02-24 | 2017-05-09 | Google Inc. | Method, medium, and system for creating an electronic book with an umbrella policy |
US20120218282A1 (en) * | 2011-02-25 | 2012-08-30 | Research In Motion Limited | Display Brightness Adjustment |
US9513799B2 (en) * | 2011-06-05 | 2016-12-06 | Apple Inc. | Devices, methods, and graphical user interfaces for providing control of a touch-based user interface absent physical touch capabilities |
US9086794B2 (en) * | 2011-07-14 | 2015-07-21 | Microsoft Technology Licensing, Llc | Determining gestures on context based menus |
-
2011
- 2011-12-29 US US13/339,569 patent/US9086794B2/en active Active
- 2011-12-30 US US13/341,074 patent/US9250766B2/en active Active
-
2012
- 2012-01-13 US US13/349,691 patent/US9021398B2/en active Active
- 2012-07-06 US US13/542,962 patent/US20130019204A1/en not_active Abandoned
- 2012-07-09 US US13/543,976 patent/US20130019208A1/en not_active Abandoned
- 2012-07-13 US US13/549,397 patent/US9116602B2/en active Active
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5392388A (en) * | 1992-12-04 | 1995-02-21 | International Business Machines Corporation | Method and system for viewing graphic images in a data processing system |
US6448987B1 (en) * | 1998-04-03 | 2002-09-10 | Intertainer, Inc. | Graphic user interface for a digital content delivery system using circular menus |
US7853877B2 (en) * | 2001-06-29 | 2010-12-14 | Microsoft Corporation | Gallery user interface controls |
US7735020B2 (en) * | 2002-03-15 | 2010-06-08 | Apple Inc. | Method and apparatus for determining font attributes |
US20050285965A1 (en) * | 2004-06-24 | 2005-12-29 | Apple Computer, Inc. | User-interface design |
US8239784B2 (en) * | 2004-07-30 | 2012-08-07 | Apple Inc. | Mode-based graphical user interfaces for touch sensitive input devices |
US20060092177A1 (en) * | 2004-10-30 | 2006-05-04 | Gabor Blasko | Input method and apparatus using tactile guidance and bi-directional segmented stroke |
US8040142B1 (en) * | 2006-03-31 | 2011-10-18 | Cypress Semiconductor Corporation | Touch detection techniques for capacitive touch sense systems |
US20080143681A1 (en) * | 2006-12-18 | 2008-06-19 | Xiaoping Jiang | Circular slider with center button |
US8578294B2 (en) * | 2008-01-11 | 2013-11-05 | Sungkyunkwan University Foundation For Corporate Collaboration | Menu user interface providing device and method thereof |
US20110246943A1 (en) * | 2008-12-18 | 2011-10-06 | Toshihiko Fujibayashi | Slide bar display control device and slide bar display control method |
US20100185983A1 (en) * | 2009-01-16 | 2010-07-22 | Corel Corporation | Curved Slider Control |
US20110013203A1 (en) * | 2009-07-20 | 2011-01-20 | Aryk Erwin Grosz | Method for Determining Effective Core Aspect Ratio for Display of Content Created in an Online Collage-Based Editor |
US20110157046A1 (en) * | 2009-12-30 | 2011-06-30 | Seonmi Lee | Display device for a mobile terminal and method of controlling the same |
US20120036480A1 (en) * | 2010-08-09 | 2012-02-09 | Peter Warner | Two-dimensional slider control |
Cited By (114)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10268367B2 (en) | 2010-02-19 | 2019-04-23 | Microsoft Technology Licensing, Llc | Radial menus with bezel gestures |
US20130076642A1 (en) * | 2011-09-22 | 2013-03-28 | Microsoft Corporation | User interface for editing a value in place |
US10133466B2 (en) | 2011-09-22 | 2018-11-20 | Microsoct Technology Licensing, LLC | User interface for editing a value in place |
US9720583B2 (en) * | 2011-09-22 | 2017-08-01 | Microsoft Technology Licensing, Llc | User interface for editing a value in place |
US10691219B2 (en) | 2012-01-17 | 2020-06-23 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US9495613B2 (en) | 2012-01-17 | 2016-11-15 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging using formed difference images |
US9767345B2 (en) | 2012-01-17 | 2017-09-19 | Leap Motion, Inc. | Systems and methods of constructing three-dimensional (3D) model of an object using image cross-sections |
US10565784B2 (en) | 2012-01-17 | 2020-02-18 | Ultrahaptics IP Two Limited | Systems and methods for authenticating a user according to a hand of the user moving in a three-dimensional (3D) space |
US11720180B2 (en) | 2012-01-17 | 2023-08-08 | Ultrahaptics IP Two Limited | Systems and methods for machine control |
US11308711B2 (en) | 2012-01-17 | 2022-04-19 | Ultrahaptics IP Two Limited | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US10410411B2 (en) | 2012-01-17 | 2019-09-10 | Leap Motion, Inc. | Systems and methods of object shape and position determination in three-dimensional (3D) space |
US10366308B2 (en) | 2012-01-17 | 2019-07-30 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US9436998B2 (en) | 2012-01-17 | 2016-09-06 | Leap Motion, Inc. | Systems and methods of constructing three-dimensional (3D) model of an object using image cross-sections |
US10699155B2 (en) | 2012-01-17 | 2020-06-30 | Ultrahaptics IP Two Limited | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US11994377B2 (en) | 2012-01-17 | 2024-05-28 | Ultrahaptics IP Two Limited | Systems and methods of locating a control object appendage in three dimensional (3D) space |
US9626591B2 (en) | 2012-01-17 | 2017-04-18 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging |
US9741136B2 (en) | 2012-01-17 | 2017-08-22 | Leap Motion, Inc. | Systems and methods of object shape and position determination in three-dimensional (3D) space |
US9652668B2 (en) | 2012-01-17 | 2017-05-16 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US9672441B2 (en) | 2012-01-17 | 2017-06-06 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US9679215B2 (en) | 2012-01-17 | 2017-06-13 | Leap Motion, Inc. | Systems and methods for machine control |
US9934580B2 (en) | 2012-01-17 | 2018-04-03 | Leap Motion, Inc. | Enhanced contrast for object detection and characterization by optical imaging based on differences between images |
US9697643B2 (en) | 2012-01-17 | 2017-07-04 | Leap Motion, Inc. | Systems and methods of object shape and position determination in three-dimensional (3D) space |
US9778752B2 (en) | 2012-01-17 | 2017-10-03 | Leap Motion, Inc. | Systems and methods for machine control |
US20140006033A1 (en) * | 2012-06-29 | 2014-01-02 | Samsung Electronics Co., Ltd. | Method and apparatus for processing multiple inputs |
US9286895B2 (en) * | 2012-06-29 | 2016-03-15 | Samsung Electronics Co., Ltd. | Method and apparatus for processing multiple inputs |
US9285893B2 (en) | 2012-11-08 | 2016-03-15 | Leap Motion, Inc. | Object detection and tracking with variable-field illumination devices |
US20140157200A1 (en) * | 2012-12-05 | 2014-06-05 | Samsung Electronics Co., Ltd. | User terminal apparatus and method of controlling the same |
US10078421B2 (en) * | 2012-12-05 | 2018-09-18 | Samsung Electronics Co., Ltd. | User terminal apparatus and method of controlling the same |
US10915225B2 (en) | 2012-12-05 | 2021-02-09 | Samsung Electronics Co., Ltd. | User terminal apparatus and method of controlling the same |
US10609285B2 (en) | 2013-01-07 | 2020-03-31 | Ultrahaptics IP Two Limited | Power consumption in motion-capture systems |
US9465461B2 (en) | 2013-01-08 | 2016-10-11 | Leap Motion, Inc. | Object detection and tracking with audio and optical signals |
US9626015B2 (en) | 2013-01-08 | 2017-04-18 | Leap Motion, Inc. | Power consumption in motion-capture systems with audio and optical signals |
US10097754B2 (en) | 2013-01-08 | 2018-10-09 | Leap Motion, Inc. | Power consumption in motion-capture systems with audio and optical signals |
US10241639B2 (en) | 2013-01-15 | 2019-03-26 | Leap Motion, Inc. | Dynamic user interactions for display control and manipulation of display objects |
WO2014113507A1 (en) * | 2013-01-15 | 2014-07-24 | Leap Motion, Inc. | Dynamic user interactions for display control and customized gesture interpretation |
US11874970B2 (en) | 2013-01-15 | 2024-01-16 | Ultrahaptics IP Two Limited | Free-space user interface and control using virtual constructs |
US9696867B2 (en) | 2013-01-15 | 2017-07-04 | Leap Motion, Inc. | Dynamic user interactions for display control and identifying dominant gestures |
US10817130B2 (en) | 2013-01-15 | 2020-10-27 | Ultrahaptics IP Two Limited | Dynamic user interactions for display control and measuring degree of completeness of user gestures |
US9632658B2 (en) | 2013-01-15 | 2017-04-25 | Leap Motion, Inc. | Dynamic user interactions for display control and scaling responsiveness of display objects |
US10782847B2 (en) | 2013-01-15 | 2020-09-22 | Ultrahaptics IP Two Limited | Dynamic user interactions for display control and scaling responsiveness of display objects |
US11740705B2 (en) | 2013-01-15 | 2023-08-29 | Ultrahaptics IP Two Limited | Method and system for controlling a machine according to a characteristic of a control object |
US11353962B2 (en) | 2013-01-15 | 2022-06-07 | Ultrahaptics IP Two Limited | Free-space user interface and control using virtual constructs |
US10564799B2 (en) | 2013-01-15 | 2020-02-18 | Ultrahaptics IP Two Limited | Dynamic user interactions for display control and identifying dominant gestures |
US11269481B2 (en) | 2013-01-15 | 2022-03-08 | Ultrahaptics IP Two Limited | Dynamic user interactions for display control and measuring degree of completeness of user gestures |
US10042510B2 (en) | 2013-01-15 | 2018-08-07 | Leap Motion, Inc. | Dynamic user interactions for display control and measuring degree of completeness of user gestures |
US20150370469A1 (en) * | 2013-01-31 | 2015-12-24 | Qualcomm Incorporated | Selection feature for adjusting values on a computing device |
US10585193B2 (en) | 2013-03-15 | 2020-03-10 | Ultrahaptics IP Two Limited | Determining positional information of an object in space |
US9702977B2 (en) | 2013-03-15 | 2017-07-11 | Leap Motion, Inc. | Determining positional information of an object in space |
US11693115B2 (en) | 2013-03-15 | 2023-07-04 | Ultrahaptics IP Two Limited | Determining positional information of an object in space |
US10620709B2 (en) | 2013-04-05 | 2020-04-14 | Ultrahaptics IP Two Limited | Customized gesture interpretation |
US11347317B2 (en) | 2013-04-05 | 2022-05-31 | Ultrahaptics IP Two Limited | Customized gesture interpretation |
US11099653B2 (en) | 2013-04-26 | 2021-08-24 | Ultrahaptics IP Two Limited | Machine responsiveness to dynamic user movements and gestures |
US9747696B2 (en) | 2013-05-17 | 2017-08-29 | Leap Motion, Inc. | Systems and methods for providing normalized parameters of motions of objects in three-dimensional space |
US9355073B2 (en) | 2013-06-18 | 2016-05-31 | Microsoft Technology Licensing, Llc | Content attribute control interface including incremental, direct entry, and scrollable controls |
EP2816460A1 (en) * | 2013-06-21 | 2014-12-24 | BlackBerry Limited | Keyboard and touch screen gesture system |
US10152199B2 (en) * | 2013-07-16 | 2018-12-11 | Pinterest, Inc. | Object based contextual menu controls |
US20150026642A1 (en) * | 2013-07-16 | 2015-01-22 | Pinterest, Inc. | Object based contextual menu controls |
US11567578B2 (en) | 2013-08-09 | 2023-01-31 | Ultrahaptics IP Two Limited | Systems and methods of free-space gestural interaction |
US12086935B2 (en) | 2013-08-29 | 2024-09-10 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US11461966B1 (en) | 2013-08-29 | 2022-10-04 | Ultrahaptics IP Two Limited | Determining spans and span lengths of a control object in a free space gesture control environment |
US10846942B1 (en) | 2013-08-29 | 2020-11-24 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US11776208B2 (en) | 2013-08-29 | 2023-10-03 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US11282273B2 (en) | 2013-08-29 | 2022-03-22 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US11775033B2 (en) | 2013-10-03 | 2023-10-03 | Ultrahaptics IP Two Limited | Enhanced field of view to augment three-dimensional (3D) sensory space for free-space gesture interpretation |
US11010512B2 (en) | 2013-10-31 | 2021-05-18 | Ultrahaptics IP Two Limited | Improving predictive information for free space gesture control and communication |
US11568105B2 (en) | 2013-10-31 | 2023-01-31 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US9996638B1 (en) | 2013-10-31 | 2018-06-12 | Leap Motion, Inc. | Predictive information for free space gesture control and communication |
US11868687B2 (en) | 2013-10-31 | 2024-01-09 | Ultrahaptics IP Two Limited | Predictive information for free space gesture control and communication |
US9613262B2 (en) | 2014-01-15 | 2017-04-04 | Leap Motion, Inc. | Object detection and tracking for providing a virtual device experience |
US10444966B2 (en) * | 2014-04-11 | 2019-10-15 | Samsung Electronics Co., Ltd. | Method and apparatus for controlling number input in an electronic device |
US11256398B2 (en) | 2014-05-28 | 2022-02-22 | Meta Platforms, Inc. | Systems and methods for providing responses to and drawings for media content |
US10558338B2 (en) * | 2014-05-28 | 2020-02-11 | Facebook, Inc. | Systems and methods for providing responses to and drawings for media content |
US11836725B2 (en) | 2014-05-29 | 2023-12-05 | Apple Inc. | User interface for payments |
US11778159B2 (en) | 2014-08-08 | 2023-10-03 | Ultrahaptics IP Two Limited | Augmented reality with motion sensing |
US12095969B2 (en) | 2014-08-08 | 2024-09-17 | Ultrahaptics IP Two Limited | Augmented reality with motion sensing |
US20160231914A1 (en) * | 2015-02-11 | 2016-08-11 | Adobe Systems Incorporated | Dynamic text control for mobile devices |
US10042539B2 (en) * | 2015-02-11 | 2018-08-07 | Adobe Systems Incorporated | Dynamic text control for mobile devices |
US12032746B2 (en) | 2015-02-13 | 2024-07-09 | Ultrahaptics IP Two Limited | Systems and methods of creating a realistic displacement of a virtual object in virtual reality/augmented reality environments |
US12118134B2 (en) | 2015-02-13 | 2024-10-15 | Ultrahaptics IP Two Limited | Interaction engine for creating a realistic experience in virtual reality/augmented reality environments |
US9952658B2 (en) | 2015-03-17 | 2018-04-24 | Wipro Limited | System and method for improving viewing experience on a digital device |
US11734708B2 (en) | 2015-06-05 | 2023-08-22 | Apple Inc. | User interface for loyalty accounts and private label accounts |
US11321731B2 (en) | 2015-06-05 | 2022-05-03 | Apple Inc. | User interface for loyalty accounts and private label accounts |
US11783305B2 (en) | 2015-06-05 | 2023-10-10 | Apple Inc. | User interface for loyalty accounts and private label accounts for a wearable device |
EP3869311A1 (en) * | 2015-12-22 | 2021-08-25 | Daikin Industries, Ltd. | Setting value change device |
US11379109B2 (en) | 2015-12-22 | 2022-07-05 | Daikin Industries, Ltd. | Setting value change device |
US11481769B2 (en) | 2016-06-11 | 2022-10-25 | Apple Inc. | User interface for transactions |
US12002042B2 (en) | 2016-06-11 | 2024-06-04 | Apple, Inc | User interface for transactions |
US11074572B2 (en) | 2016-09-06 | 2021-07-27 | Apple Inc. | User interfaces for stored-value accounts |
US11797968B2 (en) | 2017-05-16 | 2023-10-24 | Apple Inc. | User interfaces for peer-to-peer transfers |
US11222325B2 (en) | 2017-05-16 | 2022-01-11 | Apple Inc. | User interfaces for peer-to-peer transfers |
US11221744B2 (en) | 2017-05-16 | 2022-01-11 | Apple Inc. | User interfaces for peer-to-peer transfers |
US11393258B2 (en) | 2017-09-09 | 2022-07-19 | Apple Inc. | Implementation of biometric authentication |
US11386189B2 (en) | 2017-09-09 | 2022-07-12 | Apple Inc. | Implementation of biometric authentication |
US11765163B2 (en) | 2017-09-09 | 2023-09-19 | Apple Inc. | Implementation of biometric authentication |
US11875012B2 (en) | 2018-05-25 | 2024-01-16 | Ultrahaptics IP Two Limited | Throwable interface for augmented reality and virtual reality environments |
US11100498B2 (en) | 2018-06-03 | 2021-08-24 | Apple Inc. | User interfaces for transfer accounts |
US11900355B2 (en) | 2018-06-03 | 2024-02-13 | Apple Inc. | User interfaces for transfer accounts |
US11514430B2 (en) | 2018-06-03 | 2022-11-29 | Apple Inc. | User interfaces for transfer accounts |
US11782740B2 (en) * | 2018-11-30 | 2023-10-10 | Snap Inc. | Interface to configure media content |
US20230018594A1 (en) * | 2018-11-30 | 2023-01-19 | Snap Inc. | Interface to configure media content |
US11520607B2 (en) * | 2018-11-30 | 2022-12-06 | Snap Inc. | Interface to configure media content |
US20210279082A1 (en) * | 2018-11-30 | 2021-09-09 | Snap Inc. | Interface to configure media content |
CN110083296A (en) * | 2019-03-15 | 2019-08-02 | 努比亚技术有限公司 | Wearable device and its exchange method, computer readable storage medium |
US11328352B2 (en) * | 2019-03-24 | 2022-05-10 | Apple Inc. | User interfaces for managing an account |
US11610259B2 (en) | 2019-03-24 | 2023-03-21 | Apple Inc. | User interfaces for managing an account |
US11669896B2 (en) | 2019-03-24 | 2023-06-06 | Apple Inc. | User interfaces for managing an account |
US11688001B2 (en) | 2019-03-24 | 2023-06-27 | Apple Inc. | User interfaces for managing an account |
US11681537B2 (en) | 2019-09-29 | 2023-06-20 | Apple Inc. | Account management user interfaces |
US11169830B2 (en) | 2019-09-29 | 2021-11-09 | Apple Inc. | Account management user interfaces |
US11373373B2 (en) * | 2019-10-22 | 2022-06-28 | International Business Machines Corporation | Method and system for translating air writing to an augmented reality device |
US12118562B2 (en) | 2020-05-29 | 2024-10-15 | Apple Inc. | Configuring an account for a second user identity |
US12131011B2 (en) | 2020-07-28 | 2024-10-29 | Ultrahaptics IP Two Limited | Virtual interactions for machine control |
US11983702B2 (en) | 2021-02-01 | 2024-05-14 | Apple Inc. | Displaying a representation of a card with a layered structure |
US12131374B2 (en) | 2023-04-13 | 2024-10-29 | Apple Inc. | User interfaces for managing an account |
Also Published As
Publication number | Publication date |
---|---|
US20130019174A1 (en) | 2013-01-17 |
US20130019203A1 (en) | 2013-01-17 |
US20130019206A1 (en) | 2013-01-17 |
US9086794B2 (en) | 2015-07-21 |
US20130019205A1 (en) | 2013-01-17 |
US20130019208A1 (en) | 2013-01-17 |
US9250766B2 (en) | 2016-02-02 |
US9116602B2 (en) | 2015-08-25 |
US9021398B2 (en) | 2015-04-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20130019204A1 (en) | Adjusting content attributes through actions on context based menu | |
EP2732362B1 (en) | Launcher for context based menus | |
US20130019175A1 (en) | Submenus for context based menu system | |
US9589233B2 (en) | Automatic recognition and insights of data | |
US9026944B2 (en) | Managing content through actions on context based menus | |
EP2699998B1 (en) | Compact control menu for touch-enabled command execution | |
US9003298B2 (en) | Web page application controls | |
US20150033188A1 (en) | Scrollable smart menu | |
SG191849A1 (en) | User interface interaction behavior based on insertion point | |
EP2732365A2 (en) | Dynamic context based menus | |
WO2013191928A1 (en) | Displaying documents based on author preferences | |
EP2825947A1 (en) | Web page application controls |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KOTLER, MATTHEW;GIL, EREZ KIKIN;SACHIDANANDAM, VIGNESH;REEL/FRAME:028517/0870 Effective date: 20120704 |
|
AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034544/0541 Effective date: 20141014 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |