AU2014315325B2 - User interface object manipulations in a user interface - Google Patents
User interface object manipulations in a user interface Download PDFInfo
- Publication number
- AU2014315325B2 AU2014315325B2 AU2014315325A AU2014315325A AU2014315325B2 AU 2014315325 B2 AU2014315325 B2 AU 2014315325B2 AU 2014315325 A AU2014315325 A AU 2014315325A AU 2014315325 A AU2014315325 A AU 2014315325A AU 2014315325 B2 AU2014315325 B2 AU 2014315325B2
- Authority
- AU
- Australia
- Prior art keywords
- display
- crown
- speed
- computer
- virtual object
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0362—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 1D translations or rotations of an operating part of the device, e.g. scroll wheels, sliders, knobs, rollers or belts
-
- G—PHYSICS
- G04—HOROLOGY
- G04G—ELECTRONIC TIME-PIECES
- G04G21/00—Input or output devices integrated in time-pieces
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04817—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
- Electric Clocks (AREA)
- Controls And Circuits For Display Device (AREA)
- Electromechanical Clocks (AREA)
- Measuring Pulse, Heart Rate, Blood Pressure Or Blood Flow (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
Abstract
Systems and processes for manipulating a graphical user interface are disclosed. One process can include receiving user input through a crown to rotate a virtual object. The process includes selecting a surface of the object from among the multiple surfaces of the object in response to determining that the crown rotation exceeded a speed threshold.
Description
WO 2015/034966 PCT/US2014/053958
USER INTERFACE OBJECT MANIPULATIONS IN A USER INTERFACE
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] This application claims priority to U.S. Provisional Patent Application Serial No. 61/873,356, filed September 3, 2013, entitled “CROWN INPUT FOR A WEARABLE ELECTRONIC DEVICE”; U.S. Provisional Patent Application Serial No. 61/873,359, filed September 3, 2013, entitled “USER INTERFACE OBJECT MANIPULATIONS IN A USER INTERFACE”; U.S. Provisional Patent Application Serial No. 61/959,851, filed September 3, 2013, entitled “USER INTERFACE FOR MANIPULATING USER INTERFACE OBJECTS”; U.S. Provisional Patent Application Serial No. 61/873,360, filed September 3, 2013, entitled “USER INTERFACE FOR MANIPULATING USER INTERFACE OBJECTS WITH MAGNETIC PROPERTIES”; and U.S. Non-provisional Patent Application Serial No. 14/476,657, filed September 3, 2014, entitled “USER INTERFACE FOR MANIPULATING USER INTERFACE OBJECTS WITH MAGNETIC PROPERTIES”. The content of these applications is hereby incorporated by reference in its entirety for all purposes.
[0002] This application is related to co-pending applications U.S. Non-provisional Patent Application filed September 3, 2014, concurrently herewith, entitled “CROWN INPUT FOR A WEARABLE ELECTRONIC DEVICE,” naming Nicholas Zambetti et al. as inventors; U.S. Non-provisional Patent Application filed September 3, 2014, concurrently herewith, entitled “USER INTERFACE FOR MANIPULATING USER INTERFACE OBJECTS”, naming Nicholas Zambetti et al. as inventors; U.S. Non-provisional Patent Application Serial No. 14/476,657, filed September 3, 2014, entitled “USER INTERFACE OBJECT MANIPULATIONS IN A USER INTERFACE”; and U.S. Provisional Patent Application Serial No. 61/747,278, filed December 29, 2012, entitled “Device, Method, and Graphical User Interface for Manipulating User Interface Objects with Visual and/or Haptic Feedback”. The content of these applications is hereby incorporated by reference in its entirety for all purposes.
Field [0003] This disclosure relates generally to user interfaces and, more specifically, to user interfaces using a crown input mechanism.
Background [0004] Advanced personal electronic devices can have small form factors. These personal electronic devices include, but are not limited to, tablets and smart phones. Use of such personal electronic devices involves manipulation of user interface objects on display screens which also have small form factors that complement the design of the personal electronic devices.
[0005] Exemplary manipulations that users can perform on personal electronic devices include navigating a hierarchy, selecting a user interface object, adjusting the position, size, and 1 1001774811 2014315325 06 Apr 2017 zoom of user interface objects, or otherwise manipulating user interfaces. Exemplary user interface objects include digital images, video, text, icons, maps, control elements such as buttons, and other graphics. A user can perform such manipulations in image management software, video editing software, word pressing software, software execution platforms such as 5 an operating system’s desktop, website browsing software, and other environments.
[0006] Existing methods for manipulating user interface obj ects on reduced-size touch-sensitive displays can be inefficient. Further, existing methods generally provide less precision than is preferable. .0 Summary [0007] Systems and processes for manipulating a graphical user interface are disclosed. One process can include receiving user input through a crown to rotate a virtual object. The process includes selecting a surface of the object from among the multiple surfaces of the object in response to determining that the crown rotation exceeded a speed threshold. , 5 [0007A] According to a first aspect of the invention, a computer-implemented method comprises: displaying a first surface of a plurality of selectable surfaces of a virtual object on a touch-sensitive display of a wearable electronic device, the first surface associated with a first data; determining a speed, wherein the speed is based on an angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a 10 speed threshold, displaying on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state.
[0007B] According to a second aspect of the invention, a computer-implemented method 25 comprises: displaying, on a touch-sensitive display of a wearable electronic device, a first surface of a plurality of selectable surfaces of a virtual object, the first surface associated with a first data; determining an angular velocity of a physical crown of the wearable electronic device; and in response to a determination that the angular velocity exceeds a predetermined threshold, displaying an animation of the virtual object rotating about an axis parallel to the display and 30 displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state, and in response to a determination that the angular velocity is below the predetermined threshold, maintaining display of the first surface in a steady state. . 2 1001774811 2014315325 06 Apr 2017 [0007C] According to a third aspect of the invention, a non-transitory computer-readable storage medium comprises instructions for: displaying a first surface of a plurality of selectable surfaces of a virtual object on a touch-sensitive display of a wearable electronic device, the first surface associated with a first data; determining a speed, wherein the speed is based on an 5 angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a speed threshold, displaying on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state. 10 [00Q7D] According to a fourth aspect of the invention, an electronic device comprises: one or more processors; a physical crown operatively coupled to the one or more processors; and a touch-sensitive display operatively coupled to the one or more processors, the one or more processors configured to: display a first surface of a plurality of selectable surfaces of a virtual object on a touch-sensitive display of a wearable electronic device, the first surface associated 15 with a first data; determine a speed, wherein the speed is based on an angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a speed threshold, display on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second 20 surface displayed parallel to the display while in a steady state.
[0007E] As used herein, except where the context requires otherwise, the term "comprise" and variations of the term, such as "comprising", "comprises" and "comprised", are not intended to exclude further additives, components, integers or steps. 25 Brief Description of the Drawings [0008] The present application can be best understood by reference to the following description taken in conjunction with the accompanying drawing figures, in which like parts may be referred to by like numerals.
[0009] FIG. 1 illustrates an exemplary wearable electronic device according to various 30 examples.
[0010] FIG. 2 illustrates a block diagram of an exemplary wearable electronic device according to various examples.
[0011] FIGs. 3-12 illustrate an exemplary graphical user interface showing the selection of a surface of a two-sided object in response to a rotation of a crown. 2a 1001774811 2014315325 06 Apr 2017 [0012] FIG. 13 illustrates an exemplary process for selecting a surface of a two-sided obj ect in response to a rotation of a crown.
[0013] FIGs. 14-23 illustrate an exemplary graphical user interface showing the selection of 5 a surface of an object in response to a rotation of a crown.
[0014] FIG. 24 illustrates an exemplary process for selecting a surface of an object in response to a rotation of a crown.
[0015] FIG. 25 illustrates an exemplary multi-sided object in a graphical user interface.
[0016] FIG. 26 illustrates an exemplary computing system for manipulating a user interface 0 in response to a rotation of a crown according to various examples. 2b PCT/U S2014/053958 WO 2015/034966
Detailed Description [0017] In the following description of the disclosure and examples, reference is made to the accompanying drawings in which it is shown by way of illustration specific examples that can be practiced. It is to be understood that other examples can be practiced and structural changes can be made without departing from the scope of the disclosure.
[0018] Many personal electronic devices have graphical user interfaces with options that can be activated in response to user inputs. Typically, a user can select and activate a particular option from among multiple options. For example, a user may select an option by placing a mouse cursor over the desired option using a pointing device. The user may activate the option by clicking a button of the pointing device while the option is selected. In another example, a user may select and activate an option displayed on a touch-sensitive display (also known as a touch screen) by touching the touch-sensitive display at the location of the displayed option. Given the inefficiency of existing methods for selecting options on reduced-size touch-sensitive displays, there is a need for methods that enable users to more efficiently and conveniently select a desired option in a graphical user interface environment.
[0019] The examples below describe improved techniques for selecting a surface of a user interface object in a graphical user interface using user inputs. More specifically, these techniques use a physical crown as an input device to enable a user to select a desired option by selecting a surface of the user interface object. As a result, the examples described below allow a user to more efficiently and conveniently select a desired option.
[0020] FIG. 1 illustrates exemplary personal electronic device 100. In the illustrated example, device 100 is a watch that generally includes body 102 and strap 104 for affixing device 100 to the body of a user. That is, device 100 is wearable. Body 102 can designed to couple with straps 104. Device 100 can have touch-sensitive display screen (hereafter touchscreen) 106 and crown 108. Device 100 can also have buttons 110, 112, and 114.
[0021] Conventionally, the term ‘crown,’ in the context of a watch, refers to the cap atop a stem for winding the watch. In the context of a personal electronic device, the crown can be a physical component of the electronic device, rather than a virtual crown on a touch sensitive display. Crown 108 can be mechanical meaning that it can be connected to a sensor for converting physical movement of the crown into electrical signals. Crown 108 can rotate in two directions of rotation (e.g., forward and backward). Crown 108 can also be pushed in towards the body of device 100 and/or be pulled away from device 100. Crown 108 can be touch-sensitive, for example, using capacitive touch technologies that can detect whether a user is touching the crown. Moreover, crown 108 can further be rocked in one or more directions or translated along a track along an edge or at least partially around a perimeter of body 102. In 3 PCT/US2014/053958 WO 2015/034966 some examples, more than one crown 108 can be used. The visual appearance of crown 108 can, but need not, resemble crowns of conventional watches. Buttons 110, 112, and 114, if included, can each be a physical or a touch-sensitive button. That is, the buttons may be, for example, physical buttons or capacitive buttons. Further, body 102, which can include a bezel, may have predetermined regions on the bezel that act as buttons.
[0022] Display 106 can include a display device, such as a liquid crystal display (LCD), light-emitting diode (LED) display, organic light-emitting diode (OLED) display, or the like, positioned partially or fully behind or in front of a touch sensor panel implemented using any desired touch sensing technology, such as mutual-capacitance touch sensing, self-capacitance touch sensing, resistive touch sensing, projection scan touch sensing, or the like. Display 106 can allow a user to perform various functions by touching over hovering near the touch sensor panel using one or more fingers or other object.
[0023] In some examples, device 100 can further include one or more pressure sensors (not shown) for detecting a force or pressure applied to the display. The force or pressure applied to display 106 can be used as an input to device 100 to perform any desired operation, such as making a selection, entering or exiting a menu, causing the display of additional options/actions, or the like. In some examples, different operations can be performed based on the amount of force or pressure being applied to display 106. The one or more pressure sensors can further be used to determine a position that the force is being applied to display 106.
[0024] FIG. 2 illustrates a block diagram of some of the components of device 100. As shown, crown 108 can be coupled to encoder 204, which can be configured to monitor a physical state or change of state of crown 108 (e.g., the position of the crown), convert it to an electrical signal (e.g., convert it to an analog or digital signal representation of the position or change in position of crown 108), and provide the signal to processor 202. For instance, in some examples, encoder 204 can be configured to sense the absolute rotational position (e.g., an angle between 0-360°) of crown 108 and output an analog or digital representation of this position to processor 202. Alternatively, in other examples, encoder 204 can be configured to sense a change in rotational position (e.g., a change in rotational angle) of crown 108 over some sampling period and to output an analog or digital representation of the sensed change to processor 202. In these examples, the crown position information can further indicate a direction of rotation of the crown (e.g., a positive value can correspond to one direction and a negative value can correspond to the other). In yet other examples, encoder 204 can be configured to detect a rotation of crown 108 in any desired manner (e.g., velocity, acceleration, or the like) and can provide the crown rotational information to processor 202. In alternative examples, instead of providing information to processor 202, this information can be provided to other components of device 100. While the 4 PCT/US2014/053958 WO 2015/034966 examples described herein refer to the use of rotational position of crown 108 to control scrolling, scaling, or an objects position, it should be appreciated that any other physical state of crown 108 can be used.
[0025] In some examples, the physical state of the crown can control physical attributes of display 106. For example, if crown 108 is in a particular position (e.g., rotated forward), display 106 can have limited z-axis traversal ability. In other words, the physical state of the crown can represent physical modal functionality of display 106. In some examples, a temporal attribute of the physical state of crown 108 can be used as an input to device 100. For example, a fast change in physical state can be interpreted differently than a slow change in physical state.
[0026] Processor 202 can be further coupled to receive input signals from buttons 110, 112, and 114, along with touch signals from touch-sensitive display 106. The buttons may be, for example, physical buttons or capacitive buttons. Further, body 102, which can include a bezel, may have predetermined regions on the bezel that act as buttons. Processor 202 can be configured to interpret these input signals and output appropriate display signals to cause an image to be produced by touch-sensitive display 106. While a single processor 202 is shown, it should be appreciated that any number of processors or other computational devices can be used to perform the general functions discussed above.
[0027] FIGs. 3-12 illustrate an exemplary user interface 300 displaying a two-sided user interface object 302. Object 302 has a first surface 304 and a second surface 306. Each surface of object 302 is a selectable surface associated with corresponding data. The data may be, for example, text, an image, an application icon, an instruction, a binary ON or OFF option, and the like. A user can select a surface from among the multiple selectable surfaces of object 302 by using a physical crown of a wearable electronic device to rotate object 302 to align the desired selection surface such that the surface is parallel to the display 106 of the device 100 and is displayed on the display 106. The system is designed to transition between one surface to another, rather than stopping in between surfaces. Although examples are described with respect to object surfaces (or planes) being parallel to display 106, the examples can also be modified to instead be described with respect to object surfaces (or planes) facing the viewer of display 106. This modification may be particularly helpful when object surfaces or display 106 is not plane surface.
[0028] Crown 108 of device 100 is a user rotatable user interface input. The crown 108 can be turned in two distinct directions: clockwise and counterclockwise. FIGs. 3-12 include rotation direction arrows illustrating the direction of crown rotation and movement direction arrows illustrating the direction of rotation of a user interface object, where applicable. The rotation direction arrows and movement direction arrows are typically not part of the displayed 5 PCT/US2014/053958 WO 2015/034966 user interface, but are provided to aid in the interpretation of the figures. In this example, a clockwise direction rotation of crown 108 is illustrated by a rotation direction arrow pointing in the up direction. Similarly, a counterclockwise direction rotation of crown 108 is illustrated by a rotation direction arrow pointing in the down direction. The characteristics of the rotation direction arrow are not indicative of the distance, speed, or acceleration with which crown 108 is rotated by a user. Instead, the rotation direction arrow is indicative of the direction of rotation of crown 108 by the user.
[0029] At FIG. 3, first surface 304 of object 302 is aligned parallel to display 106 and is displayed, indicating selection of first surface 304. The selected first surface 304 can be activated through, for example, an additional user input. At FIG. 4, device 100 determines a change in the position of crown 108 in the clockwise direction, as indicated by rotation direction arrow 308. Device 100 determines a rotational speed and a direction based on the determined change in the position of crown 108. In response to determining the change in the position of crown 108, the device rotates object 302, as indicated by movement direction arrow 310 and illustrated in FIG. 4. The rotation of object 302 is based on the determined rotational speed and direction. Rotational speed may be expressed in numerous ways. For example, rotational speed may be expressed as hertz, as rotations per unit of time, as rotations per frame, as revolutions per unit of time, as revolutions per frame, as a change in angle per unit of time, and the like. In one example, object 302 may be associated with a mass or may have a calculated rotational inertia.
[0030] At FIGs. 5-7, device 100 continues to determine a change in the position of crown 108 in the clockwise direction, as indicated by rotation direction arrow 308. Device 100 determines a rotational speed and a direction based on the determined change in the position of crown 108. In response to determining the change in the position of crown 108, the device continues to rotate object 302, as indicated by movement direction arrow 310 and illustrated in FIG. 5-6. The rotation of object 302 is based on the determined rotational speed and direction.
[0031] In one example, the degrees of rotation of object 302, as measured from the object’s position while parallel to display 106, is based on the determined speed. For easier visualization, object 302 can be thought of as having some similar qualities as an analog tachometer. As the determined speed increases, the degree of rotation of object 302 increases. In this example, if the rotation of crown 108 is maintained at a constant speed, object 302 will stay at a static rotated position that is not parallel to display 106. If the speed of the rotation of crown 108 is increased, the determined speed will increase and object 302 will rotate an additional amount.
[0032] In some examples, object 302 is configured to become perpendicular to display 106 in response to the determined speed being at a speed threshold. When the determined speed exceeds the speed threshold, object 302 exceeds a total rotation of 90 degrees, causing first 6 PCT/US2014/053958 WO 2015/034966 surface 304 of object 302 to no longer be displayed and instead causing second surface 306 of object 302 to be displayed. This transition between the display of first surface 304 and second surface 306 is illustrated as the transition between FIGs. 7 and 8. Thus, as the determined speed exceeds the speed threshold the object 302 flips from one side to another side.
[0033] At FIGs. 9-12, device 100 determines that there is no further change in the position of crown 108. Asa result of this determination, the rotation of object 302 is changed such that a surface of object 302 is parallel to display 106. This change may be animated, as illustrated in FIGs. 9-12. Device 100 will rotate object 302 such that the surface of object 302 partially facing display 106 when device 100 determines that there is no change in the position of crown 108 is the surface that will be displayed as being parallel to display 106. When a surface of object 302 is parallel to display 106 and no change in the position of crown 108 is detected, object 302 is in a steady state. An object is in a steady state when the object is not being translated, rotated, or scaled.
[0034] In some examples, when object 302 is in a steady state, the displayed surface of object 302 that is parallel to display 106 can be activated with an additional input. The displayed surface that is parallel to display 106 in a steady state is determined to be selected even prior to activation. For example, object 302 may be used as an ON/OFF switch or toggle. First surface 304 is associated with an ON instruction and second surface 306 is associated with an OFF instruction. A user can transition between the ON and OFF states by rotating crown 108 at above a speed threshold, causing object 302 to flip and display a desired surface. The desired surface is determined to be selected when the desired surface is displayed on display 106, is parallel to display 106, and no change in the position of crown 108 is detected.
[0035] While a surface is selected, the user can activate the selected surface by one or more of many techniques. For example, the user may press on touch-sensitive display 106, press on touch-sensitive display with a force greater than a predetermined threshold, press button 112, or simply allow the surface to remain selected for a predetermined amount of time. In another example, when the displayed surface is parallel to display 106, the action can be interpreted as both a selection and an activation of the data associated with the displayed surface.
[0036] FIG. 13 illustrates an exemplary process for selecting a surface of a two-sided graphical user interface object in response to a rotation of a crown. Process 1300 is performed at a wearable electronic device (e.g., device 100 in FIG. 1) having a physical crown. In some examples, the electronic device also includes a touch-sensitive display. The process provides an efficient technique for selecting a surface of a two-sided, two-dimensional object.
[0037] At block 1302, the device causes a display of a two-sided object on a touch-sensitive display of a wearable electronic device. In some examples, the object is two-dimensional. In 7 PCT/US2014/053958 WO 2015/034966 other examples, the object is three dimensional but only two surfaces are selectable. Each selectable surface of the object is associated with a corresponding data value. The data may be, for example, text, an image, an application icon, an instruction, a binary ON or OFF option, and the like.
[0038] At block 1304, the device receives crown position information. The crown position information may be received as a series of pulse signals, real values, integer values, and the like.
[0039] At block 1306, the device determines whether a change has occurred in a crown distance value. The crown distance value is based on an angular displacement of the physical crown of the wearable electronic device. A change in the crown distance value is indicative of a user providing input to the wearable electronic device by, for example, turning the physical crown. If the device determines that a change in the crown distance value has not occurred, the system returns to block 1304 and continues receiving crown position information. If the device determines that a change in the crown distance value has occurred, the system continues to block 1308, though the system may continue to receive crown position information.
[0040] At block 1308, the device determines a direction and a crown speed. The crown speed is based on the speed of rotation of the physical crown of the wearable electronic device. For example, the determined crown speed may be expressed as hertz, as rotations per unit of time, as rotations per frame, as revolutions per unit of time, as revolutions per frame, and the like. The determined direction is based on a direction of rotation of the physical crown of the wearable electronic device. For example, an up direction can be determined based on a clockwise rotation of the physical crown. Similarly, a down direction can be determined based on a counterclockwise rotation of the physical crown. In other examples, a down direction can be determined based on a clockwise rotation of the physical crown and an up direction can be determined based on a counterclockwise rotation of the physical crown.
[0041] At block 1310, in response to determining the change in the crown distance value, the device causes an initial rotation of the two-sided object on the display. The amount of the rotation is based on the determined crown speed. The direction of rotation is based on the determined direction. The rotation may be animated.
[0042] At block 1312, the device determines whether the determined crown speed exceeds a speed threshold. If the device determines that the determined crown speed exceeds the speed threshold, the device continues to block 1314. For example, the speed threshold may be thought of as an escape velocity (or escape speed). An escape velocity is the speed at which the kinetic energy plus the gravitational potential energy of an object is zero. If the device determines that the determined crown speed does not exceed the speed threshold, the device transitions to block 1316. 8 PCT/US2014/053958 WO 2015/034966 [0043] In some examples, the minimum angular velocity of crown rotation that is necessary to reach escape velocity corresponds directly to the instantaneous angular velocity of crown 108 (FIG. 1), meaning that the user interface of device 100, in essence, responds when crown 108 reaches a sufficient angular velocity. In some embodiments, the minimum angular velocity of crown rotation necessary for reaching the escape velocity is a calculated velocity that is based on, but not directly equal to, the instantaneous (“current”) angular velocity of crown 108. In these examples, device 100 can maintain a calculated crown (angular) velocity V in discrete moments in time T according to equation 1: VT=V(T-i) + AVcrown AVdrag· (EQ. 1) [0044] In equation 1, Vt represents a calculated crown velocity (speed and direction) at time T, V(T-i) represents the previous velocity (speed and direction) at time T-l, AVcrown represents the change in velocity caused by the force being applied through the rotation of the crown at time T, and AVdrag represents the change in velocity due to a drag force. The force being applied, which is reflected through AVcrown, can depend on the current velocity of angular rotation of the crown. Thus, AVcrown can also depend on the current angular velocity of the crown. In this way, device 100 can provide user interface interactions based not only on instantaneous crown velocity but also based on user input in the form of crown movement over multiple time intervals, even if those intervals are finely divided. Note, typically, in the absence of user input in the form of AVcrown, Vt will approach (and become) zero based on AVdrag in accordance with EQ. 1, but Vt would not change signs without user input in the form of crown rotation (AVcrown)· [0045] Typically, the greater the velocity of angular rotation of the crown, the greater the value of AVcrown will be. However, the actual mapping between the velocity of angular rotation of the crown and AVcrown can be varied depending on the desired user interface effect. For example, various linear or non-linear mappings between the velocity of angular rotation of the crown and AVcrown can be used.
[0046] Also, AVdrag can take on various values. For example, AVdrag can depend on the velocity of crown rotation such that at greater velocities, a greater opposing change in velocity (AVDrag) can be produced. In another example, AVDrag can have a constant value. It should be appreciated that the above-described requirements of AVcrown and AVdrag can be changed to produce desirable user interface effects.
[0047] As can be seen from EQ. 1, the maintained velocity (Vt) can continue to increase as long as AVcrown is greater than AVdrag· Additionally, VT can have non-zero values even when no AVcrown input is being received, meaning that user interface objects can continue to change 9 PCT/U S2014/053958 WO 2015/034966 without the user rotating the crown. When this occurs, objects can stop changing based on the maintained velocity at the time the user stops rotating the crown and the AVDRAg component.
[0048] In some examples, when the crown is rotated in a direction corresponding to a rotation direction that is opposite the current user interface changes, the V(t-d component can be reset to a value of zero, allowing the user to quickly change the direction of the object without having to provide a force sufficient to offset the VT.
[0049] At block 1314, the device causes the object to flip past a transition position between a first surface that was last selected and a second surface. For example, the object has flipped past the transition position when the object will not return to having the first surface displayed parallel to the display without receiving additional user input. In the example of a two-sided object, the transition position may be when the surface is perpendicular to the display.
[0050] Once the object reaches a steady state, the displayed surface that is parallel to the display can be activated by a designated user input. The displayed surface that is parallel to the display in a steady state is determined to be selected even prior to activation. An object is in a steady state when the object is not being translated, rotated, or scaled. This may result in the first surface of the object no longer being displayed, in the case of a cube-shaped object.
[0051] At block 1316, because the escape velocity has not been reached, the device causes the object to at least partially return to the object’s initial position at the time of block 1302.
For example, part of the initial rotation of the object caused at block 2410 can be negated. To achieve this, the device animates a rotation of the object that is in an opposite direction of the initial rotation at block 1310.
[0052] FIGs. 14-23 illustrate an exemplary graphical user interface showing the selection of a surface of a cube object in response to a rotation of a crown. Object 1402 is a cube with six surfaces. In this example, four of the six surfaces are selectable. These four selectable surfaces include surface 1404 of object 1402, which is facing a viewer of display 106, the top surface of object 1402, the bottom surface of object 1402, and the back surface of object 1402. In this example, the left and right surfaces of object 1402 are not selectable. However, the left and right surfaces of object 1402 may be selectable in other examples. Although examples are described with respect to object surfaces (or planes) being parallel to display 106, the examples can also be modified to instead be described with respect to object surfaces (or planes) facing the viewer of display 106. This modification may be particularly helpful when object surfaces or display 106 is not plane surface.
[0053] Each selectable surface of object 1402 is associated with corresponding data. The data may be, for example, text, an image, an application icon, an instruction, a quad-state setting (such as Off/Low/Medium/High), and the like. A user can select a surface from among the 10 PCT/U S2014/053958 WO 2015/034966 multiple selectable surfaces of the object 1402 by using a physical crown of a wearable electronic device to rotate object 1402 to align the desired selection surface such that it is parallel to the display 106 and displayed on display 106.
[0054] Crown 108 of device 100 is a user rotatable user interface input. The crown 108 can be turned in two distinct directions: clockwise and counterclockwise. FIGs. 14-23 include rotation direction arrows illustrating the direction of crown rotation and movement direction arrows illustrating the direction of rotation of a user interface object, where applicable. The rotation direction arrows and movement direction arrows are typically not part of the displayed user interface, but are provided to aid in the interpretation of the figures. In this example, a clockwise direction rotation of crown 108 is illustrated by a rotation direction arrow pointing in the up direction. Similarly, a counterclockwise direction rotation of crown 108 is illustrated by a rotation direction arrow pointing in the down direction. The characteristics of the rotation direction arrow are not indicative of the distance, speed, or acceleration with which crown 108 is rotated by a user. Instead, the rotation direction arrow is indicative of the direction of rotation of crown 108 by the user.
[0055] At FIG. 14, first surface 1404 of object 1402 is aligned parallel to display 106 and is displayed, indicating selection of first surface 1404. At FIG. 15, device 100 determines a change in the position of crown 108 in the counterclockwise direction, as indicated by rotation direction arrow 1502. Device 100 determines a rotational speed and a direction based on the determined change in the position of crown 108. In response to determining the change in the position of crown 108, the device rotates object 1402, as indicated by movement direction arrow 1504 and illustrated in FIG. 15. The rotation of object 1402 is based on the determined rotational speed and direction. Rotational speed may be expressed in numerous ways. For example, rotational speed may be expressed as hertz, as rotations per unit of time, as rotations per frame, as revolutions per unit of time, as revolutions per frame, and the like. In one example, object 1402 may be associated with a mass or may have a calculated rotational inertia.
[0056] At FIG. 16, device 100 continues to determine a change in the position of crown 108 in the counterclockwise direction, as indicated by rotation direction arrow 1502. Device 100 determines a rotational speed and a direction based on the determined change in the position of crown 108. In response to determining the change in the position of crown 108, the device continues to rotate object 1402, as indicated by movement direction arrow 1504 and illustrated in FIG. 16. The rotation of object 1402 is based on the determined rotational speed and direction.
[0057] In one example, the degrees of rotation of object 1402 is based on the determined speed. As the determined speed increases, the degree of rotation of object 1402 increases. In this example, if the rotation of crown 108 is maintained at a constant speed, object 1402 will stay 11 PCT/US2014/053958 WO 2015/034966 at a static rotated position where no surface of object 1402 is parallel to display 106. If the speed of the rotation of crown 108 is increased, the determined speed will increase and object 1402 will rotate an additional amount.
[0058] In some examples, object 1402 is configured to rotate to have a surface parallel to display 106 in response to the determined speed being above a speed threshold. When the determined speed exceeds the speed threshold, object 1402 exceeds a rotation of 45 degrees, causing first surface 1404 of object 1402 to rotate away from the display to no longer be displayed and instead causing second surface 1406 of object 1404 rotate toward the display to be displayed. This transition between the display of first surface 1404 and second surface 1406 is illustrated as the transition between FIGs. 16 and 17. Thus, as the determined speed exceeds the speed threshold, the object 1402 flips from one surface to another surface.
[0059] At FIGs. 17-18, device 100 determines that there is no change in the position of crown 108. As a result of this determination, object 1402 is rotated such that a displayed surface of object 1402 is parallel to display 106. This rotation may be animated, as illustrated in FIGs. 17-18. Device 100 will rotate object 1402 such that the displayed surface of object 1402 that has the smallest angle with respect to the display is made parallel to the display 106. In other words, the object’s surface that best faces the display 106 or is closest to parallel to display 106 is made parallel to the display 106. When a surface of object 1402 is parallel to display 106 and no change in the position of crown 108 is detected, object 1402 is in a steady state. An object is in a steady state when the object is not being translated, rotated, or scaled.
[0060] In some examples, when object 1402 is in a steady state, the surface of object 1402 that is parallel to display 106 and displayed on display 106 is determined to be selected. For example, object 1402 may be used as four-phase selection switch. First surface 1404 is associated with a LOW setting instruction and second surface 1406 is associated with a MEDIUM instruction setting. The remaining two selectable surfaces are associated with HIGH and OFF instruction settings. A user can transition between the four settings by rotating crown 108 at above a speed threshold, causing object 1402 to flip and display a desired surface. The desired surface is determined to be selected when the displayed surface is parallel to display 106 and no change in the position of crown 108 is detected.
[0061] While a surface is selected, the user can activate the selected surface by one or more of many techniques. For example, the user may press on touch-sensitive display 106, press button 112, or simply allow the surface to remain selected for a predetermined amount of time.
In another example, when the displayed surface is parallel to display 106, the action can be interpreted as both a selection and an activation of the data associated with the displayed surface. 12 PCT/US2014/053958 WO 2015/034966 [0062] FIGs. 20-23 illustrate a second flip of object 1402 to select third surface 2002 of object 1402. In FIGs. 21-22, device 100 determines a change in the position of crown 108 in the counterclockwise direction, as indicated by rotation direction arrow 1502. Device 100 determines a rotational speed and a direction based on the determined change in the position of crown 108. In response to determining the change in the position of crown 108, the device rotates object 1402, as indicated by movement direction arrow 1504 and illustrated in FIG. 21-22. The rotation of object 1402 is based on the determined rotational speed and direction.
[0063] In response to the rotational speed exceeding a threshold, object 1402 flips to cause third surface 2002 to be parallel to display 106 and to be displayed on display 106, as illustrated in FIG. 23. An object is in a steady state when the object is not being translated, rotated, or scaled. When object 1402 is in a steady state, the surface of object 1402 that is parallel to display 106 and displayed on display 106 is determined to be selected. In this example, third surface 2002 is selected.
[0064] FIG. 24 illustrates an exemplary process for selecting a surface of a multi-sided graphical user interface object in response to a rotation of a crown. Process 2400 is performed at a wearable electronic device (e.g., device 100 in FIG. 1) having a physical crown. In some examples, the electronic device also includes a touch-sensitive display. The process provides an efficient technique for selecting a surface of a multi-sided, three-dimensional object.
[0065] At block 2402, the device causes a display of a multi-sided object on a touch-sensitive display of a wearable electronic device. Each selectable surface of the object is associated with a corresponding data value. The data may be, for example, text, an image, an application icon, an instruction, and the like.
[0066] At block 2404, the device receives crown position information. The crown position information may be received as a series of pulse signals, real values, integer values, and the like.
[0067] At block 2406, the device determines whether a change has occurred in a crown distance value. The crown distance value is based on an angular displacement of the physical crown of the wearable electronic device. A change in the crown distance value is indicative of a user providing input to the wearable electronic device by, for example, turning the physical crown. If the device determines that a change in the crown distance value has not occurred, the system returns to block 2404 and continues receiving crown position information. If the device determines that a change in the crown distance value has occurred, the system continues to block 2408, though the system may continue to receive crown position information.
[0068] At block 2408, the device determines a direction and a crown speed. The crown speed is based on the speed of rotation of the physical crown of the wearable electronic device. For example, the determined crown speed may be expressed as hertz, as rotations per unit of 13 PCT/US2014/053958 WO 2015/034966 time, as rotations per frame, as revolutions per unit of time, as revolutions per frame, and the like. The determined direction is based on a direction of rotation of the physical crown of the wearable electronic device. For example, an up direction can be determined based on a clockwise rotation of the physical crown. Similarly, a down direction can be determined based on a counterclockwise rotation of the physical crown. In other examples, a down direction can be determined based on a clockwise rotation of the physical crown and an up direction can be determined based on a counterclockwise rotation of the physical crown.
[0069] At block 2410, in response to determining the change in the crown distance value, the device causes an initial rotation of the multi-sided object on the display. The amount of the rotation is based on the determined crown speed. The direction of rotation is based on the determined direction. The rotation may be animated.
[0070] At block 2412, the device determines whether the determined crown speed exceeds a speed threshold. If the device determines that the determined crown speed exceeds the speed threshold, the device continues to block 2414. For example, the speed threshold may be thought of as an escape velocity (or escape speed). An escape velocity is the speed at which the kinetic energy plus the gravitational potential energy of an object is zero. If the device determines that the determined speed does not exceed the speed threshold, the device continues to block 2416.
[0071] In some examples, the minimum angular velocity of crown rotation that is necessary to reach escape velocity corresponds directly to the instantaneous angular velocity of crown 108 (FIG. 1), meaning that the user interface of device 100, in essence, responds when crown 108 reaches a sufficient angular velocity. In some embodiments, the minimum angular velocity of crown rotation necessary for reaching the escape velocity is a calculated velocity that is based on, but not directly equal to, the instantaneous (“current”) angular velocity of crown 108. In these examples, device 100 can maintain a calculated crown (angular) velocity V in discrete moments in time T according to equation 1: VT=V(T-i) + AVcrown - AVdrag· (EQ. 1) [0072] In equation 1, Vt represents a calculated crown velocity (speed and direction) at time T, V(T-i) represents the previous velocity (speed and direction) at time T-l, AVcrown represents the change in velocity caused by the force being applied through the rotation of the crown at time T, and AVdrag represents the change in velocity due to a drag force. The force being applied, which is reflected through AVcrown, can depend on the current velocity of angular rotation of the crown. Thus, AVcrown can also depend on the current angular velocity of the crown. In this way, device 100 can provide user interface interactions based not only on instantaneous crown velocity but also based on user input in the form of crown movement over multiple time intervals, even if those intervals are finely divided. Note, typically, in the absence of user input 14 PCT/US2014/053958 WO 2015/034966 in the form of AVcrown, Vt will approach (and become) zero based on AVdrag in accordance with EQ. 1, but Vt would not change signs without user input in the form of crown rotation (AVcrown)· [0073] Typically, the greater the velocity of angular rotation of the crown, the greater the value of AVcrown will be. However, the actual mapping between the velocity of angular rotation of the crown and AVcrown can be varied depending on the desired user interface effect. For example, various linear or non-linear mappings between the velocity of angular rotation of the crown and AVcrown can be used.
[0074] Also, AVDrag can take on various values. For example, AVDRAg can depend on the velocity of crown rotation such that at greater velocities, a greater opposing change in velocity (AVdrag) can be produced. In another example, AVdrag can have a constant value. It should be appreciated that the above-described requirements of AVcrown and AVdrag can be changed to produce desirable user interface effects.
[0075] As can be seen from EQ. 1, the maintained velocity (Vt) can continue to increase as long as AVcrown is greater than AVdrag· Additionally, Vt can have non-zero values even when no AVcrown input is being received, meaning that user interface objects can continue to change without the user rotating the crown. When this occurs, objects can stop changing based on the maintained velocity at the time the user stops rotating the crown and the AVdrag component.
[0076] In some examples, when the crown is rotated in a direction corresponding to a rotation direction that is opposite the current user interface changes, the V(x-q component can be reset to a value of zero, allowing the user to quickly change the direction of the object without having to provide a force sufficient to offset the Vt.
[0077] At block 2414, the device causes the object to flip past a transition position between a first surface that was last selected and a new surface. For example, the object has flipped past the transition position when the object will not return to having the first surface displayed parallel to the display without receiving additional user input.
[0078] Once the object reaches a steady state, the displayed surface that is parallel to the display can be activated through a designated user input. The displayed surface parallel to the display in the steady state is determined to be selected even before activation. An object is in a steady state when the object is not being translated, rotated, or scaled. This may result in the first surface of the object no longer being displayed, in the case of a cube-shaped object.
[0079] At block 2416, because the escape velocity has not been reached, the device causes the object to at least partially return to the object’s initial position at the time of block 2408. For example, part of the initial rotation of the object caused at block 2410 can be negated. To 15 PCT/US2014/053958 WO 2015/034966 achieve this, the device animates a rotation of the object that is in an opposite direction of the initial rotation at block 2410.
[0080] FIG. 25 illustrates a graphical user interface 2500 showing the selection of a surface 2506 of a multi-sided object in response to a rotation of a crown. Object 2502 is a 12-sided rotatable dial, shaped similar to a wheel. Object 2502 is rotatable along a fixed axis. In this example, all 12 surfaces of object 2502 are selectable. These 12 selectable surfaces include surface 2504, surface 2506, surface 2508, surface 2510, and surface 2512. In FIG. 25, surface 2508 is selected because surface 2508 is parallel to display 106 and is displayed on display 106. The selectable surfaces of object 2505 can be selected according to the processes and techniques described in other examples.
[0081] In some examples, device 100 can provide haptic feedback based on the content displayed on the display 106. When a user interface object is displayed on display 106, the device can modify the appearance of the object based on a change in a crown distance value received at the device 100 based on a rotation of crown 108. When a criterion is satisfied, a tactile output is output at the device 100.
[0082] In one example, the object is a rotatable multi-sided object, such as is described above. The criterion is satisfied when a surface of the multi-sided object is selected. In another example, the criterion is satisfied each time a displayed surface of the multi-sided object passes through a plane parallel to the display.
[0083] One or more of the functions relating to a user interface can be performed by a system similar or identical to system 2600 shown in FIG. 26. System 2600 can include instructions stored in a non-transitory computer readable storage medium, such as memory 2604 or storage device 2602, and executed by processor 2606. The instructions can also be stored and/or transported within any non-transitory computer readable storage medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “non-transitory computer readable storage medium” can be any medium that can contain or store the program for use by or in connection with the instruction execution system, apparatus, or device. The non-transitory computer readable storage medium can include, but is not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus or device, a portable computer diskette (magnetic), a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM) (magnetic), a portable optical disc such a CD, CD-R, CD-RW, DVD, DVD-R, or DVD-RW, or 16 PCT/US2014/053958 WO 2015/034966 flash memory such as compact flash cards, secured digital cards, USB memory devices, memory sticks, and the like.
[0084] The instructions can also be propagated within any transport medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based 5 system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions. In the context of this document, a “transport medium” can be any medium that can communicate, propagate or transport the program for use by or in connection with the instruction execution system, apparatus, or device. The transport medium can include, but is not limited to, an electronic, 10 magnetic, optical, electromagnetic or infrared wired or wireless propagation medium.
[0085] In some examples, system 2600 can be included within device 100. In these examples, processor 2606 can be the same or a different process than processor 202. Processor 2606 can be configured to receive the output from encoder 204, buttons 110, 112, and 114, and from touch-sensitive display 106. Processor 2606 can process these inputs as described above 15 with respect to the processes described and illustrated. It is to be understood that the system is not limited to the components and configuration of FIG. 26, but can include other or additional components in multiple configurations according to various examples.
[0086] Although the disclosure and examples have been fully described with reference to the accompanying drawings, it is to be noted that various changes and modifications will become 20 apparent to those skilled in the art. Such changes and modifications are to be understood as being included within the scope of the disclosure and examples as defined by the appended claims. 17
Claims (18)
- CLAIMS What is claimed is:1. A computer-implemented method comprising: displaying a first surface of a plurality of selectable surfaces of a virtual obj ect on a touch-sensitive display of a wearable electronic device, the first surface associated with a first data; determining a speed, wherein the speed is based on an angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a speed threshold, displaying on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state.
- 2. The computer-implemented method of claim 1, further comprising: determining a selection of the second surface in response to the display of the second surface parallel to the display in a steady state.
- 3. The computer-implemented method of claim 2, further comprising: in response to determining the selection of the second surface, generating a haptic output at the wearable electronic device.
- 4. The computer-implemented method of claim 1, further comprising: in response to a determination that the speed is below the speed threshold, displaying on the display an animation of the virtual object rotating about the axis parallel to the display, wherein the animation comprises displaying, in a steady state, the first surface of the plurality of selectable surfaces of the virtual object on the display parallel to the display.
- 5. The computer-implemented method of claim 4, fijrther comprising: determining a selection of the first surface in response to the display of the first surface, in a steady state, parallel to the display.
- 6. The computer-implemented method of claim 5, further comprising: in response to a determining a selection of the first surface, generating a haptic output at the wearable electronic device.
- 7. The computer-implemented method of any of claims 1-6, wherein the virtual object is a cube.
- 8. The computer-implemented method of any of claims 1 -6, wherein the virtual object is a multi-sided rotatable dial.
- 9. The computer-implemented method of any of claims 1-7, further comprising: associating the second surface with a second data, wherein the first data and the second data are different.
- 10. The computer-implemented method of any of claims 1 -8, further comprising determining a selection of the second surface responsive to one or more of detecting a tap gesture on the touch-sensitive display and detecting on the touch-sensitive display a touch with force greater than a predetermined threshold.
- 11. The computer-implemented method of any of claims 1-8, further comprising determining a selection of the second surface responsive to one or more of detecting a touch on the physical crown and detected a press on the physical crown.
- 12. The computer-implemented method of any of claims 1-8, further comprising determining a selection of the second surface responsive to detecting a touch on a touch-sensitive surface of the wearable electronic device.
- 13. The computer-implemented method of any of claims 1-12, wherein the speed is the rate of rotation of the virtual object.
- 14. A computer-implemented method comprising: displaying, on a touch-sensitive display ot a wearable electronic device, a first surface of a plurality of selectable surfaces of a virtual object, the first surface associated with a first data; determining an angular velocity of a physical crown of the wearable electronic device; and in response to a determination that the angular velocity exceeds a predetermined threshold, displaying an animation of the virtual object rotating about an axis parallel to the display and displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state, and in response to a determination that the angular velocity is below the predetermined threshold, maintaining display of the first surface in a steady state.
- 15. The computer-implemented method of claim 14, further comprising: in response to a determination that the angular velocity is below the predetermined threshold: displaying an animation of the first surface beginning parallel to the display, moving to an angle to the display, and returning to a steady state parallel to the display.
- 16. The computer-implemented method of claims 1 -15, wherein the physical crown is a mechanical crown.
- 17. A non-transitory computer-readable storage medium comprising instructions for: displaying a first surface of a plurality of selectable surfaces of a virtual object on a touch-sensitive display of a wearable electronic device, the first surface associated with a first data; determining a speed, wherein the speed is based on an angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a speed threshold, displaying on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state.
- 18. An electronic device comprising: one or more processors; a physical crown operatively coupled to the one or more processors; and a touch-sensitive display operatively coupled to the one or more processors, the one or more processors configured to: display a first surface of a plurality of selectable surfaces of a virtual object on a touch-sensitive display of a wearable electronic device, the first surface associated with a first data; determine a speed, wherein the speed is based on an angular velocity of a physical crown of the wearable electronic device; and in response to the speed being determined to exceed a speed threshold, display on the display an animation of the virtual object rotating about an axis parallel to the display, wherein the animation comprises displaying a second surface of the plurality of selectable surfaces of the virtual object on the display, the second surface displayed parallel to the display while in a steady state.
Applications Claiming Priority (11)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361873359P | 2013-09-03 | 2013-09-03 | |
US201361959851P | 2013-09-03 | 2013-09-03 | |
US201361873356P | 2013-09-03 | 2013-09-03 | |
US201361873360P | 2013-09-03 | 2013-09-03 | |
US61/873,356 | 2013-09-03 | ||
US61/873,359 | 2013-09-03 | ||
US61/873,360 | 2013-09-03 | ||
US61/959,851 | 2013-09-03 | ||
US201414476657A | 2014-09-03 | 2014-09-03 | |
PCT/US2014/053958 WO2015034966A1 (en) | 2013-09-03 | 2014-09-03 | User interface object manipulations in a user interface |
US14/476,657 | 2014-09-03 |
Publications (2)
Publication Number | Publication Date |
---|---|
AU2014315325A1 AU2014315325A1 (en) | 2016-04-21 |
AU2014315325B2 true AU2014315325B2 (en) | 2017-05-04 |
Family
ID=51589515
Family Applications (12)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2014315325A Active AU2014315325B2 (en) | 2013-09-03 | 2014-09-03 | User interface object manipulations in a user interface |
AU2014315324A Active AU2014315324B2 (en) | 2013-09-03 | 2014-09-03 | User interface for manipulating user interface objects |
AU2014315319A Active AU2014315319B2 (en) | 2013-09-03 | 2014-09-03 | Crown input for a wearable electronic device |
AU2017276285A Active AU2017276285B2 (en) | 2013-09-03 | 2017-12-14 | User interface for manipulating user interface objects |
AU2018200289A Active AU2018200289B2 (en) | 2013-09-03 | 2018-01-12 | Crown input for a wearable electronic device |
AU2019206101A Active AU2019206101B2 (en) | 2013-09-03 | 2019-07-18 | User interface for manipulating user interface objects |
AU2019257521A Abandoned AU2019257521A1 (en) | 2013-09-03 | 2019-11-01 | Crown input for a wearable electronic device |
AU2021201748A Active AU2021201748C1 (en) | 2013-09-03 | 2021-03-19 | User interface for manipulating user interface objects |
AU2021212114A Active AU2021212114B9 (en) | 2013-09-03 | 2021-08-06 | Crown input for a wearable electronic device |
AU2022235585A Abandoned AU2022235585A1 (en) | 2013-09-03 | 2022-09-21 | User interface for manipulating user interface objects |
AU2023237127A Pending AU2023237127A1 (en) | 2013-09-03 | 2023-09-28 | Crown input for a wearable electronic device |
AU2024205135A Pending AU2024205135A1 (en) | 2013-09-03 | 2024-07-26 | User interface for manipulating user interface objects |
Family Applications After (11)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2014315324A Active AU2014315324B2 (en) | 2013-09-03 | 2014-09-03 | User interface for manipulating user interface objects |
AU2014315319A Active AU2014315319B2 (en) | 2013-09-03 | 2014-09-03 | Crown input for a wearable electronic device |
AU2017276285A Active AU2017276285B2 (en) | 2013-09-03 | 2017-12-14 | User interface for manipulating user interface objects |
AU2018200289A Active AU2018200289B2 (en) | 2013-09-03 | 2018-01-12 | Crown input for a wearable electronic device |
AU2019206101A Active AU2019206101B2 (en) | 2013-09-03 | 2019-07-18 | User interface for manipulating user interface objects |
AU2019257521A Abandoned AU2019257521A1 (en) | 2013-09-03 | 2019-11-01 | Crown input for a wearable electronic device |
AU2021201748A Active AU2021201748C1 (en) | 2013-09-03 | 2021-03-19 | User interface for manipulating user interface objects |
AU2021212114A Active AU2021212114B9 (en) | 2013-09-03 | 2021-08-06 | Crown input for a wearable electronic device |
AU2022235585A Abandoned AU2022235585A1 (en) | 2013-09-03 | 2022-09-21 | User interface for manipulating user interface objects |
AU2023237127A Pending AU2023237127A1 (en) | 2013-09-03 | 2023-09-28 | Crown input for a wearable electronic device |
AU2024205135A Pending AU2024205135A1 (en) | 2013-09-03 | 2024-07-26 | User interface for manipulating user interface objects |
Country Status (5)
Country | Link |
---|---|
JP (11) | JP6333387B2 (en) |
KR (12) | KR102072614B1 (en) |
AU (12) | AU2014315325B2 (en) |
DK (1) | DK179231B1 (en) |
WO (3) | WO2015034965A1 (en) |
Families Citing this family (99)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9100493B1 (en) * | 2011-07-18 | 2015-08-04 | Andrew H B Zhou | Wearable personal digital device for facilitating mobile device payments and personal use |
TWI439960B (en) | 2010-04-07 | 2014-06-01 | Apple Inc | Avatar editing environment |
US9753436B2 (en) | 2013-06-11 | 2017-09-05 | Apple Inc. | Rotary input mechanism for an electronic device |
KR102231031B1 (en) | 2013-08-09 | 2021-03-23 | 애플 인크. | Tactile switch for an electronic device |
US10394325B2 (en) | 2013-12-10 | 2019-08-27 | Apple Inc. | Input friction mechanism for rotary inputs of electronic devices |
US10048802B2 (en) | 2014-02-12 | 2018-08-14 | Apple Inc. | Rejection of false turns of rotary inputs for electronic devices |
EP3195096B1 (en) | 2014-08-02 | 2020-08-12 | Apple Inc. | Context-specific user interfaces |
US10599101B2 (en) | 2014-09-02 | 2020-03-24 | Apple Inc. | Wearable electronic device |
US10145712B2 (en) | 2014-09-09 | 2018-12-04 | Apple Inc. | Optical encoder including diffuser members |
US9829350B2 (en) | 2014-09-09 | 2017-11-28 | Apple Inc. | Magnetically coupled optical encoder |
US9651405B1 (en) | 2015-03-06 | 2017-05-16 | Apple Inc. | Dynamic adjustment of a sampling rate for an optical encoder |
WO2016144385A1 (en) | 2015-03-08 | 2016-09-15 | Apple Inc. | Sharing user-configurable graphical constructs |
EP3251139B1 (en) | 2015-03-08 | 2021-04-28 | Apple Inc. | Compressible seal for rotatable and translatable input mechanisms |
ES2890451T3 (en) * | 2015-03-27 | 2022-01-19 | Saronikos Trading & Services Unipessoal Lda | Electronic wristwatch or pocket watch comprising a rotating crown |
WO2016171467A1 (en) | 2015-04-23 | 2016-10-27 | Samsung Electronics Co., Ltd. | Electronic device including rotary member and display method thereof |
KR102406102B1 (en) * | 2015-04-24 | 2022-06-10 | 삼성전자주식회사 | Electronic apparatus and method for displaying object thereof |
KR20160131275A (en) * | 2015-05-06 | 2016-11-16 | 엘지전자 주식회사 | Watch type terminal |
KR102356449B1 (en) | 2015-05-13 | 2022-01-27 | 삼성전자주식회사 | Apparatus and method for providing additional information according to rotary input |
CN105137746B (en) * | 2015-07-28 | 2018-03-27 | 广东欧珀移动通信有限公司 | A kind of receives frequency adjusting method and intelligent watch |
CN105068738A (en) * | 2015-07-28 | 2015-11-18 | 广东欧珀移动通信有限公司 | Control method for smartwatch and smartwatch |
CN105116996A (en) * | 2015-07-28 | 2015-12-02 | 广东欧珀移动通信有限公司 | Control method for smart watch and smart watch |
CN105141755A (en) * | 2015-07-28 | 2015-12-09 | 广东欧珀移动通信有限公司 | Information reply method, smart watch, terminal equipment and system |
CN105117002B (en) * | 2015-07-28 | 2017-07-11 | 广东欧珀移动通信有限公司 | The table hat and the operating method of intelligent watch of a kind of intelligent watch |
CN105117001B (en) * | 2015-07-28 | 2017-07-11 | 广东欧珀移动通信有限公司 | The table hat and the operating method of intelligent watch of a kind of intelligent watch |
CN105138116B (en) * | 2015-07-28 | 2018-07-06 | 广东欧珀移动通信有限公司 | A kind of information displaying method, smartwatch, terminal device and system |
CN105137819B (en) * | 2015-07-28 | 2019-07-02 | Oppo广东移动通信有限公司 | A kind of method and smartwatch of music |
CN105117118B (en) * | 2015-07-28 | 2019-02-01 | Oppo广东移动通信有限公司 | A kind of method and smartwatch controlling video playing |
CN105117120B (en) * | 2015-07-28 | 2017-07-11 | 广东欧珀移动通信有限公司 | The table hat and the operating method of intelligent watch of a kind of intelligent watch |
CN105116997B (en) * | 2015-07-28 | 2018-05-29 | 广东欧珀移动通信有限公司 | A kind of data encryption, the method for decryption and smartwatch |
CN106708379B (en) * | 2015-07-28 | 2020-01-10 | Oppo广东移动通信有限公司 | Interface operation method and device and smart watch |
CN105005479B (en) * | 2015-07-28 | 2018-06-29 | 广东欧珀移动通信有限公司 | A kind of alarm clock method for closing and smartwatch |
CN105117143B (en) * | 2015-07-28 | 2020-07-03 | Oppo广东移动通信有限公司 | Information display method, smart watch, server and system |
CN105116998B (en) * | 2015-07-28 | 2019-05-21 | Oppo广东移动通信有限公司 | A kind of method and smartwatch of fastopen |
CN113154645B (en) * | 2015-07-28 | 2022-07-26 | Oppo广东移动通信有限公司 | Air conditioner control method and smart watch |
CN105117121B (en) * | 2015-07-28 | 2019-04-02 | Oppo广东移动通信有限公司 | A kind of method that smartwatch is sought help and smartwatch |
CN105025630A (en) * | 2015-07-28 | 2015-11-04 | 广东欧珀移动通信有限公司 | Brightness adjusting method and intelligent watch |
CN105117119B (en) * | 2015-07-28 | 2018-12-11 | 广东欧珀移动通信有限公司 | A kind of method and smartwatch of rotation of screen picture |
CN105013175A (en) * | 2015-07-28 | 2015-11-04 | 广东欧珀移动通信有限公司 | Game motion control method and intelligent watch |
CN105389074A (en) * | 2015-07-28 | 2016-03-09 | 广东欧珀移动通信有限公司 | Control method for smart watch and smart watch |
CN105022947B (en) * | 2015-07-28 | 2019-02-22 | Oppo广东移动通信有限公司 | A kind of fingerprint identification method and smartwatch of smartwatch |
CN105025629B (en) * | 2015-07-28 | 2019-11-29 | Oppo广东移动通信有限公司 | A kind of control method and smartwatch of smartwatch |
EP4327731A3 (en) | 2015-08-20 | 2024-05-15 | Apple Inc. | Exercise-based watch face |
CN105208675B (en) * | 2015-08-26 | 2018-09-04 | 广东欧珀移动通信有限公司 | A kind of wireless connection method and smartwatch based on smartwatch |
CN105117010B (en) * | 2015-08-26 | 2018-12-11 | 广东欧珀移动通信有限公司 | A kind of method and smartwatch starting application program |
CN105068847B (en) * | 2015-08-26 | 2016-12-28 | 广东欧珀移动通信有限公司 | A kind of application program launching method and intelligent watch |
CN105117013B (en) * | 2015-08-26 | 2018-03-27 | 广东欧珀移动通信有限公司 | The unlocking method and intelligent watch of a kind of intelligent watch |
CN105224208B (en) * | 2015-08-26 | 2018-07-06 | 广东欧珀移动通信有限公司 | The method and smartwatch that a kind of page is shown |
CN105224193B (en) * | 2015-08-26 | 2018-05-29 | 广东欧珀移动通信有限公司 | The control method and smartwatch of a kind of smartwatch |
CN105204893B (en) * | 2015-08-26 | 2018-07-06 | 广东欧珀移动通信有限公司 | A kind of application control method and smartwatch |
CN105224072B (en) * | 2015-08-26 | 2018-07-06 | 广东欧珀移动通信有限公司 | The control method and smartwatch of a kind of music |
CN105068742B (en) * | 2015-08-26 | 2018-03-27 | 广东欧珀移动通信有限公司 | The control method and intelligent watch of a kind of intelligent watch |
CN105117014B (en) * | 2015-08-26 | 2018-03-27 | 广东欧珀移动通信有限公司 | A kind of friend-making management method and intelligent watch |
CN105227201B (en) * | 2015-08-26 | 2018-03-27 | 广东欧珀移动通信有限公司 | A kind of communication information answering method and intelligent watch |
CN105117011B (en) * | 2015-08-26 | 2017-08-29 | 广东欧珀移动通信有限公司 | A kind of method for operating application program, device and intelligent watch |
CN105068412B (en) * | 2015-08-26 | 2017-10-17 | 广东欧珀移动通信有限公司 | A kind of intelligent watch and operating method |
CN105117012B (en) * | 2015-08-26 | 2018-06-29 | 广东欧珀移动通信有限公司 | A kind of display interface method of adjustment and smartwatch |
CN105117129A (en) * | 2015-08-26 | 2015-12-02 | 广东欧珀移动通信有限公司 | Interface operation method and device and smart watch |
US10503271B2 (en) * | 2015-09-30 | 2019-12-10 | Apple Inc. | Proximity detection for an input mechanism of an electronic device |
US9983029B2 (en) | 2015-09-30 | 2018-05-29 | Apple Inc. | Integrated optical encoder for tilt able rotatable shaft |
KR102204682B1 (en) | 2016-01-14 | 2021-01-19 | 후아웨이 테크놀러지 컴퍼니 리미티드 | Electronic devices and methods of operating such electronic devices |
WO2017126727A1 (en) * | 2016-01-22 | 2017-07-27 | 엘지전자 주식회사 | Watch-type mobile terminal and method of operation thereof |
US10048837B2 (en) | 2016-02-16 | 2018-08-14 | Google Llc | Target selection on a small form factor display |
WO2017152139A1 (en) * | 2016-03-04 | 2017-09-08 | Apple Inc. | Input with haptic feedback |
CN107203261B (en) * | 2016-03-16 | 2022-05-24 | Lg电子株式会社 | Watch type mobile terminal and control method thereof |
US10025399B2 (en) * | 2016-03-16 | 2018-07-17 | Lg Electronics Inc. | Watch type mobile terminal and method for controlling the same |
US10551798B1 (en) | 2016-05-17 | 2020-02-04 | Apple Inc. | Rotatable crown for an electronic device |
JP6927670B2 (en) * | 2016-05-26 | 2021-09-01 | 株式会社アイ・オー・データ機器 | Operation reception device, program, and operation reception method |
US10061399B2 (en) | 2016-07-15 | 2018-08-28 | Apple Inc. | Capacitive gap sensor ring for an input device |
US10019097B2 (en) * | 2016-07-25 | 2018-07-10 | Apple Inc. | Force-detecting input structure |
KR102607562B1 (en) * | 2016-08-30 | 2023-11-30 | 삼성전자주식회사 | Method for providing visual effects according to interaction based on bezel and electronic device for the same |
US10324620B2 (en) | 2016-09-06 | 2019-06-18 | Apple Inc. | Processing capacitive touch gestures implemented on an electronic device |
CN114740963B (en) | 2016-09-23 | 2024-06-28 | 苹果公司 | Film watching mode |
DK179412B1 (en) | 2017-05-12 | 2018-06-06 | Apple Inc | Context-Specific User Interfaces |
DK179555B1 (en) | 2017-05-16 | 2019-02-13 | Apple Inc. | User interface for a flashlight mode on an electronic device |
US10962935B1 (en) | 2017-07-18 | 2021-03-30 | Apple Inc. | Tri-axis force sensor |
US10203662B1 (en) | 2017-09-25 | 2019-02-12 | Apple Inc. | Optical position sensor for a crown |
DK180078B1 (en) | 2018-05-07 | 2020-03-31 | Apple Inc. | USER INTERFACE FOR AVATAR CREATION |
US12033296B2 (en) | 2018-05-07 | 2024-07-09 | Apple Inc. | Avatar creation user interface |
US11327650B2 (en) | 2018-05-07 | 2022-05-10 | Apple Inc. | User interfaces having a collection of complications |
US11722764B2 (en) | 2018-05-07 | 2023-08-08 | Apple Inc. | Creative camera |
US11360440B2 (en) | 2018-06-25 | 2022-06-14 | Apple Inc. | Crown for an electronic watch |
US11561515B2 (en) | 2018-08-02 | 2023-01-24 | Apple Inc. | Crown for an electronic watch |
CN211293787U (en) | 2018-08-24 | 2020-08-18 | 苹果公司 | Electronic watch |
US11181863B2 (en) | 2018-08-24 | 2021-11-23 | Apple Inc. | Conductive cap for watch crown |
CN209625187U (en) | 2018-08-30 | 2019-11-12 | 苹果公司 | Electronic watch and electronic equipment |
US11194299B1 (en) | 2019-02-12 | 2021-12-07 | Apple Inc. | Variable frictional feedback device for a digital crown of an electronic watch |
US11960701B2 (en) | 2019-05-06 | 2024-04-16 | Apple Inc. | Using an illustration to show the passing of time |
DK180684B1 (en) | 2019-09-09 | 2021-11-25 | Apple Inc | Techniques for managing display usage |
DK181103B1 (en) | 2020-05-11 | 2022-12-15 | Apple Inc | User interfaces related to time |
US11921998B2 (en) | 2020-05-11 | 2024-03-05 | Apple Inc. | Editing features of an avatar |
CN115904596B (en) | 2020-05-11 | 2024-02-02 | 苹果公司 | User interface for managing user interface sharing |
KR102503135B1 (en) * | 2020-05-11 | 2023-02-23 | 애플 인크. | User interfaces related to time |
US11550268B2 (en) | 2020-06-02 | 2023-01-10 | Apple Inc. | Switch module for electronic crown assembly |
US11694590B2 (en) | 2020-12-21 | 2023-07-04 | Apple Inc. | Dynamic user interface with time indicator |
US11720239B2 (en) | 2021-01-07 | 2023-08-08 | Apple Inc. | Techniques for user interfaces related to an event |
US11921992B2 (en) | 2021-05-14 | 2024-03-05 | Apple Inc. | User interfaces related to time |
US11776190B2 (en) | 2021-06-04 | 2023-10-03 | Apple Inc. | Techniques for managing an avatar on a lock screen |
US12092996B2 (en) | 2021-07-16 | 2024-09-17 | Apple Inc. | Laser-based rotation sensor for a crown of an electronic watch |
US20230236547A1 (en) | 2022-01-24 | 2023-07-27 | Apple Inc. | User interfaces for indicating time |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1052566A1 (en) * | 1999-05-14 | 2000-11-15 | Alcatel | Graphical user interface |
US6266098B1 (en) * | 1997-10-22 | 2001-07-24 | Matsushita Electric Corporation Of America | Function presentation and selection using a rotatable function menu |
US6661438B1 (en) * | 2000-01-18 | 2003-12-09 | Seiko Epson Corporation | Display apparatus and portable information processing apparatus |
Family Cites Families (73)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5530455A (en) * | 1994-08-10 | 1996-06-25 | Mouse Systems Corporation | Roller mouse for implementing scrolling in windows applications |
US6047301A (en) * | 1996-05-24 | 2000-04-04 | International Business Machines Corporation | Wearable computer |
JP3673425B2 (en) * | 1999-04-16 | 2005-07-20 | 松下電器産業株式会社 | Program selection execution device and data selection execution device |
US7469381B2 (en) | 2007-01-07 | 2008-12-23 | Apple Inc. | List scrolling and document translation, scaling, and rotation on a touch-screen display |
US6809724B1 (en) * | 2000-01-18 | 2004-10-26 | Seiko Epson Corporation | Display apparatus and portable information processing apparatus |
US6522347B1 (en) * | 2000-01-18 | 2003-02-18 | Seiko Epson Corporation | Display apparatus, portable information processing apparatus, information recording medium, and electronic apparatus |
US6556222B1 (en) * | 2000-06-30 | 2003-04-29 | International Business Machines Corporation | Bezel based input mechanism and user interface for a smart watch |
US7081905B1 (en) * | 2000-06-30 | 2006-07-25 | International Business Machines Corporation | Method and apparatus for dynamically controlling scroller speed employed for a user interface of a wearable appliance |
JP2002175139A (en) * | 2000-12-07 | 2002-06-21 | Sony Corp | Information processor, menu display method and program storage medium |
JP3762243B2 (en) * | 2001-03-26 | 2006-04-05 | 陣山 俊一 | Information processing method, information processing program, and portable information terminal device |
US7312785B2 (en) * | 2001-10-22 | 2007-12-25 | Apple Inc. | Method and apparatus for accelerated scrolling |
JP2003248544A (en) * | 2002-02-25 | 2003-09-05 | Sony Corp | Graphical user interface, method for operating information processor, the information processor, and program |
WO2003075146A1 (en) * | 2002-03-05 | 2003-09-12 | Sony Ericsson Mobile Communications Japan, Inc. | Image processing device, image processing program, and image processing method |
JP3761165B2 (en) * | 2002-05-13 | 2006-03-29 | 株式会社モバイルコンピューティングテクノロジーズ | Display control device, portable information terminal device, program, and display control method |
JP2004021522A (en) * | 2002-06-14 | 2004-01-22 | Sony Corp | Apparatus, method, and program for information processing |
JP2004070654A (en) * | 2002-08-06 | 2004-03-04 | Matsushita Electric Ind Co Ltd | Portable electronic equipment |
JP2004184396A (en) * | 2002-10-09 | 2004-07-02 | Seiko Epson Corp | Display device, clock, method for controlling display device, control program, and recording medium |
JP2004178584A (en) | 2002-11-26 | 2004-06-24 | Asulab Sa | Input method of security code by touch screen for accessing function, device or specific place, and device for executing the method |
US20040130581A1 (en) * | 2003-01-03 | 2004-07-08 | Microsoft Corporation | Interaction model |
JP2004259063A (en) * | 2003-02-26 | 2004-09-16 | Sony Corp | Device and method for display processing for three dimensional object and computer program |
US8046705B2 (en) * | 2003-05-08 | 2011-10-25 | Hillcrest Laboratories, Inc. | Systems and methods for resolution consistent semantic zooming |
US20040264301A1 (en) * | 2003-06-30 | 2004-12-30 | Microsoft Corporation | Calendar user interface |
US7454713B2 (en) * | 2003-12-01 | 2008-11-18 | Sony Ericsson Mobile Communications Ab | Apparatus, methods and computer program products providing menu expansion and organization functions |
WO2005055034A1 (en) * | 2003-12-01 | 2005-06-16 | Research In Motion Limited | Previewing a new event on a small screen device |
US8082382B2 (en) * | 2004-06-04 | 2011-12-20 | Micron Technology, Inc. | Memory device with user configurable density/performance |
US7778671B2 (en) * | 2004-10-08 | 2010-08-17 | Nokia Corporation | Mobile communications terminal having an improved user interface and method therefor |
JP2006140990A (en) * | 2004-10-13 | 2006-06-01 | Olympus Corp | Image display apparatus, camera, display methods of image display apparatus and camera |
KR100630154B1 (en) * | 2005-08-31 | 2006-10-02 | 삼성전자주식회사 | Method for controlling display according to declension degree using a terrestrial magnetism sensor and the mobile terminal thereof |
US20070063995A1 (en) * | 2005-09-22 | 2007-03-22 | Bailey Eric A | Graphical user interface for use with a multi-media system |
JP2007170995A (en) * | 2005-12-22 | 2007-07-05 | Casio Comput Co Ltd | Electronic equipment and electronic timepiece |
KR100678963B1 (en) * | 2005-12-28 | 2007-02-06 | 삼성전자주식회사 | Portable device and operation method comprising input button to enable revolution |
KR100754674B1 (en) * | 2006-03-10 | 2007-09-03 | 삼성전자주식회사 | Method and apparatus for selecting menu in portable terminal |
KR100896055B1 (en) * | 2007-01-15 | 2009-05-07 | 엘지전자 주식회사 | Mobile terminal having a rotating input device and display method thereof |
KR20080073868A (en) * | 2007-02-07 | 2008-08-12 | 엘지전자 주식회사 | Terminal and method for displaying menu |
TW200734916A (en) * | 2007-05-03 | 2007-09-16 | Ying-Chu Lee | Method of using mouse wheel to operate picture |
WO2009053606A2 (en) * | 2007-10-12 | 2009-04-30 | France Telecom | Device for displaying a plurality of multimedia documents |
JP4462338B2 (en) * | 2007-11-27 | 2010-05-12 | セイコーエプソン株式会社 | Electronic clock, electronic clock time correction method, electronic clock control program |
JP5356713B2 (en) * | 2008-03-28 | 2013-12-04 | 京セラ株式会社 | Mobile phone |
JP2009265793A (en) | 2008-04-23 | 2009-11-12 | Sony Ericsson Mobilecommunications Japan Inc | Display and operation device, operation device and program |
KR101512041B1 (en) * | 2008-07-01 | 2015-04-14 | 엘지전자 주식회사 | Mobile terminal and control method thereof |
KR101546774B1 (en) * | 2008-07-29 | 2015-08-24 | 엘지전자 주식회사 | Mobile terminal and operation control method thereof |
KR101555055B1 (en) * | 2008-10-10 | 2015-09-22 | 엘지전자 주식회사 | Mobile terminal and display method thereof |
US20110055752A1 (en) * | 2009-06-04 | 2011-03-03 | Rubinstein Jonathan J | Method and Apparatus for Displaying and Auto-Correcting an Over-Scroll State on a Computing Device |
JP5513071B2 (en) * | 2009-10-26 | 2014-06-04 | 株式会社プロフィールド | Information processing apparatus, information processing method, and program |
CH701440A2 (en) | 2009-07-03 | 2011-01-14 | Comme Le Temps Sa | Wrist touch screen and method for displaying on a watch with touch screen. |
KR101608764B1 (en) * | 2009-07-14 | 2016-04-04 | 엘지전자 주식회사 | Mobile terminal and method for controlling display thereof |
KR101595384B1 (en) * | 2009-07-20 | 2016-02-18 | 엘지전자 주식회사 | Watch type mobile terminal |
JP5333068B2 (en) | 2009-08-31 | 2013-11-06 | ソニー株式会社 | Information processing apparatus, display method, and display program |
KR101649646B1 (en) * | 2010-02-11 | 2016-08-19 | 엘지전자 주식회사 | Portable terminal |
US8930841B2 (en) * | 2010-02-15 | 2015-01-06 | Motorola Mobility Llc | Methods and apparatus for a user interface configured to display event information |
CH702862B1 (en) * | 2010-03-30 | 2024-06-14 | Smart Communications Sa | Wristwatch with electronic display. |
US20110252376A1 (en) | 2010-04-07 | 2011-10-13 | Imran Chaudhri | Device, Method, and Graphical User Interface for Managing Concurrently Open Software Applications |
US20110316888A1 (en) * | 2010-06-28 | 2011-12-29 | Invensense, Inc. | Mobile device user interface combining input from motion sensors and other controls |
JP5676952B2 (en) * | 2010-07-26 | 2015-02-25 | キヤノン株式会社 | Display control apparatus, display control method, program, and storage medium |
JP5745241B2 (en) * | 2010-09-08 | 2015-07-08 | 任天堂株式会社 | Information processing program, information processing apparatus, information processing system, and information processing method |
US9104211B2 (en) * | 2010-11-19 | 2015-08-11 | Google Inc. | Temperature controller with model-based time to target calculation and display |
JP5762718B2 (en) | 2010-10-20 | 2015-08-12 | シャープ株式会社 | Image forming apparatus |
US9244606B2 (en) * | 2010-12-20 | 2016-01-26 | Apple Inc. | Device, method, and graphical user interface for navigation of concurrently open software applications |
KR101740439B1 (en) * | 2010-12-23 | 2017-05-26 | 엘지전자 주식회사 | Mobile terminal and method for controlling thereof |
US9423951B2 (en) * | 2010-12-31 | 2016-08-23 | Microsoft Technology Licensing, Llc | Content-based snap point |
KR101785323B1 (en) * | 2011-01-05 | 2017-10-17 | 삼성전자주식회사 | Method and apparatus for providing a user interface in a portable terminal |
TWI441051B (en) * | 2011-01-25 | 2014-06-11 | Compal Electronics Inc | Electronic device and information display method thereof |
US20120246678A1 (en) * | 2011-03-24 | 2012-09-27 | Tobe Barksdale | Distance Dependent Scalable User Interface |
JP2012252384A (en) | 2011-05-31 | 2012-12-20 | Camelot:Kk | Screen control system, screen control method, and screen control program |
JP2013003718A (en) * | 2011-06-14 | 2013-01-07 | Mitsubishi Electric Information Systems Corp | Information processing device, scroll display method of information processing device, and scroll display program |
EP2551784A1 (en) * | 2011-07-28 | 2013-01-30 | Roche Diagnostics GmbH | Method of controlling the display of a dataset |
US20130097566A1 (en) * | 2011-10-17 | 2013-04-18 | Carl Fredrik Alexander BERGLUND | System and method for displaying items on electronic devices |
US20130117698A1 (en) | 2011-10-31 | 2013-05-09 | Samsung Electronics Co., Ltd. | Display apparatus and method thereof |
JP6159078B2 (en) * | 2011-11-28 | 2017-07-05 | 京セラ株式会社 | Apparatus, method, and program |
JP2013152693A (en) * | 2011-12-27 | 2013-08-08 | Nintendo Co Ltd | Information processing program, information processing device, image display method, and image display system |
CN103460164B (en) * | 2012-02-03 | 2017-02-08 | 松下知识产权经营株式会社 | Tactile sense presentation device and method for driving tactile sense presentation device |
JP2013164700A (en) * | 2012-02-10 | 2013-08-22 | Samsung Electronics Co Ltd | Scroll method and scroll device for portable terminal |
KR20130094054A (en) * | 2012-02-15 | 2013-08-23 | 삼성전자주식회사 | Apparatus and method for managing object in portable electronic device |
-
2014
- 2014-09-03 KR KR1020167008488A patent/KR102072614B1/en active IP Right Grant
- 2014-09-03 KR KR1020197007748A patent/KR102143895B1/en active IP Right Grant
- 2014-09-03 KR KR1020197028736A patent/KR20190114034A/en active Application Filing
- 2014-09-03 WO PCT/US2014/053957 patent/WO2015034965A1/en active Application Filing
- 2014-09-03 KR KR1020167008474A patent/KR20160048967A/en active Application Filing
- 2014-09-03 KR KR1020187013265A patent/KR102131228B1/en active IP Right Grant
- 2014-09-03 AU AU2014315325A patent/AU2014315325B2/en active Active
- 2014-09-03 WO PCT/US2014/053951 patent/WO2015034960A1/en active Application Filing
- 2014-09-03 KR KR1020167008449A patent/KR102045111B1/en active IP Right Grant
- 2014-09-03 JP JP2016537946A patent/JP6333387B2/en active Active
- 2014-09-03 JP JP2016537945A patent/JP6397918B2/en active Active
- 2014-09-03 KR KR1020217001918A patent/KR20210010661A/en not_active Application Discontinuation
- 2014-09-03 KR KR1020207019035A patent/KR102305362B1/en active IP Right Grant
- 2014-09-03 KR KR1020187032106A patent/KR102029303B1/en active IP Right Grant
- 2014-09-03 KR KR1020207022802A patent/KR102263620B1/en active IP Right Grant
- 2014-09-03 JP JP2016537947A patent/JP6170250B2/en active Active
- 2014-09-03 KR KR1020187010872A patent/KR102111452B1/en active IP Right Grant
- 2014-09-03 AU AU2014315324A patent/AU2014315324B2/en active Active
- 2014-09-03 KR KR1020217017259A patent/KR20210070395A/en not_active IP Right Cessation
- 2014-09-03 WO PCT/US2014/053958 patent/WO2015034966A1/en active Application Filing
- 2014-09-03 AU AU2014315319A patent/AU2014315319B2/en active Active
-
2016
- 2016-02-29 DK DKPA201670117A patent/DK179231B1/en not_active IP Right Cessation
-
2017
- 2017-12-14 AU AU2017276285A patent/AU2017276285B2/en active Active
-
2018
- 2018-01-12 AU AU2018200289A patent/AU2018200289B2/en active Active
- 2018-04-24 JP JP2018083313A patent/JP6564493B2/en active Active
- 2018-05-08 JP JP2018090084A patent/JP6547039B2/en active Active
-
2019
- 2019-06-24 JP JP2019116590A patent/JP7128153B2/en active Active
- 2019-07-18 AU AU2019206101A patent/AU2019206101B2/en active Active
- 2019-07-26 JP JP2019138053A patent/JP6924802B2/en active Active
- 2019-11-01 AU AU2019257521A patent/AU2019257521A1/en not_active Abandoned
-
2021
- 2021-03-19 AU AU2021201748A patent/AU2021201748C1/en active Active
- 2021-07-05 JP JP2021111630A patent/JP7471262B2/en active Active
- 2021-08-02 JP JP2021126843A patent/JP7223081B2/en active Active
- 2021-08-06 AU AU2021212114A patent/AU2021212114B9/en active Active
-
2022
- 2022-09-21 AU AU2022235585A patent/AU2022235585A1/en not_active Abandoned
-
2023
- 2023-02-03 JP JP2023015606A patent/JP7532568B2/en active Active
- 2023-06-09 JP JP2023095734A patent/JP2023126783A/en active Pending
- 2023-09-28 AU AU2023237127A patent/AU2023237127A1/en active Pending
-
2024
- 2024-07-26 AU AU2024205135A patent/AU2024205135A1/en active Pending
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6266098B1 (en) * | 1997-10-22 | 2001-07-24 | Matsushita Electric Corporation Of America | Function presentation and selection using a rotatable function menu |
EP1052566A1 (en) * | 1999-05-14 | 2000-11-15 | Alcatel | Graphical user interface |
US6661438B1 (en) * | 2000-01-18 | 2003-12-09 | Seiko Epson Corporation | Display apparatus and portable information processing apparatus |
Also Published As
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11829576B2 (en) | User interface object manipulations in a user interface | |
AU2014315325B2 (en) | User interface object manipulations in a user interface | |
US10275117B2 (en) | User interface object manipulations in a user interface | |
EP3042271B1 (en) | User interface object manipulations in a user interface | |
US12050766B2 (en) | Crown input for a wearable electronic device | |
US10691230B2 (en) | Crown input for a wearable electronic device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FGA | Letters patent sealed or granted (standard patent) |