WO2023219959A1 - Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application - Google Patents

Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application Download PDF

Info

Publication number
WO2023219959A1
WO2023219959A1 PCT/US2023/021376 US2023021376W WO2023219959A1 WO 2023219959 A1 WO2023219959 A1 WO 2023219959A1 US 2023021376 W US2023021376 W US 2023021376W WO 2023219959 A1 WO2023219959 A1 WO 2023219959A1
Authority
WO
WIPO (PCT)
Prior art keywords
user interface
wake
displaying
representation
version
Prior art date
Application number
PCT/US2023/021376
Other languages
English (en)
Inventor
Graham R. CLARKE
Eric Lance Wilson
Caelan G. Stack
Christopher P. Foss
Andrew P. CLYMER
William M. TYLER
Gregory M. Apodaca
William A. SORRENTINO, III.
Christian X. Dalonzo
Andre Souza Dos Santos
Alan C. Dye
Taylor G. Carrigan
Noah A. Witherspoon
Bobby MARTIN, Jr.
Aaron M. Melim
Original Assignee
Apple Inc.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US17/949,128 external-priority patent/US11868601B2/en
Application filed by Apple Inc. filed Critical Apple Inc.
Publication of WO2023219959A1 publication Critical patent/WO2023219959A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus

Definitions

  • This relates generally to electronic devices with touch-sensitive surfaces, including but not limited to electronic devices with touch-sensitive surfaces that display notifications and application information for applications of the electronic device.
  • Example touch-sensitive surfaces include touchpads and touch-screen displays. Such surfaces are widely used to manipulate user interfaces and objects therein on a display.
  • Example manipulations include adjusting the position and/or size of one or more user interface objects or activating buttons or opening files/applications represented by user interface objects, as well as associating metadata with one or more user interface objects or otherwise manipulating user interfaces.
  • Example user interface objects include digital images, video, text, icons, control elements such as buttons and other graphics.
  • the device is a desktop computer.
  • the device is portable (e.g., a notebook computer, tablet computer, or handheld device).
  • the device is a personal electronic device (e.g., a wearable electronic device, such as a watch).
  • the device has a touchpad.
  • the device has a touch-sensitive display (also known as a “touch screen” or “touch-screen display”).
  • the device has a graphical user interface (GUI), one or more processors, memory and one or more modules, programs or sets of instructions stored in the memory for performing multiple functions.
  • GUI graphical user interface
  • the user interacts with the GUI primarily through stylus and/or finger contacts and gestures on the touch-sensitive surface.
  • the functions optionally include image editing, drawing, presenting, word processing, spreadsheet making, game playing, telephoning, video conferencing, e-mailing, instant messaging, workout support, digital photographing, digital videoing, web browsing, digital music playing, note taking, and/or digital video playing. Executable instructions for performing these functions are, optionally, included in a non- transitory computer readable storage medium or other computer program product configured for execution by one or more processors.
  • a method is performed at a computer system that is in communication with a display generation component.
  • the method includes displaying, via the display generation component, a first version of a first user interface that corresponds to a restricted state of the computer system.
  • Displaying the first version of the first user interface includes displaying a first plurality of user interface objects displayed concurrently with a first background in accordance with a first configuration.
  • the first plurality of user interface objects correspond to a first plurality of applications and include respective content from the first plurality of applications and are updated periodically as information represented by the first plurality of user interface objects changes.
  • the method further includes, while displaying the first version of the first user interface, detecting a first input.
  • the method further includes, in response to detecting the first input: in accordance with a determination that the first input meets first criteria, wherein the first criteria require that the first input includes first movement in a first direction in order for the first criteria to be met, replacing display of the first version of the first user interface with display of a second user interface that includes respective representations of a second plurality of applications, wherein the respective representations of the second plurality of applications, when activated, cause the computer system to launch corresponding applications of the respective representations; and in accordance with a determination that the first input meets second criteria, wherein the second criteria require that the first input includes second movement in a second direction, different from the first direction, in order for the second criteria to be met, replacing display of the first version of the first user interface with display of a second version of the first user interface, wherein displaying the second version of the first user interface includes displaying a second plurality of user interface objects concurrently with a second background in accordance with a second configuration.
  • the second plurality of user interface objects correspond to a third plurality of applications and include respective content from the third plurality of applications and are updated periodically as information represented by the second plurality of user interface objects changes.
  • the first background is different from the second background, the first plurality of user interface objects is different from the second plurality of user interface objects, and/or the first configuration is different from the second configuration.
  • a method is performed at a computer system that is in communication with a display generation component.
  • the method includes detecting a request to change a wake user interface of the computer system.
  • the wake user interface is a user interface that is displayed when the computer system is turned on or transitioned from a low power state to a higher power state and corresponds to a restricted mode of operation for the computer system.
  • the method further includes, in response to detecting the request to change the wake user interface of the computer system, displaying, via the display generation component, a first user interface for changing the wake user interface for the computer system.
  • the method further includes, while displaying the first user interface, concurrently displaying a first representation of the wake user interface, and a first representation of a home user interface, where the home user interface is a user interface that is displayed when the wake user interface is dismissed and the computer system has exited the restricted mode of operation.
  • the first representation of the wake user interface corresponds to a first set of one or more wake user interface settings including a first wake user interface background.
  • the first representation of the home user interface corresponds to a first set of one or more home user interface settings including a first home user interface background.
  • the method further includes displaying a second representation of the wake user interface.
  • the second representation of the wake user interface corresponds to a second set of one or more wake user interface settings including a second wake user interface background that is different from the first wake user interface background.
  • the method further includes detecting a sequence of one or more inputs corresponding to selection of a respective representation of the wake user interface for the computer system from the first user interface.
  • the method further includes, in response to detecting the sequence of one or more inputs: in accordance with a determination that the first representation of the wake user interface was selected based on the sequence of one or more inputs, setting the wake user interface of the computer system based on the first set of one or more wake user interface settings associated with the first representation of the wake user interface, including using the first wake user interface background as a respective background for the wake user interface and setting the home user interface of the computer system based on the first set of one or more home user interface settings, including using the first home user interface background as a respective background for the home user interface; and in accordance with a determination that the second representation of the wake user interface was selected based on the sequence of one or more inputs, setting the wake user interface of the computer system based on the second set of one or more wake user interface settings associated with the second representation of the wake user interface, including using the second wake user interface background as the background for the wake user interface.
  • a method is performed at a computer system that is in communication with a display generation component.
  • the method includes displaying, via the display generation component, a first user interface for configuring a wake user interface.
  • a respective version of the wake user interface includes a respective background and a respective plurality of editable user interface objects overlaying the respective background.
  • the first user interface displays at least a first representation of a first version of the wake user interface illustrating a first plurality of editable user interface objects overlaying a first background.
  • the method further includes, while displaying the first user interface, detecting a first input directed to the first user interface.
  • the method further includes, in response to detecting the first input directed to the first user interface: in accordance with a determination that the first input meets first criteria, displaying a second user interface for editing a first user interface object of the first plurality of editable user interface objects, wherein the first user interface object is selected in accordance with a location of the first input; and in accordance with a determination that the first input meets second criteria different from the first criteria, updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of a second representation of a second version of the wake user interface.
  • the second representation of the second version of the wake user interface includes a second plurality of editable user interface objects overlaying a second background that is different from the first background.
  • the second plurality of editable user interface objects is different from the first plurality of editable user interface objects.
  • a method is performed at a computer system with a display generation component and one or more input devices.
  • the method includes displaying, via the display generation component, a first user interface that includes a plurality of notifications including: in accordance with a determination that the computer system has a first mode for displaying notifications enabled, displaying a representation of the plurality of notifications in a first configuration in a first region of the first user interface; and in accordance with a determination that the computer system has a second mode for displaying notifications enabled, displaying the representation of the plurality of notifications in a second configuration in a second region of the first user interface that is smaller than the first region of the first user interface.
  • the method further includes, while displaying the first user interface, detecting a first user input at a respective location on the first user interface corresponding to the representation of the plurality of notifications.
  • the method further includes, in response to detecting the first user input, and while continuing to detect the first user input: in accordance with a determination that the first user input meets first criteria and in accordance with a determination that the representation of the plurality of notifications is displayed with the first configuration, scrolling notifications in the plurality of notifications in the first region of the first user interface in accordance with the first user input; and in accordance with a determination that the first user input meets the first criteria and in accordance with a determination that the representation of the plurality of notifications is displayed with the second configuration, scrolling the notifications in the plurality of notifications in a third region of the first user interface, in accordance with the first user input.
  • a method is performed at a computer system that is in communication with a display generation component.
  • the method includes detecting one or more inputs to subscribe to updates from a first application for a first event, and to subscribe to updates from a second application for a second event.
  • the method further includes displaying a first user interface.
  • the first user interface includes a first region at a first location in the first user interface.
  • Displaying the first user interface includes: in accordance with a determination that the first event is active and that the second event is not active, displaying a first representation of the first event in the first region of the first user interface, and updating first information contained in the first representation of the first event in accordance with updates received from the first application for the first event; and in accordance with a determination that the second event is active and that the first event is not active, displaying a second representation of the second event in the first region of the first user interface, and updating second information contained in the second representation of the second event in accordance with updates received from the second application for the second event.
  • a method is performed at a computer system with a display generation component and one or more input devices.
  • the method includes, while displaying a wake user interface that includes a representation of a first plurality of notifications in a first configuration, wherein the wake user interface is a user interface that is displayed when the computer system wakes from a low power state, detecting, via the one or more input devices, a first user input.
  • the method further includes, in response to detecting the first user input: in accordance with a determination that the first user input meets first criteria, displaying the representation of the first plurality of notifications in a second configuration on the wake user interface, wherein the second configuration is different from the first configuration; and in accordance with a determination that the first user input does not meet the first criteria, maintaining display of the representation of the first plurality of notifications in the first configuration on the wake user interface.
  • the method further includes, after detecting the first user input, detecting an occurrence of a condition that causes the computer system to redisplay the wake user interface.
  • the method further includes, in response to detecting the occurrence of the condition that causes the computer system to redisplay the wake user interface: in accordance with a determination that the first user input met the first criteria, displaying a representation of a second plurality of notifications in the second configuration; and in accordance with a determination that the first user input did not meet the first criteria, displaying the representation of the second plurality of notifications in the first configuration.
  • a method is performed at a computer system with a display generation component.
  • the method includes displaying, via the display generation component, a first user interface for configuring a system user interface that has a first background and a first set of one or more system user interface objects overlaying the first background, wherein: while the system user interface is displayed, the computer system automatically shuffles through two or more media items selected from a collection of media items in the first background over time; the first user interface includes respective selectable representations of a plurality of categories for media items associated with the computer system, including at least a first selectable representation of a first category and a second selectable representation of a second category; a first plurality of media items associated with the computer system are automatically selected for inclusion in the first category based on the first plurality of media items containing automatically detected content of a first type; and a second plurality of media items associated with the computer system are automatically selected for inclusion in the second category based on the second plurality of media items containing automatically detected content of
  • the method further includes, while displaying the first user interface for configuring the system user interface, detecting a first input selecting a set of one or more of the plurality of categories; and after the set of one or more of the plurality of categories were selected by the first input, displaying the system user interface, wherein displaying the system user interface includes, over time displaying the system user interface with a plurality of versions of the first background that respectively include media items selected from media items in respective categories in the set of one or more of the plurality of categories, wherein: in accordance with a determination that the set of one or more of the plurality of categories includes the first category, without including the second category, the plurality of versions of the first background include media items from the first category without including media items from the second category; in accordance with a determination that the set of one or more of the plurality of categories includes the second category, without including the first category, the plurality of versions of the first background include media items from the second category without including media items from the first category; and in accordance with a determination that the set of one or more
  • a method is performed at a computer system with a display generation component.
  • the method includes displaying, via the display generation component, a first representation of a system user interface, wherein a respective version of the system user interface includes a respective background and a respective set of one or more system user interface objects overlaying the respective background, and wherein the first representation of the system user interface corresponds to a first version of the system user interface illustrating a first set of one or more system user interface objects overlaying a first background.
  • the method further includes, while displaying the first representation of the system user interface that corresponds to the first version of the system user interface, detecting occurrence of a first condition that causes the computer system to change an appearance of the system user interface based on a first combination of a first background media item and a first filter for the system user interface.
  • the method includes in response to detecting the occurrence of the first condition that causes the computer system to change the appearance of the system user interface based on the first combination of the first background media item and the first filter for the system user interface: in accordance with a determination that the first combination of the first background media item and the first filter meets first criteria, wherein the first criteria require that a first set of one or more visual properties of the first background media item meets a first requirement in order for the first combination of the first background media item and the first filter to meet the first criteria, applying a first version of the first filter to the first background media item to create a second version of the system user interface by modifying the first background media item in a first manner; and in accordance with a determination that the first combination of the first background media item and the first filter meets second criteria, wherein the second criteria require that the first set of one or more visual properties of the first background media item meets a second requirement different from the first requirement in order for the first background media item to meet the second criteria, applying a second version of the first filter to the first background media item to
  • a method is performed at a computer system with a display generation component.
  • the method includes displaying, via the display generation component, a wake screen user interface that corresponds to a restricted state of the computer system, including displaying a first background and a plurality of system user interface objects overlaying at least a portion of the first background, wherein the first background includes a plurality of graphical elements arranged in accordance with a first spatial configuration.
  • the method further includes, while displaying the wake screen user interface that corresponds to the restricted state of the computer system, detecting a first user input, including a request to dismiss the wake screen user interface.
  • the method includes, in response to detecting the first user input that includes the request to dismiss the wake screen user interface: moving the plurality of graphical elements in a first direction in accordance with the first user input, while increasing a spatial gap between the plurality of graphical elements; and in accordance with a determination that the request to dismiss the wake screen user interface included in the first user input meets first criteria, replacing display of the wake screen user interface that corresponds to the restricted state of the computer system with display of a second user interface different from the wake screen user interface, including displaying the plurality of graphical elements in the second user interface while reducing the spatial gap between the plurality of graphical elements.
  • an electronic device includes a display, a touch-sensitive surface, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, optionally one or more tactile output generators, one or more processors, and memory storing one or more programs; the one or more programs are configured to be executed by the one or more processors and the one or more programs include instructions for performing or causing performance of the operations of any of the methods described herein.
  • a computer readable storage medium has stored therein instructions that, when executed by an electronic device with a display, a touch-sensitive surface, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, and optionally one or more tactile output generators, cause the device to perform or cause performance of the operations of any of the methods described herein.
  • a graphical user interface on an electronic device with a display, a touch-sensitive surface, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, optionally one or more tactile output generators, a memory, and one or more processors to execute one or more programs stored in the memory includes one or more of the elements displayed in any of the methods described herein, which are updated in response to inputs, as described in any of the methods described herein.
  • an electronic device includes: a display, a touch-sensitive surface, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, and optionally one or more tactile output generators; and means for performing or causing performance of the operations of any of the methods described herein.
  • an information processing apparatus for use in an electronic device with a display, a touch-sensitive surface, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, and optionally one or more tactile output generators, includes means for performing or causing performance of the operations of any of the methods described herein.
  • electronic devices with displays, touch-sensitive surfaces, optionally one or more sensors to detect intensities of contacts with the touch-sensitive surface, optionally one or more tactile output generators, optionally one or more device orientation sensors, and optionally an audio system are provided with improved methods and interfaces for modifying user interfaces and displaying notifications and/or status information, thereby increasing the effectiveness, efficiency, and user satisfaction with such devices.
  • Such methods and interfaces may complement or replace conventional methods for modifying user interfaces and displaying notifications and/or status information.
  • Figure 1 A is a block diagram illustrating a portable multifunction device with a touch-sensitive display in accordance with some embodiments.
  • Figure IB is a block diagram illustrating example components for event handling in accordance with some embodiments.
  • Figure 2 illustrates a portable multifunction device having a touch screen in accordance with some embodiments.
  • Figure 3 is a block diagram of an example multifunction device with a display and a touch-sensitive surface in accordance with some embodiments.
  • Figure 4A illustrates an example user interface for a menu of applications on a portable multifunction device in accordance with some embodiments.
  • Figure 4B illustrates an example user interface for a multifunction device with a touch-sensitive surface that is separate from the display in accordance with some embodiments.
  • Figures 5A-5CB illustrate example user interfaces for navigating and editing respective types of user interfaces in accordance with some embodiments.
  • Figures 6A-6AR illustrate example user interfaces for an expanded face switcher in accordance with some embodiments.
  • Figures 7A-7V-3 illustrate a representation of a plurality of notifications in different configurations, and user inputs for switching between the different configurations in accordance with some embodiments.
  • Figures 8A-8AN illustrate example user interfaces for displaying a session region for ongoing events in accordance with some embodiments.
  • Figures 9A-9I are flow diagrams of a process for changing a user interface based on a user input in accordance with some embodiments.
  • Figures 10A-10I are flow diagrams of a process for selecting and/or modifying a user interface in accordance with some embodiments.
  • Figures 11 A-l 1H are flow diagrams of a process for modifying a user interface in accordance with some embodiments.
  • Figures 12A-12E are flow diagrams of a process for displaying a representation of a plurality of notifications in different configurations in accordance with some embodiments.
  • Figures 13A-13G are flow diagrams of a process for displaying status information in a session region based on subscriptions in accordance with some embodiments.
  • Figures 14A-14G are flow diagrams of a process for changing between different configurations in which a representation of a plurality of notifications can be displayed in accordance with some embodiments.
  • Figures 15A-15I are flow diagrams of a process for automatically shuffling through media items as a background in accordance with some embodiments.
  • Figures 16A-16J are flow diagrams of a process for applying different versions of a filter to a background media item in accordance with some embodiments.
  • Figures 17A-17D are flow diagrams of a process for providing feedback on a system user interface in accordance with some embodiments.
  • Many electronic devices have graphical user interfaces that allow a user to navigate between application user interfaces and/or system user interfaces.
  • Some methods for navigating between user interfaces enable multitasking, such that a respective application continues to update in the background even after navigating away from the respective application user interface.
  • Some methods for providing a system user interface limit customizations made to the system user interface, which can obscure certain elements and/or status information displayed on the system user interface. For example, with these methods, a user may need to navigate back to the respective application user interface in order to view the updates.
  • an improved method for providing status updates for a plurality of applications within a persistent session region is provided. This method streamlines the user’s ability to view real-time status information for active sessions, thereby eliminating the need for extra, separate steps to navigate back to the respective user interface of the respective application to view a status update.
  • the methods, devices, and GUIs described herein use haptic feedback to improve user interface interactions in multiple ways. For example, they make it easier to indicate hidden thresholds and indicate user interface components that represent selectable options.
  • the processes described below enhance the operability of the devices and make the user-device interfaces more efficient (e.g., by helping the user to provide proper inputs and reducing user mistakes when operating/interacting with the device) through various techniques, including by providing improved visual, audio, and/or tactile feedback to the user, reducing the number of inputs needed to perform an operation, providing additional control options without cluttering the user interface with additional displayed controls, performing an operation when a set of conditions has been met without requiring further user input, and/or additional techniques. These techniques also reduce power usage and improve battery life of the device by enabling the user to use the device more quickly and efficiently.
  • Figures 1A-1B, 2, and 3 provide a description of example devices.
  • Figures 4A-4B and 5A-5CB illustrate example user interfaces for navigating and editing respective types of user interfaces.
  • Figures 6A-6AR illustrate example user interfaces for an expanded face switcher.
  • Figures 7A-7V-3 illustrate a representation of a plurality of notifications in different configurations, and user inputs for switching between the different configurations.
  • Figures 8A-8AN illustrate example user interfaces for displaying a session region for ongoing events.
  • Figures 9A-9I illustrate a flow diagram of a method of changing a user interface based on a user input.
  • Figures 10A-10I illustrate a flow diagram of a method of selecting and/or modifying a user interface.
  • Figures 11 A-l 1H illustrate a flow diagram of a method of modifying a user interface.
  • Figure 12A-12E illustrate a flow diagram of a method of displaying a representation of a plurality of notifications in different configurations.
  • Figures 13A-13G illustrate a flow diagram of a method of displaying status information in a session region based on subscriptions.
  • Figures 14A-14G illustrate a flow diagram of a method of changing between different configurations in which a representation of a plurality of notifications can be displayed.
  • Figures 15A-15I are flow diagrams of a process for automatically shuffling through media items as a background in accordance with some embodiments.
  • Figures 16A-16J are flow diagrams of a process for applying different versions of a filter to a background media item in accordance with some embodiments.
  • Figures 17A- 17D are flow diagrams of a process for providing feedback in a system user interface in accordance with some embodiments.
  • FIGS 5A-5CB, 6A-6AR, 7A- 7V-3 and 8A-8AN are used to illustrate the processes in Figures 9A-9I, 10A-10I, 11 A-l 1H, 12A-12E, 13A-13G, 14A-14G, 15A-15I, 16A-16J, and 17A-17D.
  • first, second, etc. are, in some instances, used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first contact could be termed a second contact, and, similarly, a second contact could be termed a first contact, without departing from the scope of the various described embodiments. The first contact and the second contact are both contacts, but they are not the same contact, unless the context clearly indicates otherwise.
  • the term “if’ is, optionally, construed to mean “when” or “upon” or “in response to determining” or “in response to detecting,” depending on the context.
  • the phrase “if it is determined” or “if [a stated condition or event] is detected” is, optionally, construed to mean “upon determining” or “in response to determining” or “upon detecting [the stated condition or event]” or “in response to detecting [the stated condition or event],” depending on the context.
  • the device is a portable communications device, such as a mobile telephone, that also contains other functions, such as PDA and/or music player functions.
  • portable multifunction devices include, without limitation, the iPhone®, iPod Touch®, and iPad® devices from Apple Inc. of Cupertino, California.
  • Other portable electronic devices such as laptops or tablet computers with touch-sensitive surfaces (e.g., touch-screen displays and/or touchpads), are, optionally, used.
  • the device is not a portable communications device, but is a desktop computer with a touch- sensitive surface (e.g., a touch-screen display and/or a touchpad).
  • an electronic device that includes a display and a touch-sensitive surface is described. It should be understood, however, that the electronic device optionally includes one or more other physical user-interface devices, such as a physical keyboard, a mouse and/or a joystick.
  • the device typically supports a variety of applications, such as one or more of the following: a note taking application, a drawing application, a presentation application, a word processing application, a website creation application, a disk authoring application, a spreadsheet application, a gaming application, a telephone application, a video conferencing application, an e-mail application, an instant messaging application, a workout support application, a photo management application, a digital camera application, a digital video camera application, a web browsing application, a digital music player application, and/or a digital video player application.
  • applications such as one or more of the following: a note taking application, a drawing application, a presentation application, a word processing application, a website creation application, a disk authoring application, a spreadsheet application, a gaming application, a telephone application, a video conferencing application, an e-mail application, an instant messaging application, a workout support application, a photo management application, a digital camera application, a digital video camera application, a web browsing application, a digital music player application
  • the various applications that are executed on the device optionally use at least one common physical user-interface device, such as the touch-sensitive surface.
  • One or more functions of the touch-sensitive surface as well as corresponding information displayed on the device are, optionally, adjusted and/or varied from one application to the next and/or within a respective application.
  • a common physical architecture (such as the touch- sensitive surface) of the device optionally supports the variety of applications with user interfaces that are intuitive and transparent to the user.
  • FIG. 1 A is a block diagram illustrating portable multifunction device 100 with touch-sensitive display system 112 in accordance with some embodiments.
  • Touch- sensitive display system 112 is sometimes called a “touch screen” for convenience, and is sometimes simply called a touch-sensitive display.
  • Device 100 includes memory 102 (which optionally includes one or more computer readable storage mediums), memory controller 122, one or more processing units (CPUs) 120, peripherals interface 118, RF circuitry 108, audio circuitry 110, speaker 111, microphone 113, input/output (VO) subsystem 106, other input or control devices 116, and external port 124.
  • Device 100 optionally includes one or more optical sensors 164.
  • Device 100 optionally includes one or more intensity sensors 165 for detecting intensities of contacts on device 100 (e.g., a touch-sensitive surface such as touch-sensitive display system 112 of device 100).
  • Device 100 optionally includes one or more tactile output generators 167 for generating tactile outputs on device 100 (e.g., generating tactile outputs on a touch-sensitive surface such as touch-sensitive display system 112 of device 100 or touchpad 355 of device 300). These components optionally communicate over one or more communication buses or signal lines 103.
  • the term “tactile output” refers to physical displacement of a device relative to a previous position of the device, physical displacement of a component (e.g., a touch-sensitive surface) of a device relative to another component (e.g., housing) of the device, or displacement of the component relative to a center of mass of the device that will be detected by a user with the user’s sense of touch.
  • a component e.g., a touch-sensitive surface
  • another component e.g., housing
  • the tactile output generated by the physical displacement will be interpreted by the user as a tactile sensation corresponding to a perceived change in physical characteristics of the device or the component of the device.
  • a touch-sensitive surface e.g., a touch-sensitive display or trackpad
  • the user is, optionally, interpreted by the user as a “down click” or “up click” of a physical actuator button.
  • a user will feel a tactile sensation such as an “down click” or “up click” even when there is no movement of a physical actuator button associated with the touch-sensitive surface that is physically pressed (e.g., displaced) by the user’s movements.
  • movement of the touch-sensitive surface is, optionally, interpreted or sensed by the user as “roughness” of the touch-sensitive surface, even when there is no change in smoothness of the touch-sensitive surface. While such interpretations of touch by a user will be subject to the individualized sensory perceptions of the user, there are many sensory perceptions of touch that are common to a large majority of users.
  • a tactile output when a tactile output is described as corresponding to a particular sensory perception of a user (e.g., an “up click,” a “down click,” “roughness”), unless otherwise stated, the generated tactile output corresponds to physical displacement of the device or a component thereof that will generate the described sensory perception for a typical (or average) user.
  • Using tactile outputs to provide haptic feedback to a user enhances the operability of the device and makes the user-device interface more efficient (e.g., by helping the user to provide proper inputs and reducing user mistakes when operating/interacting with the device) which, additionally, reduces power usage and improves battery life of the device by enabling the user to use the device more quickly and efficiently.
  • a tactile output pattern specifies characteristics of a tactile output, such as the amplitude of the tactile output, the shape of a movement waveform of the tactile output, the frequency of the tactile output, and/or the duration of the tactile output.
  • the tactile outputs may invoke different haptic sensations in a user holding or touching the device. While the sensation of the user is based on the user’s perception of the tactile output, most users will be able to identify changes in waveform, frequency, and amplitude of tactile outputs generated by the device. Thus, the waveform, frequency and amplitude can be adjusted to indicate to the user that different operations have been performed.
  • tactile outputs with tactile output patterns that are designed, selected, and/or engineered to simulate characteristics (e.g., size, material, weight, stiffness, smoothness, etc.); behaviors (e.g., oscillation, displacement, acceleration, rotation, expansion, etc.); and/or interactions (e.g., collision, adhesion, repulsion, attraction, friction, etc.) of objects in a given environment (e.g., a user interface that includes graphical features and objects, a simulated physical environment with virtual boundaries and virtual objects, a real physical environment with physical boundaries and physical objects, and/or a combination of any of the above) will, in some circumstances, provide helpful feedback to users that reduces input errors and increases the efficiency of the user’s operation of the device.
  • characteristics e.g., size, material, weight, stiffness, smoothness, etc.
  • behaviors e.g., oscillation, displacement, acceleration, rotation, expansion, etc.
  • interactions e.g., collision, adhesion, repulsion, attraction, friction, etc.
  • tactile outputs are, optionally, generated to correspond to feedback that is unrelated to a simulated physical characteristic, such as an input threshold or a selection of an object. Such tactile outputs will, in some circumstances, provide helpful feedback to users that reduces input errors and increases the efficiency of the user’s operation of the device.
  • a tactile output with a suitable tactile output pattern serves as a cue for the occurrence of an event of interest in a user interface or behind the scenes in a device.
  • the events of interest include activation of an affordance (e.g., a real or virtual button, or toggle switch) provided on the device or in a user interface, success or failure of a requested operation, reaching or crossing a boundary in a user interface, entry into a new state, switching of input focus between objects, activation of a new mode, reaching or crossing an input threshold, detection or recognition of a type of input or gesture, etc.
  • an affordance e.g., a real or virtual button, or toggle switch
  • tactile outputs are provided to serve as a warning or an alert for an impending event or outcome that would occur unless a redirection or interruption input is timely detected.
  • Tactile outputs are also used in other contexts to enrich the user experience, improve the accessibility of the device to users with visual or motor difficulties or other accessibility needs, and/or improve efficiency and functionality of the user interface and/or the device.
  • Tactile outputs are optionally accompanied with audio outputs and/or visible user interface changes, which further enhance a user’s experience when the user interacts with a user interface and/or the device, and facilitate better conveyance of information regarding the state of the user interface and/or the device, and which reduce input errors and increase the efficiency of the user’s operation of the device.
  • device 100 is only one example of a portable multifunction device, and that device 100 optionally has more or fewer components than shown, optionally combines two or more components, or optionally has a different configuration or arrangement of the components.
  • the various components shown in Figure 1 A are implemented in hardware, software, firmware, or a combination thereof, including one or more signal processing and/or application specific integrated circuits.
  • Memory 102 optionally includes high-speed random access memory and optionally also includes non-volatile memory, such as one or more magnetic disk storage devices, flash memory devices, or other non-volatile solid-state memory devices. Access to memory 102 by other components of device 100, such as CPU(s) 120 and the peripherals interface 118, is, optionally, controlled by memory controller 122.
  • Peripherals interface 118 can be used to couple input and output peripherals of the device to CPU(s) 120 and memory 102.
  • the one or more processors 120 run or execute various software programs and/or sets of instructions stored in memory 102 to perform various functions for device 100 and to process data.
  • peripherals interface 118, CPU(s) 120, and memory controller 122 are, optionally, implemented on a single chip, such as chip 104. In some other embodiments, they are, optionally, implemented on separate chips.
  • RF (radio frequency) circuitry 108 receives and sends RF signals, also called electromagnetic signals.
  • RF circuitry 108 converts electrical signals to/from electromagnetic signals and communicates with communications networks and other communications devices via the electromagnetic signals.
  • RF circuitry 108 optionally includes well-known circuitry for performing these functions, including but not limited to an antenna system, an RF transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a CODEC chipset, a subscriber identity module (SIM) card, memory, and so forth.
  • an antenna system an RF transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a CODEC chipset, a subscriber identity module (SIM) card, memory, and so forth.
  • SIM subscriber identity module
  • RF circuitry 108 optionally communicates with networks, such as the Internet, also referred to as the World Wide Web (WWW), an intranet and/or a wireless network, such as a cellular telephone network, a wireless local area network (LAN) and/or a metropolitan area network (MAN), and other devices by wireless communication.
  • networks such as the Internet, also referred to as the World Wide Web (WWW), an intranet and/or a wireless network, such as a cellular telephone network, a wireless local area network (LAN) and/or a metropolitan area network (MAN), and other devices by wireless communication.
  • networks such as the Internet, also referred to as the World Wide Web (WWW), an intranet and/or a wireless network, such as a cellular telephone network, a wireless local area network (LAN) and/or a metropolitan area network (MAN), and other devices by wireless communication.
  • WLAN wireless local area network
  • MAN metropolitan area network
  • the wireless communication optionally uses any of a plurality of communications standards, protocols and technologies, including but not limited to Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), high-speed downlink packet access (HSDPA), high-speed uplink packet access (HSUPA), Evolution, Data-Only (EV-DO), HSPA, HSPA+, Dual-Cell HSPA (DC-HSPA), long term evolution (LTE), near field communication (NFC), wideband code division multiple access (W-CDMA), code division multiple access (CDMA), time division multiple access (TDMA), Bluetooth, Wireless Fidelity (Wi-Fi) (e.g., IEEE 802.
  • GSM Global System for Mobile Communications
  • EDGE Enhanced Data GSM Environment
  • HSDPA high-speed downlink packet access
  • HUPA high-speed uplink packet access
  • Evolution, Data-Only (EV-DO) Evolution, Data-Only
  • HSPA HSPA+
  • DC-HSPA Dual-Cell HSPA
  • LTE long term evolution
  • I la IEEE 802.1 lac, IEEE 802.1 lax, IEEE 802.1 lb, IEEE 802.11g and/or IEEE 802.1 In
  • VoIP voice over Internet Protocol
  • Wi-MAX a protocol for e-mail (e.g., Internet message access protocol (IMAP) and/or post office protocol (POP)), instant messaging (e.g., extensible messaging and presence protocol (XMPP), Session Initiation Protocol for Instant Messaging and Presence Leveraging Extensions (SIMPLE), Instant Messaging and Presence Service (IMPS)), and/or Short Message Service (SMS), or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • IMAP Internet message access protocol
  • POP post office protocol
  • instant messaging e.g., extensible messaging and presence protocol (XMPP), Session Initiation Protocol for Instant Messaging and Presence Leveraging Extensions (SIMPLE), Instant Messaging and Presence Service (IMPS)
  • SMS Short Message Service
  • Audio circuitry 110, speaker 111, and microphone 113 provide an audio interface between a user and device 100.
  • Audio circuitry 110 receives audio data from peripherals interface 118, converts the audio data to an electrical signal, and transmits the electrical signal to speaker 111.
  • Speaker 111 converts the electrical signal to human-audible sound waves.
  • Audio circuitry 110 also receives electrical signals converted by microphone 113 from sound waves.
  • Audio circuitry 110 converts the electrical signal to audio data and transmits the audio data to peripherals interface 118 for processing. Audio data is, optionally, retrieved from and/or transmitted to memory 102 and/or RF circuitry 108 by peripherals interface 118.
  • audio circuitry 110 also includes a headset jack (e.g., 212, Figure 2).
  • the headset jack provides an interface between audio circuitry 110 and removable audio input/output peripherals, such as output-only headphones or a headset with both output (e.g., a headphone for one or both ears) and input (e.g., a microphone
  • I/O subsystem 106 couples input/output peripherals on device 100, such as touch-sensitive display system 112 and other input or control devices 116, with peripherals interface 118.
  • I/O subsystem 106 optionally includes display controller 156, optical sensor controller 158, intensity sensor controller 159, haptic feedback controller 161, and one or more input controllers 160 for other input or control devices.
  • the one or more input controllers 160 receive/send electrical signals from/to other input or control devices 116.
  • the other input or control devices 116 optionally include physical buttons (e.g., push buttons, rocker buttons, etc.), dials, slider switches, joysticks, click wheels, and so forth.
  • input controlled s) 160 are, optionally, coupled with any (or none) of the following: a keyboard, infrared port, USB port, stylus, and/or a pointer device such as a mouse.
  • the one or more buttons optionally include an up/down button for volume control of speaker 111 and/or microphone 113.
  • the one or more buttons optionally include a push button (e.g., 206, Figure 2).
  • Touch-sensitive display system 112 provides an input interface and an output interface between the device and a user.
  • Display controller 156 receives and/or sends electrical signals from/to touch-sensitive display system 112.
  • Touch-sensitive display system 112 displays visual output to the user.
  • the visual output optionally includes graphics, text, icons, video, and any combination thereof (collectively termed “graphics”).
  • graphics optionally includes graphics, text, icons, video, and any combination thereof (collectively termed “graphics”).
  • some or all of the visual output corresponds to user interface objects.
  • the term “affordance” refers to a user-interactive graphical user interface object (e.g., a graphical user interface object that is configured to respond to inputs directed toward the graphical user interface object). Examples of user-interactive graphical user interface objects include, without limitation, a button, slider, icon, selectable menu item, switch, hyperlink, or other user interface control.
  • Touch-sensitive display system 112 has a touch-sensitive surface, sensor or set of sensors that accepts input from the user based on haptic and/or tactile contact.
  • Touch-sensitive display system 112 and display controller 156 (along with any associated modules and/or sets of instructions in memory 102) detect contact (and any movement or breaking of the contact) on touch-sensitive display system 112 and converts the detected contact into interaction with user-interface objects (e.g., one or more soft keys, icons, web pages or images) that are displayed on touch-sensitive display system 112.
  • user-interface objects e.g., one or more soft keys, icons, web pages or images
  • a point of contact between touch-sensitive display system 112 and the user corresponds to a finger of the user or a stylus.
  • Touch-sensitive display system 112 optionally uses LCD (liquid crystal display) technology, LPD (light emitting polymer display) technology, or LED (light emitting diode) technology, although other display technologies are used in other embodiments.
  • Touch-sensitive display system 112 and display controller 156 optionally detect contact and any movement or breaking thereof using any of a plurality of touch sensing technologies now known or later developed, including but not limited to capacitive, resistive, infrared, and surface acoustic wave technologies, as well as other proximity sensor arrays or other elements for determining one or more points of contact with touch-sensitive display system 112.
  • projected mutual capacitance sensing technology is used, such as that found in the iPhone®, iPod Touch®, and iPad® from Apple Inc. of Cupertino, California.
  • Touch-sensitive display system 112 optionally has a video resolution in excess of 100 dpi. In some embodiments, the touch screen video resolution is in excess of 400 dpi (e.g., 500 dpi, 800 dpi, or greater).
  • the user optionally makes contact with touch-sensitive display system 112 using any suitable object or appendage, such as a stylus, a finger, and so forth.
  • the user interface is designed to work with finger-based contacts and gestures, which can be less precise than stylus-based input due to the larger area of contact of a finger on the touch screen.
  • the device translates the rough finger-based input into a precise pointer/cursor position or command for performing the actions desired by the user.
  • device 100 in addition to the touch screen, device 100 optionally includes a touchpad for activating or deactivating particular functions.
  • the touchpad is a touch-sensitive area of the device that, unlike the touch screen, does not display visual output.
  • the touchpad is, optionally, a touch-sensitive surface that is separate from touch-sensitive display system 112 or an extension of the touch-sensitive surface formed by the touch screen.
  • Device 100 also includes power system 162 for powering the various components.
  • Power system 162 optionally includes a power management system, one or more power sources (e.g., battery, alternating current (AC)), a recharging system, a power failure detection circuit, a power converter or inverter, a power status indicator (e.g., a lightemitting diode (LED)) and any other components associated with the generation, management and distribution of power in portable devices.
  • a power management system one or more power sources (e.g., battery, alternating current (AC)), a recharging system, a power failure detection circuit, a power converter or inverter, a power status indicator (e.g., a lightemitting diode (LED)) and any other components associated with the generation, management and distribution of power in portable devices.
  • power sources e.g., battery, alternating current (AC)
  • AC alternating current
  • a recharging system e.g., a recharging system
  • a power failure detection circuit e.g., a
  • Device 100 optionally also includes one or more optical sensors 164 (e.g., as part of one or more cameras).
  • Figure 1 A shows an optical sensor coupled with optical sensor controller 158 in I/O subsystem 106.
  • Optical sensor(s) 164 optionally include charge-coupled device (CCD) or complementary metal-oxide semiconductor (CMOS) phototransistors.
  • CMOS complementary metal-oxide semiconductor
  • Optical sensor(s) 164 receive light from the environment, projected through one or more lens, and converts the light to data representing an image.
  • imaging module 143 also called a camera module
  • optical sensor(s) 164 optionally capture still images and/or video.
  • an optical sensor is located on the back of device 100, opposite touch-sensitive display system 112 on the front of the device, so that the touch screen is enabled for use as a viewfinder for still and/or video image acquisition.
  • another optical sensor is located on the front of the device so that the user's image is obtained (e.g., for selfies, for videoconferencing while the user views the other video conference participants on the touch screen, etc.).
  • Device 100 optionally also includes one or more contact intensity sensors 165.
  • Figure 1A shows a contact intensity sensor coupled with intensity sensor controller 159 in I/O subsystem 106.
  • Contact intensity sensor(s) 165 optionally include one or more piezoresistive strain gauges, capacitive force sensors, electric force sensors, piezoelectric force sensors, optical force sensors, capacitive touch-sensitive surfaces, or other intensity sensors (e.g., sensors used to measure the force (or pressure) of a contact on a touch-sensitive surface).
  • Contact intensity sensor(s) 165 receive contact intensity information (e.g., pressure information or a proxy for pressure information) from the environment.
  • contact intensity information e.g., pressure information or a proxy for pressure information
  • At least one contact intensity sensor is collocated with, or proximate to, a touch-sensitive surface (e.g., touch-sensitive display system 112). In some embodiments, at least one contact intensity sensor is located on the back of device 100, opposite touch-screen display system 112 which is located on the front of device 100.
  • Device 100 optionally also includes one or more proximity sensors 166.
  • Figure 1 A shows proximity sensor 166 coupled with peripherals interface 118. Alternately, proximity sensor 166 is coupled with input controller 160 in I/O subsystem 106. In some embodiments, the proximity sensor turns off and disables touch-sensitive display system 112 when the multifunction device is placed near the user's ear (e.g., when the user is making a phone call).
  • Device 100 optionally also includes one or more tactile output generators 167.
  • Figure 1 A shows a tactile output generator coupled with haptic feedback controller 161 in VO subsystem 106.
  • tactile output generator(s) 167 include one or more electroacoustic devices such as speakers or other audio components and/or electromechanical devices that convert energy into linear motion such as a motor, solenoid, electroactive polymer, piezoelectric actuator, electrostatic actuator, or other tactile output generating component (e.g., a component that converts electrical signals into tactile outputs on the device).
  • Tactile output generator(s) 167 receive tactile feedback generation instructions from haptic feedback module 133 and generates tactile outputs on device 100 that are capable of being sensed by a user of device 100.
  • At least one tactile output generator is collocated with, or proximate to, a touch-sensitive surface (e.g., touch-sensitive display system 112) and, optionally, generates a tactile output by moving the touch- sensitive surface vertically (e.g., in/out of a surface of device 100) or laterally (e.g., back and forth in the same plane as a surface of device 100).
  • at least one tactile output generator sensor is located on the back of device 100, opposite touch-sensitive display system 112, which is located on the front of device 100.
  • Device 100 optionally also includes one or more accelerometers 168.
  • Figure 1 A shows accelerometer 168 coupled with peripherals interface 118.
  • accelerometer 168 is, optionally, coupled with an input controller 160 in I/O subsystem 106.
  • information is displayed on the touch-screen display in a portrait view or a landscape view based on an analysis of data received from the one or more accelerometers.
  • Device 100 optionally includes, in addition to accelerometer(s) 168, a magnetometer and a GPS (or GLONASS or other global navigation system) receiver for obtaining information concerning the location and orientation (e.g., portrait or landscape) of device 100.
  • GPS or GLONASS or other global navigation system
  • the software components stored in memory 102 include operating system 126, communication module (or set of instructions) 128, contact/motion module (or set of instructions) 130, graphics module (or set of instructions) 132, haptic feedback module (or set of instructions) 133, text input module (or set of instructions) 134, Global Positioning System (GPS) module (or set of instructions) 135, and applications (or sets of instructions) 136.
  • memory 102 stores device/global internal state 157, as shown in Figures 1A and 3.
  • Device/global internal state 157 includes one or more of: active application state, indicating which applications, if any, are currently active; display state, indicating what applications, views or other information occupy various regions of touch-sensitive display system 112; sensor state, including information obtained from the device’s various sensors and other input or control devices 116; and location and/or positional information concerning the device’s location and/or attitude.
  • Operating system 126 e.g., iOS, Darwin, RTXC, LINUX, UNIX, OS X, WINDOWS, or an embedded operating system such as VxWorks
  • Operating system 126 includes various software components and/or drivers for controlling and managing general system tasks (e.g., memory management, storage device control, power management, etc.) and facilitates communication between various hardware and software components.
  • general system tasks e.g., memory management, storage device control, power management, etc.
  • Communication module 128 facilitates communication with other devices over one or more external ports 124 and also includes various software components for handling data received by RF circuitry 108 and/or external port 124.
  • External port 124 e.g., Universal Serial Bus (USB), FIREWIRE, etc.
  • USB Universal Serial Bus
  • FIREWIRE FireWire
  • the external port is a multi-pin (e.g., 30-pin) connector that is the same as, or similar to and/or compatible with the 30-pin connector used in some iPhone®, iPod Touch®, and iPad® devices from Apple Inc. of Cupertino, California.
  • the external port is a Lightning connector that is the same as, or similar to and/or compatible with the Lightning connector used in some iPhone®, iPod Touch®, and iPad® devices from Apple Inc. of Cupertino, California.
  • the external port is a USB Type-C connector that is the same as, or similar to and/or compatible with the USB Type-C connector used in some electronic devices from Apple Inc. of Cupertino, California.
  • Contact/motion module 130 optionally detects contact with touch-sensitive display system 112 (in conjunction with display controller 156) and other touch- sensitive devices (e.g., a touchpad or physical click wheel).
  • Contact/motion module 130 includes various software components for performing various operations related to detection of contact (e.g., by a finger or by a stylus), such as determining if contact has occurred (e.g., detecting a finger-down event), determining an intensity of the contact (e.g., the force or pressure of the contact or a substitute for the force or pressure of the contact), determining if there is movement of the contact and tracking the movement across the touch-sensitive surface (e.g., detecting one or more finger-dragging events), and determining if the contact has ceased (e.g., detecting a finger-up event or a break in contact).
  • Contact/motion module 130 receives contact data from the touch-sensitive surface.
  • Determining movement of the point of contact which is represented by a series of contact data, optionally includes determining speed (magnitude), velocity (magnitude and direction), and/or an acceleration (a change in magnitude and/or direction) of the point of contact. These operations are, optionally, applied to single contacts (e.g., one finger contacts or stylus contacts) or to multiple simultaneous contacts (e.g., “multitouch”/multiple finger contacts).
  • contact/motion module 130 and display controller 156 detect contact on a touchpad.
  • Contact/motion module 130 optionally detects a gesture input by a user. Different gestures on the touch-sensitive surface have different contact patterns (e.g., different motions, timings, and/or intensities of detected contacts). Thus, a gesture is, optionally, detected by detecting a particular contact pattern. For example, detecting a finger tap gesture includes detecting a finger-down event followed by detecting a finger-up (lift off) event at the same position (or substantially the same position) as the finger-down event (e.g., at the position of an icon).
  • detecting a finger swipe gesture on the touch- sensitive surface includes detecting a finger-down event followed by detecting one or more finger-dragging events, and subsequently followed by detecting a finger-up (lift off) event.
  • tap, swipe, drag, and other gestures are optionally detected for a stylus by detecting a particular contact pattern for the stylus.
  • detecting a finger tap gesture depends on the length of time between detecting the finger-down event and the finger-up event, but is independent of the intensity of the finger contact between detecting the finger-down event and the finger-up event.
  • a tap gesture is detected in accordance with a determination that the length of time between the finger-down event and the finger-up event is less than a predetermined value (e.g., less than 0.1, 0.2, 0.3, 0.4 or 0.5 seconds), independent of whether the intensity of the finger contact during the tap meets a given intensity threshold (greater than a nominal contact-detection intensity threshold), such as a light press or deep press intensity threshold.
  • a finger tap gesture can satisfy particular input criteria that do not require that the characteristic intensity of a contact satisfy a given intensity threshold in order for the particular input criteria to be met.
  • the finger contact in a tap gesture typically needs to satisfy a nominal contact-detection intensity threshold, below which the contact is not detected, in order for the finger-down event to be detected.
  • a similar analysis applies to detecting a tap gesture by a stylus or other contact.
  • the nominal contact-detection intensity threshold optionally does not correspond to physical contact between the finger or stylus and the touch sensitive surface.
  • a swipe gesture, a pinch gesture, a depinch gesture, and/or a long press gesture are optionally detected based on the satisfaction of criteria that are either independent of intensities of contacts included in the gesture, or do not require that contact(s) that perform the gesture reach intensity thresholds in order to be recognized.
  • a swipe gesture is detected based on an amount of movement of one or more contacts;
  • a pinch gesture is detected based on movement of two or more contacts towards each other;
  • a depinch gesture is detected based on movement of two or more contacts away from each other;
  • a long press gesture is detected based on a duration of the contact on the touch-sensitive surface with less than a threshold amount of movement.
  • the statement that particular gesture recognition criteria do not require that the intensity of the contact(s) meet a respective intensity threshold in order for the particular gesture recognition criteria to be met means that the particular gesture recognition criteria are capable of being satisfied if the contact(s) in the gesture do not reach the respective intensity threshold, and are also capable of being satisfied in circumstances where one or more of the contacts in the gesture do reach or exceed the respective intensity threshold.
  • a tap gesture is detected based on a determination that the finger-down and finger-up event are detected within a predefined time period, without regard to whether the contact is above or below the respective intensity threshold during the predefined time period, and a swipe gesture is detected based on a determination that the contact movement is greater than a predefined magnitude, even if the contact is above the respective intensity threshold at the end of the contact movement.
  • detection of a gesture is influenced by the intensity of contacts performing the gesture (e.g., the device detects a long press more quickly when the intensity of the contact is above an intensity threshold or delays detection of a tap input when the intensity of the contact is higher), the detection of those gestures does not require that the contacts reach a particular intensity threshold so long as the criteria for recognizing the gesture can be met in circumstances where the contact does not reach the particular intensity threshold (e.g., even if the amount of time that it takes to recognize the gesture changes).
  • Contact intensity thresholds, duration thresholds, and movement thresholds are, in some circumstances, combined in a variety of different combinations in order to create heuristics for distinguishing two or more different gestures directed to the same input element or region so that multiple different interactions with the same input element are enabled to provide a richer set of user interactions and responses.
  • the statement that a particular set of gesture recognition criteria do not require that the intensity of the contact(s) meet a respective intensity threshold in order for the particular gesture recognition criteria to be met does not preclude the concurrent evaluation of other intensity-dependent gesture recognition criteria to identify other gestures that do have criteria that are met when a gesture includes a contact with an intensity above the respective intensity threshold.
  • first gesture recognition criteria for a first gesture - which do not require that the intensity of the contact(s) meet a respective intensity threshold in order for the first gesture recognition criteria to be met - are in competition with second gesture recognition criteria for a second gesture - which are dependent on the contact(s) reaching the respective intensity threshold.
  • the gesture is, optionally, not recognized as meeting the first gesture recognition criteria for the first gesture if the second gesture recognition criteria for the second gesture are met first. For example, if a contact reaches the respective intensity threshold before the contact moves by a predefined amount of movement, a deep press gesture is detected rather than a swipe gesture.
  • a swipe gesture is detected rather than a deep press gesture.
  • the first gesture recognition criteria for the first gesture still do not require that the intensity of the contact(s) meet a respective intensity threshold in order for the first gesture recognition criteria to be met because if the contact stayed below the respective intensity threshold until an end of the gesture (e.g., a swipe gesture with a contact that does not increase to an intensity above the respective intensity threshold), the gesture would have been recognized by the first gesture recognition criteria as a swipe gesture.
  • particular gesture recognition criteria that do not require that the intensity of the contact(s) meet a respective intensity threshold in order for the particular gesture recognition criteria to be met will (A) in some circumstances ignore the intensity of the contact with respect to the intensity threshold (e.g. for a tap gesture) and/or (B) in some circumstances still be dependent on the intensity of the contact with respect to the intensity threshold in the sense that the particular gesture recognition criteria (e.g., for a long press gesture) will fail if a competing set of intensitydependent gesture recognition criteria (e.g., for a deep press gesture) recognize an input as corresponding to an intensity-dependent gesture before the particular gesture recognition criteria recognize a gesture corresponding to the input (e.g., for a long press gesture that is competing with a deep press gesture for recognition).
  • a competing set of intensitydependent gesture recognition criteria e.g., for a deep press gesture
  • Graphics module 132 includes various known software components for rendering and displaying graphics on touch-sensitive display system 112 or other display, including components for changing the visual impact (e.g., brightness, transparency, saturation, contrast or other visual property) of graphics that are displayed.
  • graphics includes any object that can be displayed to a user, including without limitation text, web pages, icons (such as user-interface objects including soft keys), digital images, videos, animations and the like.
  • graphics module 132 stores data representing graphics to be used. Each graphic is, optionally, assigned a corresponding code. Graphics module 132 receives, from applications etc., one or more codes specifying graphics to be displayed along with, if necessary, coordinate data and other graphic property data, and then generates screen image data to output to display controller 156.
  • Haptic feedback module 133 includes various software components for generating instructions (e.g., instructions used by haptic feedback controller 161) to produce tactile outputs using tactile output generator(s) 167 at one or more locations on device 100 in response to user interactions with device 100.
  • instructions e.g., instructions used by haptic feedback controller 161 to produce tactile outputs using tactile output generator(s) 167 at one or more locations on device 100 in response to user interactions with device 100.
  • Text input module 134 which is, optionally, a component of graphics module 132, provides soft keyboards for entering text in various applications (e.g., contacts 137, e-mail 140, IM 141, browser 147, and any other application that needs text input).
  • GPS module 135 determines the location of the device and provides this information for use in various applications (e.g., to telephone 138 for use in location-based dialing, to camera 143 as picture/video metadata, and to applications that provide locationbased services such as weather widgets, local yellow page widgets, and map/navigation widgets).
  • Applications 136 optionally include the following modules (or sets of instructions), or a subset or superset thereof:
  • contacts module 137 (sometimes called an address book or contact list);
  • camera module 143 for still and/or video images
  • calendar module 148 • calendar module 148;
  • widget modules 149 which optionally include one or more of: weather widget 149-1, stocks widget 149-2, calculator widget 149-3, alarm clock widget 149-4, dictionary widget 149-5, and other widgets obtained by the user, as well as user-created widgets 149-6;
  • widget creator module 150 for making user-created widgets 149-6;
  • search module 151 • search module 151;
  • video and music player module 152 which is, optionally, made up of a video player module and a music player module;
  • Examples of other applications 136 that are, optionally, stored in memory 102 include other word processing applications, other image editing applications, drawing applications, presentation applications, JAVA-enabled applications, encryption, digital rights management, voice recognition, and voice replication.
  • contacts module 137 includes executable instructions to manage an address book or contact list (e.g., stored in application internal state 192 of contacts module 137 in memory 102 or memory 370), including: adding name(s) to the address book; deleting name(s) from the address book; associating telephone number(s), e-mail address(es), physical address(es) or other information with a name; associating an image with a name; categorizing and sorting names; providing telephone numbers and/or e-mail addresses to initiate and/or facilitate communications by telephone 138, video conference 139, e-mail 140, or IM 141; and so forth.
  • an address book or contact list e.g., stored in application internal state 192 of contacts module 137 in memory 102 or memory 370
  • telephone module 138 includes executable instructions to enter a sequence of characters corresponding to a telephone number, access one or more telephone numbers in address book 137, modify a telephone number that has been entered, dial a respective telephone number, conduct a conversation and disconnect or hang up when the conversation is completed.
  • the wireless communication optionally uses any of a plurality of communications standards, protocols and technologies.
  • videoconferencing module 139 includes executable instructions to initiate, conduct, and terminate a video conference between a user and one or more other participants in accordance with user instructions.
  • e-mail client module 140 includes executable instructions to create, send, receive, and manage e-mail in response to user instructions.
  • e-mail client module 140 makes it very easy to create and send e-mails with still or video images taken with camera module 143.
  • the instant messaging module 141 includes executable instructions to enter a sequence of characters corresponding to an instant message, to modify previously entered characters, to transmit a respective instant message (for example, using a Short Message Service (SMS) or Multimedia Message Service (MMS) protocol for telephony -based instant messages or using XMPP, SIMPLE, Apple Push Notification Service (APNs) or IMPS for Internet-based instant messages), to receive instant messages, and to view received instant messages.
  • SMS Short Message Service
  • MMS Multimedia Message Service
  • APIs Apple Push Notification Service
  • IMPS Internet Messaging Protocol
  • transmitted and/or received instant messages optionally include graphics, photos, audio files, video files and/or other attachments as are supported in a MMS and/or an Enhanced Messaging Service (EMS).
  • EMS Enhanced Messaging Service
  • instant messaging refers to both telephony-based messages (e.g., messages sent using SMS or MMS) and Internet-based messages (e.g., messages sent using XMPP, SIMPLE, APNs, or IMPS).
  • workout support module 142 includes executable instructions to create workouts (e.g., with time, distance, and/or calorie burning goals); communicate with workout sensors (in sports devices and smart watches); receive workout sensor data; calibrate sensors used to monitor a workout; select and play music for a workout; and display, store and transmit workout data.
  • camera module 143 includes executable instructions to capture still images or video (including a video stream) and store them into memory 102, modify characteristics of a still image or video, and/or delete a still image or video from memory 102.
  • image management module 144 includes executable instructions to arrange, modify (e.g., edit), or otherwise manipulate, label, delete, present (e.g., in a digital slide show or album), and store still and/or video images.
  • modify e.g., edit
  • present e.g., in a digital slide show or album
  • browser module 147 includes executable instructions to browse the Internet in accordance with user instructions, including searching, linking to, receiving, and displaying web pages or portions thereof, as well as attachments and other files linked to web pages.
  • calendar module 148 includes executable instructions to create, display, modify, and store calendars and data associated with calendars (e.g., calendar entries, to do lists, etc.) in accordance with user instructions.
  • widget modules 149 are mini-applications that are, optionally, downloaded and used by a user (e.g., weather widget 149-1, stocks widget 149-2, calculator widget 149-3, alarm clock widget 149-4, and dictionary widget 149-5) or created by the user (e.g., user-created widget 149-6).
  • a widget includes an HTML (Hypertext Markup Language) file, a CSS (Cascading Style Sheets) file, and a JavaScript file.
  • a widget includes an XML (Extensible Markup Language) file and a JavaScript file (e.g., Yahoo! Widgets).
  • the widget creator module 150 includes executable instructions to create widgets (e.g., turning a user-specified portion of a web page into a widget).
  • search module 151 includes executable instructions to search for text, music, sound, image, video, and/or other files in memory 102 that match one or more search criteria (e.g., one or more user-specified search terms) in accordance with user instructions.
  • search criteria e.g., one or more user-specified search terms
  • video and music player module 152 includes executable instructions that allow the user to download and play back recorded music and other sound files stored in one or more file formats, such as MP3 or AAC files, and executable instructions to display, present or otherwise play back videos (e.g., on touch-sensitive display system 112, or on an external display connected wirelessly or via external port 124).
  • device 100 optionally includes the functionality of an MP3 player, such as an iPod (trademark of Apple Inc.).
  • notes module 153 includes executable instructions to create and manage notes, to do lists, and the like in accordance with user instructions.
  • map module 154 includes executable instructions to receive, display, modify, and store maps and data associated with maps (e.g., driving directions; data on stores and other points of interest at or near a particular location; and other location-based data) in accordance with user instructions.
  • maps e.g., driving directions; data on stores and other points of interest at or near a particular location; and other location-based data
  • online video module 155 includes executable instructions that allow the user to access, browse, receive (e.g., by streaming and/or download), play back (e.g., on the touch screen 112, or on an external display connected wirelessly or via external port 124), send an e-mail with a link to a particular online video, and otherwise manage online videos in one or more file formats, such as H.264.
  • instant messaging module 141 rather than e-mail client module 140, is used to send a link to a particular online video.
  • modules and applications correspond to a set of executable instructions for performing one or more functions described above and the methods described in this application (e.g., the computer-implemented methods and other information processing methods described herein).
  • modules e.g., sets of instructions
  • memory 102 optionally stores a subset of the modules and data structures identified above.
  • memory 102 optionally stores additional modules and data structures not described above.
  • device 100 is a device where operation of a predefined set of functions on the device is performed exclusively through a touch screen and/or a touchpad.
  • a touch screen and/or a touchpad as the primary input control device for operation of device 100, the number of physical input control devices (such as push buttons, dials, and the like) on device 100 is, optionally, reduced.
  • the predefined set of functions that are performed exclusively through a touch screen and/or a touchpad optionally include navigation between user interfaces.
  • the touchpad when touched by the user, navigates device 100 to a main, home, or root menu from any user interface that is displayed on device 100.
  • a “menu button” is implemented using a touchpad.
  • the menu button is a physical push button or other physical input control device instead of a touchpad.
  • Figure IB is a block diagram illustrating example components for event handling in accordance with some embodiments.
  • memory 102 in Figures 1A or 370 ( Figure 3) includes event sorter 170 (e.g., in operating system 126) and a respective application 136-1 (e.g., any of the aforementioned applications 136, 137-155, 380- 390).
  • event sorter 170 e.g., in operating system 126
  • application 136-1 e.g., any of the aforementioned applications 136, 137-155, 380- 390.
  • Event sorter 170 receives event information and determines the application 136-1 and application view 191 of application 136-1 to which to deliver the event information.
  • Event sorter 170 includes event monitor 171 and event dispatcher module 174.
  • application 136-1 includes application internal state 192, which indicates the current application view(s) displayed on touch- sensitive display system 112 when the application is active or executing.
  • device/global internal state 157 is used by event sorter 170 to determine which application(s) is (are) currently active, and application internal state 192 is used by event sorter 170 to determine application views 191 to which to deliver event information.
  • application internal state 192 includes additional information, such as one or more of: resume information to be used when application 136-1 resumes execution, user interface state information that indicates information being displayed or that is ready for display by application 136-1, a state queue for enabling the user to go back to a prior state or view of application 136-1, and a redo/undo queue of previous actions taken by the user.
  • Event monitor 171 receives event information from peripherals interface 118.
  • Event information includes information about a sub-event (e.g., a user touch on touch- sensitive display system 112, as part of a multi-touch gesture).
  • Peripherals interface 118 transmits information it receives from I/O subsystem 106 or a sensor, such as proximity sensor 166, accelerometer(s) 168, and/or microphone 113 (through audio circuitry 110).
  • Information that peripherals interface 118 receives from I/O subsystem 106 includes information from touch-sensitive display system 112 or a touch-sensitive surface.
  • event monitor 171 sends requests to the peripherals interface 118 at predetermined intervals.
  • peripherals interface 118 transmits event information.
  • peripheral interface 118 transmits event information only when there is a significant event (e.g., receiving an input above a predetermined noise threshold and/or for more than a predetermined duration).
  • event sorter 170 also includes a hit view determination module 172 and/or an active event recognizer determination module 173.
  • Hit view determination module 172 provides software procedures for determining where a sub-event has taken place within one or more views, when touch- sensitive display system 112 displays more than one view. Views are made up of controls and other elements that a user can see on the display.
  • FIG. 1 Another aspect of the user interface associated with an application is a set of views, sometimes herein called application views or user interface windows, in which information is displayed and touch-based gestures occur.
  • the application views (of a respective application) in which a touch is detected optionally correspond to programmatic levels within a programmatic or view hierarchy of the application. For example, the lowest level view in which a touch is detected is, optionally, called the hit view, and the set of events that are recognized as proper inputs are, optionally, determined based, at least in part, on the hit view of the initial touch that begins a touch-based gesture.
  • Hit view determination module 172 receives information related to sub-events of a touch-based gesture.
  • hit view determination module 172 identifies a hit view as the lowest view in the hierarchy which should handle the sub-event. In most circumstances, the hit view is the lowest level view in which an initiating sub-event occurs (e.g., the first sub-event in the sequence of subevents that form an event or potential event). Once the hit view is identified by the hit view determination module, the hit view typically receives all sub-events related to the same touch or input source for which it was identified as the hit view.
  • Active event recognizer determination module 173 determines which view or views within a view hierarchy should receive a particular sequence of sub-events. In some embodiments, active event recognizer determination module 173 determines that only the hit view should receive a particular sequence of sub-events. In other embodiments, active event recognizer determination module 173 determines that all views that include the physical location of a sub-event are actively involved views, and therefore determines that all actively involved views should receive a particular sequence of sub-events. In other embodiments, even if touch sub-events were entirely confined to the area associated with one particular view, views higher in the hierarchy would still remain as actively involved views.
  • Event dispatcher module 174 dispatches the event information to an event recognizer (e.g., event recognizer 180). In embodiments including active event recognizer determination module 173, event dispatcher module 174 delivers the event information to an event recognizer determined by active event recognizer determination module 173. In some embodiments, event dispatcher module 174 stores in an event queue the event information, which is retrieved by a respective event receiver module 182.
  • an event recognizer e.g., event recognizer 180.
  • event dispatcher module 174 delivers the event information to an event recognizer determined by active event recognizer determination module 173.
  • event dispatcher module 174 stores in an event queue the event information, which is retrieved by a respective event receiver module 182.
  • operating system 126 includes event sorter 170.
  • application 136-1 includes event sorter 170.
  • event sorter 170 is a stand-alone module, or a part of another module stored in memory 102, such as contact/motion module 130.
  • application 136-1 includes a plurality of event handlers 190 and one or more application views 191, each of which includes instructions for handling touch events that occur within a respective view of the application’s user interface.
  • Each application view 191 of the application 136-1 includes one or more event recognizers 180.
  • a respective application view 191 includes a plurality of event recognizers 180.
  • one or more of event recognizers 180 are part of a separate module, such as a user interface kit or a higher level object from which application 136-1 inherits methods and other properties.
  • a respective event handler 190 includes one or more of: data updater 176, object updater 177, GUI updater 178, and/or event data 179 received from event sorter 170.
  • Event handler 190 optionally utilizes or calls data updater
  • one or more of the application views 191 includes one or more respective event handlers 190. Also, in some embodiments, one or more of data updater 176, object updater
  • GUI updater 178 is included in a respective application view 191.
  • a respective event recognizer 180 receives event information (e.g., event data 179) from event sorter 170, and identifies an event from the event information.
  • Event recognizer 180 includes event receiver 182 and event comparator 184.
  • event recognizer 180 also includes at least a subset of: metadata 183, and event delivery instructions 188 (which optionally include sub-event delivery instructions).
  • Event receiver 182 receives event information from event sorter 170.
  • the event information includes information about a sub-event, for example, a touch or a touch movement.
  • the event information also includes additional information, such as location of the sub-event.
  • the event information optionally also includes speed and direction of the sub-event.
  • events include rotation of the device from one orientation to another (e.g., from a portrait orientation to a landscape orientation, or vice versa), and the event information includes corresponding information about the current orientation (also called device attitude) of the device.
  • Event comparator 184 compares the event information to predefined event or sub-event definitions and, based on the comparison, determines an event or sub-event, or determines or updates the state of an event or sub-event.
  • event comparator 184 includes event definitions 186.
  • Event definitions 186 contain definitions of events (e.g., predefined sequences of sub-events), for example, event 1 (187-1), event 2 (187- 2), and others.
  • sub-events in an event 187 include, for example, touch begin, touch end, touch movement, touch cancellation, and multiple touching.
  • the definition for event 1 (187-1) is a double tap on a displayed object.
  • the double tap for example, comprises a first touch (touch begin) on the displayed object for a predetermined phase, a first lift-off (touch end) for a predetermined phase, a second touch (touch begin) on the displayed object for a predetermined phase, and a second lift-off (touch end) for a predetermined phase.
  • the definition for event 2 (187-2) is a dragging on a displayed object.
  • the dragging for example, comprises a touch (or contact) on the displayed object for a predetermined phase, a movement of the touch across touch- sensitive display system 112, and lift-off of the touch (touch end).
  • the event also includes information for one or more associated event handlers 190.
  • event definition 187 includes a definition of an event for a respective user-interface object.
  • event comparator 184 performs a hit test to determine which user-interface object is associated with a sub-event. For example, in an application view in which three user-interface objects are displayed on touch- sensitive display system 112, when a touch is detected on touch-sensitive display system 112, event comparator 184 performs a hit test to determine which of the three user-interface objects is associated with the touch (sub-event). If each displayed object is associated with a respective event handler 190, the event comparator uses the result of the hit test to determine which event handler 190 should be activated. For example, event comparator 184 selects an event handler associated with the sub-event and the object triggering the hit test.
  • the definition for a respective event 187 also includes delayed actions that delay delivery of the event information until after it has been determined whether the sequence of sub-events does or does not correspond to the event recognizer’s event type.
  • a respective event recognizer 180 determines that the series of subevents do not match any of the events in event definitions 186, the respective event recognizer 180 enters an event impossible, event failed, or event ended state, after which it disregards subsequent sub-events of the touch-based gesture. In this situation, other event recognizers, if any, that remain active for the hit view continue to track and process subevents of an ongoing touch-based gesture.
  • a respective event recognizer 180 includes metadata 183 with configurable properties, flags, and/or lists that indicate how the event delivery system should perform sub-event delivery to actively involved event recognizers.
  • metadata 183 includes configurable properties, flags, and/or lists that indicate how event recognizers interact, or are enabled to interact, with one another.
  • metadata 183 includes configurable properties, flags, and/or lists that indicate whether sub-events are delivered to varying levels in the view or programmatic hierarchy.
  • a respective event recognizer 180 activates event handler 190 associated with an event when one or more particular sub-events of an event are recognized.
  • a respective event recognizer 180 delivers event information associated with the event to event handler 190.
  • Activating an event handler 190 is distinct from sending (and deferred sending) sub-events to a respective hit view.
  • event recognizer 180 throws a flag associated with the recognized event, and event handler 190 associated with the flag catches the flag and performs a predefined process.
  • event delivery instructions 188 include sub-event delivery instructions that deliver event information about a sub-event without activating an event handler. Instead, the sub-event delivery instructions deliver event information to event handlers associated with the series of sub-events or to actively involved views. Event handlers associated with the series of sub-events or with actively involved views receive the event information and perform a predetermined process.
  • data updater 176 creates and updates data used in application 136-1. For example, data updater 176 updates the telephone number used in contacts module 137, or stores a video file used in video and music player module 152.
  • object updater 177 creates and updates objects used in application 136-1. For example, object updater 177 creates a new user-interface object or updates the position of a user-interface object.
  • GUI updater 178 updates the GUI. For example, GUI updater 178 prepares display information and sends it to graphics module 132 for display on a touch- sensitive display.
  • event handler(s) 190 includes or has access to data updater 176, object updater 177, and GUI updater 178.
  • data updater 176, object updater 177, and GUI updater 178 are included in a single module of a respective application 136-1 or application view 191. In other embodiments, they are included in two or more software modules.
  • event handling of user touches on touch-sensitive displays also applies to other forms of user inputs to operate multifunction devices 100 with input-devices, not all of which are initiated on touch screens.
  • mouse movement and mouse button presses optionally coordinated with single or multiple keyboard presses or holds; contact movements such as taps, drags, scrolls, etc., on touch-pads; pen stylus inputs; movement of the device; oral instructions; detected eye movements; biometric inputs; and/or any combination thereof are optionally utilized as inputs corresponding to sub-events which define an event to be recognized.
  • Figure 2 illustrates a portable multifunction device 100 having a touch screen (e.g., touch-sensitive display system 112, Figure 1A) in accordance with some embodiments.
  • the touch screen optionally displays one or more graphics within user interface (UI) 200.
  • UI user interface
  • the device detects input(s) that corresponds to a request to select one or more of the graphics by detecting a gesture on the graphics, for example, with one or more fingers 202 (not drawn to scale in the figure) or one or more styluses 203 (not drawn to scale in the figure).
  • selection of one or more graphics occurs when the user breaks contact with the one or more graphics.
  • the gesture optionally includes one or more taps, one or more swipes (from left to right, right to left, upward and/or downward) and/or a rolling of a finger (from right to left, left to right, upward and/or downward) that has made contact with device 100.
  • inadvertent contact with a graphic does not select the graphic.
  • a swipe gesture that sweeps over an application icon optionally does not select the corresponding application when the gesture corresponding to selection is a tap.
  • Device 100 optionally also includes one or more physical buttons, such as “home” or menu button 204.
  • menu button 204 is, optionally, used to navigate to any application 136 in a set of applications that are, optionally executed on device 100.
  • the menu button is implemented as a soft key in a GUI displayed on the touch-screen display.
  • device 100 includes the touch-screen display, menu button 204 (sometimes called home button 204), push button 206 for powering the device on/off and locking the device, volume adjustment button(s) 208, Subscriber Identity Module (SIM) card slot 210, head set jack 212, and docking/charging external port 124.
  • Push button 206 is, optionally, used to turn the power on/off on the device by depressing the button and holding the button in the depressed state for a predefined time interval; to lock the device by depressing the button and releasing the button before the predefined time interval has elapsed; and/or to unlock the device or initiate an unlock process.
  • device 100 also accepts verbal input for activation or deactivation of some functions through microphone 113.
  • Device 100 also, optionally, includes one or more contact intensity sensors 165 for detecting intensities of contacts on touch-sensitive display system 112 and/or one or more tactile output generators 167 for generating tactile outputs for a user of device 100.
  • FIG. 3 is a block diagram of an example multifunction device with a display and a touch-sensitive surface in accordance with some embodiments.
  • Device 300 need not be portable.
  • device 300 is a laptop computer, a desktop computer, a tablet computer, a multimedia player device, a navigation device, an educational device (such as a child’s learning toy), a gaming system, or a control device (e.g., a home or industrial controller).
  • Device 300 typically includes one or more processing units (CPU’s) 310, one or more network or other communications interfaces 360, memory 370, and one or more communication buses 320 for interconnecting these components.
  • Communication buses 320 optionally include circuitry (sometimes called a chipset) that interconnects and controls communications between system components.
  • Device 300 includes input/output (I/O) interface 330 comprising display 340, which is typically a touch-screen display.
  • I/O interface 330 also optionally includes a keyboard and/or mouse (or other pointing device) 350 and touchpad 355, tactile output generator 357 for generating tactile outputs on device 300 (e.g., similar to tactile output generator(s) 167 described above with reference to Figure 1 A), sensors 359 (e.g., optical, acceleration, proximity, touch-sensitive, and/or contact intensity sensors similar to contact intensity sensor(s) 165 described above with reference to Figure 1 A).
  • sensors 359 e.g., optical, acceleration, proximity, touch-sensitive, and/or contact intensity sensors similar to contact intensity sensor(s) 165 described above with reference to Figure 1 A).
  • Memory 370 includes high-speed random access memory, such as DRAM, SRAM, DDR RAM or other random access solid state memory devices; and optionally includes nonvolatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid state storage devices. Memory 370 optionally includes one or more storage devices remotely located from CPU(s) 310. In some embodiments, memory 370 stores programs, modules, and data structures analogous to the programs, modules, and data structures stored in memory 102 of portable multifunction device 100 ( Figure 1A), or a subset thereof. Furthermore, memory 370 optionally stores additional programs, modules, and data structures not present in memory 102 of portable multifunction device 100.
  • memory 370 of device 300 optionally stores drawing module 380, presentation module 382, word processing module 384, website creation module 386, disk authoring module 388, and/or spreadsheet module 390, while memory 102 of portable multifunction device 100 ( Figure 1A) optionally does not store these modules.
  • Each of the above identified elements in Figure 3 are, optionally, stored in one or more of the previously mentioned memory devices.
  • Each of the above identified modules corresponds to a set of instructions for performing a function described above.
  • the above identified modules or programs (e.g., sets of instructions) need not be implemented as separate software programs, procedures or modules, and thus various subsets of these modules are, optionally, combined or otherwise re-arranged in various embodiments.
  • memory 370 optionally stores a subset of the modules and data structures identified above.
  • memory 370 optionally stores additional modules and data structures not described above.
  • Figure 4A illustrates an example user interface for a menu of applications on portable multifunction device 100 in accordance with some embodiments. Similar user interfaces are, optionally, implemented on device 300.
  • user interface 400 includes the following elements, or a subset or superset thereof:
  • Tray 408 with icons for frequently used applications such as: o Icon 416 for telephone module 138, labeled “Phone,” which optionally includes an indicator 414 of the number of missed calls or voicemail messages; o Icon 418 for e-mail client module 140, labeled “Mail,” which optionally includes an indicator 410 of the number of unread e-mails; o Icon 420 for browser module 147, labeled “Browser”; and o Icon 422 for video and music player module 152, labeled “Music”; and
  • Icons for other applications such as: o Icon 424 for IM module 141, labeled “Messages”; o Icon 426 for calendar module 148, labeled “Calendar”; o Icon 428 for image management module 144, labeled “Photos”; o Icon 430 for camera module 143, labeled “Camera”; o Icon 432 for online video module 155, labeled “Online Video”; o Icon 434 for stocks widget 149-2, labeled “Stocks”; o Icon 436 for map module 154, labeled “Maps”; o Icon 438 for weather widget 149-1, labeled “Weather”; o Icon 440 for alarm clock widget 149-4, labeled “Clock”; o Icon 442 for workout support module 142, labeled “Workout Support”; o Icon 444 for notes module 153, labeled “Notes”; and o Icon 446 for for
  • a label for a respective application icon includes a name of an application corresponding to the respective application icon.
  • a label for a particular application icon is distinct from a name of an application corresponding to the particular application icon.
  • Figure 4B illustrates an example user interface on a device (e.g., device 300, Figure 3) with a touch-sensitive surface 451 (e.g., a tablet or touchpad 355, Figure 3) that is separate from the display 450.
  • a touch-sensitive surface 451 e.g., a tablet or touchpad 355, Figure 3
  • the device detects inputs on a touch-sensitive surface that is separate from the display, as shown in FIG. 4B.
  • the touch-sensitive surface (e.g., 451 in Figure 4B) has a primary axis (e.g., 452 in Figure 4B) that corresponds to a primary axis (e.g., 453 in Figure 4B) on the display (e.g., 450).
  • the device detects contacts (e.g., 460 and 462 in Figure 4B) with the touch-sensitive surface 451 at locations that correspond to respective locations on the display (e.g., in Figure 4B, 460 corresponds to 468 and 462 corresponds to 470).
  • user inputs e.g., contacts 460 and 462, and movements thereof
  • the device on the touch-sensitive surface e.g., 451 in Figure 4B
  • the device on the touch-sensitive surface e.g., 451 in Figure 4B
  • the device on the touch-sensitive surface e.g., 451 in Figure 4B
  • the device on the touch-sensitive surface e.g., 451 in Figure 4B
  • the device on the touch-sensitive surface e.g., 451 in Figure 4B
  • Similar methods are, optionally, used for other user interfaces described herein.
  • finger inputs e.g., finger contacts, finger tap gestures, finger swipe gestures, etc.
  • one or more of the finger inputs are replaced with input from another input device (e.g., a mouse based input or a stylus input).
  • a swipe gesture is, optionally, replaced with a mouse click (e.g., instead of a contact) followed by movement of the cursor along the path of the swipe (e.g., instead of movement of the contact).
  • a tap gesture is, optionally, replaced with a mouse click while the cursor is located over the location of the tap gesture (e.g., instead of detection of the contact followed by ceasing to detect the contact).
  • a tap gesture is, optionally, replaced with a mouse click while the cursor is located over the location of the tap gesture (e.g., instead of detection of the contact followed by ceasing to detect the contact).
  • multiple user inputs it should be understood that multiple computer mice are, optionally, used simultaneously, or a mouse and finger contacts are, optionally, used simultaneously.
  • a gesture includes an air gesture.
  • An air gesture is a gesture that is detected without the user touching (or independently of) an input element that is part of a device (e.g., computer system 101, one or more input device 125, and/or hand tracking device 140) and is based on detected motion of a portion (e.g., the head, one or more arms, one or more hands, one or more fingers, and/or one or more legs) of the user’s body through the air including motion of the user’s body relative to an absolute reference (e.g., an angle of the user’s arm relative to the ground or a distance of the user’s hand relative to the ground), relative to another portion of the user’s body (e.g., movement of a hand of the user relative to a shoulder of the user, movement of one hand of the user relative to another hand of the user, and/or movement of a finger of the user relative to another finger or portion of a hand of the user), and/or absolute motion of a portion of the user
  • input gestures used in the various examples and embodiments described herein include air gestures performed by movement of the user’s finger(s) relative to other finger(s) or part(s) of the user’s hand) for interacting with an XR environment (e.g., a virtual or mixed-reality environment), in accordance with some embodiments.
  • XR environment e.g., a virtual or mixed-reality environment
  • an air gesture is a gesture that is detected without the user touching an input element that is part of the device (or independently of an input element that is a part of the device) and is based on detected motion of a portion of the user’s body through the air including motion of the user’s body relative to an absolute reference (e.g., an angle of the user’s arm relative to the ground or a distance of the user’s hand relative to the ground), relative to another portion of the user’s body (e.g., movement of a hand of the user relative to a shoulder of the user, movement of one hand of the user relative to another hand of the user, and/or movement of a finger of the user relative to another finger or portion of a hand of the user), and/or absolute motion of a portion of the user’s body (e.g., a tap gesture that includes movement of a hand in a predetermined pose by a predetermined amount and/or speed, or a shake gesture that includes a predetermined speed or amount of rotation of a portion of the user
  • the input gesture is an air gesture (e.g., in the absence of physical contact with an input device that provides the computer system with information about which user interface element is the target of the user input, such as contact with a user interface element displayed on a touchscreen, or contact with a mouse or trackpad to move a cursor to the user interface element), the gesture takes into account the user’s attention (e.g., gaze) to determine the target of the user input (e.g., for direct inputs, as described below).
  • attention e.g., gaze
  • the input gesture is, for example, detected attention (e.g., gaze) toward the user interface element in combination (e.g., concurrent) with movement of a user’s finger(s) and/or hands to perform a pinch and/or tap input, as described in more detail below.
  • detected attention e.g., gaze
  • the user interface element in combination (e.g., concurrent) with movement of a user’s finger(s) and/or hands to perform a pinch and/or tap input, as described in more detail below.
  • input gestures that are directed to a user interface object are performed directly or indirectly with reference to a user interface object.
  • a user input is performed directly on the user interface object in accordance with performing the input gesture with the user’s hand at a position that corresponds to the position of the user interface object in the three-dimensional environment (e.g., as determined based on a current viewpoint of the user).
  • the input gesture is performed indirectly on the user interface object in accordance with the user performing the input gesture while a position of the user’s hand is not at the position that corresponds to the position of the user interface object in the three-dimensional environment while detecting the user’s attention (e.g., gaze) on the user interface object.
  • attention e.g., gaze
  • the device responds to the user’s input to the user interface object when the user initiates the gesture at, or near, a position corresponding to the displayed position of the user interface object (e.g., within 0.5 cm, 1 cm, 5 cm, or a distance between 0-5 cm, as measured from an outer edge of the option or a center portion of the option).
  • a position corresponding to the displayed position of the user interface object e.g., within 0.5 cm, 1 cm, 5 cm, or a distance between 0-5 cm, as measured from an outer edge of the option or a center portion of the option.
  • the device responds to the user’s input to the user interface object when the user directs his or her attention to the user interface object (e.g., by gazing at the user interface object) and, while paying attention to the option, the user initiates the input gesture (e.g., at any position that is detectable by the computer system) (e.g., at a position that does not correspond to the displayed position of the user interface object).
  • input gestures used in the various examples and embodiments described herein include pinch inputs and tap inputs, for interacting with a virtual or mixed-reality environment, in accordance with some embodiments.
  • the pinch inputs and tap inputs described below are performed as air gestures.
  • a pinch input is part of an air gesture that includes one or more of: a pinch gesture, a long pinch gesture, a pinch and drag gesture, or a double pinch gesture.
  • a pinch gesture that is an air gesture includes movement of two or more fingers of a hand to make contact with one another, that is, optionally, followed by an immediate (e.g., within 0-1 seconds) break in contact from each other.
  • a long pinch gesture that is an air gesture includes movement of two or more fingers of a hand to make contact with one another for at least a threshold amount of time (e.g., at least 1 second), before detecting a break in contact with one another.
  • a long pinch gesture includes the user holding a pinch gesture (e.g., with the two or more fingers making contact), and the long pinch gesture continues until a break in contact between the two or more fingers is detected.
  • a double pinch gesture that is an air gesture comprises two (e.g., or more) pinch inputs (e.g., performed by the same hand) detected in immediate (e.g., within a predefined time period) succession of each other.
  • the user performs a first pinch input (e.g., a pinch input or a long pinch input), releases the first pinch input (e.g., breaks contact between the two or more fingers), and performs a second pinch input within a predefined time period (e.g., within 1 second or within 2 seconds) after releasing the first pinch input.
  • a first pinch input e.g., a pinch input or a long pinch input
  • releases the first pinch input e.g., breaks contact between the two or more fingers
  • a second pinch input within a predefined time period (e.g., within 1 second or within 2 seconds) after releasing the first pinch input.
  • a pinch and drag gesture that is an air gesture includes a pinch gesture (e.g., a pinch gesture or a long pinch gesture) performed in conjunction with (e.g., followed by) a drag input that changes a position of the user’s hand from a first position (e.g., a start position of the drag) to a second position (e.g., an end position of the drag).
  • a pinch gesture e.g., a pinch gesture or a long pinch gesture
  • the user maintains the pinch gesture while performing the drag input, and releases the pinch gesture (e.g., opens their two or more fingers) to end the drag gesture (e.g., at the second position).
  • the pinch input and the drag input are performed by the same hand (e.g., the user pinches two or more fingers to make contact with one another and moves the same hand to the second position in the air with the drag gesture).
  • the pinch input is performed by a first hand of the user and the drag input is performed by the second hand of the user (e.g., the user’s second hand moves from the first position to the second position in the air while the user continues the pinch input with the user’s first hand.
  • an input gesture that is an air gesture includes inputs (e.g., pinch and/or tap inputs) performed using both of the user’s two hands.
  • the input gesture includes two (e.g., or more) pinch inputs performed in conjunction with (e.g., concurrently with, or within a predefined time period of) each other.
  • two pinch inputs performed in conjunction with (e.g., concurrently with, or within a predefined time period of) each other.
  • a first pinch gesture performed using a first hand of the user (e.g., a pinch input, a long pinch input, or a pinch and drag input), and, in conjunction with performing the pinch input using the first hand, performing a second pinch input using the other hand (e.g., the second hand of the user’s two hands).
  • movement between the user’s two hands e.g., to increase and/or decrease a distance or relative orientation between the user’s two hands).
  • a tap input (e.g., directed to a user interface element) performed as an air gesture includes movement of a user's finger(s) toward the user interface element, movement of the user's hand toward the user interface element optionally with the user’s finger(s) extended toward the user interface element, a downward motion of a user's finger (e.g., mimicking a mouse click motion or a tap on a touchscreen), or other predefined movement of the user’s hand.
  • a tap input that is performed as an air gesture is detected based on movement characteristics of the finger or hand performing the tap gesture movement of a finger or hand away from the viewpoint of the user and/or toward an object that is the target of the tap input followed by an end of the movement.
  • the end of the movement is detected based on a change in movement characteristics of the finger or hand performing the tap gesture (e.g., an end of movement away from the viewpoint of the user and/or toward the object that is the target of the tap input, a reversal of direction of movement of the finger or hand, and/or a reversal of a direction of acceleration of movement of the finger or hand).
  • a change in movement characteristics of the finger or hand performing the tap gesture e.g., an end of movement away from the viewpoint of the user and/or toward the object that is the target of the tap input, a reversal of direction of movement of the finger or hand, and/or a reversal of a direction of acceleration of movement of the finger or hand.
  • the term “focus selector” refers to an input element that indicates a current part of a user interface with which a user is interacting.
  • the cursor acts as a “focus selector,” so that when an input (e.g., a press input) is detected on a touch-sensitive surface (e.g., touchpad 355 in Figure 3 or touch-sensitive surface 451 in Figure 4B) while the cursor is over a particular user interface element (e.g., a button, window, slider or other user interface element), the particular user interface element is adjusted in accordance with the detected input.
  • a touch-sensitive surface e.g., touchpad 355 in Figure 3 or touch-sensitive surface 451 in Figure 4B
  • a particular user interface element e.g., a button, window, slider or other user interface element
  • a detected contact on the touch-screen acts as a “focus selector,” so that when an input (e.g., a press input by the contact) is detected on the touch-screen display at a location of a particular user interface element (e.g., a button, window, slider or other user interface element), the particular user interface element is adjusted in accordance with the detected input.
  • a particular user interface element e.g., a button, window, slider or other user interface element
  • focus is moved from one region of a user interface to another region of the user interface without corresponding movement of a cursor or movement of a contact on a touch-screen display (e.g., by using a tab key or arrow keys to move focus from one button to another button); in these implementations, the focus selector moves in accordance with movement of focus between different regions of the user interface.
  • the focus selector is generally the user interface element (or contact on a touch-screen display) that is controlled by the user so as to communicate the user’s intended interaction with the user interface (e.g., by indicating, to the device, the element of the user interface with which the user is intending to interact).
  • a focus selector e.g., a cursor, a contact, or a selection box
  • a press input is detected on the touch-sensitive surface (e.g., a touchpad or touch screen) will indicate that the user is intending to activate the respective button (as opposed to other user interface elements shown on a display of the device).
  • UI user interfaces
  • portable multifunction device 100 or device 300 with a display, a touch-sensitive surface, (optionally) one or more tactile output generators for generating tactile outputs, and (optionally) one or more sensors to detect intensities of contacts with the touch-sensitive surface.
  • Figures 5A-5CB illustrate example user interfaces for navigating and editing respective types of user interfaces
  • Figures 6A-6AR illustrate example user interfaces for an expanded face switcher
  • Figures 7A-7V-3 illustrate a representation of a plurality of notifications in different configurations, and user inputs for switching between the different configurations
  • Figures 8A-8AN illustrate example user interfaces for displaying a session region for ongoing events, in accordance with some embodiments.
  • the user interfaces in these figures are used to illustrate the processes described below, including the processes in Figures 9A-9I, 10A-10I, 11A-11H, 12A-12E, 13A-13G, 14A-14G, 15A-15I, 16A-16J, and 17A-17D.
  • the focus selector is, optionally: a respective finger or stylus contact, a representative point corresponding to a finger or stylus contact (e.g., a centroid of a respective contact or a point associated with a respective contact), or a centroid of two or more contacts detected on the touch-sensitive display system 112.
  • analogous operations are, optionally, performed on a device with a display 450 and a separate touch- sensitive surface 451 in response to detecting the contacts on the touch-sensitive surface 451 while displaying the user interfaces shown in the figures on the display 450, along with a focus selector.
  • Figures 5A-5CB illustrate example user interfaces for navigating and editing respective types of user interfaces in accordance with some embodiments.
  • Figure 5 A illustrates a user interface 500 of a device 100 while the display of device 100 is in a low power state.
  • an “always on” display optionally displays an indication of a time and/or date.
  • the display of device 100 in the low power state, does not display any information (e.g., the display is off).
  • the display enters a low power state after a predetermined time of inactivity during which no user interactions with device 100 have been detected.
  • the display enters a low power state in response to a user input to lock and/or put the device 100 to sleep, such as by pressing a physical or virtual button on device 100.
  • the display of device 100 leaves the low power state in accordance with receiving an incoming alert (e.g., a system alert and/or a notification from an application). For example, the display of device 100 illuminates, and leaves the low power state, to display an indication of the alert. In some embodiments, the display of device 100 illuminates in response to a user input, such as a user input detected on the display, a user picking up the device, and/or a user input on a physical or virtual button of the device.
  • a user input such as a user input detected on the display, a user picking up the device, and/or a user input on a physical or virtual button of the device.
  • a wake screen user interface is a user interface that is displayed after the display of device 100 has entered a low power state during which the display is at least partially off.
  • a wake screen user interface is also referred to herein as a face.
  • actions described as being performed with respect to a wake screen user interface may also be described as being performed with respect to a face (e.g., “switching between wake screen user interfaces” may also be stated as “switching between faces” and “editing a wake screen user interface” may also be stated as “editing a face”).
  • an “expanded face switcher” user interface includes display of one or more faces (e.g., one or more wake screen user interfaces), wherein a size of a respective face (e.g., wake screen user interface) is less than a full size of the display area (e.g., as illustrated in Figure 5U).
  • the display in the low power state, the display optionally displays an “always on” indicator of a time and/or date and the device displays the wake screen user interface when the device is prompted to come out of the low power state.
  • the device optionally in response to a user input and/or in response to a threshold amount of time elapsing, the device enters a locked state in which a password, passcode and/or biometric authentication is required to unlock the device, wherein the device has limited functionality in the locked state and must be unlocked before accessing respective applications and/or data stored on device 100.
  • the wake screen user interface is displayed regardless of whether the device is in the locked state or has already been unlocked (e.g., the wake screen user interface is displayed upon waking the device before the user accesses a home screen user interface and/or other application user interfaces).
  • a cover sheet user interface includes the same characteristics as the wake screen user interface, but the device displays the cover sheet user interface optionally without entering the locked and/or low power state of the device.
  • one or more alerts are displayed on the wake screen user interface and/or the cover sheet user interface, optionally in response to a user input (e.g., a swipe gesture upward in the middle of the display or another gesture).
  • a home screen user interface includes icons for navigating to a plurality of applications that are executed by the device 100.
  • the device 100 detects and responds to interaction with the home screen user interface using one or more gestures, including touch inputs.
  • a tap input or other selection input on a respective application icon causes the respective application to launch, or otherwise open a user interface for the respective application, on the display area of device 100.
  • a plurality of views for the home screen user interface is available.
  • the device detects and responds to user inputs such as swipe gestures or other inputs (e.g., inputs directed to the currently displayed view of the home screen user interface) that correspond to requests to navigate between the plurality of views, wherein each view of the home screen user interface includes different application icons for different applications.
  • the application icons are different sizes, such as an application widget that displays information for the respective application, wherein the application widget is larger than the application icons.
  • mini-application objects are user interface objects that provide a limited subset of functions and/or information available from their corresponding applications without requiring the applications to be launched.
  • mini-application objects contain application content that is dynamically updated based on the current context.
  • a tap input or other selection input on a mini-application object causes the corresponding application to be launched.
  • a respective mini application object operates as a standalone application residing in memory of the device, distinct from an associated application also residing in the memory of the device.
  • a respective mini application object operates as an extension or component of an associated application on the device.
  • a respective mini application object has a dedicated memory portion for temporary storage of information. In some embodiments, the memory portion is accessible by a corresponding full-featured application of the respective mini application object. In some embodiments, a mini application object is configured to perform a subset, less than all, of the functions of a corresponding application. In some embodiments, a mini application object displays an identifier for the corresponding application. In some embodiments, a mini application object displays a portion of the content from the corresponding application. For example, a map mini application object displays a portion of a map that is displayed in a map application that corresponds to the map mini application object. For example, a calendar mini application object displays a portion of a calendar that is displayed in a corresponding calendar application.
  • a predefined input on a mini application object launches the corresponding application.
  • a mini application object operates as a standalone application residing in memory of the device, distinct from an associated application also residing in the memory of the device.
  • a mini application object corresponding to a social networking application operates as a single-purpose or streamlined application with a subset, less than all, of the functionality of the corresponding application, but is associated with the full-featured social networking application.
  • the mini application object operates independently of the social networking application, and in a scenario where the social networking application is not running, the mini application object continues to operate.
  • a mini application object operates as an extension or component of an associated application on the device.
  • a mini application object for a calendar application is a single feature or operational component of the full-featured calendar application.
  • the calendar mini application object does not operate either.
  • a mini application object has a dedicated memory portion for temporary storage of information. In some embodiments, this memory portion can be accessed by the corresponding full-featured application.
  • a mini application object for an instant messaging application has a memory portion for temporary storage of partially written reply messages. In this example, if the user opens the corresponding application in the middle of writing a reply message, the contents of the reply message are retrieved from the temporary storage location and used by the full-featured application to allow the user to complete his reply message.
  • an animation is displayed to transition the device from being off (e.g., displaying a black background optionally with the “always on” time and/or date) to displaying a wake screen user interface 501 (also referred to herein as a lock screen user interface).
  • the wake screen user interface 501 is a user interface that includes an indication of a time and/or date, optionally one or more complications (e.g., workout complication 502-1, weather complication 502-2 and calendar complication 502-3) that correspond to respective applications and update with current status information of the respective applications, optionally one or more alerts (e.g., messages notification 503-1 and social media notification 503-2), optionally a shortcut to one or more applications (e.g., a flashlight and/or a camera), and a background (e.g., an image, a pattern, a color, and/or a photograph).
  • complications e.g., workout complication 502-1, weather complication 502-2 and calendar complication 502-3
  • alerts e.g., messages notification 503-1 and social media notification 503-2
  • a shortcut to one or more applications e.g., a flashlight and/or a camera
  • a background e.g., an image, a pattern, a color, and/or
  • the device detects and responds to user input(s) (e.g., inputs directed to the editing user interface 565-a in 5W1 or 565 in 5X) that correspond to request(s) to change one or more settings of the wake screen user interface, for example to change a font color and/or style of the time and/or date indication and/or to change the complications that are included in the wake screen user interface.
  • the device 100 stores a plurality of wake screen user interfaces and rotates and/or shuffles (optionally periodically and/or upon entering/leaving the low power state) through the plurality of wake screen user interfaces.
  • the animation that is displayed to transition the device from being off to display the wake screen user interface is based on the wake screen user interface that is to be displayed.
  • Figures 5E-5L illustrate examples of different animated transitions for different wake screen user interfaces.
  • Figure 5B illustrates a wake screen user interface 501 with a gradient background, having a dark grey at the top of the user interface background, a medium grey in the middle of the user interface background, and a light grey at the bottom of the user interface background.
  • a user input is detected for replacing display of the wake screen user interface with a home screen user interface, wherein the home screen user interface includes representations of a plurality of applications that the device 100 executes.
  • the wake screen user interface 505 in response to the user input 504, is displayed as sliding up and off of the display area while maintaining one or more features of the wake screen user interface (e.g., the indication of the time and/or date, one or more complications, and/or the flashlight and camera shortcuts), as illustrated in Figure 5C.
  • the gradient background is cycled through, as if on a carousel, such that the dark grey that was at the top of the user interface background is displayed as sliding up from the bottom of the screen, while the medium grey in the middle of the user interface background slides up and off the display, and the light grey slides up in the display area.
  • the gradient background completes a full cycle in accordance with the home screen user interface 506 being displayed.
  • Figure 5D illustrates the home screen user interface 506 with the gradient background having the different shades of grey arranged in the same arrangement as on the wake screen user interface. Accordingly, the animation that shifts the shades of grey completes a full cycle in response to the user swiping up to navigate to the home screen user interface.
  • Figures 5E-5F illustrate an example of an animated transition while the device 100 transitions from the low power state to displaying the wake screen user interface 507-2.
  • user interface 507-1 is an intermediary wake screen user interface that is displayed during the animated transition.
  • the intermediary wake screen user interface 507-1 does not include one or more features of the wake screen user interface 507-2 (e.g., no complications, notifications, and/or flashlight or camera shortcuts).
  • the wake screen user interface 507-2 comprises an image, such as a plurality of colored lines (e.g., a rainbow or a pride flag).
  • the animated transition includes increasing a size of the image.
  • the animated transition includes changing one or more font properties of the indication of the time and/or date. In some embodiments, the animated transition includes displaying one or more complications, optionally by fading-in the one or more complications, and displaying one or more alerts, optionally by fading-in the one or more complications.
  • Figures 5G-5H illustrate an example of an animated transition for a wake screen user interface 508-2 that includes a portrait-style photo.
  • a portrait-style photo includes a subject (e.g., an individual and/or a face) that is optionally displayed with a simulated depth effect such that the subject appears closer in view than one or more background objects (e.g., the cloud) in the photo.
  • the animated transition for a wake screen user interface with a portrait-style photo includes increasing a size of the subject from the user interface 508-1 to the wake screen user interface 508-2, and optionally displaying the subject overlaid with the indication of the time and/or date.
  • the animated transition includes applying the depth effect (e.g., the photo appears flat in Figure 5G and appears with the depth effect in Figure 5H).
  • the wake screen user interface that includes a portrait-style photo does not include one or more complications (optionally, the device allows the user to add one or more complications by editing the wake screen user interface (e.g., using selectable options in user interface element 570 in the editing user interface 565 shown in Figure 5X-5Y)).
  • Figures 5I-5J illustrate an example of an animated transition for a wake screen user interface 509-2 that includes one or more emoji.
  • the wake screen user interface 509-2 includes one or more emoji arranged in a predefined pattern.
  • the smiley face emoji is displayed in a grid pattern.
  • the animated transition includes changing a shadow effect applied to the smiley face emoji, as displayed in user interface 509-1, to a different shadow effect as displayed in wake screen user interface 509-2.
  • no shadow effect is displayed in at least one of the user interfaces 509-1 or 509-2. For example, the size and/or opacity of the shadow increases or decreases during the animation.
  • Figures 5K-5L illustrate an example of an animated transition for a wake screen user interface 510-2 that includes an astronomy background.
  • the astronomy background includes a celestial and/or planetary body (e.g., the earth, a moon, stars, solar system and/or a galaxy).
  • the animated transition includes changing a crop and/or changing a view angle of the astronomy background from the user interface 510-1 to the wake screen user interface 510-2.
  • the animated transition includes applying a visual effect, such as decreasing and/or moving a shadow displayed over the background.
  • the astronomy background 510-2 further includes an indication of a current location of the device 100 (e.g., a dot appears on the displayed earth in the user interface 510-2 at the device’s location).
  • Figures 5L2-5L3 illustrate an example of an animated transition from a wake screen user interface 510-3 that includes an astronomy background to a home screen user interface 510-4 that includes an astronomy background.
  • wake screen user interface 510-3 includes a representation of one or more celestial bodies, for example the earth and the sun are illustrated in Figure 5L2.
  • the device 100 in response to user input 515 requesting to dismiss the wake screen user interface 510-3, displays an animated transition and displays a home screen user interface 510-4.
  • home screen user interface 510-4 is displayed with a background that includes representation of the one or more celestial bodies that is distinct from the representation of the one or more celestial bodies displayed on wake screen user interface 510-3.
  • the representation of the earth is displayed as rotating and shifting to a different portion of the display, and the representation of the sun is displayed as shifting in accordance with the rotation of the representation of the earth, as illustrated in Figure 5L3.
  • the representation of the sun is shifted to be displayed relatively closer to the representation of the earth and/or is shifted to be displayed at a position directly behind the representation of the earth (e.g., as opposed to up and to the left, as illustrated in Figure 5L2).
  • the representation of the earth is shifted to increase in size (e.g., as if getting closer) and the representation of the sun updates in size relative to the increase in size of the representation of the earth.
  • Figure 5M illustrates an example of a wake screen user interface 511.
  • the wake screen user interface 511 is an example of an astronomy background (e.g., that includes a moon).
  • the device detects user interactions with the wake screen user interface 511 that cause the device 100 to perform respective operations (e.g., selecting a respective complication to displaying its corresponding application, navigating to the home screen user interface, displaying a plurality of previously received and/or stored notification, and/or navigating to another system user interface, such as a system search user interface, a widget screen, or a control panel user interface).
  • respective operations e.g., selecting a respective complication to displaying its corresponding application, navigating to the home screen user interface, displaying a plurality of previously received and/or stored notification, and/or navigating to another system user interface, such as a system search user interface, a widget screen, or a control panel user interface.
  • the wake screen user interface 511 includes music complication 502-4, health complication 502-5, weather complication 502-6 and breathe complication 502-7.
  • music complication 502-4 updates with progress information (e.g., in a circular, curved, or linear shape) indicating an amount of time that a currently playing media item is played back
  • weather complication 502-6 updates to indicate a real-time AQI level (e.g., “32”) based on the status of the weather application.
  • the device 100 in response to a first type of user input, such as user input 512, selecting a respective complication, displays an application user interface for the application associated with the respective complication.
  • the first type of user input is a tap input.
  • the device 100 in response to user input 512 directed to health complication 502-5, displays a user interface for a health/fitness application associated with health complication 502-5.
  • the device 100 in response to a second type of user input 514, such as a swipe input or a drag gesture in a first direction (e.g., upward, downward, rightward or leftward) that is detected within a predefined portion of the wake screen user interface 511 (e.g., within a middle of the wake screen and/or not on an edge of the wake screen), the device 100 displays a plurality of notifications, such as in user interface 532 illustrated in Figure 50.
  • a second type of user input 514 such as a swipe input or a drag gesture in a first direction (e.g., upward, downward, rightward or leftward) that is detected within a predefined portion of the wake screen user interface 511 (e.g., within a middle of the wake screen and/or not on an edge of the wake screen)
  • a third type of user input 516 such as a swipe input or a drag gesture in a second direction (e.g., upward, downward, rightward, or leftward) that is optionally the same direction is the direction of second type of user input 514
  • the device displays a home screen user interface 518.
  • the third type of user input 516 is received on a user interface element that, when selected (optionally by initiating a swipe gesture over the user interface element) causes the device 100 to display a home screen user interface 518.
  • Figure 5N illustrates home screen user interface 518 (also referred to as a home user interface).
  • the home screen user interface includes icons for navigating to a plurality of applications that are executed by the device 100.
  • the device detects and responds to user interactions with the home screen user interface using one or more gestures, including touch inputs, that cause the device to perform respective operations (e.g., selecting a respective application icon to display a corresponding application, and/or navigating to another system user interface such as the cover sheet user interface, the system search user interface, the control panel user interface, or the wake screen user interface).
  • a tap input e.g., user input 524 on a respective application icon causes the respective application to launch, or otherwise open a user interface for the respective application, on the display area of device 100.
  • a plurality of views for the home screen user interface is available.
  • the device detects one or more swipe gestures or other types of navigation inputs directed to a currently displayed view for the home screen user interface to navigate between the plurality of views, wherein each view of the home screen user interface includes different application icons for different applications.
  • the application icons are different sizes, such as an application widget that displays information for the respective application, wherein the application widget is larger than the application icons.
  • a user input 520 is a swipe user input in a third direction (e.g., downward or another direction) that is initiated at a top corner edge of the display of device 100.
  • the device 100 in response to user input 520, displays a control user interface for modifying one or more settings.
  • a user input 522 is a swipe user input in a fourth direction (e.g., from right to left) and corresponds to a request to switch to another view of the home screen user interface. For example, different representations for different applications are displayed on different views of the home screen user interface.
  • a user input 524 is a user input (e.g., a tap input or other selection input) that selects application icon 422 for a music application.
  • the device 100 in response to user input 524, displays a user interface for the application associated with application icon 422 (e.g., a music application user interface).
  • a user input 525 (e.g., a tap input or other selection input) on another application icon 440 for a clock application causes device 100 to open the clock application and display a user interface for the clock application.
  • a user input 526 is detected at a predefined portion of the user interface (e.g., a swipe up gesture that is initiated at an edge of the display of device 100).
  • the device in response to user input 526, displays a multitasking user interface with indications of one or more open applications that are optionally executing in the background on device 100.
  • a different type of input such as user input 528 (e.g., a swipe input from left to right that is initiated at a left edge of the display of device 100 or other gesture), causes the device 100 to display a user interface that includes a search bar and optionally one or more widgets and/or shortcuts that display information for a subset of applications.
  • user input 528 e.g., a swipe input from left to right that is initiated at a left edge of the display of device 100 or other gesture
  • a user input 530 e.g., a swipe input downward that is initiated at a top edge of the display of device 100 or other gesture
  • a user interface 532 Figure 50
  • device 100 detects different types of inputs and/or gestures (e.g., inputs 530, 520, 525, 528, 522, 526, and 524 directed to home screen user interface 518 in Figure 5N) that cause the device 100 to perform sets of operations, including opening applications, navigating between views of the home screen user interface, and access other user interfaces for modifying settings and/or accessing shortcuts.
  • Figure 50 illustrates a coversheet user interface 532 that includes information that is also displayed on wake screen user interface 511.
  • user interface 532 optionally includes an indication of the time and/or date, one or more complications, shortcut options to a flashlight and/or camera, and/or a plurality of notifications.
  • the device detects and responds to user inputs directed to user interface 532 in the same ways as user inputs directed to wake screen user interface 511, such as detecting a tap input on a complication and opening the respective application associated with the complication, or detecting a swipe input that is initiated on the right edge of the display of device 100, such as user input 534, and displaying a user interface 536 (Figure 5P) for a camera application, wherein the device provides functions to capture and/or view photos using the camera application.
  • Figure 5P illustrates a user input 538 corresponding to a request to return to a wake screen user interface.
  • user interface 540-1 in response to user input 538, user interface 540-1 is displayed, as illustrated in Figure 5Q.
  • wake screen user interface 540-1 includes one or more properties that are distinct from the wake screen user interface 511 (Figure 5M).
  • at least one complication is changed from wake screen user interface 511 to wake screen user interface 540-1. For example, music complication 502-4 ( Figure 5M) is replaced with weather complication 502-9 ( Figure 5Q), while optionally maintaining one or more other complications (e.g., health complication 502- 5, weather complication 502-6 and breathe complication 502-7).
  • the complications update in accordance with a status of the application.
  • weather complication 502-6 indicates an air quality index (AQI) of 32
  • AQI air quality index
  • weather complication 502-6 indicates an AQI of 30.
  • the background of the wake screen is maintained (e.g., an image of a moon).
  • a user input 541 such as a swipe input from right to left that is detected within a predefined region of the display of device 100 (e.g., within a predefined distance of the bottom edge of the display), causes the device 100 to cycle the wake screen user interface 540-1 to another wake screen user interface 540-2.
  • the device 100 cycles through a predefined set of wake screen user interfaces.
  • the predefined set of wake screen user interfaces are related. For example, for an astronomy wake screen user interface (e.g., a moon image), cycling through the predefined set of astronomy wake screen user interfaces includes displaying different phases of the moon.
  • the predefined set of wake screen user interfaces are defined by the user, such that the user selects a collection of wake screen user interfaces to cycle through, regardless of whether the wake screens relate. For example, the user selects a photo album as a background image, and the wake screen user interfaces cycle through the photos in the photo album or photos automatically selected by the device based on criteria specified by the user.
  • the wake screen user interface automatically changes periodically (e.g., every 2 minutes, every day, or every week).
  • the wake screen user interface automatically changes after the device has entered and/or exited a low power state (e.g., after the display has been off, the wake screen user interface updates the next time the display wakes up).
  • FIG. 5R illustrates wake screen user interface 540-2 that replaces wake screen user interface 540-1 in response to user input 541.
  • wake screen user interface 540-2 includes at least one feature that is different than wake screen user interface 540-1.
  • the font of the indication of the time and/or date is changed to a bolded font, and the background image and color of the background changes.
  • one or more other features of the wake screen user interface do not change, such as the type of complications. It will be understood that any combination of features may be updated between a first wake screen user interface and a second wake screen user interface, including a subset, less than all, or all of the complications changing, and/or font properties of the indication of the time and/or date.
  • Figure 5R further illustrates a user input 544 corresponding to a request to switch to another wake screen user interface via a wake screen selector user interface.
  • user input 544 is a long press or other selection input, wherein the user input maintains contact with the display for a predetermined threshold amount of time (e.g., 2 seconds, 3 seconds, or 5 seconds) at a predefined portion of the user interface 540-2 (e.g., at a portion of the user interface that does not correspond to a complication, a shortcut to the flashlight and/or camera, or to the indication of the time and/or date).
  • the device 100 in response to user input 544, displays a wake screen selector user interface 548-1.
  • the user input 544 maintains contact with the device while the wake screen selector user interface 548-1 is displayed, and the timing and/or position of liftoff of user input 544 causes the device 100 to perform a respective operation (e.g., a long press that is let go after a timer without moving position, causes the device to display the expanded face switcher user interface 561 ( Figure 5U) and a long press with a change in position of the contact (e.g., a swipe to an adjacent face without lifting off) causes the device to go to the adjacent wake screen user interface without entering the expanded face switcher user interface 561).
  • a respective operation e.g., a long press that is let go after a timer without moving position
  • a long press with a change in position of the contact e.g., a swipe to an adjacent face without lifting off
  • the device 100 in accordance with a determination that the device 100 is in a locked state, before displaying the wake screen selector user interface 548-1, the device 100 requires authentication to unlock the device, and displays a passcode user interface 547, as illustrated in Figure 5S. In some embodiments, the device 100 prompts the user to enter a passcode and/or authenticate using a fingerprint, facial, or other biometric identification.
  • Figure 5T illustrates the wake screen selector user interface 548-1
  • the wake screen selector user interface 548 includes representations for a plurality of wake screen user interfaces, including a representation 552 of a portion of a first wake screen user interface (e.g., wake screen user interface 540-1), a representation 550 of a second wake screen user interface (e.g., wake screen user interface 540-2), and a representation 554 of a portion of a third wake screen user interface (e.g., wake screen user interface 507-1 ( Figure 5W)).
  • the wake screen selector user interface 548-1 displays representations for a set of wake screen interfaces that the device 100 stores, and optionally cycles through, as described above.
  • the wake screen selector user interface 548-1 provides a plurality of options for the user to interact with the representations of wake screen user interfaces. For example, the device initiates a process to add a new wake screen user interface to the set of wake screen user interfaces in response to user input 556 (e.g., a tap input or other selection input) on the “+” button.
  • user input 556 e.g., a tap input or other selection input
  • the device 100 in response to detecting a user input that selects a respective representation of a respective wake screen user interface in a wake screen selector interface displaying respective representations of multiple instances of the wake screen user interface, displays the respective wake screen user interface as the current wake screen user interface. For example, in response to a user input 560, such as a tap input or other selection input, on or directed to the representation 550 of the second wake screen user interface, the device ceases display of the wake screen selector user interface 548-1 and redisplays the wake screen user interface 540-2 ( Figure 5R). In some embodiments, in response to a user input 558, such as a tap input or other input, the device scrolls to the right to bring the representation 554 to the center of the wake screen selector user interface 548-2 ( Figure 5V).
  • a user input 560 such as a tap input or other selection input
  • the device in response to a user input 558, such as a tap input or other input, the device scrolls to the right to bring the representation 554 to the center
  • the device 100 in accordance with a determination that the user input 544 is maintained for a threshold amount of time (e.g., a long press user input that is held from the wake screen user interface 540-2 and while displaying wake screen selector user interface 548-1), and is lifted off without moving (e.g., swiping to the left or right), the device 100 displays an expanded face switcher user interface 561, as illustrated in Figure 5U.
  • a threshold amount of time e.g., a long press user input that is held from the wake screen user interface 540-2 and while displaying wake screen selector user interface 548-1
  • the device 100 displays an expanded face switcher user interface 561, as illustrated in Figure 5U.
  • Figure 5U illustrates that the expanded face switcher user interface 561 includes a representation 550 of a wake screen user interface (e.g., in the center of the expanded face switcher user interface 561) and at least a portion of a representation of a home screen user interface 551 that is associated with the representation 550 of the wake screen user interface.
  • a representation 550 of a wake screen user interface e.g., in the center of the expanded face switcher user interface 561
  • a representation of a home screen user interface 551 that is associated with the representation 550 of the wake screen user interface.
  • the respective wake screen user interfaces have a corresponding respective home screen user interface.
  • changing display of a wake screen user interface from a first wake screen to a second wake screen automatically updates the home screen user interface from a first home screen associated with the first wake screen to a second home screen associated with or corresponding to the second wake screen.
  • the displayed home screen is related to the currently displayed wake screen.
  • a color, pattern, and/or image of the home screen is selected in accordance with a color, pattern and/or image of the wake screen with which the home screen is associated.
  • expanded face switcher user interface 561 optionally also displays at least a portion of a representation 552 and/or representation 554 of other wake screen user interfaces.
  • only the wake screen user interface that is currently centered in the expanded face switcher user interface 561 is displayed with a representation of its related home screen user interface.
  • representations 552 and 554 of wake screen user interfaces are not displayed with corresponding representations of home screen user interfaces.
  • expanded face switcher user interface 561 includes a user-selectable button 564 for adding a new wake screen and/or home screen to the set of wake screens and home screens that are stored and displayed in the expanded face switcher user interface 561.
  • the computer system detects and responds to user inputs directed to the expanded face switcher user interface 561 that correspond to requests to scroll to the left and/or right to navigate between wake screen and home screen options in the set of wake screens and home screens in the expanded face switcher user interface 561.
  • expanded face switcher user interface 561 further includes a user- selectable option 553 to customize the wake screen that is centered in the expanded face switcher user interface 561.
  • a user selection input on option 553 would open the editing user interface (e.g., editing user interface 565, Figure 5X) for the wake screen corresponding to representation 550.
  • expanded face switcher user interface 561 includes an option 555 for changing a background image of the background of the wake screen corresponding to representation 550.
  • a photo picker is displayed in which the computer system detects and responds to user inputs that correspond to requests to select a photo or image to be used as the background of the wake screen corresponding to representation 550.
  • Figure 5 V illustrates updating the wake screen selector user interface 548-1 to wake screen selector user interface 548-2, which brings the representation 554 of a wake screen user interface into the center of the wake screen selector user interface.
  • the wake screen selector user interface 548-2 is displayed in response to a user input 558, such as a tap input or other selection input, on the representation 554, or a swipe input or other scroll input from right to left, which causes the device to shift the representations of wake screen user interfaces that are within view to bring representation 554 of a wake screen user interface into the center, and move representation 550 of the previously centered wake screen user interface to the left (optionally sliding a portion of the representation 550 off the display), of the wake screen selector user interface 548-2.
  • a user input 558 such as a tap input or other selection input
  • a swipe input or other scroll input from right to left causes the device to shift the representations of wake screen user interfaces that are within view to bring representation 554 of a wake screen user interface into the center, and move
  • wake screen selector user interface 548-2 includes a user-selectable option to edit the currently centered wake screen. For example, in response to user input 568 (e.g., a tap input or other selection input) on the Edit button, the device 100 displays an editing user interface 565 for the wake screen, as illustrated in Figure 5X.
  • user input 568 e.g., a tap input or other selection input
  • Figure 5W illustrates a wake screen user interface 563 that is displayed in response to a user input (e.g., user input 562) that corresponds to a request to quick switch the wake screen user interface.
  • the user input to request to quick switch the wake screen user interface comprises a long press input that is initiated on a displayed wake screen user interface 540-2 (e.g., user input 544, Figure 5R), wherein the long press input is maintained (e.g., the contact of the input is continued) while displaying the wake screen selector user interface 548-1 ( Figure 5T), and while the long press input is ongoing, the input changes position (e.g., swipes in a first direction), such as user input 562 that slides from right to left.
  • a user input e.g., user input 562
  • the user input to request to quick switch the wake screen user interface comprises a long press input that is initiated on a displayed wake screen user interface 540-2 (e.g., user input 544, Figure 5R), wherein the long press input is maintained (
  • the wake screen that is displayed in the center of wake screen selector user interface 548-1 is switched to, such that the wake screen is displayed as the current wake screen. For example, in Figure 5T, the user swipes left until the representation 554 is centered (or substantially centered) in the wake screen selector user interface 548-1, then lifts off the user input to display the wake screen 563 that is associated with the representation 554, as illustrated in Figure 5W.
  • the quick switch of the wake screen user interface does not display the expanded face switcher user interface 561.
  • the device 100 displays the editing user interface 565-a ( Figure 5W1) or editing user interface 565 (Figure 5X) in response to a user input 564 on the indication of a time and/or date displayed in the wake screen user interface 563.
  • the user input 564 is a tap input, a long press input, or another selection input.
  • editing user interface 565-a, illustrated in Figure 5W1 includes reticle 572 for complications that is displayed behind, or at least partially occluded by, a subject of the background image (e.g., the star).
  • reticle 572 is displayed in front of the subject of the background image (e.g., the star) in response to a user interaction to edit one or more visual properties of the wake screen user interface, such as user input 571 -a directed to the reticle 572.
  • the background image e.g., the star
  • Figure 5X illustrates editing user interface 565, which as explained above, can be accessed by a user input 564 on the wake screen user interface 563, or by selecting the Edit button via user input 568.
  • the editing user interface 565 is also accessible via a settings application on device 100, or while a user is creating a wallpaper, which is optionally displayed in a share menu that appears in response to a user requesting to share a selected image (e.g., from a photos application).
  • editing user interface 565 includes a plurality of reticles indicating portions of the wake screen user interface that are customizable. For example, reticle 568 around the date and reticle 569 around the time indicate the date and time are editable (e.g., in text font and/or the type of information displayed), and reticle 572 indicates that one or more complications are customizable (e.g., the computer system detects and responds to a user input directed to reticle 572 that corresponds to a request to initiate a process to add, remove, and edit the complications displayed on wake screen user interface 563).
  • reticle 568 around the date and reticle 569 around the time indicate the date and time are editable (e.g., in text font and/or the type of information displayed)
  • reticle 572 indicates that one or more complications are customizable (e.g., the computer system detects and responds to a user input directed to reticle 572 that corresponds to a request to initiate a process to add, remove, and edit the complications displayed on wake
  • an indication 574 that other views are available for the selected wake screen user interface 563.
  • editing user interface 565 includes a user-selectable option to cancel 567 editing the wake screen user interface and/or a user-selectable option for saving any changes to the wake screen user interface and exiting the editing user interface 565 (in response to selection of “Done” option 566).
  • Figure 5X illustrates a user input 571, such as a tap input or other selection input, to add complications to the wake screen user interface 563.
  • the device in response to user input 571, displays a user interface element 570 that includes a plurality of available and/or recommended complications as illustrated in Figure 5Y, and the device detects and responds to user inputs selecting the displayed complications to add the selected complications to the wake screen user interface 563.
  • the device in response to user input 572 (e.g., a tap input or other selection input) selecting the recommended set of complications, including a health complication, a breathe complication, and a calendar complication), the device adds the set of complications to a predefined portion of the wake screen user interface, optionally below the time indication.
  • user input 572 e.g., a tap input or other selection input
  • the device adds the set of complications to a predefined portion of the wake screen user interface, optionally below the time indication.
  • user interface element 570 includes a plurality of complications, wherein each complication is associated with an application that executes on device 100.
  • an indication of the application associated with each complication is displayed (e.g., “Calendar” application, “Health” application, “Weather” application and “Breathe” application).
  • one application provides a plurality of complications (e.g., a plurality of options having different designs and/or displaying different status information for a complication for the respective application).
  • a third-party provider that provides a respective application optionally designs one or more complications for the respective application (e.g., using an API, programming guidelines, and/or toolkits) that are receives permission for display on one or more system user interfaces by the operating system.
  • each complication includes status information for a respective application.
  • the calendar complication includes an upcoming event that is saved on the user’s calendar (“11 :00 AM Event”).
  • a health complication includes a distance of a current workout, and/or an indication of a daily amount of activity.
  • a weather application is associated with a plurality of complications, including a complication that provides a current air quality index (AQI) and/or a complication that indicates a current weather (e.g., a sun for sunny weather or a cloud with rain for rainy weather).
  • AQI current air quality index
  • a complication that indicates a current weather e.g., a sun for sunny weather or a cloud with rain for rainy weather.
  • the computer system detects and responds to user inputs directed to user interface element 570 that correspond to requests to select one or more individual complications, as opposed to a recommended set, for example user input 574 selects a calendar complication 502-10, as shown in Figure 5Y.
  • the device in response to user input 574, the device adds the calendar complication 502-10 to the predefined area in the wake screen user interface for complications, as illustrated in Figure 5Z.
  • the device detects and responds to user input(s) directed to editing user interface 565 that correspond to request(s) to change a location of the predefined area (e.g., from below the time indication to above the date indication) in the wake screen user interface.
  • the indication to add complications (e.g., button 572 and/or button 576) is optionally displayed if there is additional space in the predefined area. In some embodiments, the indication to add complications is optionally displayed only if there are no complications in the predefined area.
  • Figure 5Z illustrates a user input 578 selecting a health complication 502-11.
  • Figure 5AA illustrates that, in response to the user input 578, the device 100 adds the health complication 502-11 to the predefined area in the wake screen user interface for complications (e.g., next to calendar complication 502-10).
  • the device detects and responds to user input(s) directed to the predefined area that correspond to request(s) to change a position of the complications, for example an input 580 directed to complication 502-11 to arrange the health complication 502-11 to be on the left of the calendar complication 502-10, as illustrated in Figures 5AA-5AB.
  • Figure 5AA illustrates a user input 580, such as a tap input or other selection input, selecting health complication 502-11.
  • user input 580 further indicates movement of health complication 502-11 (e.g., via a drag and/or swipe gesture) from right to left.
  • the complications are rearranged in user interface 565, with health complication 502-11 to the left of calendar complication 502-10, as illustrated in Figure 5AB.
  • Figure 5AB illustrates a user input 582 selecting a breathe complication 502- 12, associated with a breathe application for reminding a user to periodically take deep breaths, from the set of complications displayed in user interface element 570.
  • user interface element 570 is scrollable, and the device detects and responds to user input(s) directed to user interface element 570 that correspond to request(s) to scroll and view additional complications that are available to add to the wake screen user interface.
  • Figure 5AC illustrates that, in response to user input 582, the device 100 replaces display of calendar complication 502-10 with breathe complication 502-12.
  • the predefined area that displays complication has a maximum size and/or a maximum number of complications that fit within the predefined area.
  • the predefined area fits up to four complications of a first size (e.g., a single-slot size), or two complications of a second size (e.g., each having a double-slot size).
  • a first size e.g., a single-slot size
  • two complications of a second size e.g., each having a double-slot size.
  • one or more complications are automatically removed in response to a user input to add a complication to the predefined area.
  • user interface element 570 is optionally minimized or closed. In some embodiments, the device reopens user interface element 570 in response to detecting a user input tapping or otherwise selecting a complication in the editing user interface 565.
  • Figure 5 AD illustrates a user interface element 586 that is displayed in response to detecting a user input 584 at a position within the reticle 569 around the time.
  • user interface element 586 includes options for modifying properties of the time. For example, the device detects user input(s) directed to different font colors, fonts, and display styles in the user interface element 586 and modifies a text style (e.g., font, font size and/or font color) and/or modifies how the time is displayed (e.g., as a 24-hour indication or as a 12-hour indication).
  • reticle 568 around the date is optionally not displayed.
  • the device in response to a user input to change a text style, the text style of both the date and the time is updated.
  • the device detects separate user input(s) that select different text styles for the date and for the time (e.g., selecting the text style for the date while the reticle around the date is selected, and selecting the text style for the time while the reticle around the time is selected).
  • the device detects and responds to user input(s) directed to user interface element 586 that correspond to request(s) to modify a color of the time.
  • recommended colors of the time are provided in response to a tap 592 on the time element (e.g., as illustrated in Figure 5AE).
  • the recommended colors include a vibrant material, black and/or white (e.g., depending on a background color), and/or one or more colors derived from the background (e.g., a color that matches, or complements, a color in a photo used as the background).
  • the device provides one or more user interface objects in the user interface element 586 to customize a color (e.g., modifying one or more aspects (e.g., tint, tone, and/or saturation) of a currently displayed color option) and/or to select another color that is not included in the recommended colors (e.g., by selecting the color from a color wheel, and/or sampling the color from the background media item) to be applied to the time.
  • a color e.g., modifying one or more aspects (e.g., tint, tone, and/or saturation) of a currently displayed color option
  • select another color that is not included in the recommended colors e.g., by selecting the color from a color wheel, and/or sampling the color from the background media item
  • the device in response to detecting the user’s selection of a user interface object that corresponds to a user request to customize the color by sampling a portion of the background, the device ceases to display the user interface element 586 and reveals the background of the wake screen user interface.
  • the device displays a color sampling object that moves across the currently displayed background in accordance with a user’s movement input and updates the color of the color sampling object according to the color of the background at the current location of the color sampling object.
  • the user interface element 586 is redisplayed with the sampled color included in the set of available font colors, and, optionally, the selected color (and the displayed color of the time element) is matched to the sampled color obtained from the background.
  • a text style of the date and time are updated to be displayed with text style 5, as illustrated in Figure 5AE.
  • a color of the text is updated to the color selected by user input 590, as illustrated in Figure 5AE.
  • a color scheme of one or more other wake screen user interface elements e.g., one or more complications and/or a background image
  • the health complication 502-13 and the breathe complication 502-14 are displayed with a different color scheme.
  • a color of the complications is selected to match and/or complement a selected color for the text.
  • Figure 5AE further illustrates a user input 592 (e.g., a tap input or other selection input) selecting the date indication, optionally without displaying reticle 568 around the date.
  • a user interface element 594 is displayed for changing a type of information displayed for the date indication.
  • reticle 568 is optionally redisplayed to indicate that the user is modifying one or more properties of the date indication, as illustrated in Figure 5AF.
  • Figure 5AF illustrates a user input 596 selecting to modify the information displayed as the date indication to include the day of the week (e.g., instead of the day of the week and the current month and day).
  • user interface element 594 optionally includes one or more text properties (e.g., font and/or color options) that when selected, cause the device to modify the text properties of the date indication (e.g., separately from modifying the text properties of the time indication described above).
  • Figure 5AG illustrates, in response to user input 596, the indication of the date is updated to display the day of the week (e.g., “Friday”) without displaying the current month and date.
  • a user input 598 selects done option 566 to exit the editing user interface 565.
  • the device 100 in response to the user input 598 selecting done option 566, the device 100 ceases displaying the editing user interface 565, and displays a wake screen user interface 5001 having the properties, as selected by the user in the editing user interface, as illustrated in Figure 5AH.
  • Figure 5 Al illustrates a portrait-style wake screen user interface 5001 that optionally includes a plurality of complications, including health complication 502-15, weather complication 502-16, breathe complication 502-17 and sports complication 502-18, and includes a background photo with a subject (e.g., an individual).
  • a portrait-style wake screen user interface 5001 is displayed with a simulated depth effect, as described with reference to Figure 5H.
  • a wake screen user interface that includes a photo as the background image (e.g., such as the portrait-style background image illustrated in Figure 5 Al), displays the photo with a different aspect ratio than the aspect ratio of the stored photo (e.g., as stored in a photos application).
  • the device 100 automatically centers, zooms and/or crops a photo stored in the photos application to be displayed as the background of the wake screen user interface.
  • a photo-style wake screen user interface comprises a live photo that animates over time, or optionally in response to a user input on the wake screen user interface.
  • a photo-style wake screen user interface includes a plurality of interpolated frames such that the photo appears animated (e.g., to show movement of the individual displayed in wake screen user interface 5001).
  • the device 100 in response to a user input 5002 of a first type, for example tap input or a press and hold input that satisfies a threshold amount of time (e.g., at least 1 second, 3 seconds or 5 seconds) over the indication of the time and/or date (or optionally over a complication of the plurality of complications), the device 100 displays an editing user interface 565-2 for the wake screen user interface 5001.
  • a threshold amount of time e.g., at least 1 second, 3 seconds or 5 seconds
  • editing user interface 565- 2 includes a same set of features as editing user interface 565, for example reticles 568 and/or 569 for the date and/or time indications, cancel option 567, done option 566, and an indication 574 that additional views of the user interface are available.
  • a crop indication 5003 is optionally displayed to indicate that the device will resize and/or scale a background image for the wake screen user interface in response to user input(s) directed to the crop indication 5003.
  • crop indication 5003 optionally includes a textual indication such as, “pinch to crop” to inform the user that a pinch gesture, or other type of gesture, will cause the device 100 to crop the photo used as the background of wake user interface 5001.
  • Figure 5AJ further illustrates detecting a pinch gesture 5004, or other user input, for changing a crop of the background photo.
  • pinch gesture 5004 includes two points of contact that move apart and/or closer together.
  • an inward pinch gesture that moves the points of contact closer together causes a size of the subject in the background photo to decrease (e.g., as if zooming out)
  • an outward pinch gesture that moves the points of contact farther apart causes a size of the subject in the background photo to increase (e.g., as if zooming in) and/or optionally cropping the photo.
  • pinch gesture 5004 is an outward pinch gesture that zooms in the subject, as illustrated in Figure 5AK
  • pinch gesture 5006 e.g., an inward pinch gesture causes the subject to zoom out, as illustrated in Figure 5 AL.
  • the reticles 568 and/or 569 are optionally not displayed.
  • the subject in accordance with a determination that the subject of a portrait-style user interface has a size such that the subject overlaps with a portion of the time and/or date indication, the subject is optionally displayed overlaying the time and/or date indication.
  • the subject is optionally displayed as overlaying the plurality of complications.
  • the plurality of complications is optionally displayed over the subject of the portrait-style wake screen user interface.
  • Figure 5AL illustrates a user input 5008 selecting reticle 569 for the time indication.
  • the user interface element 586 for editing the time indication is displayed.
  • a user input 5010 is detected for changing a text style of the time and/or date indications to “Style 2.”
  • device 100 changes the time and optionally date indication to be displayed with text style “Style 2,” as illustrated in Figure 5AN.
  • Figure 5AN illustrates a user input 5012 corresponding to a request to minimize, or close, user interface element 586.
  • the user input 5012 comprises a swipe down gesture, or another input selecting a close option.
  • Figure 5AO illustrates the editing user interface 565-2 after minimizing, or closing, user interface element 586 in response to user input 5012.
  • a plurality of views of the wake screen user interface are available.
  • each view in the plurality of views corresponds to a different set of visual properties, optionally without changing content of the background of the wake screen user interface.
  • the plurality of views includes different patterns, or arrangements of one or more selected emoji used in the background and/or changes a color effect (e.g., monochrome, sepia, or neon) and/or a level of shadow displayed with the emoji in the background of the wake screen user interface.
  • the plurality of views include the background image with different visual effects applied (e.g., greyscale, sepia, or another filter).
  • the plurality of views include different versions of the background image, for example different phases of the moon.
  • the plurality of views available optionally include different sets of complications, different styles of text for the time and/or date indication (e.g., including a font style and/or a color), and/or different background colors for a photo (e.g., replacing a background color while maintaining a subject in a photo).
  • the plurality of views are views that are automatically, without user input, generated and/or selected by the device 100.
  • the device 100 while the device 100 displays the editing user interface 565-2, the device detects user inputs directed to a currently displayed view of the plurality of views to scroll the plurality of views in the editing user interface 565-2.
  • Figure 5AO illustrates a user input 5014-1, such as a swipe input or other input (e.g., a tap on a respective side of the user interface to scroll in the respective direction).
  • user input 5014-1 is initiated with a contact detected at a first position, and the contact moves in a first direction (e.g., from right to left or left to right).
  • the device 100 updates the view in the editing user interface to display at least a portion of a second view of the wake screen user interface.
  • a portion of the second view is gradually displayed, to appear as if the second view is sliding onto the display, while the first view is gradually removed from display, as if sliding off the display.
  • Figure 5AP illustrates a portion of the second view, sliding onto the display while the first view slides off the display.
  • one or more visual elements in the first view are optionally maintained in both the first view and the second view.
  • the one or more visual elements that are maintained optionally do not move within the user interface.
  • the one or more visual elements that are maintained correspond to visual elements of the wake screen user interface that the user has previously modified (e.g., in current editing session and/or in a previous editing session). For example, the user modified the date and time indications, and the user modified a size of the subject in the background photo. In some embodiments, the date and time indications, and the subject in the background photo, do not move while the second view slides onto the display.
  • a set of visual elements optionally that were not modified by the user in the current editing session and/or in a previous editing session, move off the display with the first view and are replaced by another set of visual elements.
  • the complications and a background color appear to slide off the display in accordance with user input 5014-2.
  • Figure 5AQ illustrates user input 5014-3 swiping in an opposite direction from the direction of user input 5014-2.
  • user input 5014-3 is optionally a continuation of user input 5014-1 and 5014-2, wherein the user input 5014-3 has maintained contact with the display and changes to swipe in another direction.
  • the portion of the second view that was sliding onto the display in Figure 5AP moves in the opposite direction, in accordance with user input 5014-3.
  • user input 5014 e.g., 5014-1, 5014-2 and 5014-3) control bringing view onto and/or removing views from the editing user interface 565-2.
  • Figure 5AR illustrates user input 5014-4, optionally as a continuation of user input 5014-3, swiping in the same direction as user inter 5014-2 in order to slide the second view of the wake screen user interface onto the display.
  • Figure 5AS as the user input 5014-5 continues to swipe from right to left, a larger portion of the second view of the wake screen user interface slides onto the display, optionally until the second view of the wake screen user interface is displayed in the editing user interface 565-2 without displaying the first view of the wake screen user interface, as illustrated in Figure 5AT.
  • Figure 5AT illustrates the second view of the wake screen user interface, wherein the complications have been changed relative to the first view of the wake screen user interface, and/or a background color of the second view has been changed.
  • the background color(s) selected for the views in a portrait-style wake screen user interface are selected based on content in the photo. For example, an original background of the photo is replaced with a single color or colored gradient, wherein the color(s) are selected to match and/or complement colors that appear in the original background of the photo and/or that appear in the subject of the portrait-style wake screen user interface.
  • one or more editing options are optionally not displayed.
  • the reticles, done option, cancel option and/or indication of additional views are optionally not displayed until a single view is displayed.
  • Figure 5AT the second view is displayed, and the first view is no longer displayed, and the one or more editing options are redisplayed.
  • a user input 5016 is detected on a complication displayed in the second view of the editing user interface 565-2.
  • user interface element 570 for changing complications is displayed, as illustrated in Figure 5 AU.
  • user input 5018 selects a health complication, and device 100 replaces display of the weather complication with the selected health complication.
  • Figure 5AV illustrates user input 5020 requesting to switch views in the editing user interface after editing the complications.
  • Figure 5AW illustrates that, in response to user input 5020, the view switches from the second view to a third view (e.g., with a different background color and/or pattern) without changing the time indication, the date indication, the subject, or the complications in accordance with each of these features having been manually edited by the user during the current (or optionally, during a previous) editing session.
  • Figure 5AW illustrates a user input 5022 selecting the “Done” option to exit out of the editing user interface.
  • Figure 5AX illustrates wake screen user interface 5024, displayed with the visual features, as selected by the user during the editing process.
  • Figure 5AX2 illustrates an editing user interface 565-3 for editing a wake screen user interface that includes a textual indication of the date, a textual indication of the time, and a plurality of complications.
  • a plurality of the complications are displayed with corresponding affordances for removing the complications. For example, a minus symbol, an “x” or another removal affordance is displayed, optionally at a corner of a complication that partially overlaps the complication; and selection of the minus symbol, “x” or other removal affordance causes the device to remove the complication from the wake screen user interface.
  • device 100 detects a user input 5080 (e.g., a tap input or other selection input) selecting complication 5089-1.
  • device 10 in response to user input 5080, device 10 displays a user interface element 5082 for changing a size of the selected complication 5089-1.
  • device 100 provides a plurality of size options for displaying the information of complication 5089-1.
  • the plurality of size options correspond to different text sizes.
  • user input 5084 is detected as selecting a first size option for the complication 5089-1, and in response to the user input, the complication 5089-1 is updated to the selected size, as illustrated in Figure 5AX4.
  • Figure 5AX3 illustrates device 100 detecting user input 5086 selecting the affordance for removing complication 5089-2.
  • the complication 5089-2 is removed from the wake screen user interface, as illustrated in Figure 5AX4.
  • Figure 5AX3 further illustrates user input 5088 corresponding to selection of the reticle for the textual indication of the date.
  • device 100 displays user interface element 5090 for changing the content that is displayed in the area above the textual indication of the time. For example, the device displays a plurality of complications in user interface element 5090 and detects user input that corresponds to a request to select a complication from the plurality of complications displayed in user interface element 5090.
  • the plurality of complications that are selectable for display above the textual indication of the time is a distinct set of complications than the set of complications that are selectable for display below the textual indication of the time (e.g., the set of complications displayed in user interface element 570, as described with reference to Figures 5AA-5AC).
  • the complications to be displayed above the textual indication of the time are text-based complications and/or have a size (e.g., height and/or length) that is distinct from the complications to be displayed below the textual indication of the time.
  • Figure 5AX4 illustrates user input 5092 selecting a complication to be displayed above the textual indication of the time.
  • the selected complication in response to user input 5092, is displayed in reticle 5094, as illustrated in Figure 5AX5, above the textual indication of the time, and replaces display of the textual indication of the date.
  • Figure 5AF more details about selection of a complication for displayed above the time and/or selecting a different set of information to display in a complication are described with reference to Figure 5AF.
  • device 100 optionally displays the content of user interface element 594 (Figure 5AF) and the content of user interface element 5090, such that the user can select the textual information and/or complications to be displayed above the indication of the time.
  • Figures 5AY-5BD illustrate another example of editing a wake screen user interface, including switching between a plurality of views for a respective wake screen user interface in the editing user interface.
  • Figure 5 AY illustrates device 100 displaying a wake screen user interface 5026 and receiving a user input 5028 selecting calendar complication.
  • the user input 5028 is a first type of input (e.g., a tap input or other selection input)
  • the device 100 displays an application user interface for the application associated with the selected complication (e.g., the calendar application).
  • the user input 5028 is a second type of input (e.g., a press and hold or another gesture)
  • the device 100 displays editing user interface 565-3 for editing the visual features of the wake screen user interface 5026.
  • Figure 5AZ illustrates user interface element 570 for editing complications is displayed in the editing user interface 565-3, in response to the user input 5028.
  • a user input 5030 for minimizing, or closing, user interface element 570 is detected, optionally without detecting a user input to change or select a complication.
  • a user input 5032 is detected as selecting a complication (e.g., a weather complication) from user interface element 570, and the device 100 updates the wake screen user interface to include the selected complication, before minimizing the user interface element 570.
  • a complication e.g., a weather complication
  • Figure 5BA illustrates the editing user interface 565-3 after the user interface element 570 has been minimized or closed.
  • a user input 5034 to change the current view of the wake screen user interface is detected.
  • user input 5034 is a swipe gesture (e.g., from left to right or from right to left).
  • the editing user interface 565-3 displays a second view of the wake screen user interface.
  • Figure 5BB illustrates the second view of the wake screen user interface.
  • the background image of the wake screen user interface updates, for example, for an astronomy wake screen user interface, different phases of the moon are included in different views of the wake screen user interface.
  • a color palette, or other virtual effect changes between different views of the wake screen user interface.
  • the second view of the wake screen user interface includes a different text style for the date and/or time indications.
  • a subset, less than all, of the complications are updated from the first view to the second view.
  • the complication that was manually selected by the user does not change between the first view and the second view, while the other complications change.
  • an animated transition is displayed while changing display from the first view to the second view, such as the sliding animation described with reference to Figure 5AP.
  • the first view of the wake screen user interface is redisplayed in the editing user interface 565-3, as illustrated in Figure 5BC.
  • the device 100 in response to detecting a user input 5038, such as a user input in a direction distinct from user input 5034, the device 100 displays a third view of the wake screen user interface, as illustrated in Figure 5BD.
  • the third view of the wake screen user interface optionally displays a different phase of the moon and/or optionally displays the same complications as the first view of the wake screen user interface.
  • the device 100 in response to user input 5040 selecting the “Done” option, the device 100 ceases to display the editing user interface 565-3 and displays the view of the wake screen user interface that is displayed while the user input 5040 is detected as the current wake screen user interface.
  • Figures 5BE-5BF illustrate an example of navigating from a wake screen user interface to a home screen user interface.
  • a user input 5042 e.g., a swipe input initiated over a home icon or other selection input, such as on a home button
  • the device 100 displays an animated transition 5043 in which the wake screen user interface is visually deemphasized (e.g., with a blur effect and/or by changing a level of translucency) as the wake screen user interface is animated as sliding off the display, optionally in an upward direction.
  • one or more application icons on the home screen user interface optionally animate as entering onto the display and/or are at least partially visible behind the visually deemphasized wake screen user interface during the animation.
  • Figures 5BG-5BH illustrate an example of navigating from a wake screen user interface to a user interface of an application.
  • the user interface that is displayed in response to a request to navigate away from the wake user interface (e.g., via user input 5042 and/or user input 5044), the user interface that is displayed is based on a user interface that was displayed before displaying the wake screen user interface (e.g., before entering a low power state and/or a locked mode). For example, if the device 100 displays an application user interface as the most recently displayed user interface prior to the device entering a low power state and/or a locked mode, in response to a user input requesting to navigate away from the wake screen user interface 5041, the previously displayed application user interface is redisplayed.
  • the device 100 displays a home screen user interface as the most recently displayed user interface prior to the device entering a low power state and/or a locked mode, in response to a user input requesting to navigate away from the wake screen user interface 5041, the previously displayed home screen user interface is redisplayed.
  • the device 100 in response to a user input 5044 (e.g., a swipe input initiated over a home icon or other selection input, such as on a home button) requesting to navigate away from wake screen user interface 5041 to a home screen user interface, the device 100 displays an animated transition 5045 in which the wake screen user interface is optionally not visually deemphasized, or is visually deemphasized in a distinct manner than the animated transition 5043 (e.g., with a different blur effect and/or with a different level of translucency), while the wake screen user interface is animated as sliding off the display, optionally in an upward direction.
  • the user interface of the application is optionally displayed on the portion of the display that is not covered by the wake screen user interface as it slides off the display.
  • Figure 5BI illustrates home screen user interface 5046.
  • home screen user interface 5046 includes one or more application icons and is displayed with a background image.
  • the background image displayed in home screen user interface 5046 is based at least in part on the wake screen user interface that the user navigated away from.
  • wake screen user interface 5041 includes a star image as the background, and the related home screen user interface 5046 also includes star images.
  • the backgrounds of the home screen and wake screen need not share content, but a respective home screen is stored in associated with a respective wake screen such that, if the current wake screen changes from a first wake screen to a second wake screen, the home screen that is displayed in response to a user input on the respective wake screen is associated with the respective wake screen.
  • Figure 5BI illustrates a user input 5048 corresponding to a request to view a coversheet user interface, which optionally includes the same visual features as the wake screen user interface.
  • a user input 5050 is detected on a power button (e.g., push button 206) that causes the device 100 to enter a low-power state and/or a locked mode.
  • the wake screen user interface 5041 is optionally redisplayed.
  • the wake screen user interface upon waking the device 100 from the low-power state and/or locked mode, is updated to a next wake screen user interface in the stored set of wake screen user interfaces (e.g., wherein the device 100 cycles through the stored set of wake screen user interfaces, optionally periodically and/or upon waking).
  • Figure 5BJ illustrates a user input 5051, optionally a swipe user input in a first direction (e.g., from left to right or from right to left).
  • user input 5051 is detected within a predefined area of the wake screen user interface 5041.
  • an input within the predefined area of the wake screen user interface 5041 causes the device 100 to change to another wake screen user interface.
  • the wake screen user interface is updated to display wake screen user interface 5052.
  • the device 100 displays a home screen user interface 5056 that is related to the wake screen user interface 5052.
  • Figure 5BL illustrates user interactions with the home screen user interface in manners similarly to those described with respect to the home screen user interface 5046 ( Figure 5BI).
  • user input 5058 causes the device 100 to display a coversheet user interface (e.g., having the visual properties of wake screen user interface 5052) and user input 5060 causes the device 100 to enter a lower-power state and/or a locked mode.
  • coversheet user interface e.g., having the visual properties of wake screen user interface 5052
  • user input 5060 causes the device 100 to enter a lower-power state and/or a locked mode.
  • Figures 5BM-5BN illustrates an example of a user input selecting a complication displayed on a wake screen user interface 5062.
  • user input 5064 is a tap input or other selection input on a breathe complication, and causes the device 100 to display an application user interface for the breathe application associated with the breathe complication.
  • user input 5066 e.g., a tap input or other selection input
  • the device 100 displays an application user interface 5068 for the weather application associated with the weather complication, as illustrated in Figure 5BN.
  • Figures 5BO-5BP illustrate an example of an animated transition while the device 100 transitions from the low power state to displaying the wake screen user interface 5070-2.
  • user interface 5070-1 is displayed in an always-on (or dimmed) display (e.g., in the low power state).
  • user interface 5070-1 includes a plurality of colored stripes (or lines) that are animated as the device transitions between different user interfaces.
  • the colored stripes are representations of ribbons, ropes, strings, or fabrics.
  • each stripe is a different color (e.g., like a rainbow, a pride flag, or another set of stripes).
  • the colored stripes are arranged as if wrapped around an object, such as a cylinder.
  • the animated transition includes increasing a size of the stripes and/or decreasing a distance between the stripes. For example, in Figure 5BO, each of the colored stripes are of a first size and there is space between the stripes.
  • the stripes are animated as coming together (e.g., such that there is no space, or less space, between the stripes) and the size of the stripes increases to a second size greater than the first size.
  • Figures 5BQ-5BS illustrate an example of animating the background, including the colored stripes, in response to one or more user inputs.
  • the device 100 detects user input 5072 (e.g., a swipe input or other input in a first direction).
  • the colored stripes are animated as being pulled in a same direction (or opposite direction) as the user input 5072.
  • the animation includes decreasing a size of one or more of the colored stripes and/or increasing a distance between at least two of the colored stripes.
  • a top-most stripe is animated as separating from, and increasing a distance from, its neighboring stripe. For example, first the top-most stripe separates from the set of stripes, then a second top-most stripe separates from the remaining stripes, optionally until there is space between each of the stripes in the set of stripes. For example, in Figure 5BR, the top three stripes are illustrated as being separated, as indicated by the black space, from the remaining lower three stripes in the set of stripes.
  • the animation is displayed in accordance with the user input 5072.
  • user input 5074 is detected as being in an opposite direction as user input 5072.
  • user input 5074 is a swipe input in a downward direction
  • user input 5072 is a swipe input in an upward direction.
  • the wake screen user interface 5070-3 is not dismissed, and in response to user input 5074, the animation reverses, as illustrated in Figure 5BS.
  • the stripes have increased in size, relative to the stripes in Figure 5BR, and the top stripes have shifted back down to remove the distance between the set of stripes.
  • the top-most stripe is the last stripe to rejoin the set of stripes.
  • the user input is no longer detected, for example in response to a lift off 5076 of the user input.
  • the reverse animation continues until the wake screen user interface 5078 is back to its original state (e.g., the state illustrated in Figure 5BQ before detecting user input 5072), as illustrated in Figure 5BT.
  • Figure 5BU illustrates a user input 5080-1, such as a swipe input in the upward direction.
  • user input 5080-1 satisfies dismiss criteria for dismissing the wake screen user interface 5070-7.
  • the dismiss criteria is satisfied in accordance with a determination that user input 5080-1 satisfies a threshold swipe velocity and/or a threshold swipe distance.
  • the dismiss criteria is satisfied in accordance with user input 5080-2 (e.g., a continuation of input 5080-1) continuing to be detected as the user continues the swipe input upward.
  • the device 100 displays the animation described with reference to Figure 5BR, including separating the colored stripes and/or decreasing a size of the colored stripes, as shown in Figure 5BV.
  • the device 100 displays a complete animation as the device 100 transitions from displaying the wake screen user interface 5070-7 to displaying home screen user interface 5070-10.
  • an intermediary user interface 5070-9 is displayed that includes the colored stripes moving up and twisting (e.g., around an invisible object, such as a cylinder) while the colored stripes are displayed as separated, as indicated by the black space between the stripes, and/or are displayed with a smaller width that the stripes displayed on the wake screen user interface 5070-7.
  • the stripes are displayed with the smaller width, as illustrated in Figure 5BX, optionally the most narrow width displayed during the animation, as the background of the home screen user interface 5070-10, which also includes one or more application icons in the foreground.
  • the stripes displayed in the background of the home screen user interface 5070-10 are optionally displayed with a same width as the stripes displayed in the background of the wake screen user interface 5070-7 (Figure 5BU) and optionally displayed as shifted upward from the stripes displayed in the wake screen user interface 5070-7.
  • the stripes are displayed as a same size but in a different position as the background of the home screen user interface 5070-10 as compared to the stripes of the background of the wake screen user interface 5070-7.
  • Figures 5BY and 5BZ illustrate examples of changing an orientation of the stripes of the wake screen user interface 5070-11 to wake screen user interface 5070-12.
  • a direction in which the stripes are wrapped around the invisible objects is updated to be in an opposite direction (e.g., clockwise or counter-clockwise).
  • the orientation of the stripes of the wake screen user interface 5070 is selected in accordance with a user input. For example, an option is provided to the user to select an orientation of the stripes to be used as the background of the wake screen user interface.
  • the orientation of the stripes is automatically determined, without user input.
  • the orientation of the stripes is alternated periodically, for example, each time the wake screen user interface is invoked, the orientation of the stripes changes from the previous orientation that was displayed the previous time the wake screen user interface was invoked. For example, upon a lock and unlock cycle and/or upon dismissing a low-power state, the wake screen user interface alternates the orientation of the stripes.
  • Figure 5CA-5CB illustrate an example of changing a background color of the wake screen user interface 5070-13 to 5070-14.
  • the background color of the wake screen user interface is automatically selected in accordance with a current mode of operation of the device 100. For example, while the device 100 is in a first mode of operation, such as a light mode, the background is selected for the light mode of operation (e.g., a white background is selected), as illustrated in Figure 5CA. In some embodiments, while the device 100 is in a second mode of operation, such as a dark mode, the background is selected for the dark mode of operation (e.g., a black background is selected), as illustrated in Figure 5CB.
  • a first mode of operation such as a light mode
  • the background is selected for the light mode of operation (e.g., a white background is selected), as illustrated in Figure 5CA.
  • a second mode of operation such as a dark mode
  • the background is selected for the dark mode of operation (e.g., a black background is selected), as illustrated
  • the current mode of operation is selected in accordance with a time of day (e.g., light mode in the daytime and/or between a predefined time period (e.g., from 6 AM to 8 PM or another time period) and dark mode in the nighttime and/or outside of the predefined time period (e.g., from 8PM to 6AM)).
  • a time of day e.g., light mode in the daytime and/or between a predefined time period (e.g., from 6 AM to 8 PM or another time period) and dark mode in the nighttime and/or outside of the predefined time period (e.g., from 8PM to 6AM)
  • the light mode and the dark mode have respective sets of visual characteristics that either increase or decrease the overall luminance of the user interface displayed by the display generation component.
  • user interface elements and/or background platters of user interface elements displayed in a user interface are made darker (e.g., reduced in luminance) than their counterparts displayed in the light mode, to reduce the average luminance of the user interface (e.g., to suit a particular time of day, function, and/or mood).
  • text shown in a dark color or black on a light background in the light mode is optionally converted to text shown in a light color or white on a dark background, when the device switches from the light mode to the dark mode.
  • text is shown in a dark color or black on a light background in the dark mode is optionally converted to text shown in a light color or white on a dark background, when the device switches from the dark mode to the light mode.
  • the user is provided with an option to manually select the background color.
  • Figure 6A illustrates detecting a user input 602 (e.g., a tap input, a long press input, or another selection input) on wake screen user interface 600.
  • user input 602 is detected at a portion of the wake screen user interface 600 that does not correspond to the date and/or time indication, a complication, or a flashlight and/or camera shortcut.
  • the device 100 in accordance with a determination that device 100 is in a locked mode (e.g., optionally indicated by the lock indicator above the date), the device 100 displays user interface 604 ( Figure 6B) for entering a passcode, or otherwise prompts the user to authenticate before allowing the user to access applications or data stored on device 100.
  • a locked mode e.g., optionally indicated by the lock indicator above the date
  • the device 100 displays user interface 604 ( Figure 6B) for entering a passcode, or otherwise prompts the user to authenticate before allowing the user to access applications or data stored on device 100.
  • the user authenticates using a passcode entered in user interface 604, as illustrated in Figure 6C.
  • the device 100 displays the expanded face switcher user interface 606, as illustrated in Figure 6D.
  • the expanded face switcher user interface e.g., user interface 561
  • the expanded face switcher user interface includes a plurality of representations of wake screen user interfaces (e.g., representation of wake screen user interface 611), with the centered wake screen user interface 615 being concurrently displayed with a portion of a related home screen user interface.
  • Figure 6D illustrates a plurality of user interactions with expanded face switcher user interface 606.
  • user input 608 selecting the plus button causes the device 100 to display a user interface (e.g., an editing user interface) for creating a new wake screen user interface to add to the set of wake screen user interfaces.
  • user input 610 selecting a portion of a representation of a second wake screen user interface 611 causes the device 100 to display an editing user interface for the second wake screen user interface corresponding to the representation of the second wake screen user interface 611.
  • a user input selecting user interface element 609a (“Focus”) enables the user to set a restricted notification mode (e.g., in which certain types of notifications are suppressed and/or delayed) for the wake screen corresponding to representation 615 of a wake screen.
  • a user input selecting option 609b enables a user to view a gallery view (e.g., user interface 652, Figure 6K) for creating a new wake screen user interface to add to the set of user interfaces.
  • user input 614 such as a swipe input (e.g., from right to left), causes the device to bring the representation of the second wake screen user interface 611 to the center of the expanded face switcher user interface 606, as illustrated in Figure 6F.
  • the device detects and responses to user inputs directed to the expanded face switcher user interface 606 that correspond to requests to rearrange an order of the set of wake screen user interfaces in the expanded face switcher user interface 606. For example, in response to a long press user input, a touch-hold and drag input, or another type of user input directed to a respective representation of a wake screen user interface in the expanded face switcher user interface 606, the device selects and optionally drag the respective representation of the wake screen user interface to the left and/or right of the other representations of wake screen user interfaces to change the order in which the device 100 cycles through the set of wake screen user interface.
  • one or more wake screen user interfaces in the set of wake screen user interfaces are associated with a respective type, or theme, of wake screen user interface.
  • a respective wake screen user interface is identified as a photostyle, an emoji-style, a portrait-style, or another style that is optionally pre-generated.
  • the device 100 optionally generates one or more themes, or styles, for wake screen user interfaces, such as a smart album that rotates through images and/or photos stored on device 100, a weather-style that includes a representation of a weather forecast at a current location of device 100, an astronomy, globe, or other celestial body style, and/or a style that represents a lifestyle or other event (e.g., Pride, Women’s History Month, or other event).
  • Examples of wake user interfaces that have a theme, or style, and that are generated automatically by device 100, optionally without user input, are illustrated in Figure 6K.
  • user input 612 such as a swipe gesture (e.g., upward or downward) causes the device 100 to provide the user with an option to delete the centered wake screen user interface 615.
  • the device 100 displays a delete icon 616.
  • the device 100 in response to the user input 612 continuing (e.g., the swipe gesture continues in a same direction), the device 100 optionally displays the representation of the wake screen user interface 615 sliding off the display and removes the wake screen user interface 600 from the set of wake screen user interfaces.
  • a second user input 618 selects the delete icon 616, and in response to user input 618, the device 100 removes the wake screen user interface 600 from the set of wake screen user interfaces and ceases display of the representation of the wake screen user interface 615.
  • Figure 6F illustrates the representation of the second wake screen user interface 611 centered on the display (e.g., in response to user input 614).
  • a portion, less than all, of a representation 620 of an associated home screen user interface is optionally displayed as partially occluded behind the representation of second wake screen user interface 611.
  • representation of the second wake screen user interface 611 is concurrently displayed with a portion, less than all, of two or more other representations of wake screen user interfaces, for example representation 615 of a wake screen user interface on the left and representation 613 of a wake screen user interface on the right.
  • Figure 6F illustrates a plurality of user inputs that enable the user to interact with the expanded face switcher user interface 606.
  • the device 100 in response to user input 608-1 on the “+” button, the device 100 initiates a process for adding (optionally including designing and/or editing) a new wake screen user interface to be added to the set of wake screen user interfaces, as illustrated in Figure 6K.
  • the device in response to user input 610-1 (e.g., a tap input or other selection input), the device brings the representation 613 of a wake screen user interface to the center of the expanded face switcher user interface.
  • user input 610-1 corresponds to a swipe input that causes the device 100 to scroll through the available representations of wake screen user interfaces that are in the set of wake screen user interface.
  • the device 100 in response to user input 622 selecting the “Customize” button, displays the editing user interface 626 ( Figures 6G-6H) for the representation of the wake screen that is currently in the center of the expanded face switcher user interface 606 when the user input 622 is detected.
  • the expanded face switcher user interface displays the representation of the home screen user interface in a center region of the user interface 638, as illustrated in Figure 61.
  • Figures 6G-6H illustrate an editing user interface 626 for a wake screen user interface corresponding to representation 611 of a wake screen user interface.
  • editing user interface 626 is optionally enabled with one or more of the same functionalities described with respect to editing user interface 565 ( Figure 5X).
  • editing user interface 626 includes user-selectable options for changing a font color of the user interface.
  • editing user interface 626 includes a user- selectable option for applying a visual filter (e.g., applying a color filter, a blur, a transparency filter and/or a gradient filter) to the wake screen user interface.
  • a visual filter e.g., applying a color filter, a blur, a transparency filter and/or a gradient filter
  • a font color of the date and/or time indication is updated to the selected first color, as illustrated in Figure 6H, where the date and/or time indication is displayed with a white font color.
  • a user interface element 630 in response a user input 628 and/or in response to a user input on the “Font Color” button in Figure 6G, a user interface element 630, as illustrated in Figure 6H is displayed.
  • the background of the wake screen user interface is an emoji-style background (e.g., a smiley face emoji), in which one or more emojis are arranged in a pattern (e.g., a geometric pattern).
  • different views of the emoji user interface include changing a size and/or arrangement (e.g., pattern) of the emojis in the background of the wake screen user interface.
  • an option 628b for selecting additional and/or alternative emojis is displayed in editing user interface 626.
  • the device displays a number of slots for chosen emojis to allow the user to select from the emoji keyboard 629 up to a threshold number of emojis (e.g., 3 or 4 emojis), and the device displays the selected emojis in a predefined pattern (e.g., a grid pattern, a swirl pattern, or another pattern) in the background of the wake screen user interface.
  • a threshold number of emojis e.g., 3 or 4 emojis
  • a predefined pattern e.g., a grid pattern, a swirl pattern, or another pattern
  • user input 627 is detected as selecting option 628b to select an emoji, and in response to user input 627, an emoji keyboard 629 is displayed in the editing user interface 626, as illustrated in Figure 6H2.
  • emoji keyboard 629 includes a user interface element that displays the currently selected emoji that are used in the pattern. For example, in Figure 6H2, a smiley face emoji is selected, and the grid pattern includes the smiley face emoji.
  • the device 100 while displaying the editing user interface 626 with the emoji-style background, the device 100 detects a user input 631-1, such as a swipe input in a first direction (e.g., from right to left).
  • the device 100 in response to user input 631-1, updates the editing user interface 626 to display the emoji-style background with a different pattern (and, optionally, with the same emoji(s) arranged in a different pattern).
  • the emoji are arranged in a first pattern (e.g., a grid pattern) in Figure 6H2 and are arranged in a second pattern (e.g., a swirl pattern) in Figure 6H3.
  • Figure 6H3 illustrates a user input 633-1 selecting a first emoji (e.g., a thumbs- up emoji) and a user input 633-2 selecting a second emoji (e.g., a heart emoji).
  • a first emoji e.g., a thumbs- up emoji
  • a second emoji e.g., a heart emoji
  • the emoji displayed in the background are updated in accordance with the user selection.
  • an order in which the emoji are selected is the order the emoji are shown (e.g., and alternated in the pattern of the background).
  • the thumbs-up emoji and the heart emoji are added after the smiley face emoji, and the pattern displays the selected emoji in an alternated manner.
  • the swirl pattern in Figure 6H4 includes the smiley face emoji, the thumbs-up emoji, and the heart emoji.
  • the device sets an upper limit on the number of emojis that a user is allowed to select (e.g., 3 emoji, 5 emoji, or 10 emoji) to include in a same background, e.g., by showing the threshold number of input slots at the top of the emoji keyboard 629.
  • emojis are optionally selected by the user to generate the emoji pattern of the wake screen user interface, and/or a large variety of different patterns is made available for user selection to arrange the selected emojis in the background of the wake screen user interface.
  • the set of emoji patterns made available for user selection is automatically updated in response to and in accordance with the set of emojis that have been selected by the user for inclusion in the background of the wake screen user interface.
  • the device optionally displays a first set of available patterns when a first set of emojis have been selected by the user, and the device optionally displays a second set of available patterns different from the first set of available patterns when a second set of emojis have been selected by the user.
  • the device in response to one or more inputs directed to the emoji keyboard (e.g., swipe inputs, and/or a tap input on a category symbol for a category of emojis), the device scrolls or replaces the emojis currently displayed in the emoji keyboard 629 to show additional emojis that are available for selection.
  • the device 100 in response to user input 631-2 (e.g., a swipe input), updates the background to display a third pattern that is distinct from the first pattern (e.g., grid pattern) and second pattern (e.g., swirl pattern).
  • a third pattern that is distinct from the first pattern (e.g., grid pattern) and second pattern (e.g., swirl pattern).
  • the selected emoji including the smiley face emoji, the thumbs-up emoji, and the heart emoji, are arranged in a geometric pattern that includes diagonal arrangement of the emojis.
  • Figure 6H displays user interface element 630 for viewing and/or selecting a font color.
  • user interface element 630 includes additional options than the color options displayed in user interface 626 in Figure 6G.
  • the device selects and/or changes a font color for the system generated text (e.g., the date and time elements) in response to detecting the user selecting a color sample in the list of color samples displayed in user interface element 630. For example, in response to user input 632, the device 100 updates the color to the color of the sample color selected by user input 632, and in response to user input 634, the device 100 updates the color of the font to the sample color selected by user input 634.
  • the device 100 in response to user input 636 on a “Done” button, the device 100 exits the editing user interface 626 and optionally returns to the expanded face switcher user interface 606, or displays the wake screen user interface (e.g., with the edited wake screen user interface as the current wake screen user interface).
  • Figure 6H6 illustrates that, in response to a user input 631-3 selecting the Done affordance, device 100 displays respective user-selectable options to also use the newly edited background of the wake screen user interface as a background a home screen user interface associated with the wake screen user interface (e.g., “Use as Home Screen”) or to customize the home screen user interface (e.g., to modify one or more aspects of the home screen user interface and/or use a distinct background for the home screen user interface from the wake screen user interface).
  • a home screen user interface associated with the wake screen user interface e.g., “Use as Home Screen”
  • customize the home screen user interface e.g., to modify one or more aspects of the home screen user interface and/or use a distinct background for the home screen user interface from the wake screen user interface.
  • device 100 sets the emoji user interface that includes the smiley face emoji, thumbs up emoji, and heart emoji arranged in the pattern illustrated in Figure 6H5 as the background for the home screen user interface and displays representations of a plurality of applications overlaying the background of the home screen user interface.
  • the device displays the wake screen user interface and the home screen user interface respectively with the same sets of selected emojis arranged in the same or substantially similar patterns in their backgrounds.
  • the device 100 in response to user input 635-1, exits the editing user interface(s) and displays a home screen user interface and/or a wake screen user interface with the newly edited background.
  • the device 100 in response to user input 635-2 selecting a second affordance to customize the home screen user interface (e.g., “Customize the Home Screen”), the device 100 does not use the newly edited background of the currently selected wake screen user interface as the background of the home screen user interface associated with this wake screen user interface right away; instead the device provides an opportunity for the user to further customize one or more aspects of the home screen user interface, including but not limited to the background of the home screen user interface.
  • the device 100 in response to a user request to customize the home screen user interface, displays an editing user interface for the home screen user interface (e.g., analogous to the editing user interface 642 illustrated in Figure 6 J) for the user to customize the home screen user interface (e.g., without first navigating to the user interface 638 in Figure 61, and without requiring the user to select “Customize” button in user interface 638).
  • an editing user interface for the home screen user interface e.g., analogous to the editing user interface 642 illustrated in Figure 6 J
  • the user to customize the home screen user interface e.g., without first navigating to the user interface 638 in Figure 61, and without requiring the user to select “Customize” button in user interface 638.
  • the editing user interface for the home screen user interface is initially displayed having the properties of the newly edited background of the currently selected wake screen user interface (e.g., having the same set of selected emojis, and/or having the same or substantially the same emoji pattern as the background of the wake screen user interface), such that the customization of the home screen user interface uses the newly edited background of the wake screen user interface as a starting point.
  • the device provides one or more selectable options for editing one or more aspects of the background of the home screen user interface, such as the background media item, filters, emoji patterns, selected emojis, colors, font colors, fonts, legibility blur of the background of the home screen user interface.
  • the editing user interface 642 of the home screen user interface optionally changes the visual properties of the application icons and/or widgets (e.g., making them more translucent, dimmed, and/or less saturated) to indicate that the editing user interface does not support selection and arrangement of the application icons and/or widgets on the home screen user interface.
  • representations of application icons and/or widgets are displayed in the editing user interface 642 in lieu of the application icons and widgets actually present in the home screen user interface.
  • Figure 61 illustrates the representation of the home screen user interface in a center region of the expanded face switcher user interface 638 (e.g., in response to user input 624).
  • the device initiates a process to edit, or customize the home screen user interface in response to a user input selecting, via user input 640, the “Customize” button in Figure 61.
  • Figure 6 J illustrates editing user interface 642 for editing the home screen user interface.
  • user input 644 selects an option to change a color and/or a gradient of a background of the home screen user interface.
  • a user interface 668 ( Figure 6N) is displayed, wherein the device change a photo to include in the home screen user interface in response to a user input selecting the photo from the user interface 668.
  • the editing user interface 642 illustrated in Figure 6J is displayed in response to a user’s request to further edit the home screen user interface after selecting the Done affordance in the editing user interface for customizing the wake screen user interface (e.g., in response to user input 635-2 selecting the Customize the Home Screen affordance, as described with reference to Figure 6H6).
  • the editing user interface 642 includes the newly edited background of the wake screen user interface if the editing user interface 642 was displayed in response to an input selecting the “Customize the Home Screen Affordance” after the end of the wake screen editing process.
  • the editing user interface 642 includes a default background that corresponds to a background of the currently selected wake screen user interface in the user interface 638, if the editing user interface 642 was displayed in response to an input selecting the “Customize” affordance in the user interface 638.
  • one or more of the application icons and/or widgets displayed in the home screen user interface that is being edited in Figure 6J are replaced with translucent objects, optionally that do not include indications of applications associated with the objects.
  • the application icons and/or widgets are optionally displayed in a default arrangement (e.g., the application icons and/or widgets optionally do not correspond to an arrangement of the home screen user interface that is currently set for the device 100).
  • the application icons and/or widgets are replaced with translucent shapes that correspond to the shapes of icons and/or widgets but maintain a same arrangement (e.g., in size and/or position) as the arrangement of application icons and/or widgets in the home screen user interface that is currently set for the device 100.
  • the device toggles “legibility blur” on and/or off in response to user inputs (e.g., tap inputs) directed to the legibility blur toggle control shown in the editing user interface 642.
  • legibility blur when activated, provides a visual deemphasis (e.g., a blurred effect) on the background, such that text, for example textual labels that are optionally displayed with the icons for applications, are more easily read (e.g., legible) when displayed on top of the background.
  • the editing user interface 642 includes an indication 647 that additional views of the home screen user interface are available.
  • the additional views correspond to one or more different visual effects applied to the background, including changing a color, a tone, or changing another visual effect of the background.
  • the additional views include one or more different patterns, or arrangements, of the background of the home screen user interface. For example, the smiley face emoji are rearranged in a different pattern and/or displayed at a different size in one or more additional views.
  • user input 648 corresponds to a request to change the currently displayed view of the home screen user interface, and in response to user input 648, the device 100 displays a different view (e.g., changes the background) of the home screen user interface.
  • the device 100 in response to user input 650 selecting the “Done” button, the device 100 exits out of the editing user interface 642 for the home screen user interface.
  • Figure 6K illustrates a user interface 652 for creating a new wake screen user interface (also referred to herein as a face) to add to the set of wake screen user interfaces.
  • user interface 652 includes a plurality of options for creating a new face that is based on a portrait (e.g., a photo that includes a main subject or individual), based on a photo (optionally selected from a photos library stored on device 100), and/or based on one or more emoji (optionally up to a threshold number of emoji may be included in the face).
  • a portrait e.g., a photo that includes a main subject or individual
  • a photo optionally selected from a photos library stored on device 100
  • emoji optionally up to a threshold number of emoji may be included in the face.
  • the device 100 in response to user input 656 selecting to create a new face based on a photo, displays a photo picker user interface 668 (Figure 6N), for the user to select a photo to be used as the background of the new face, and enables the user to, after selecting the photo, modify other features of the face, including font features.
  • a photo picker user interface 668 Figure 6N
  • user interface 652 includes one or more “Featured Faces” that correspond to wake screen user interfaces that have been automatically generated (e.g., by device 100), without user input.
  • the Featured Faces optionally includes a “Smart Album” that identifies a plurality of images (e.g., photos) to include in the wake screen user interface, and optionally rotates through the plurality of images while the “Smart Album” face is selected as the wake screen user interface.
  • the device 100 enables the user to select a subset of individuals, pets, locations, and/or photo albums to include in a “Smart Album” wake screen user interface.
  • the device 100 provides the user (e.g., in an editing user interface) an option to set a frequency of changing between respective photos in the “Smart Album” to be used as the current wake screen user interface.
  • device 100 in response to user input 6002 selecting the “Smart Album” face, displays user interface 6004 ( Figure 6U) for the user to select and/or deselect categories to include in the Smart Album and optionally to set a frequency of changing between the photos.
  • the Featured Faces includes an emoji user interface, which includes a preselected emoji and/or pattern of emoji.
  • the Featured Faces includes a weather user interface that creates a visual effect that corresponds to a current weather forecast, a globe user interface that includes an image of a glob, an astronomy user interface that includes one or more celestial bodies, and/or a Pride user interface that includes a symbol representing Pride.
  • each of the automatically generated user interfaces optionally includes a set of complications that are automatically selected and included in the respective wake screen user interface.
  • the device 100 displays an editing user interface 660 for editing the emoji user interface having the set of properties generated by device 100, as illustrated in Figure 6L.
  • the complications, date indication, and font properties of the time and date are selected by the device 100 without user input.
  • the device provides one or more selectable options that allow the user to edit the emoji user interface by change one or more of the properties in the set of properties (e.g., in manners as described with respect to Figures 6G-6H5).
  • the device provides one or more selectable options to change the complications and/or change a font of the date and/or time indication (e.g., in manners as described with respect to Figures 5X-5AB and Figures 5AC-5AH).
  • the device 100 in response to user input 662 selecting the “Done” button, the device 100 ceases display of the editing user interface and displays the emoji user interface as the current wake screen user interface, and optionally adds the emoji user interface to the set of wake screen user interfaces.
  • the device 100 in response to user input 658 (Figure 6K) selecting the weather user interface, displays an editing user interface 664 for editing the weather user interface.
  • the weather user interface is initially displayed with the set of properties as generated by device 100, and the device provides one or more selectable options to edit the weather user interface to change one or more properties in the set of properties (e.g., in manners as described with respect to Figures 6G-6H5, Figures 5X- 5AB, and Figures 5AC-5AH).
  • the device 100 In response to user input 666 selecting the “Done” button, the device 100 ceases display of the editing user interface and displays the weather user interface as the current wake screen user interface, and optionally adds the weather user interface to the set of wake screen user interfaces.
  • Figure 6N illustrates photo picker user interface 668 for selecting a photo to use as a background image for the wake screen user interface.
  • the device 100 in response to user input 670 selecting a photo 672, displays user interface 674 ( Figure 60) for viewing a preview of the photo 672 as the background of the wake screen user interface.
  • the device provides user interface elements and selectable options (e.g., in an editing user interface 678-a in Figure 602) that allow the user to further edit, or customize, the wake screen user interface in response to detecting a user input selecting the “Customize” button via user input 676 in Figure 60.
  • the device 100 displays editing user interface 678-a, including displaying at least one reticle (e.g., a reticle 572 for editing one or more complications, and optionally a reticle for the time and/or date indication) behind, or at least partially occluded by, a portion of the background image, such as the subject of the background image (e.g., the airplanes).
  • the reticles in response to user input 682-a directed to a reticle 569 for editing the time indication, the reticles, including reticle 572, reticle 569 and reticle 568, are displayed in front of the subject of the background image, as illustrated in Figure 6P.
  • Figure 6P illustrates editing user interface 678 for editing the wake screen user interface with the photo 672 as the background, in response to user input 676.
  • the device edits the wake screen user interface in response to a user’s request to adding one or more complications to the wake screen user interface, as illustrated in Figures 6P-6Q.
  • the device 100 displays user interface element 680 for adding complications as shown in Figure 6Q.
  • Figure 6Q illustrates user input 684 selecting a set of complications 686 to add to the wake screen user interface.
  • the device 100 displays the set of complications 686 under the time indication ( Figure 6R).
  • the device detects a respective user input directed to the set of recommended complications 686 displayed below the time indication and moves the set of complications 686, for example to above and/or to the side of the date and/or time indication, in response to the respective user input.
  • Figure 6Q further illustrates user input 685 selecting a reticle for the time indication, and in response to user input 685, the device 100 displays user interface element 687 for modifying one or more properties (e.g., color and/or font style) of the time and/or date indication.
  • properties e.g., color and/or font style
  • Figure 6R illustrates user input 688 selecting a different style for the font of the time.
  • the device 100 updates the time and/or date indication to be displayed with “Style 2.”
  • Figure 6R further illustrates user input 690 selecting the “Done” button, which causes the device 100 to exit the editing user interface 678, and display the user interface with the photo 672 as the background as the current user interface, and optionally adds the new user interface with photo 672 as the background to the set of wake screen user interfaces.
  • Figures 6S-6T illustrate examples of portrait-style wake screen user interfaces.
  • the portrait-style wake screen user interface 692 and the portrait-style wake screen user interface 694 are included in the set of wake screen user interfaces that device 100 automatically cycles through over time.
  • the device 100 automatically changes the wake screen user interface from a first wake screen user interface to a second wake screen user interface from the set of wake screen user interfaces, optionally without detecting user input.
  • the device 100 automatically changes the wake screen user interface periodically (e.g., every 5 minutes, every day, every week).
  • the device 100 automatically changes the wake screen user interfaces in accordance with a determination that the device 100 has entered a locked state and/or a low-power state. For example, upon bringing the device 100 out of a low-power state, the wake screen user interface is updated to a next wake screen user interface in the set of wake screen user interfaces.
  • Figure 6T illustrates user input 696 requesting to dismiss the wake screen user interface 694, and, in response to user input 696, the device 100 displays a home screen user interface that is related to (or otherwise associated with) the wake screen user interface 694.
  • Figure 6U illustrates a user interface 6004 for selecting one or more categories for photos to be included in a Smart Album, and an option for selecting a frequency for changing the background while a Smart Album is selected as the background of the wake screen user interface.
  • the Smart Album identifies a set of images, or photos, optionally that are stored on device 100 or otherwise associated with the user of device 100, and periodically rotates through the images as the wake screen user interface.
  • the visual properties e.g., complications and textual indications of time and/or date
  • the frequency of how often the images are refreshed, or rotated, is selected by the user.
  • user interface 6004 provides an option to select, or deselect, one or more categories of photos, including a people category (e.g., corresponding to media items with a person or optionally a person’s face as prominent subject(s)), a pets category (e.g., corresponding to media items with one or more animals or specific types of animals such as cats and dogs as prominent subject(s)), a nature category (e.g., corresponding to media items with one or more items found in nature as prominent subject(s)), and/or an urban category (e.g., corresponding to media items with one or more items found in an urban setting as prominent subject(s)).
  • photos are determined as belonging to a respective category based on the content of the photo, which is optionally automatically identified by device 100 without a user manually tagging the photo with a category.
  • the user interface 6004 includes representations of wake screen user interfaces that are included in the Smart Album, including representation 6003-1.
  • the device displays selectable representations that corresponds to different categories of background media items, and the device updates a select/unselected state of a respective selectable representation in accordance with a user input directed to the respective representation. For example in Figure 6U, user input 6006 selects the check mark to unselect the category of “Nature.”
  • the Smart Album in accordance with a determination that the category “Nature” is deselected, the Smart Album does not include photos that have been identified as having content related to nature (e.g., trees, flowers, sunsets, or other landscapes).
  • the categories that are currently selected are indicated by a check mark next to the category.
  • the representation 6003-1 that includes a flower image is no longer included in the Smart Album, and device 100 optionally ceases to display the representation 6003-1, as illustrated in user interface 6008 in Figure 6V.
  • the device detects and/or processes media items such as photos and videos stored in a user’s media library and/or online account to determine whether they should be included in the “Nature” category.
  • the device detects and/or processes icons and/or stock images (e.g., stored on the device or online) in addition to media items stored in the user’s media library and/or online account.
  • the device includes one or more categories in the set of user-selected categories in accordance with user input 6010 directed to one or more currently unselected categories, such as the category “Urban”.
  • the device 100 displays a representation 6003-2 of a wake screen user interface that is included in the urban category (e.g., an image that includes a house or building).
  • images, or photos that include content related to a city (e.g., skyscrapers, buildings, skyline views, cityscapes, billboards, or other urban images).
  • the device ceases to display the user interface 6008 in response to detecting user input 6012 (e.g., a swipe down in Figure 6V) that corresponds to a request to close, or otherwise cease display of, the category selection for the Smart Album.
  • user input 6012 e.g., a swipe down in Figure 6V
  • the device 100 in response to user input 6012, returns to the user interface displayed in Figure 6K.
  • Figure 6W illustrates user input 6016 selecting a “Choose...” option that enables the user to select which people to include in the Smart Album.
  • the device 100 displays a user interface 6018 ( Figure 6X) for selecting one or more people to include in the Smart Album.
  • the one or more people displayed in user interface 6018 are automatically selected based on a media library of device 100 (e.g., a private or personal media library associated with a user of the device, including photos and/or videos taken by the user (e.g., with the device or another device different from the device)).
  • a media library of device 100 e.g., a private or personal media library associated with a user of the device, including photos and/or videos taken by the user (e.g., with the device or another device different from the device).
  • Figure 6X illustrates a plurality of user inputs selecting one or more people to include in the Smart Album.
  • user input 6020 e.g., a tap input or other selection input
  • user input 6022 selects representation 6026-3 of a second person
  • user input 6024 selects representation 6026-1 of a third person.
  • one or more photos that include the respective person are included in the set of photos for the Smart Album.
  • an option for viewing and selecting additional people that are not displayed in user interface 6018 is provided (e.g., “More...”).
  • the representations illustrated in user interface 6028 include representations of a wake screen user interface that includes a background of a respective selected person, as illustrated in Figure 6Y.
  • representation 6003-3 includes a representation of the person associated with representation 6026-2.
  • Figure 6Y illustrates user input 6030 detected on an option to select a frequency of changing the photos in the Smart Album.
  • a menu of user-selectable options is displayed for the user.
  • the menu of user-selectable options includes one or more options to change frequency based on a time period (e.g., every 1 hour, every 5 hours, or every 1 day), and/or includes one or more options to change frequency based on a detected event of device 100 (e.g., on unlock, in a respective focus mode, or on initiating charge).
  • Figure 6Y further illustrates user input 6032 selecting the option to automatically select images that correspond to the currently selected categories.
  • device 100 identifies one or more images associated with a respective category that is selected (e.g., people, pets, and/or urban) and displays the one or more images as the background media item of the wake screen user interface.
  • the device provides an option for the user to manually select the images.
  • the device 100 displays a photo library, or photo picker (e.g., user interface 668, Figure 6N) in which the device responds to user inputs that corresponds to requests to browse and select particular photos to include in the Smart Album background.
  • a photo library or photo picker (e.g., user interface 668, Figure 6N) in which the device responds to user inputs that corresponds to requests to browse and select particular photos to include in the Smart Album background.
  • device 100 in response to user input 6032, ceases display of the user interface 6028.
  • device 100 optionally displays an editing user interface for a first wake screen user interface of the Smart Album, as illustrated in Figure 6Z.
  • device 100 displays editing user interface 6034-1 ( Figure 6Z) in response to one or more interactions as described with reference to Figures 6F- 6G.
  • the device provides access to editing user interface 6034-1 using any of the methods described herein.
  • editing user interface 6034-1 is accessible from a settings user interface that provides an option for editing a wake screen user interface, and/or an option for editing a home screen user interface.
  • Figure 6Z illustrates an editing user interface 6034-1 for a first wake screen user interface that includes a first person, optionally a person selected in a Smart Album.
  • an original photo or “Natural” view
  • an option to shuffle, or view another photo in the Smart Album is provided (e.g., “Tap to Shuffle”).
  • the original photo or “Natural” view) is processed by device 100 and an initial tone is determined.
  • the initial tone represents an amount of luminance and/or brightness of a background (and/or a foreground) of the original photo.
  • the original photo in Figure 6Z is identified as a low-key level of luminance (e.g., a low amount of luminance, such as in a relatively dark photo) in accordance with a calculated level of luminance satisfying a threshold amount of luminance.
  • editing user interface 6034-1 includes a photo picker 6033 to select a different background photo.
  • the device 100 displays a photo library or photo picker user interface (e.g., as described with reference to Figure 6N).
  • editing user interface 6034-1 includes an indicator 6035 of additional views of the wake screen user interface, for example a same foreground (e.g., subject or individual) of the photo and a different background and/or a different filter applied to the photo (e.g., as described with reference to swipe input 6040, Figure 6AA).
  • a same foreground e.g., subject or individual
  • a different background e.g., a different filter applied to the photo
  • a menu 6037 in response to user input 6036 selecting a menu option, a menu 6037 is displayed.
  • menu 6037 provides a plurality of options for modifying the wake screen user interface displayed in editing user interface 6034-2, including an option to radar, an option to change to low-key and/or high-key, an option to change a frequency (e.g., of switching background photos in a Smart Album), and/or an options to disable a depth effect of the photo.
  • user input 6038 selecting the option to change a frequency of the background enables the user to modify when device 100 changes the wake screen user interface (e.g., on a time basis and/or based on a device event).
  • device 100 in response to detecting a swipe input 6040, displays editing user interface 6034-3 (Figure 6AB), which includes another view (e.g., a duotone view) of the photo.
  • the duotone view is generated as a low- key version of duotone in accordance with a determination that the original photo is identified as a low-key level of luminance. For example, device 100 automatically displays the set of views with a level of luminance that is selected based on a level of luminance of the original photo.
  • Figure 6AB illustrates user input 6046, and in response to user input 6046, displaying menu 6037.
  • the device 100 detects user input 6044 selecting to switch to a high-key version of the currently displayed view.
  • device 100 in response to user input 6044, updates the display to display the duotone view (also referred to herein as the duotone filter) as the high-key version, as illustrated in Figure 6AC.
  • the duotone view is represented by horizontal stripes, but the high-key version of the duotone view (illustrated in Figure 6AC) has a greater level of luminance than the low-key version of the duotone view (illustrated in Figure 6AB).
  • device 100 automatically generates a respective version of each view based on an identified tone of the original photo.
  • a low-key photo e.g., with less than a threshold amount of luminance
  • a plurality of low-key views are provided for the photo (e.g., low-key duotone view, low-key studio view, and/or low-key black and white view).
  • Figure 6AC illustrates user input 6048 (e.g., a swipe input) corresponding to a request to switch to another view.
  • a low-key black and white view of the photo is displayed in user interface 6034-5, as illustrated in Figure 6AD.
  • device 100 continues to recommend, or automatically provide, the low-key versions of the other views. For example, in Figure 6AD, the device 100 provides the low-key version of the black and white view (e.g., because the original photo is a low-key photo).
  • the user is enabled the manually change a particular view, or filter, to switch between low-key and/or high-key, without changing the recommended views that are automatically generated and presented, which are likely to be visually pleasing because the views maintain a luminance (or other color tone) that is present in the original photo.
  • Figure 6AE illustrates an example of an original high-key photo that is displayed in editing user interface 6050-1.
  • the original photo satisfies a threshold level of luminance that makes the photo relatively bright, or high-key.
  • the device 100 in response to user input 6052, displays another view of the photo.
  • the next view illustrated in user interface 6050-2 is a high-key version of the view, for example the high-key version of the duotone view, as illustrated in Figure 6AF.
  • the device 100 in response to user input 6054, replaces display of the user interface with another view, a high-key version of a studio view, as illustrated in user interface 6050-3 in Figure 6AG.
  • Figure 6AG illustrates user input 6056 selecting a menu icon, and in response to user input 6056, the device 100 displays the menu 6037.
  • user input 6058 selects the option, within menu 6037, to switch the current view to the low-key version.
  • the device 100 displays user interface 6050-4 in Figure 6AH, corresponding to the low-key version of the studio view.
  • Figure 6AH illustrates user input 6060 requesting to switch to another view.
  • device 100 displays the high-key version of a black and white view (e.g., because the original photo is determined as a high-key photo), as illustrated in user interface 6050-5 in Figure 6AI.
  • Figure 6AI illustrates a user input 6062 corresponding to a request to switch to another view.
  • device 100 displays a color view (also referred to herein as a color filter), as illustrated in user interface 6050-6 in Figure 6AJ.
  • a tone of the color selected for the color view matches a tone identified in the original photo. For example, an overall tone (or luminance) of the original photo is calculated, and the color selected for the color view has a same overall tone (or luminance) as the original photo.
  • a background of the photo is updated to include the selected color.
  • the selected color is applied to the entire photo (e.g., background and foreground), optionally including applying a different level (e.g., with a different transparency or luminance) of the selected color to various portions of the photo.
  • the selected color is optionally applied to the entire photo, wherein the foreground that includes the subject of the photo only faintly applies the selected color (e.g., as a filter), while the background is replaced with a solid fill, or more opaque version, of the selected color.
  • Figure 6AJ illustrates user input 6064 corresponding to a request to modify a background of the wake screen user interface.
  • device 100 displays a color picker 6066 for changing a color applied to the photo.
  • the initially selected color is selected to have a color tone that matches the tone of the original photo.
  • the selected color has a high level of luminance (e.g., for a high-key original photo), as indicated by the selector being to the left on the slider bar 6068 that indicates a level of luminance that defines the tone of the currently selected color.
  • the tone of the color would automatically be set to a low-key tone (e.g., on the right side of the slider 6068).
  • user input 6070 corresponds to a request to change a tone of the selected color.
  • the background of user interface 6050-8 is displayed with a darker (e.g., more low-key) tone of the same color of the background in user interface 6050-7 in Figure 6AK.
  • each of the colors provided in the color picker 6066 are displayed with a same tone (e.g., that matches the tone of the original photo). For example, while the colors are distinct (e.g., purple, green, red, yellow, or another color), the level of luminance and/or the tone of each color is optionally automatically selected in accordance with the tone of the original photo. In some embodiments, one or more colors displayed as options in the color picker are not selected as having the tone of the original photo.
  • Figure 6AL illustrates a user input 6069 selecting another color that is displayed in the color picker 6066.
  • the background of the user interface is updated to display the selected color.
  • one or more of the views are used in the Smart Album as the wake screen user interface, as illustrated in Figure 6AM.
  • Figure 6AM illustrates a wake screen user interface 6072-1 and a case 6073 for device 100 that is near the.
  • case 6073 is a protective case that is physically placed on at least a portion of device 100.
  • device 100 detects one or more properties associated with case 6073 while case 6073 is on the device 100. For example, device 100 identifies a color of the case 6073.
  • Figure 6AN illustrates case 6073 while placed on device 100.
  • the device 100 determines that the case 6073 is placed on device 100, and, in response to determining that case 6073 is on the device 100, device 100 optionally displays a prompt 6074 that asks the user if device 100 should update the background color (e.g., of wake screen user interface 6072-2) to match the case color (e.g., as detected by device 100).
  • the prompt 6074 includes an option for selecting yes and an option for selecting no.
  • device 100 detects user input 6076 selecting yes to the prompt 6074, and in response to user input 6076, the device 100 updates the user interface 6072-3 to be displayed with a color (e.g., a color filter) that matches (or complements) the color of case 6073.
  • a color e.g., a color filter
  • the color is applied to the background of the wake screen user interface 6072-3 only.
  • the color is applied to the foreground and background of the wake screen user interface 6072-3 (optionally with different portions being displayed with different levels of transparency or luminance of the color).
  • Figure 6AP illustrates a user interface 652 (e.g., as described with reference to Figure 6K).
  • the user interface 652 for creating a new wake screen includes a recommended face 6078 that is based on a color of a case that is detected as being on device 100.
  • the user interface 652 is updated to automatically include a featured wake screen user interface that matches (or complements) the color of the case. It will be understood that additional recommendations may be provided to the user based at least in part on a color of a case that is detected as being on device 100.
  • Figure 6AQ-6AR illustrate an example of displaying the Smart Album as the wake screen user interface.
  • one or more views of the wake screen user interfaces described with reference to Figures 6AA-6AJ are displayed as the wake screen user interface.
  • a first view of a first photo is displayed as wake screen user interface 6080, and the wake screen user interface is automatically updated, without additional user input, to display a second wake screen user interface 6084 that includes a second view of a second photo, in accordance with the selected frequency of the Smart Album.
  • a user input 6082 is optionally detected, and in response to user input 6082, the wake screen user interface is updated from wake screen user interface 6080 to wake screen user interface 6084.
  • Figures 7A-7V-3 illustrate a representation of a plurality of notifications in different configurations, as well as user inputs that can be used to switch between the different configurations.
  • Figure 7A shows a representation of a first plurality of notifications in a first configuration 7000-a.
  • the representation of the first plurality of notifications will be referred to as “the representation 7000” (e.g., regardless of what configuration the representation of the first plurality of notifications is displayed with), and the letter after the label 7000 (e.g., 7000-a, 7000-b, or 7000-c) refers to the particular configuration that the representation 7000 is displayed in (e.g., 7000-a refers to a first configuration, 7000-b refers to a second configuration different from the first configuration 7000-a, and 7000-c refers to a third configuration different from both the first configuration 7000-a and the second configuration 7000-b).
  • the representation 7000 may represent (e.g., include) different notifications (e.g., based on notification settings for an active mode of the portable multifunction device 100).
  • the plurality of notifications represented by the representation 7000 include at least: a notification 7002 associated with an application A, a notification 7004 associated with an application Z, a notification 7006 associated with an application S, a notification 7008 associated with an application D, and a notification 7010 associated with an application M. While the representation 7000 is displayed in the first configuration 7000-a (e.g., a “list” configuration), each notification of the plurality of notifications is displayed separately from other notification in the plurality of notifications (e.g., in a list format, without overlay between adjacent notifications).
  • the representation 7000 is aligned with the bottom of a display of the portable multifunction device 100 (e.g., if notifications 7006, 7008, and 7010 were the only notifications to display, they would still be displayed at the same locations as shown in Figure 7A, but without notifications 7002 and 7004).
  • the notifications are displayed in reverse chronological order (e.g., as shown in Figure 7A), but in other embodiments, the order of the notifications may be different (e.g., chronological order or a user-defined order).
  • the first configuration 7000-a is a default configuration for displaying notifications.
  • a user can interact with the notifications in the first plurality of notifications through different gestures. For example, in response to a tap input 7012 on the notification 7002, the portable multifunction device 100 opens the application A corresponding to the notification 7002. Alternatively, in response to tap input 7012, the portable multifunction device 100 instead displays additional content associated with the notification 7002 (e.g., by expanding the area occupied by the notification 7002).
  • the portable multifunction device 100 In response to a rightward swipe input 7014, the portable multifunction device 100 displays one or more affordances for interacting with the notification 7006, and in response to a leftward swipe input 7016, the portable multifunction device 100 displays one or more affordances for configuring display options (e.g., for dismissing, deferring, and/or adjusting a prominence of) for the notification 7010.
  • display options e.g., for dismissing, deferring, and/or adjusting a prominence of
  • Figure 7B shows the representation 7000 in a second configuration 7000-b, while a “Work” mode of the portable multifunction device 100 is active.
  • the portable multifunction device 100 displays a visual indication, such as a “Work” indicator 7022, while a respective mode (e.g., the “Work” mode) is active.
  • the “Work” mode of the portable multifunction device 100 may control how notifications are displayed. For example, while the “Work” mode is active, some notifications for certain applications (e.g., a social media application S, associated with notification 7006) are suppressed, while work-relevant applications (e.g., a messaging application A, associated with the notification 7002) are not suppressed.
  • the “Work” mode of the portable multifunction device 100 controls which configuration notifications are displayed in (e.g., activation of the “Work” mode activates the second configuration 7000-c and/or changes the portable multifunction device 100 from displaying notifications in the first configuration 7000-b to the second configuration 7000-c).
  • the representation 7000 is displayed in the second configuration 7000-b (e.g., a “stack” configuration), some notifications (e.g., notification 7002) partially overlay other notifications.
  • the notifications 7002 partially overlays the notification 7018
  • the notification 7018 partially overlays the notification 7020.
  • the representation 7000 is aligned with the bottom of the display of the portable multifunction device 100. In such embodiments, some portions of the display are kept available for display of user interface elements other than notifications, thus increasing visibility of a background image or wallpaper (e.g., as shown in Figure 7B).
  • the configuration 7000-b also condenses the display area for notifications to allow better visibility (e.g., of contextually relevant or urgent notifications and/or system user interfaces, as described below with reference to Figure 7N), or to position user interface elements closer to the user’s hands (e.g., as a user’s hands are frequently positioned closer to the bottom of the display portable multifunction device 100).
  • Figure 7C is analogous to Figure 7B, but shows the representation 7000 in the second configuration 7000-b, while a “Personal” mode of the portable multifunction device 100 is active (as shown by the “Personal” indicator 7026).
  • the “Personal” mode may have different notification settings than the “Work” mode (e.g., allowing notifications from personal contacts but not work contacts, while the “Work” mode allows notifications from work contacts but not personal contacts), and visually, the representation 7000 includes a different set of notifications (e.g., as compared to Figure 7B).
  • a notification count 7024 displays additional information regarding the notifications that are not displayed (e.g., information regarding the number of notifications that are represented by the first representation of the plurality of notifications but not currently displayed, information regarding applications associated with the notifications that are not displayed and/or information regarding a sender associated with the notifications that are not displayed). In some embodiments, the notification count 7024 does not include a count of notifications that are suppressed (e.g., due to settings of the “Personal” mode).
  • the notification count 7024 includes a count of all notifications represented by the representation 7000 (e.g., even notifications that are not visually displayed due to notification settings of the “Personal” mode). In some embodiments the notification count 7024 includes separate indications for the number of notifications allowed by the notification settings of the “Personal” mode, and for the number of notifications suppressed in accordance with the notification settings of the “Personal” mode.
  • Figure 7D shows the representation 7000 in a third configuration 7000-c.
  • the first representation of the plurality of notifications includes a count of the number of notifications in the plurality of notifications, without displaying individual visual representations of notifications in the plurality of notifications.
  • Figure 7E shows a side-by-side view of the representation 7000 in the first configuration 7000-a, the second configuration 7000-b, and the third configuration 7000-c.
  • Each configuration is shown as representing a different set of notifications, to highlight different contexts in which different configurations may be beneficial and/or desirable.
  • the plurality of notifications includes 5 notifications.
  • the portable multifunction device 100 can display the entire representation 7000 in the first configuration 7000-a, which allows content from all of the notifications to be displayed simultaneously. Even in cases where there are more notifications (e.g., 6 to 8 notifications) than can fit on the display at the same time, the portable multifunction device 100 can still display notification content for at least some of the notifications in the plurality of notifications.
  • the first configuration 7000-a allows the user to efficiently view notifications, as no additional inputs are needed to expand notifications for viewing, and, often, the number of notifications is low enough that no (or minimal) scrolling is needed to view additional notification content.
  • the plurality of notifications includes 9 notifications.
  • the first configuration 7000-a may become less desirable, as it is difficult to display the larger number of notifications on the display of the portable multifunction device 100 (e.g., without changing a text size, without changing an overall size of the representation of each respective notification, and/or without scrolling).
  • the second configuration 7000-b allows for some notification content to remain displayed, without cluttering the display with a large amount of detail (e.g., as would be the case if the representation 7000 were displayed in the first configuration 7000- a).
  • the plurality of notifications includes 11 notifications.
  • the third configuration 7000-c can be useful (e.g., in combination with a Do Not Disturb or other reduced notification mode) in order to avoid distractions (e.g., from incoming notifications that would be displayed in the first configuration 7000-a or the second configuration 7000-b), or simply because the number of notifications is so high that neither the first configuration 7000-a or the second configuration 7000-b is desirable due to the large amount of scrolling that would be required to view a particular notification (e.g., the oldest notification at the end of, or at the bottom of, the representation 7000).
  • the user can use different user inputs in order to efficiently switch between configurations.
  • These user inputs allow users to select a suitable configuration as the circumstances change, and are described in further detail below, with reference to Figures 7F-7U.
  • the plurality of notifications includes 7 notifications (with 2 notifications that are not displayed in Figure 7F), and the representation 7000 is displayed in the first configuration 7000-a.
  • the portable multifunction device scrolls the displayed notifications.
  • the notification 7002 and the notification 7004 have been scrolled off the display.
  • a notifications 7030 associated with an application W, and a notifications 7032 associated with an application N, which were not displayed in Figure 7F, are now displayed in Figure 7G.
  • the portable multifunction device 100 transitions to displaying the representation 7000 in the second configuration 7000- b, as shown in Figure 7H.
  • Figure 7H shows that, while the first representation of the plurality of notifications is displayed in the second configuration 7000-b, in response to an upward swipe gesture 7038, the notifications in the plurality of notifications can be scrolled in the second configuration 7000-b.
  • the notification 7002 has been scrolled off the display and the notification 7004 is now displayed at the top of the representation 7000 (e.g., at the top of the “stack” of notifications).
  • the notification count 7024 updates to reflect that 4 additional notifications can be displayed if the user continue scrolling, and at least one notification of those 4 additional notifications is associated with the application D.
  • Figure 7J shows that a user can interact with displayed notifications, for example, through a tap gesture 7040 on a notification 7004.
  • the notification 7004 is the topmost notification in the plurality of notifications, as shown in Figure 7K
  • the portable multifunction device 100 in response to detecting the tap gesture 7040, displays an affordance 7042 for opening the application Z, associated with the notification 7004.
  • the portable multifunction device 100 in response to detecting the tap gesture 7040, the portable multifunction device 100 instead launches the application Z and displays a user interface for the application Z.
  • Figure 7L is analogous to Figure 7H, but shows a tap input 7044 on the notification 7006 instead of the notification 7004.
  • the portable multifunction device 100 scrolls the notifications in the plurality of notifications such that the notification 7006 is displayed as the topmost notification (e.g., the portable multifunction device 100 scrolls notification 7004 off the display).
  • the portable multifunction device 100 displays an affordance 7046 for opening the application S that is associated with the notification 7006.
  • the portable multifunction device 100 does not respond to the tap gesture 7044 (e.g., because the notification 7006 is not the topmost notification). In such embodiments, only the topmost notification (e.g., the notification 7004) is available for interaction.
  • neither the notification 7004 nor the notification 7006 are available for interaction while the representation 7000 is displayed in the second configuration 7000-b.
  • none of the notifications visible in the representation 7000 are available for interaction while the representation 7000 is displayed in the second configuration 7000-b.
  • the user changes the configuration for the representation 7000 (e.g., via various user inputs, as described in further detail with reference to Figures 7G, 7P, 7R and 7S) to interact with notifications represented in the representation 7000.
  • Figure 7N shows that the user-selected configuration is persistent.
  • the first representation of the plurality of notifications remains displayed in the second configuration 7000-b (e.g., the next time the portable multifunction device is woken).
  • FIG. 7N also shows that a session 7048 has been received/generated.
  • the session 7048 is a user interface that is contextually relevant, and is displayed separately from the representation of the plurality of notifications in the second configuration 7000-b (e.g., for clarity and/or ease of access).
  • Examples of contextually relevant user interfaces include urgent (e.g., time sensitive) notifications, boarding passes, music player controls (e.g., while music is being played on the portable multifunction device 100), or a user interface associated with an ongoing event (e.g., a user interface that displays status information for a food delivery order that is in progress, a user interface that displays the score of an ongoing sports game, or a user interface that displays workout information while a user is exercising).
  • the session 7048 is visually distinct from other notifications (e.g., notifications represented by the representation 7000).
  • notifications e.g., notifications represented by the representation 7000.
  • the session 7048 has a bolded border to draw attention to the session 7048, and to distinguish the session 7048 from other user interface elements (e.g., non-urgent notifications, notifications that are not time sensitive, and/or old or older notifications) that are not contextually relevant.
  • the user has ordered food from an application F, and the session 7048 displays information related to the user’s food delivery order.
  • the session 7048 remains “contextually relevant,” as the session 7048 can provide updates to the user throughout the food delivery process.
  • the session 7048 is displayed at the top of the displayed notifications (e.g., above, and separate from, the representation 7000 in the second configuration 7000-b), and provides relevant information to the user regarding the user’s food delivery.
  • Figure 70 shows that the session 7048 updates with new information as the time advances to 11 :45.
  • Figure 70 also shows that a new notification 7050 associated with an application C has been received/generated. Because the session 7048 is contextually relevant, it remains displayed at the top of the displayed notifications, even though the notification 7050 is more recent (e.g., would normally be displayed above older notifications based on reverse chronological order).
  • a newly received notification e.g., a notification received after a user has changed the configuration for the representation 7000
  • the notification 7050 is initially displayed separate from the first representation of the plurality of notifications in the second configuration 7000-b. This provides visual feedback regarding which notifications are new (e.g., that the user has not previously viewed and/or interacted with).
  • the notification 7050 is not displayed in the stack with the notification 7002 and the notification 7004 (e.g., does not at least partially overlay the notification 7002 and the notification 7004).
  • the portable multifunction device 100 transitions to displaying the representation 7000 in the third configuration 7000-c.
  • the representation 7000 in the third configuration 7000-c includes a count of the number of notifications represented by the first representation. As shown in Figure 7Q, the count of notifications is 8 notifications, including the most recently generated notification 7050.
  • the session 7048 remains displayed separately from the representation 7000 of the plurality of notifications in the third configuration 7000-c (e.g., and the session 7048 is displayed separately, regardless of what configuration the representation 7000 of the plurality of notifications is displayed in). As the session 7048 is displayed separately from (e.g., not represented by) the representation 7000 of the plurality of notifications in the third configuration 7000-c, the session 7048 is not included in the count of notifications.
  • Figure 7G-7Q show changing from the first configuration 7000-a to the second configuration 7000-b, then changing from the second configuration 7000-b to the third configuration 7000-c
  • the user can change from the first configuration 7000-a directly to the third configuration 7000-c.
  • a characteristic of the user input e.g., a size of the movement of the user input determines which configuration is selected.
  • the portable multifunction device 100 transitions to displaying the representation 7000 in the second configuration 7000-b (e.g., as described above). If, however, the pinch gesture 7034 is a large (or larger) pinch gesture (e.g., includes a large or larger amount of movement of the two contacts of the pinch gesture 7034 towards one another), the portable multifunction device 100 instead transitions to displaying the representation 7000 in the third configuration 7000-c (e.g., as shown in Figure 7Q) directly, and without first displaying the representation 7000 in the second configuration 7000-b. This allows the user to quickly switch to a desired configuration (e.g., with a single gesture), without being forced to switch through other undesired configurations (e.g., to switch through configurations in a predefined order).
  • a desired configuration e.g., with a single gesture
  • undesired configurations e.g., to switch through configurations in a predefined order
  • Figure 7R shows a depinch gesture 7056, or an upward swipe gesture 7058, at a location corresponding to the representation 7000 in the third configuration 7000-c.
  • Figure 7S in response to detecting the depinch gesture 7056, or the upward swipe gesture 7058, the portable multifunction device 100 transitions from displaying the representation 7000 in the third configuration 7000-c, to displaying the representation 7000 in the second configuration 7000-b.
  • Figure 7R shows two possible gestures, but the exemplary gestures can be replaced by any suitable gesture.
  • a pinch gesture can be used to switch from the first, to the second, to the third configuration (e.g., switching from configurations that occupy a larger amount of space on the display, to configurations that occupy less space on the display compared to the previous configuration), and a depinch gesture can be used to switch in the opposite direction (e.g., from the third, to the second, to the first configuration). Pairing gestures that have opposite types of movement provides an intuitive way for users to switch between configurations.
  • Figure 7S shows a depinch gesture 7060, or an upward swipe gesture 7062, at a location corresponding to the representation 7000 in the second configuration 7000-b. While the representation 7000 is displayed in the second configuration 7000-b, the user can switch back to the third configuration 7000-c (e.g., via a user input such a pinch gesture or downward swipe gesture analogous to the gestures in Figure 7P).
  • the portable multifunction device 100 transitions from displaying the representation 7000 in the second configuration 7000-b, to displaying the representation 7000 of the plurality of notifications in an expanded configuration 7000-d, which is an expanded version of the second configuration 7000-b.
  • the expanded configuration 7000-d is temporary.
  • the portable multifunction device 100 reverts to displaying (e.g., redisplays) the representation 7000 in the second configuration 7000-b (e.g., the state shown in Figure 7S).
  • the temporary nature of the expanded configuration 7000-d allows a user to view additional details regarding the notifications in the plurality of notifications, without having to switch to the first configuration 7000-a (e.g., and then needing to switch back to the second configuration 7000-b, after viewing the additional details).
  • the notifications can be scrolled in the expanded version of the second configuration 7000-b (e.g., in response to detecting an upward swipe gesture 7063) (e.g., in the same manner as described above with reference to Figures 7H and 71).
  • the notifications in the plurality of notifications are displayed in reverse chronological order, with the oldest notification being displayed at the bottom of the displayed notifications.
  • the portable multifunction device 100 scrolls notifications (e.g., such that at least one of the one or more older notifications is now displayed, while maintaining display of the representation 7000 in the expanded configuration 7000-d).
  • the portable multifunction device instead transitions to displaying the representation 7000 in the first configuration 7000-a (e.g., as shown in Figure 7U).
  • the representation 7000 is displayed in the expanded configuration 7000-d in response to a user request to scroll notifications (e.g., the upward swipe gesture 7038 in Figure 7H).
  • the expanded configuration 7000-d is optionally a temporary configuration (e.g., for scrolling notifications).
  • the portable multifunction device 100 If the portable multifunction device 100 ceases to display the representation 7000 (e.g., in response to detecting a user input, or when the portable multifunction device transitions to a low power or sleep state), in response to a first event (e.g., that transitions the portable multifunction device 100 out of the low power or sleep state), the portable multifunction device 100 redisplays the representation 7000 in the second configuration 7000-b (e.g., and not in the expanded configuration 7000-d).
  • a first event e.g., that transitions the portable multifunction device 100 out of the low power or sleep state
  • Figure 7U shows that, in response to detecting the depinch gesture 7060, or the upward swipe gesture 7062 (shown in Figure 7S), the portable multifunction device 100 instead transitions from displaying representation 7000 in the second configuration 7000-b, to displaying the first representation 7000 in the first configuration 7000-a.
  • the session 7048 remains displayed at the top of the displayed notifications (e.g., even though it is out of reverse chronological order with the other notifications).
  • Figure 7R-7U show changing from the third configuration 7000-c to the second configuration 7000-b, then changing from the second configuration 7000-b to the first configuration 7000-a
  • the user can change from the third configuration 7000-c directly to the first configuration 7000-a.
  • a characteristic of the user input e.g., a size of the movement of the user input
  • determines which configuration is selected e.g., as described above with reference to Figures 7R-7U).
  • the portable multifunction device 100 transitions to displaying the representation 7000 in the second configuration 7000-b (e.g., as described above).
  • the upward swipe gesture 7058 is a large (or larger) swipe gesture (e.g., includes a large or larger amount of upward movement)
  • the portable multifunction device 100 instead transitions to displaying the representation 7000 in the first configuration 7000-a (e.g., as shown in Figure 7U) directly, and without needing to first display the representation 7000 in the second configuration 7000-b.
  • Figures 7V-1 through 7V-3 shows a region 7076 for invoking a system user interface of the portable multifunction device 100 (e.g., a system user interface for accessing functions of the portable multifunction device 100, or a camera user interface).
  • a size of the region 7076 changes depending on which configuration the representation 7000 is displayed in.
  • the region 7076 is confined to an upper portion of the display, as shown in Figure 7V-1.
  • a first leftward swipe gesture 7064 and a first rightward swipe gesture 7066 are shown in the region 7076.
  • the portable multifunction device 100 displays a system user interface for accessing functions of the portable multifunction device 100.
  • the portable multifunction device 100 displays a camera user interface.
  • Figure 7V-1 also shows two additional pairs of swipe gestures: a second leftward swipe gesture 7068 and a second rightward swipe gesture 7070, and a third leftward swipe gesture 7072 and a third rightward swipe gesture 7074.
  • the second leftward swipe gesture 7068 and the second rightward swipe gesture 7070 are not in the region 7076, and so the portable multifunction device does not display a system user interface.
  • the portable multifunction device 100 forgoes displaying any affordances (e.g., for opening an application associated with a notification, for dismissing a notification, for deferring the notification, and/or for configuring display settings for the notification) for interacting with a notification, and forgoes performing any actions associated with a notification.
  • any affordances e.g., for opening an application associated with a notification, for dismissing a notification, for deferring the notification, and/or for configuring display settings for the notification
  • the third set of swipe gestures are detected at a location corresponding to the notification 7010.
  • the portable multifunction device 100 dismisses the notification 7072 (e.g., if the movement of the third leftward swipe gesture meets a distance threshold), or displays one or more affordances for interacting with the notification (e.g., an affordance for deferring the notification, an affordance for dismissing the notification, and/or an affordance for configuring display settings for the notification 7010).
  • the portable multifunction device 100 opens the application M associated with the notification 7010 (e.g., if the movement of the third rightward swipe gesture 7074 meets a distance threshold), or displays one or more affordances for interacting with the notification (e.g., one or more affordances different than the one or more affordances displayed in response to the third leftward swipe gesture 7072, and/or an affordance such as the affordance 7042 described above with reference to Figure 7K).
  • Figure 7V-2 shows the same six swipe gestures as Figure 7V-1, but the region 7076 has a larger size (e.g., because the representation 7000 in the second configuration 7000-b occupies less space on the display).
  • the first leftward swipe gesture and the first rightward swipe gesture are both within the region 7076, and the portable multifunction device performs the same functions in response to these gestures, as described above with reference to Figure 7V-1.
  • the second leftward swipe gesture 7068 and the second rightward swipe gesture 7070 are also within the region 7076, and so behave in an analogous way to the first leftward swipe gesture 7064 and the first rightward swipe gesture 7066 (e.g., the portable multifunction device 100 performs the same functions in response to detecting the first leftward swipe gesture 7064 and the second leftward swipe gesture 7068, and in response to detecting the first rightward swipe gesture 7066 and the second leftward swipe gesture 7070).
  • the third leftward swipe gesture 7072 and the third rightward swipe gesture 7074 are still located outside the region 7076, and so the portable multifunction device 100 forgoes displaying a system user interface in response to detecting the third leftward swipe gesture 7072 or the third rightward swipe gesture 7074.
  • the third set of swipe gestures is detected at a location over the notification 7020.
  • the portable multifunction device 100 forgoes performing any actions in response to detecting the third leftward swipe gesture 7072 or the third rightward swipe gesture 7074, as the notification 7018 is not displayed at the top of the representation 7000.
  • the portable multifunction device 100 scrolls notifications such that the notification 7018 is displayed at the top of the stack of notifications (e.g., in an analogous fashion to the scrolling described above with reference to Figure 7L and 7M). In such embodiments, the portable multifunction device 100 performs the same functions, or displays the same affordances, in response to detecting the third leftward swipe gesture 7072 or the third rightward swipe gesture 7074, as described above with reference to Figure 7V-1.
  • Figure 7V-3 shows the same six swipe gestures as Figure 7V-1 and Figure 7V- 2, but the region 7076 has an even larger size (e.g., e.g., because the representation 7000 in the third configuration 7000-c occupies less space on than the first configuration 7000-a and the second configuration 7000-b on the display).
  • all three sets of swipe gestures are located within the region 7064.
  • the portable multifunction device 100 displays a system user interface for accessing functions of the portable multifunction device 100.
  • the portable multifunction device 100 displays a camera user interface.
  • Figures 8A-8AN illustrate example user interfaces for displaying one or more sessions in a session region of a user interface in accordance with some embodiments.
  • Figure 8A illustrates a wake screen user interface 800 that optionally includes a date indication, a time indication, one or more complications, and/or icons for accessing a camera and/or flashlight.
  • indications of one or more notifications are optionally displayed on wake screen user interface 800. For example, indications for notification 804, notification 806-1, and “+6 more notifications” are displayed.
  • an option 802 for subscribing to the event of the respective notification 804 is optionally displayed.
  • notification 804 corresponds to a notification for a food delivery application
  • the device 100 subscribes to the food delivery order indicated in notification 804.
  • the device displays a session 816-1 for the food delivery order, as described with reference to Figure 8D.
  • option 802 is not displayed with the notification 804.
  • the device displays options for the notification, for example in response to a swipe user input 810 (or other gesture) directed to the notification 804.
  • the device 100 displays one or more options for the notification 804, for example an option to subscribe to the event associated with the notification and/or an option to dismiss the notification (e.g., to remove the notification from display in the wake screen user interface 800, as illustrated in Figure 8C).
  • Figure 8C illustrates a user input 812 selecting the option to subscribe to the event (e.g., the food delivery event corresponding to notification 804).
  • the device 100 displays a session 816-1 for the food delivery event, as illustrated in Figure 8D, optionally while continuing to display indications for notifications, such as notification 806-1, that have not been dismissed from the wake screen user interface 800.
  • session 816-1 for the food delivery event is displayed within a predefined session region of the user interface.
  • the session region is above a notification region of the user interface (e.g., that displays notifications, including notification 806-1).
  • session 816-1 updates with status information related to the event of the session. For example, session 816-1 updates an estimated delivery time (e.g., as illustrated in Figures 8E-8F).
  • session 816-1 includes additional status information for the event that is not displayed in a notification for the event.
  • session 816-1 includes additional status information, such as a map indicating a location of the delivery, and/or an estimated delivery time, that is not displayed in notification 804.
  • Figure 8E illustrates that, while displaying session 816-2, a notification 806-2 is received.
  • notification 806-2 is displayed in the wake screen user interface without ceasing display of the session 816-2.
  • session 816-2 continues to be displayed at a same position in the wake screen user interface 800. For example, session 816-2 is maintained in the session region that is above the region of the wake screen user interface 800 that displays notifications, even if one or more notifications, or other alerts, are received after the session 816-2 is displayed.
  • Figure 8F illustrates device 100 receiving another notification 806-3, and displaying the notification 806-3 while maintaining display of session 816-3.
  • session 816-3 includes updated status information for the food delivery event.
  • the session 816-3 states the delivery time is “32 minutes” whereas session 816- 3, displays the delivery time in “34 minutes.”
  • the status information displayed in the session 816 is retrieved from the application associated with the session, for example the food delivery application.
  • Figure 8F illustrates a user input 818 corresponding to a request to dismiss the wake screen user interface 800.
  • user input 818 is a swipe gesture (e.g., an upward swipe).
  • the device 100 in response to user input 818, the device 100 ceases display of the wake screen user interface 800 and optionally displays a home screen user interface or an application user interface. As such, the device 100 remains responsive to user inputs while session 816 is ongoing.
  • Figure 8G illustrates session 816-4 updating the status information for the food delivery event.
  • updating the status information includes updating a map that indicates a location of the delivery.
  • updating the status information includes updating the estimated time until delivery.
  • the type of status information displayed, and updated, in a respective session is based on the type of event.
  • status information for an event for a food delivery, package delivery, workout (e.g., workout route) and/or ride share application optionally includes a map.
  • additional and/or alternative status information is included in the status region.
  • Figure 8G further illustrates a new notification 806-4 is received and displayed in the notification region below the status region.
  • the session 816-5 continues to update, in real-time, the status information, and continues to display additional notifications, such as notification 806-5 on the wake screen user interface 800.
  • Figure 81 illustrates the session 816-6 updating the status information to indicate the food delivery has arrived.
  • the session 816 e.g., session 816-1 through 816-6
  • a size of the session is optionally maintained.
  • Figure 8J illustrates that, in accordance with a determination that the event associated with session 816 is ended, complete, or otherwise no longer active, the session 816 is automatically, without user input, removed from display on wake screen user interface 800.
  • the user is not required to acknowledge, or provide any user input directed to, the session 816, and the device 100 ceases displaying the session 816.
  • a final status of the session 816 is maintained on the wake screen user interface 800 until the device has completed at least one wake cycle, including entering and exiting a low-power state (e.g., such that the device makes the final status available for viewing upon waking device 100 out of the low-power state, before the final status disappears).
  • a low-power state e.g., such that the device makes the final status available for viewing upon waking device 100 out of the low-power state, before the final status disappears.
  • the session is persistent and maintained on the wake screen user interface 800 (e.g., in the session region), while the event corresponding to the session (e.g.,
  • Figure 8 J illustrates that, although session 816 is automatically removed from the wake screen user interface 800, indications for notifications that have not been acknowledged and/or dismissed by the user, remain displayed.
  • the device 100 detects user input 822 corresponding to a request to navigate away from the wake screen user interface 800, and in response to user input 822, the device 100 displays a home screen user interface 801 (e.g., a first view of a home screen user interface), as illustrated in Figure 8K.
  • a home screen user interface 801 e.g., a first view of a home screen user interface
  • Figure 8K illustrates user input 824 (e.g., a tap input or other selection input) requesting to open a sports application by selecting an application icon for the sports application.
  • the device 100 displays an application user interface 803 for the sports application, as illustrated in Figure 8L.
  • the application user interface 803 for the sports application includes indications for one or more upcoming basketball games.
  • each basketball game corresponds to an event that is made available to a user for subscription.
  • the device provides respective selectable user interface objects for subscribing to a particular event (e.g., game).
  • a first event corresponding to the first game is subscribed, and/or in response to a user input 828 directed to a “Subscribe” button for a second game (e.g., Los Angeles vs Phoenix), a second event corresponding to the second game is subscribed.
  • a first game e.g., Golden State vs Chicago
  • a second game e.g., Los Angeles vs Phoenix
  • the device provide selectable options to subscribe to individual events (e.g., games), and/or to a set of events.
  • the device optionally provides respective selectable options that, when selected by respective user inputs, cause the device to subscribe to all events for an application (e.g., all events in the sports application, events that include a particular team (e.g., Golden State, Chicago, or Nets) and/or events of a certain type (e.g., football games, baseball games, or basketball games).
  • device 100 automatically subscribes to one or more events, without additional user input, based on past events that the user has previously subscribed.
  • the device unsubscribes from an event (e.g., as described with reference to Figure 8AM) in response to detecting a user input that is directed to an un-subscription affordance and/or option corresponding to the event.
  • the session for the event is not displayed in the session region on the wake screen user interface until the event has started, or is otherwise ongoing.
  • the wake screen user interface 800 does not display a session for the event.
  • the wake screen user interface 800 in accordance with the event starting (e.g., at 7:00 pm), the wake screen user interface 800 automatically displays session 830-1 for the event.
  • one or more complications that are displayed in the wake screen user interface are also updated as the status of the respective application changes.
  • the music complication that indicates a playback completion of a media item is updated over time (e.g., between 6:59 and 7:00)
  • the calendar complication that displays an upcoming event ceases displaying “7:00PM Dinner at Max’s” and instead displays “10:00 AM Yoga” in accordance with a current time (e.g., at 7:00 pm the next upcoming event changes from dinner to yoga).
  • Figure 80 illustrates session 830-2 updating over time (e.g., at 7:41 pm).
  • session 830-2 includes status information about the event, including a current quarter of the basketball game, a current score, and/or an amount of time remaining in the quarter.
  • Figure 80 further illustrates a plurality of notifications that are concurrently displayed with session 830-2, including notification 806-7 and notification 806-8.
  • the device scrolls through the plurality of notifications in response to a user input that corresponds to a request to scroll through the plurality of notifications, for example user input 832 corresponds to a swipe gesture to cycle through the plurality of notifications.
  • Figure 8P illustrates that, in response to user input 832, the device 100 scrolls through the notifications to bring notification 806-8 to the top of the stack, and displays a portion of an underlying notification 806-9.
  • the device responds to user interactions with the notifications according to any of the methods described with reference to Figures 7A-7V-3.
  • Figure 8Q illustrates that, in response to user input 834, corresponding to a request to expand the stack of notifications into a list of notifications, the device 100 displays a list of notifications, including notifications 806-7 through 806-10.
  • the session region is maintained above the top-most notification, and continues to update the status information in the session 830-3.
  • one or more notifications are displayed as scrolling under session 830-3 before scrolling off of the display.
  • Figure 8Q illustrates user input 836 requesting to dismiss the wake screen user interface 800, and in response to user input 836, the device 100 displays home screen user interface 805 (e.g., a second view of the home screen user interface), as illustrated in Figure 8R.
  • home screen user interface 805 e.g., a second view of the home screen user interface
  • an indication of the session is displayed in a session region 838-1 of the home screen user interface (e.g., that is distinct from the session region displayed in the wake screen user interface), as shown in Figure 8R.
  • the session region of the home screen user interface is displayed between one or more status indicators (e.g., a time, a cellular network connection, a WiFi indicator, and/or a battery indicator).
  • session region 838-1 displays status information for the event associated with the active session.
  • the status information is different, or a subset less than all of, the status information provided in the session region on the wake screen user interface.
  • session region 838-1 displays a score of the game (e.g., without displaying the timing information).
  • the device in response to detecting a user input directed to session region 838-1, the device expands session region 838-1 to display additional status information.
  • Figure 8R illustrates user input 840 selecting an application icon for a music application.
  • the device 100 launches the music application and displays application user interface 807-1 for the music application, as illustrated in Figure 8S.
  • application user interface 807-1 and 807-2 While displaying an application user interface (e.g., application user interface 807-1 and 807-2), the session region 838 (e.g., 838-2 and 838-2) is maintained, and the status information in the session region 838 continues to be updated.
  • Figure 8S illustrates a user input 842 requesting to begin playback of a media item (“Song A”).
  • Figure 8T illustrates an application user interface 807-2 for the music application that includes information about a currently playing media item.
  • the device 100 detects user input 844, corresponding to a request to dismiss application user interface 807-2, and display the home screen user interface 805.
  • Figure 8U illustrates, in response to user input 844, the device 100 displays home screen user interface 805.
  • the session region 838 is updated to include information for two sessions, including the sports session 838-4a and the music session 838-4b.
  • Figure 8U further illustrates user input 846 corresponding to a request to display a search user interface 809-1 (e.g., as illustrated in Figure 8V).
  • search user interface 809-1 which includes a search bar and optionally search suggestions (e.g., application icons and/or widgets of recently used or often used applications).
  • the device 100 while displaying search user interface 809-1, the device 100 continues displaying, in the session region, indications of the active sessions (e.g., the sports session 838-4a and the music session 838-4b). In some embodiments, the device 100 detects user input 848 directed to the search bar and/or a user input requesting to search for “workout” (e.g., as illustrated in Figure 8W).
  • the device 100 detects user input 848 directed to the search bar and/or a user input requesting to search for “workout” (e.g., as illustrated in Figure 8W).
  • Figure 8W illustrates user interface 809-2 that displays search results responsive to user input 848 to search for workout.
  • the device 100 continues displaying, in the session region, indications of the active sessions (e.g., the sports session 838-4a and the music session 838-4b), and updates the active sessions (e.g., the score is updated from 44 ( Figure 8V) to 46 ( Figure 8W)).
  • the search results include an application icon 850 for a workout and/or health, application, and optionally includes an option for subscribing to events for the workout and/or health application.
  • user input 854 is directed to “Subscribe” button 852-1, and in response to user input 854, the device 100 subscribes the user to events (e.g., activities and/or workouts) for the workout and/or health application.
  • the user interface 809-3 provides an indication 852-2 that the user has subscribed to the respective application.
  • an option to toggle the subscription on and/or off is provided.
  • the device unsubscribes from events for the workout and/or health application.
  • Figure 8X illustrates user input 856, optionally a physical activation of a button such as a power button or another user input, requesting to put the device 100 in a low-power state.
  • Figure 8Y illustrates that, at a later time, the wake screen user interface 800 is displayed, including displaying session 830-4 in the session region (e.g., above the notifications, including collapsed notifications), and a user interface element 862 (e.g., a platter) corresponding to the active music session.
  • the notifications in response to user input 860 (e.g., a tap input or other selection input) directed to the collapsed notifications, the notifications expand into a list view, optionally under the session 830-4 and/or the user interface element 862.
  • the device 100 expands the notifications into the list view.
  • Figure 8Z illustrates another active session, for a current workout session 864- 1, displayed concurrently with session 830-5 and user interface element 862.
  • status information for the current workout is retrieved from a workout and/or health application executing on device 100, and/or is retrieved from a distinct device (e.g., a wearable device, such as a watch or fitness tracker).
  • device 100 displays session 864-1 in accordance with a determination that a workout (e.g., event or activity) has been initiated. For example, the session 864-1 is displayed on the wake screen user interface 800 optionally without the user interacting with device 100.
  • Figure 8Z illustrates example user inputs for interacting with a session and/or user interface element 862.
  • user input 870 e.g., a tap or a long press
  • session 864-1 causes the device 100 to open an application user interface for the application associated with session 864-1 (e.g., the workout and/or health application), and optionally displays an application user interface that includes session information.
  • the user interface that is displayed includes session information.
  • a user input directed to a respective control within user interface element 862 causes the device 100 to control playback of the media content in accordance with the respective control.
  • user input 868 directed to the skip forward control causes the device 100 to begin playback of a next media item.
  • a user input directed to a predefined portion of user interface element 862 for example, user input 866 directed to an image associated with the media item (e.g., cover art or album art), causes the device 100 to expand the user interface element 862 such that the wake screen user interface is displayed with an overlay that includes information about the current media playback, as illustrated in user interface 800b ( Figure 8AB).
  • the device 100 displays an animated transition, as illustrated in user interface 800a ( Figure 8AA), for expanding an overlay the information about the current media playback.
  • the overlay covers a background of the wake screen user interface (e.g., the hearts image background), optionally without covering display of the date and/or time indication and/or the complications of the wake screen user interface.
  • the icons for accessing a camera and/or flashlight are occluded by the overlay.
  • the overlay is displayed with a background color or gradient.
  • the background color or gradient is selected based on one or more colors associated with the currently playing media item, such as one or more colors that appear in the cover art for the media item.
  • the overlay is a visually deemphasized version of the cover art (e.g., a blurred version of the cover art).
  • the animated transition includes increasing a size of the overlay from a first comer of the display (e.g., the lower left comer), until the overlay is displayed over the entire display.
  • the overlay is maintained over the wake screen user interface for a threshold amount of time, before automatically returning to the wake screen user interface and displaying the user interface element 862.
  • the device 100 ceases display of the overlay in response to a user input requesting to dismiss the overlay.
  • the device detects and responds to user input directed to one or more controls displayed in the overlay of user interface 800b. For example, in response to user input 872 directed to a pause control, the device 100 ceases playback of the current media item. In some embodiments, in response to user input 874, corresponding to a request to dismiss the overlay, the device 100 optionally displays a home screen user interface 801 (or an application user interface) and dismisses the overlay over the wake screen user interface, as illustrated in Figure 8AC.
  • a home screen user interface 801 or an application user interface
  • Figure 8AC illustrates a user input 876 directed to a rideshare application icon.
  • the device 100 launches the rideshare application and displays user interface 811 for the rideshare application, as illustrated in Figure 8 AD.
  • the device 100 in response to user input 878, to request ride option 1, the device 100 updates the user interface to indicate that ride option 1 has been requested.
  • Figure 8AE also illustrates that, after requesting ride option 1, the device 100 optionally displays an option 880 for enabling automatic subscription for the rideshare application. For example, in response to a user input directed to toggle option 880, the device toggles the on and/or off state of the toggle option and subscribe and/or unsubscribe to a particular type of event (e.g., rides requested) for the rideshare application depending on whether the particular type of event is currently subscribed. In some embodiments, the device 100 optionally provides an option for subscribing only to the current event (e.g., ride option 1).
  • the current event e.g., ride option 1
  • an option to enable automatic subscription for any application is provided, for example, in a settings user interface of device 100.
  • the device 100 if the user has enabled automatic subscription for all applications, or a subset of applications, the device 100 automatically subscribes the user to certain events.
  • automatic subscription does not necessarily subscribe the user to all events from an application. For example, automatic subscription enables device 100 to determine which subset of events the user is likely interested in subscribing, such as if the user has shown an interest in a particular team for a first sport, the device 100 automatically subscribes the user to all events that the particular team participates, but not all sporting events of the first sport.
  • Figure 8AE illustrates user input 882 corresponding to a request to enter a low-power state and/or lock the device.
  • Figure 8AF illustrates wake screen user interface 800, optionally after the device 100 has transitioned out of the low-power state.
  • the three active sessions: sports session 830-6, rideshare session 884-1 and workout session 864-2 are concurrently displayed on the wake screen user interface 800.
  • an indication of notifications are optionally displayed (e.g., below the session region displaying the three active sessions).
  • Figure 8 AG illustrates that, in accordance with a determination that a fourth session is active (e.g., Los Angeles vs. Phoenix game starts at 8:00), the sessions are displayed in a condensed, or collapsed view.
  • the device 100 concurrently displays up to a threshold number of sessions (e.g., 3 sessions), and if there are more than the threshold number of sessions, the sessions are displayed in a condensed view in the session region.
  • a threshold number of sessions e.g., 3 sessions
  • sports session 888-1 for the Los Angeles vs. Phoenix game is displayed at the top of the condensed view, with session 830-6, session 884-2 and session 864-3 displayed below the sports session 888-1.
  • an order of the sessions is determined based on a respective priority of the applications associated with the sessions. For example, applications are assigned a priority, and a session associated with the application having the highest priority is displayed at the top of the sessions. In some embodiments, the sessions are organized based on a time the session was initiated. For example, the most recently initiated session is displayed at the top (or bottom) of the sessions.
  • Figure 8 AG illustrates a user input 886 directed to the condensed view of the sessions.
  • the device 100 displays the sessions concurrently, as illustrated in Figure 8AH.
  • the device scrolls through additional sessions (e.g., the rideshare session 884) that are not displayed in the user interface 800 in response to detecting one or more user inputs directed to the currently displayed sessions (e.g., in response to one or more swipe gestures directed on the displayed sessions).
  • Figure 8AH illustrates user input 890 directed to workout session 864-4.
  • the device 100 displays an application user interface 813 for the workout application, including a view of the current workout corresponding to the current workout session 864-4.
  • the device unsubscribes and/or ends the current workout session 864-4 in response to detecting a user input directed to the view of the current workout session 864-4 that meets preset criteria. For example, in response to user input 892 selecting the button “End Workout” in Figure 8AI, the current workout session 864-4 is stopped.
  • the device 100 detects user input 894 corresponding to a request to display the home screen user interface 805.
  • user input 890b detected at a different portion of the workout session 864-4 causes the device 100 to display additional information related to the session. For example, in response to user input 890b, the device 100 increases a size of session 864-4 and optionally displays additional status information, such as a map of a route of the workout.
  • Figure 8AJ illustrates, in response to user input 894, the device 100 displays home screen user interface 805.
  • user input 896 is detected, and in response to user input 896, the device enters the low-power state.
  • Figure 8AK illustrates that, at 8:30, only one session, sports session 888-3, is still active. For example, device 100 ceases displaying each respective session in accordance with the event of the session ending. For example, the Golden State vs Chicago game has ended, the user ended the current workout session ( Figure 8AH), and the rideshare session 884-1 has also ended (e.g., in accordance with completing the ride and/or in accordance with the ride arriving for pickup). Thus session 830, session 884 and session 864 are no longer displayed.
  • a user input 898 e.g., distinct from a type of user input of user input 890 (Figure 8AH) corresponds to a request to display options for session 888-3.
  • user input 898 is a swipe input (e.g., from right to left).
  • device 100 slides the session 888-3 at least partially off the display, optionally in the direction of user input 898, and displays an options button 816, as illustrated in Figure 8AL.
  • user input 899 selects the options button 816, and in response to the user input 899, the device 100 displays a menu for controlling the session 888-3, as illustrated in Figure 8AM.
  • Figure 8AM illustrates the menu for controlling the session 888-3 includes an option to mute the session (e.g., forgo providing updates and/or alerts) for 1 hour, an option to mute the session for the rest of the day, an option to view and/or edit settings for the session, and an option to unsubscribe.
  • user input 8002 selects the option to unsubscribe from the session 888-3.
  • Figure 8AN illustrates that, in response to user input 8002, the session 888-3 is removed from display in the wake screen user interface 800, even if the event associated with session 888-3 is still ongoing, because the user has unsubscribed from the session.
  • FIGS 9A-9I are flow diagrams illustrating method 900 of changing a user interface based on a user input in accordance with some embodiments.
  • Method 900 is performed (901) at a computer system (e.g., device 300, Figure 3, or portable multifunction device 80, Figure 1 A) that is in communication with a display generation component (e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like) (and optionally, the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as lidars, depth sensors, and/or distance sensors)
  • a display generation component e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like
  • method 900 provides a wake screen user interface that includes a first set of complications, whereby the user can optionally navigate to another wake screen user interface that includes a second set of complications.
  • the computer system displays (902), via the display generation component, a first version (e.g., a currently selected one of a plurality of preset and/or customizable versions) of a first user interface (e.g., a wake user interface, also referred to herein as a wake screen user interface, and/or a lock user interface) that corresponds to a restricted state of the computer system to a normal operation mode.
  • a first version e.g., a currently selected one of a plurality of preset and/or customizable versions
  • a first user interface e.g., a wake user interface, also referred to herein as a wake screen user interface, and/or a lock user interface
  • the first user interface is an initial user interface that is displayed when the computer system transitions from a powersaving mode (e.g., display is turned off, and/or in a dimmed always-on state, as illustrated in Figure 5A) and the normal operation mode corresponds to a mode wherein the display is turned on and/or restored to a higher power state, as illustrated in Figure 5B.
  • the first user interface is a basic limited function user interface that needs to be dismissed before a home screen or application launching user interface is displayed, and/or a coversheet user interface that can be displayed to block a currently displayed application user interface and/or home screen and that can be dismissed to redisplay the blocked application user interface and/or home screen.
  • displaying the first version of the first user interface includes displaying a first plurality of user interface objects displayed concurrently with (e.g., displayed adjacent to or overlaying) a first background (e.g., image and/or wallpaper) in accordance with a first configuration (e.g., a first theme, a first layout, and/or a first style), wherein the first plurality of user interface objects correspond to a first plurality of applications and include respective content from the first plurality of applications and are updated periodically as information represented by the first plurality of user interface objects changes (e.g., the first plurality of user interface objects include a first plurality of complications, widgets, and/or other similar user interface elements that correspond to different applications).
  • the wake screen user interface 501 is displayed with a gradient background and complications 502-1 through 502-3.
  • the computer system detects a first input (e.g., a touch input such as a swipe input on a touch- sensitive display, or a touch-sensitive surface, and/or an air gesture such as an air swipe gesture (e.g., movement of an input object such as a controller or finger in the air, while a gaze input is directed to a target region or while a target region has input focus)).
  • a first input e.g., a touch input such as a swipe input on a touch- sensitive display, or a touch-sensitive surface, and/or an air gesture such as an air swipe gesture (e.g., movement of an input object such as a controller or finger in the air, while a gaze input is directed to a target region or while a target region has input focus).
  • a first input e.g., a touch input such as a swipe input on a touch- sensitive display, or a touch-sensitive surface, and/or an air gesture such as an air swipe gesture (e.g., movement of an input object such
  • the computer system In response to detecting the first input (906): in accordance with a determination that the first input meets first criteria, wherein the first criteria require that the first input includes first movement in a first direction in order for the first criteria to be met (e.g., the first movement meets first speed, and/or position requirements for navigating to the home screen from the first user interface), the computer system replaces (908) display of the first version of the first user interface (e.g., a first version of the wake user interface, or a first version of the lock user interface) with display of a second user interface (e.g., a respective version of the second user interface that corresponds to the currently selected version of the first user interface) that includes respective representations (e.g., application icons) of a second plurality of applications.
  • first version of the first user interface e.g., a first version of the wake user interface, or a first version of the lock user interface
  • a second user interface e.g., a respective version of the second user interface that corresponds to
  • the first input is an upward swipe that is initiated from a bottom region of the touch-sensitive display, or an upward swipe that is started from the top edge region of the touch-sensitive display.
  • the second user interface includes a home screen or application launching user interface that includes application icons corresponding to different applications, and that, optionally, includes widgets corresponding to different applications.
  • the respective representations of the second plurality of applications when activated (e.g., by a tap input on a touch-sensitive surface, a double tap on a touch-screen display, and/or an air tap or an air flick input), cause the computer system to launch corresponding applications of the respective representations (e.g., the second user interface is a home screen user interface or application launch pad with application icons that are a distinct type of user interface objects from the first plurality of user interface objects (e.g., the complications and/or widgets) shown on the first user interface (e.g., wake user interface, lock user interface, and/or a coversheet user interface)).
  • the device 100 displays user interface 518 ( Figure 5N).
  • the computer system In response to detecting the first input (906): in accordance with a determination that the first input meets second criteria, wherein the second criteria require that the first input includes second movement in a second direction, different from the first direction, in order for the second criteria to be met (e.g., the second movement meets second speed and/or position requirements for navigating to another version of the first user interface), the computer system replaces (910) display of the first version of the first user interface with display of a second version of the first user interface (e.g., a currently unselected version of the plurality of preset and/or customizable versions of the first user interface).
  • the second version of the first user interface is a currently unselected version of the plurality of preset and/or customizable versions of the first user interface.
  • the first input is a horizontal swipe and/or a horizontal arc swipe that is within a bottom region of the touch-sensitive display, or a horizontal swipe that is within the top edge region of the touch-sensitive display.
  • displaying the second version of the first user interface includes displaying a second plurality of user interface objects concurrently with (e.g., displayed adjacent to or overlaying) a second background (e.g., image and/or wallpaper) in accordance with a second configuration (e.g., a second theme, a second preset layout, and/or a second style), wherein the second plurality of user interface objects correspond to a third plurality of applications and include respective content from the third plurality of applications (e.g., the plurality of user interface objects include a second plurality of complications, and/or widgets that correspond to different applications (e.g., the third plurality of applications)) and are updated periodically as information represented by the second plurality of user interface objects changes.
  • a second plurality of user interface objects concurrently with (e.g.,
  • the first background is different from the second background
  • the first plurality of user interface objects is different from the second plurality of user interface objects
  • the first configuration is different from the second configuration.
  • the device 100 replaces display of wake screen user interface 540-1 with wake screen user interface 540-2 ( Figure 5R), having a different background and/or a different set of complications and/or different visual properties for text than the wake screen user interface 540-1.
  • the first version of the first user interface and the second version of the first user interface differ in at least one aspect of multiple aspects of the first user interface, such as the number and/or type of complications/widgets that are included in the first user interface, the appearance of the user interface objects (e.g., time and/or date), the layout of the objects on the background, the type and/or visual properties of the background, and/or the interactions between the background and the objects overlaying the background.
  • the first plurality of user interface objects and the second plurality of user interface objects are of a distinct object type from application icons, notifications, date and time, and application shortcuts that may be displayed on a wake user interface, a lock user interface, and/or a coversheet user interface.
  • the computer system in response to detecting selection of a respective one of the first and second plurality of user interface objects (e.g., selection by a tap input, or an air selection gesture), the computer system ceases to display the currently displayed version of the first user interface and displays a user interface of the application corresponding to the selected user interface object, or optionally displays an authentication user interface if access to the application requires authentication first.
  • a respective one of the first and second plurality of user interface objects e.g., selection by a tap input, or an air selection gesture
  • the first user interface includes a plurality of user interface objects (optionally distinct from the first plurality of user interface objects, and the second plurality of user interface objects) that provide respective functions that are available on different versions of the first user interface, where the set of functions include a time object that displays the current time, a date object that displays the current date, a lock/unlock icon that indicates the current locked/unlocked status of the computer system, and/or a plurality of device status indicators (e.g., network connectivity, WIFI connectivity, battery status, mobile carrier, unread notifications, and/or shortcut to frequently accessed applications and/or device functions).
  • the set of functions include a time object that displays the current time, a date object that displays the current date, a lock/unlock icon that indicates the current locked/unlocked status of the computer system, and/or a plurality of device status indicators (e.g., network connectivity, WIFI connectivity, battery status, mobile carrier, unread notifications, and/or shortcut to frequently accessed applications and/or device functions).
  • different versions of the first user interface can be displayed according to a user's preferences, and/or new versions of the first user interface can be created according to a user's configuration inputs (e.g., inputs directed to a currently displayed version of the first user interface, inputs directed to a selection user interface that displays different versions of the first user interface, and/or inputs directed to a configuration user interface that provides customization options for the first user interface).
  • the computer system displays only one version of the first user interface at a time, unless when switching between different versions of the first user interface and/or when a wake screen switcher user interface (E.g., user interface 548-1, Figure 5T) is displayed.
  • a respective version of the first user interface has a respective background (e.g., a first background or a second background) and a respective plurality of user interface objects (e.g., a first plurality of user interface object, or a second plurality of user interface objects) that are configured according to a respective style (e.g., a first theme, or a second theme), and in the respective style, one or more visual properties and/or behaviors of the respective background, one or more visual properties, functions, type, and/or quantities of the respective plurality of user interface objects, and/or one or more visual properties and/or functions, of the plurality of user interface objects (e.g., time object and/or date object) are selected to correspond to the first style (e.g., portrait mode, text color, and/or font of time and/or plurality of icons change with the background image).
  • a respective style e.g., portrait mode, text color, and/or font of time and/or plurality of icons change with the background image.
  • one or more aspects of the first style can be modified (e.g., in editing user interface 565 ( Figure 5X)) according to user inputs to create a new style and a new version of the first user interface that is stored among the plurality of selectable versions of the first user interface or that becomes the currently selected version of the first user interface.
  • the computer system while the display generation component is in a powersaving state (e.g., a display-off state, and/or a dimmed always-on state), the computer system detects (912) a second input that corresponds to a request to display the first user interface (e.g., an input that activates a power button of the computer system, a touch input on a touchscreen display, and/or a change in the posture of the display generation component).
  • a second input that corresponds to a request to display the first user interface (e.g., an input that activates a power button of the computer system, a touch input on a touchscreen display, and/or a change in the posture of the display generation component).
  • the computer system In response to detecting the second input that corresponds to the request to display the first user interface: in accordance with a determination that the first version of the first user interface is a currently selected version for the first user interface, the computer system displays a first animated transition that corresponds to the first version of the first user interface and display the first version of the first user interface upon completion of the first animated transition; and in accordance with a determination that the second version of the first user interface is the currently selected version of the first user interface, the computer system displays a second animated transition that corresponds to the second version of the first user interface and display the second version of the first user interface upon completion of the second animated transition, wherein the first animated transition is different from the second animated transition.
  • the displayed animated transition (e.g., the first animated transition, the second animated transition, or another animated transition that corresponds to a respective version of the first user interface that is currently selected) is a wake animation from an inactive or power-saving (e.g., a display-off state, and/or a dimmed always-on state) state of the computer system.
  • the animated transition starts from a dark or dimmed user interface that is displayed.
  • a plurality of animated transitions are provided, each animated transition corresponding to a respective wake screen user interface. Displaying a first animated transition to display a first wake screen user interface, versus displaying a second animated transition to display a second wake screen user interface, causes the device to automatically adjust how the wake screen user interface is initially presented based on the current wake screen user interface, to present an additional indication of which user interface is selected and to make more efficient use of the display area.
  • the computer system in response to detecting the second input that corresponds to the request to display the first user interface (914): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a plurality of weather-based elements (e.g., weather-related background, and/or user interface objects), the computer system displays the currently selected version of the first user interface after displaying animated changes of one or more elements of the plurality of weather-based elements (e.g., showing clouds moving and/or raindrops falling, in the background, showing animations to representations of real local weather in the background, showing animations to complications related to weather (e.g., AQI and/or weather forecast) on the background).
  • a plurality of weather-based elements e.g., weather-related background, and/or user interface objects
  • the plurality of weather-based elements are selected at least in part based on a current location of the computer system.
  • the local weather is determined for a geographic region in which the computer system is currently located.
  • a weather-style wake screen user interface is illustrated in Figure 6M. Displaying an animated transition that includes changing display of weather components indicates that the user interface will be displayed as a weather-themed user interface, thereby providing feedback about a state of the device.
  • the computer system in response to detecting the second input that corresponds to the request to display the first user interface (916): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes an image of at least a portion of a planetary or celestial body (e.g., a globe, the earth, a moon, and/or a star), the computer system displays the currently selected version of the first user interface after displaying animated movement of the planetary or celestial body.
  • the animated transition shows different crops of the globe, or different phases of the moon.
  • a first crop of the globe corresponds to displaying a first portion of the globe from a first perspective (e.g., approximately % of the globe is displayed, optionally with a shadow blocking another portion of the globe), and the animated movement changes an amount of the displayed portion of the globe (optionally decreasing a size of the shadow) to a second crop of the globe that displays a second portion of the globe, optionally from the first perspective, or from another perspective (e.g., the globe appears to rotate, clockwise or counterclockwise), such that a larger portion (e.g., approximately ⁇ 2) of the globe is displayed in the second crop.
  • a first perspective e.g., approximately % of the globe is displayed, optionally with a shadow blocking another portion of the globe
  • the animated movement changes an amount of the displayed portion of the globe (optionally decreasing a size of the shadow) to a second crop of the globe that displays a second portion of the globe, optionally from the first perspective, or from another perspective (e.g., the globe appears to rotate, clockwise or
  • a current location of the computer system is indicated on the globe displayed with the second crop (e.g., a green dot is displayed on a position of the globe corresponding to the computer system's current location).
  • a green dot is displayed on a position of the globe corresponding to the computer system's current location.
  • an animation for a globe wake screen user interface is described with reference to Figures 5K-5L. Displaying an animated transition that includes changing an amount of a displayed globe or planet, indicates that the user interface will be displayed as a planetary- themed user interface, thereby providing feedback about a state of the device.
  • the computer system in response to detecting the second input that corresponds to the request to display the first user interface (918): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a photo (e.g., a scenic photo, a landscape photo, and/or a portrait), the computer system displays the currently selected version of the first user interface after changing at least a first visual property (e.g., a blur radius, a luminance level, a saturation level, and/or a translucency level) of at least a portion of the photo through a plurality of values for the first visual property (e.g., gradually changing a blur filter, a luminance filter, a saturation filter, and/or a translucency filter applied to the photo).
  • a first visual property e.g., a blur radius, a luminance level, a saturation level, and/or a translucency level
  • an animation for a wake screen user interface that includes a photo is described with reference to Figures 5G-5H.
  • Displaying an animation of a visual property of a photo displayed as a background of the user interface, such as by the photo becoming less visually deemphasized (e.g., less blurred and/or less transparent) while appearing makes the photo background increasingly prominent, thereby providing feedback about a state of the device.
  • the computer system in response to detecting the second input that corresponds to the request to display the first user interface (920): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a first color gradient, the computer system displays animated changes of the first colored gradient before displaying the currently selected version of the first user interface (e.g., changing a position at which the gradient lines are located and/or changing one or more colors in the color gradient before displaying the exact color gradient shown in the first user interface). For example, an animation for a wake screen user interface with a gradient is described with reference to Figures 5B-5D.
  • Displaying an animation of a colored gradient selected as a background of the user interface indicates to the user that the user input is causing the device to update a displayed user interface, thereby providing feedback about a state of the device.
  • the computer system in response to detecting the second input that corresponds to the request to display the first user interface (922): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes one or more graphical objects (e.g., emojis, icons, and/or avatars), the computer system displays animated movements of the one or more graphical objects before displaying the currently selected version of the first user interface (e.g., displaying the emojis, icons, and/or avatars, shifting in position, displaying a parallax effect of the graphical objects in response to detecting movement of the display generation component in the physical environment, and/or animating the one or more emojis, icons, and/or avatars bouncing into the first user interface from the edges of the display region).
  • animated movements of the one or more graphical objects before displaying the currently selected version of the first user interface (e.g., displaying the emojis, icons, and/or avatars, shifting in position,
  • the animated movements of the one or more graphical objects include displaying and/or moving the one or more graphical objects in a repeating pattern (e.g., a geometric pattern).
  • a repeating pattern e.g., a geometric pattern
  • an animation for a wake screen user interface that includes emojis is illustrated in Figures 5I-5J.
  • displaying an animation of graphical objects such as emojis, icons, and/or avatars in the selected version of the user interface including by displaying shifts in position and/or a parallax effect, indicates to the user that the user input is causing the device to update a displayed user interface, thereby providing feedback about a state of the device.
  • the computer system while displaying a currently selected version of the first user interface (924) (e.g., the first version of the first user interface, the second version of the first user interface, or another preset and/or customizable version of the first user interface), the computer system detects a third input that meets the first criteria (e.g., the third input corresponds to a request to display the second user interface (e.g., the home screen or the application launching user interface)).
  • the third input e.g., the third input corresponds to a request to display the second user interface (e.g., the home screen or the application launching user interface)).
  • the computer system replaces display of the currently selected version of the first user interface with a respective version of the second user interface that corresponds to the currently selected version of the first user interface, including: in accordance with a determination that the currently selected version of the first user interface is the first version of the first user interface, displaying a third animated transition that corresponds to the first version of the first user interface and display a first version of the second user interface upon completion of the third animated transition; and in accordance with a determination that the currently selected version of the first user interface is the second version of the first user interface, displaying a fourth animated transition that corresponds to the second version of the first user interface and display a second version of the second user interface upon completion of the fourth animated transition, wherein the third animated transition is different from the fourth animated transition.
  • different version of the first user interface have corresponding versions of the second user interface that share one or more visual characteristics, themes, and/or elements; and the animated transitions that are displayed when transitioning from displaying the first user interface to displaying the second user interface are tailored to the visual characteristics, themes, and/or elements of the currently selected versions of the first user interface and the second user interface.
  • the third input corresponds to user input 5042 (Figure 5BE) for dismissing the wake screen user interface 5041
  • the third input corresponds to user input 5044 (Figure 5BG) for dismissing the wake screen user interface 5041.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (926): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a respective color gradient (e.g., the first color gradient, or a second color gradient different from the first color gradient), displaying animated changes of the respective color gradient before displaying the respective version of the second user interface that corresponds to the currently selected version of the first user interface.
  • a respective color gradient e.g., the first color gradient, or a second color gradient different from the first color gradient
  • the computer system in accordance with a determination that the currently selected version of the first user interface includes a first color gradient, displays a first animated change of the first color gradient; and in accordance with a determination that the currently selected version of the first user interface includes a second color gradient, distinct from the first color gradient (e.g., distinct in gradient pattern and/or color), displays a second animated change, optionally distinct from the first animated change, of the second color gradient.
  • the animated changes of the respective color gradient include changing respective positions (and the gradient line positions) of colors in the respective colored gradient to create a series of new colored gradients (e.g., displaying the initial top color of the colored gradient at the bottom of a new colored gradient, and the new colored gradient corresponding to the colored gradient that is displayed halfway through the animation).
  • the respective version of the second user interface displays the last new colored gradient shown in the animation as its initial colored gradient.
  • the animated transition between the currently selected color gradient to the respective version of the second user interface includes shifting the color values in the color gradient in one direction (e.g., the first direction) in response to the input that meets the first criteria (e.g., the swipe gesture in the first direction), followed by shifting the color values in the color gradient in another direction (e.g., a direction opposite the first direction) after termination of the input that meets the first criteria to restore the original appearance of the color gradient the respective version of the second user interface is displayed .
  • the animated transition between the currently selected color gradient to the respective version of the second user interface includes shifting the color values in the color gradient in a same direction as a direction of the input, as described with reference to the gradient animation illustrated in Figures 5B-5D.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (928): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a portrait (e.g., a photo of a person's face, and/or a portrait of an animal), displaying animated increase of a scale of at least a portion of the portrait (e.g., at least a main subject in the portrait, and/or at least a central portion of the portrait) before displaying the respective version of the second user interface that corresponds to the currently selected version of the first user interface.
  • a portrait e.g., a photo of a person's face, and/or a portrait of an animal
  • animated increase of a scale of at least a portion of the portrait e.g., at least a main subject in the portrait, and/or at least a central portion of the portrait
  • the portrait includes a portrait style photo or image that includes a main subject, such as an individual and/or an animal, and the main subject of the portrait-style photo is enlarged during the animated transition, while one or more other objects in the background and/or foreground of the photo do not change in scale, or optionally, is enlarged by a smaller amount as compared to the main subject of the photo.
  • the animated transition for the portrait-style wake screen user interface illustrated in Figures 5G-5H is displayed while dismissing the wake screen user interface and displaying a home screen user interface.
  • displaying an animation of the portrait increasing in scale indicates to the user that the user interface is being updated, thereby providing feedback about a state of the device.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (930): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes one or more objects in a foreground of the first user interface (e.g., emojis, icons, avatars, and/or image of a planetary or celestial body), displaying animated movements of the one or more objects in the foreground of the first user interface (optionally without animating movement of a background of the first user interface, such that the one or more objects animate in movement relative to the background of the first user interface) before displaying the respective version of the second user interface that corresponds to the currently selected version of the first user interface, as illustrated in Figures 5I-5J.
  • the currently selected version of the first user interface includes one or more objects in a foreground of the first user interface (e.g., emojis, icons, avatars, and/or
  • displaying the animated movements include displaying the emojis, icons, and/or avatars, shifting in position, displaying a parallax effect of the graphical objects in response to detecting movement of the display generation component in the physical environment (e.g., a simulated parallax effect), and/or animating the one or more emojis, icons, and/or avatars bouncing into the first user interface from the edges of the display region.
  • displaying the animated movements include displaying the emojis, icons, and/or avatars, shifting in position, displaying a parallax effect of the graphical objects in response to detecting movement of the display generation component in the physical environment (e.g., a simulated parallax effect), and/or animating the one or more emojis, icons, and/or avatars bouncing into the first user interface from the edges of the display region.
  • the animated movement of the one or more objects that simulates a parallax effect is displayed in accordance with a determination that the currently selected version of the first user interface includes at least a portion of a planetary or celestial body (e.g., a globe, the earth, a moon, and/or a star).
  • a planetary or celestial body e.g., a globe, the earth, a moon, and/or a star.
  • the animation described with reference to Figures 5K-5L is replicated while dismissing the wake screen user interface and displaying a home screen user interface.
  • the animation described with reference to Figures 5L2-5L3 is displayed in response to a request to dismiss the wake screen user interface.
  • displaying an animation of the graphical objects moving indicates to the user that the user interface is being updated, thereby providing feedback about a state of the device.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (932): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes one or more preset objects (e.g., emojis, icons, avatars, and/or image of a planetary or celestial body), increasing respective visual depths of the one or more objects before displaying the respective version of the second user interface that corresponds to the currently selected version of the first user interface.
  • preset objects e.g., emojis, icons, avatars, and/or image of a planetary or celestial body
  • the animated transition between displaying the first user interface and displaying the second user interface shows the one or more objects in the first user interface being pushed back in the background away from the viewer.
  • the animation is a reverse animation from the animation described with reference to Figures 5I-5J while dismissing the wake screen user interface and displaying a home screen user interface.
  • displaying an animation of the graphical objects appearing to move further away visually deemphasizes the graphical objects and indicates to the user that the user interface is being changed to a different user interface, thereby providing feedback about a state of the device.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (934): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a pattern (e.g., a geometric pattern) of objects (e.g., emojis, icons, and/or avatars), moving the pattern of objects in accordance with a preset movement pattern (e.g., moving in a geometric pattern such as a spiral pattern, and/or moving along a predefined path such as a curved path).
  • a pattern e.g., a geometric pattern
  • objects e.g., emojis, icons, and/or avatars
  • moving the pattern of objects in accordance with a preset movement pattern e.g., moving in a geometric pattern such as a spiral pattern, and/or moving along a predefined path such as a curved path.
  • a plurality of patterns for a selected set of emojis are displayed as the background of the wake screen user interface.
  • the emojis are animated as moving, wherein the animation is optionally selected based on a currently selected pattern of the emojis.
  • displaying an animation of the graphical objects appearing to move in a preset movement pattern indicates to the user that the user interface is being changed to a different user interface, thereby providing feedback about a state of the device.
  • replacing display of the currently selected version of the first user interface with the respective version of the second user interface that corresponds to the currently selected version of the first user interface includes (936): in accordance with a determination that the currently selected version of the first user interface is a respective version of first user interface that includes a preset scene (e.g., a landscape, a cityscape, weather scene, and/or nature scene), displaying different views of the scene that corresponds to movement of a virtual viewpoint within the scene before displaying the respective version of the second user interface that corresponds to the currently displayed version of the first user interface.
  • a preset scene e.g., a landscape, a cityscape, weather scene, and/or nature scene
  • simulating movement of the virtual viewpoint within the scene is performed in accordance with a determination that the currently selected version of the first user interface includes one or more weather-based elements, and displaying the different views of the scene simulates a view of a user moving through the scene in weather shown by the one or more weather elements (e.g., the animated transition displays a camera view of moving through the rain and/or clouds represented by the weather elements).
  • the weather wake screen user interface (Figure 6M) includes animated raindrops and/or dynamically updates to change a viewpoint of the user, as if the user is walking through the rain.
  • a wake screen user interface that includes a preset scene (e.g., weather changes)
  • displaying an animation of the user’s viewpoint appearing to travel through the scene indicates to the user that the user interface is being changed to a different user interface and provides information about a state of the weather that is visible from farther away than other indications of weather such as a weather widget or complication, thereby providing feedback about a state of the device.
  • the computer system while displaying a currently selected version of the first user interface (e.g., the first version of the first user interface, the second version of the first user interface, another preset and/or customizable version of the first user interface), the computer system detects (938) a fourth input that corresponds to a request to dismiss the first user interface (e.g., the fourth input meets the first criteria).
  • the currently-selected version of the first user interface was displayed in response to an input that causes the display generation component to transition from a power-saving mode to a normal operation mode.
  • the currently-selected version of the first user interface was displayed in response to an input that replaces display of an application user interface of a respective application with display of the first user interface. In some embodiments, prior to detecting the fourth input, the currently-selected version of the first user interface was displayed in response to an input that replaces display of the second user interface (e.g., the home screen, or the application launching user interface) with display of the first user interface. In some embodiments, prior to detecting the fourth input, the currently-selected version of the first user interface was displayed in response to an input that replaces display of a widget screen with display of the first user interface.
  • a respective user interface that is to replace display of the first user interface in response to the fourth input includes the second user interface (e.g., a respective version of the second user interface that corresponds to the currently selected version of the first user interface) (and, optionally, in accordance with a determination that the first user interface was had been displayed as a wake screen rather than a coversheet user interface that had blocked access to the second user interface prior to the detection of the fourth input), the computer system displays a first intermediate view of the first user interface that visually obscures at least a portion of the second user interface before displaying the second user interface (e.g., the respective portion of the second user interface that corresponds to the currently selected version of the first user interface, and/or a standard version of the second user interface); and in accordance with a determination that the respective user interface that is to replace display of the first user interface in response to the fourth input includes a respective user interface of a first application (e.g., a respective version of the second user interface that corresponds to the currently selected version of the first user interface, and
  • the computer system displays the first user interface moving out of the display area (e.g., shifting in the first direction in accordance with the movement in the fourth input) and revealing the underlying user interface (e.g., the second user interface, and/or a respective user interface of the first application), wherein the first user interface is more opaque and visually obscures the underlying user interface to a greater degree when the underlying user interface is the second user interface (e.g., the home screen); and the first user interface is more translucent and visually obscure the underlying user interface to a lesser degree when the underlying user interface is a respective user interface of an application.
  • the first user interface moving out of the display area (e.g., shifting in the first direction in accordance with the movement in the fourth input) and revealing the underlying user interface (e.g., the second user interface, and/or a respective user interface of the first application), wherein the first user interface is more opaque and visually obscures the underlying user interface to a greater degree when the underlying user interface is the second user interface (e.g.
  • both the first intermediate view and the second intermediate view of the currently displayed version of the first user interface become increasingly translucent as the currently displayed version of the first user interface gradually shifts out of the display region, but the first intermediate view of the currently displayed version of the first user interface is displayed with a greater blur radius as compared to the second intermediate view of the currently displayed version of the first user interface, such that the intermediate view of the first user interface visually obscures the portion of the second user interface underlying the first user interface more than it does to the portion of application user interface underlying the first user interface.
  • the animated transition for dismissing the wake screen user interface 5041 is selected in accordance with the content that is displayed after the wake screen user interface 5041 is dismissed (e.g., the home screen user interface is to be displayed in Figure 5BF and an application user interface is to be displayed in Figure 5BH).
  • the animated transition for dismissing the wake screen user interface 5041 is selected in accordance with the content that is displayed after the wake screen user interface 5041 is dismissed (e.g., the home screen user interface is to be displayed in Figure 5BF and an application user interface is to be displayed in Figure 5BH).
  • displaying an intermediate view of the wake screen that changes one or more visual properties of the wake screen in different ways depending on which user interface will be displayed once the wake screen is dismissed e.g., whether the user is navigating to a home or application launching user interface or to an application user interface
  • the computer system detects (940) a fifth input (e.g., same as the second input, or another input that is different from the second input) that corresponds to a request to display the first user interface (e.g., an input that activates a power button of the computer system, a touch input on a touch-screen display, and/or a change in the posture of the display generation component).
  • a fifth input e.g., same as the second input, or another input that is different from the second input
  • the computer system In response to detecting the fifth input that corresponds to the request to display the first user interface: the computer system displays a respective animated transition (e.g., the first animated transition, the second animated transition, another animated transition that corresponds to another version of the first user interface).
  • a respective animated transition that corresponds to a currently selected version of the first user interface (e.g., the first version of the first user interface, the second version of the first user interface, or another preset or customizable version of the first user interface)) and the computer system displays a currently selected version of the first user interface upon completion of the respective animated transition, wherein displaying the respective animated transition includes changing an appearance of a textual element in the currently selected version of the first user interface (e.g., changing a thickness, fill, and/or size of the font of the textual element (e.g., a time element, a date element, and/or textual header in the first user interface)).
  • the text of the date and time is changed (e.g., to a bold version of the font).
  • displaying an animated transition to the requested wake screen user interface that includes a transition in an appearance of a textual element indicates to the user that the user interface is being changed from a power-saving display state user interface to a different user interface, thereby providing feedback about a state of the device.
  • displaying the first version of the first user interface includes (942) displaying an indication of a current time, and the first plurality of user interface objects are displayed proximate to (e.g., above, next to, or below) the indication of the current time.
  • the location of the indication of the current time has a fixed position on different versions of the first user interface and if a respective version of the first user interface includes complications (e.g., complications associated with weather, health, compass, fitness, and/or third-party applications), the complications are displayed adjacent to the indication of the current time on the respective version of the first user interface.
  • the date and/or time indication are displayed above the complications 502-1 through 502-3.
  • Displaying an indication of a current time, and a plurality of user interface objects that are periodically updated with content from active applications near the current time causes the device to automatically present current information that is likely to be of interest to the user in an efficient way in the display area, thereby reducing an amount of time that is needed for the user to view the information.
  • the computer system while displaying the first version of the first user interface, the computer system detects (944) a user input that corresponds to a request to update the first plurality of user interface objects that are displayed concurrently with the first background. In response to detecting the user input that corresponds to the request to update the first plurality of user interface objects, the computer system displays one or more selectable options that, when selected, modify at least one of the first plurality of user interface objects.
  • the user input is to change the applications associated with at least one user interface object of the first plurality of user interface objects.
  • the user input is to change a style (and/or icon) of the user interface object in the first plurality of user interface objects.
  • the user input is an input to change a position of at least one user interface object of the first plurality of user interface objects (e.g., to be displayed above and/or below the indication of the current time and/or to change where the first user interface object is positioned relative to the other user interface objects in the first plurality of user interface objects (e.g., move a complication to the left and/or right)).
  • the computer system while displaying the one or more selectable options, the computer system detects selection of a first option of the one or more selectable options, and in response, the computer system changes at least one aspect (e.g., absolute position, style, ordinal position, and/or application) of at least one of the first plurality of user interface objects in the first version of the first user interface.
  • the device 100 displays editing user interface 565 for the user to edit the wake screen user interface, including adding complications and/or changing a style of the complications, as described with reference to Figures 5Y-5AG.
  • the computer system detects (946) a first condition that causes the display generation component to cease display of the first user interface (e.g., an inactivity time-out or an activation of the power button that causes the display generation component to transition into a power-saving mode (e.g., a display-off state, and/or a dimmed always-on state), an input that dismisses the first user interface and displays another user interface (e.g., the home screen or an application user interface)).
  • a first condition that causes the display generation component to cease display of the first user interface e.g., an inactivity time-out or an activation of the power button that causes the display generation component to transition into a power-saving mode (e.g., a display-off state, and/or a dimmed always-on state)
  • a power-saving mode e.g., a display-off state, and/or a dimmed always-on state
  • the computer system In response to detecting the first condition that causes the display generation component to cease display of the first user interface, the computer system ceases to display the first version of the first user interface (e.g., turns off the display, dims the display to only shown a time element, or displays another user interface such as the home screen or an application user interface). While the first user interface is not displayed (e.g., while the display is turned off, the display is dimmed, or a home screen or application user interface is displayed), the computer system detects a second condition that causes the display generation component to redisplay the first user interface (e.g., activation of the power button, an input that wakes the display, or a user input for displaying the coversheet user interface to block the currently displayed user interface).
  • a second condition that causes the display generation component to redisplay the first user interface e.g., activation of the power button, an input that wakes the display, or a user input for displaying the coversheet user interface to block the currently displayed user interface.
  • the computer system redisplays the first version of the first user interface (e.g., while the first version of the first user interface is the currently selected version of the first user interface), including: redisplaying the first plurality of user interface objects with updated application content (e.g., based on updated information that is generated between occurrences of the first condition and the second condition) from the first plurality of applications, in the first version of the first user interface; and in accordance with a determination that one or more notifications have been received between occurrences of the first condition and the second condition (e.g., a first notification from a first application, and/or a second notification from a second application that were not displayed in the first version of the first user interface prior to the detection of the first condition), displaying the one or more notifications in the first version of the first user interface, along with the first plurality of user interface objects that include the updated application content.
  • updated application content e.g., based on updated information that is generated between occurrences of the first condition and the second condition
  • the device 100 optionally changes the wake screen user interface 692 to wake screen user interface 694 in accordance with a determination that the device 100 has entered, and exited, a low-power state.
  • wake screen user interface 692 in accordance with a determination that device 100 has entered and exited a low-power state, wake screen user interface 692 remains displayed (and is not replaced with wake screen user interface 694).
  • Displaying, in a wake screen user interface, a plurality of user interface objects that are periodically updated with content from active applications, including redisplaying the plurality of user interface objects with updated application content as well as interim notifications received when the wake screen user interface is redisplayed after being dismissed or turned off, causes the device to automatically present current information that is likely to be of interest to the user in an efficient way in the display area, thereby reducing an amount of time that is needed for the user to view the information.
  • displaying the first version of the first user interface including the first plurality of user interface objects includes (948): at a first time: displaying the first version of the first user interface with a first user interface object corresponding to a first application and a second user interface object corresponding to a second application, wherein the first user interface object includes first application information from the first application and the second user interface object includes second application information from the second application; and at a second time later than the first time: displaying the first version of the first user interface with the first user interface object corresponding to the first application and the second user interface object corresponding to the second application, wherein the first user interface object includes updated application information from the first application that is different from the first application information from the first application.
  • the first user interface object and the second user interface object are updated according to the same updating schedule, and the second user interface object optionally includes second updated information from the second application that is different from the second application.
  • the first user interface object and the second user interface object are automatically updated (e.g., independently of each other), when new information becomes available from their corresponding applications.
  • the first user interface object is updated in accordance with a first set of rules and conditions
  • the second user interface object is updated in accordance with a second, different set of rules and conditions from the first set of rules and conditions.
  • one or more complications on the wake screen user interface are updated (e.g., weather complication 502-6 is updated to display an AQI of 30 ( Figure 5Q) instead of an AQI of 32 ( Figure 5M)).
  • Displaying, in a wake screen user interface, a plurality of user interface objects that are updated over time with content from active applications, causes the device to automatically present current information that is likely to be of interest to the user, thereby reducing an amount of time that is needed for the user to view the information.
  • the first plurality of user interface object includes (950) at least a first user interface object that is associated with a first application published by a first third-party provider and a second user interface object that is associated with a second application published by a second third-party provider different from the first third-party provider.
  • different application vendors may utilize an API provided by the maker of the operating system that designs the first user interface and its operations, and allow their respective applications to have corresponding complications included in the first user interface, as described with reference to Figure 5Y.
  • Displaying, in a wake screen user interface, a plurality of user interface objects that are updated over time with content from active applications, including applications from third-party providers, causes the device to automatically present current information that is likely to be of interest to the user regardless of source, thereby reducing an amount of time that is needed for the user to view the information.
  • the computer system while displaying the first version of the first user interface (952): the computer system detects a sixth input that meets third criteria different from the first criteria and the second criteria.
  • the third criteria include a criterion that is satisfied in accordance with a determination that the input has been maintained for at least a threshold amount of time (e.g., a long press input for at least 1 second, 2 seconds, or 5 seconds).
  • the computer system In response to detecting the sixth input that meets the third criteria (e.g., in accordance with a determination that the sixth input corresponds to a request to launch an editing user interface (e.g., editing user interface 565 ( Figure 5X)) to edit the currently displayed version of the first user interface or select another version of the first user interface as the currently selected version of the first user interface): the computer system initiates a process to display an editing user interface (e.g., displaying the editing user interface directly, or displaying an authentication user interface (e.g., user interface 547 ( Figure 5 S)) before displaying the editing user interface), including displaying one or more selectable options for editing one or more aspects (e.g., background, theme, style, set of complications, time element, date element, notification display style, font, color scheme, and/or display properties) of the first user interface.
  • an editing user interface e.g., displaying the editing user interface directly, or displaying an authentication user interface (e.g., user interface 547 ( Figure 5 S))
  • the sixth user input is a long press that lasts a predefined amount of time on a touch-sensitive display at a location that corresponds to a portion of the first user interface (e.g., the time element, the background, and/or an unoccupied portion).
  • the editing user interface 565 ( Figure 5X) includes reticles 568 and 569 indicating that the time and/or date are editable.
  • initiating a process to edit one or more aspects of the wake screen user interface enables a transition to an editing mode of the wake screen user interface without displaying additional controls.
  • initiating the process to display the editing user interface includes (954): in accordance with a determination that the computer system is in an unauthenticated state (e.g., locked state), acquiring authentication information prior to displaying the editing user interface (e.g., displaying an authentication user interface with information about a status of acquiring authentication information, instructions for providing authentication information, and/or one or more controls to initiate a process for acquiring authentication information or inputting authentication information); and in response to acquiring the authentication information, in accordance with a determination that the authentication information is consistent with authorized authentication information (e.g., a password, passcode, unlocking gesture, and/or biometric information such as fingerprint, facial scan, iris scan, and/or voice pattern that matches an authorized password, passcode, unlocking gesture, and/or biometric information such as fingerprint, facial scan, iris scan, and/or voice pattern) required to transition the computer system from the unauthenticated state to an authenticated state (e.g., unlocked state), dismissing the authentication user interface and displaying
  • authorized authentication information e.g
  • the computer system in accordance with a determination that authentication information is not received or the authentication information that is received is not consistent with authorized authentication information, the computer system redisplays or maintains display of the first user interface and the editing user interface is not displayed.
  • the computer system displays a prompt for authentication information if the user attempts to edit the first user interface when the computer system is in an unauthenticated state. For example, before displaying editing user interface 565 ( Figure 5X), the device 100 displays a user interface 547, or another authentication user interface. Before displaying an editing user interface to edit one or more aspects of the wake screen user interface, acquiring valid authentication information if the computer system is in an unauthenticated state provides improved privacy and security by requiring user authentication to enable editing of the wake screen user interface.
  • the second criteria require that the second movement in the second direction is detected (956) at a location that corresponds to a predefined portion of the first user interface.
  • the predefined portion comprises the bottom edge region of the first user interface.
  • the second criteria require the second movement in the second direction to be detected in the bottom edge region of the touch-screen display.
  • the first direction is a vertical direction (e.g., upward or downward)
  • the second direction is a horizontal direction (e.g., leftward or rightward) relative to the first user interface.
  • the second criteria are met when the first input further includes a movement component in the first direction along with the second movement in the second direction (e.g., the first input is an arc swipe in the horizontal direction).
  • the first input is detected within a predefined edge region of the touch-sensitive surface (e.g., detecting the first input at an initial touch-down location that is within a predefined region of the device in proximity to the bottom edge of the display), and an initial portion of first movement of the first input includes movement in a vertical direction (e.g., upward) and movement in a horizontal direction (e.g., rightward) relative to a predefined edge (e.g., bottom edge) of the display (e.g., a touch-sensitive surface).
  • a predefined edge region of the touch-sensitive surface e.g., detecting the first input at an initial touch-down location that is within a predefined region of the device in proximity to the bottom edge of the display
  • an initial portion of first movement of the first input includes movement in a vertical direction (e.
  • the movement of the first input does not have to be completely vertical and can include a small horizontal component along with the vertical component in order to cause display of the second version of the first user interface.
  • the initial portion of the first movement includes the movement in the vertical direction followed by the movement in the horizontal direction. In some embodiments, the initial portion of the first movement includes the movement in the vertical direction concurrent with the movement in the horizontal direction.
  • user input 541 ( Figure 5Q) satisfies the second criteria, and in response to user input 541, the wake screen user interface changes from user interface 540- 1 to user interface 540-2 ( Figures 5Q-5R).
  • user input 5051 ( Figure 5BJ) is an arc swipe user input that satisfies the second criteria.
  • Requiring that the input for navigating to a different version or page of the wake screen user interface be directed to a predefined portion of the wake screen user interface causes the device to automatically require more intentional input as well as leaves other portions of the wake screen user interface available for performing other types of operations without displaying additional controls.
  • replacing display of the first version of the first user interface with the second user interface in accordance with the determination that the first input meets the first criteria includes (958) replacing display of the first version of the first user interface with a first version of the second user interface that corresponds to the first version of the first user interface.
  • the computer system While displaying the second version of the first user interface as a result of the first input meeting the second criteria, the computer system detects a seventh input.
  • the computer system in response to detecting the seventh input: in accordance with a determination that the seventh input meets the first criteria, the computer system replaces display of the second version of the first user interface with a second version of the second user interface that corresponds to the second version of the first user interface.
  • the currently selected version of the first user interface when the currently selected version of the first user interface is changed in response to user input (e.g., the first input), the currently selected version of the second user interface is also changed automatically without further user input.
  • a first wake user interface e.g., wake screen user interface 5041 (Figure 5BG)
  • dismissing the wake user interface or another application user interface
  • a first home screen user interface is displayed (optionally the first home screen user interface is selected as being related to the first wake user interface) (e.g., home screen user interface 5046 ( Figure 5BI)).
  • a second home screen user interface that is related to the second wake user interface is displayed (e.g., home screen user interface 5056 ( Figure 5BL)).
  • the home screen user interface is automatically updated in accordance with the current respective wake user interface.
  • displaying a version of the other type of user interface that corresponds to the version of the wake screen user interface that was displayed causes the device to automatically maintain visual consistency when transitioning between different types of user interfaces.
  • the first version of the second user interface includes (960) a third background and the second version of the second user interface includes a fourth background that is different from the third background.
  • the third background of first version of the second user interface corresponds to the first background of the first version of the first user interface; and the fourth background of the second version of the second user interface corresponds to the second background of the second version of the first user interface.
  • the background (e.g., stars background) of home screen user interface 5041 is related to the background (e.g., star background) of the wake screen user interface 5046 ( Figure 5BI)
  • the background (e.g., moon and cloud) of home screen user interface 5056 is related to the background (e.g., moon) of the wake screen user interface 5052 ( Figure 5BK).
  • displaying different backgrounds for the other type of user interface based on which version of the other type of user interface is being transitioned to causes the device to automatically maintain visual consistency when transitioning between different types of user interfaces.
  • the second version of the second user interface includes (962) respective representations of a third plurality of applications that are distinct from the respective representations of the second plurality of applications that are included in the first version of the second user interface.
  • the third plurality of applications and the second plurality of applications includes one or more same applications that are optionally arranged at different positions and/or with different sizes of icons (e.g., a weather application icon in the third plurality of applications and a weather application widget in the second plurality of applications).
  • the application icons displayed in the home screen user interface 5056 ( Figure 5BL) are distinct from one or more of the application icons displayed in the home screen user interface 5046 ( Figure 5BI).
  • the contacts, gestures, user interface objects and animations described above with reference to method 900 optionally have one or more of the characteristics of the contacts, gestures, user interface objects and animations described herein with reference to other methods described herein (e.g., methods 1000, 1100, 12000, 1300, 14000, 15000, 16000, and 17000). For brevity, these details are not repeated here.
  • Figures 10A-10I are flow diagrams illustrating method 1000 of selecting and/or modifying a user interface in accordance with some embodiments.
  • Method 1000 is performed (1002) at a computer system (e.g., device 300, Figure 3, or portable multifunction device 80, Figure 1 A) that is in communication with a display generation component (e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like) (and optionally, the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as lidars, depth sensors, and/or distance sensors)
  • a display generation component e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like
  • the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as
  • displaying a preview of the wake user interface concurrently with at least a part of a preview of a home user interface reduces the number of inputs needed to view the effect that potential changes may have on both the wake user interface and the home user interface.
  • the computer system detects (1004) a request to change a wake user interface of the computer system, wherein a wake user interface is a user interface that is displayed when the computer system is turned on or transitioned from a low power state to a higher power state (e.g., from an off state to a dimmed state, and/or from an off state or a dimmed, always-on state to a normal state) and corresponds to a restricted mode of operation for the computer system.
  • the request to change the wake user interface satisfies criteria for invoking a first user interface that allows a user to customize a wake user interface, including selecting a background, one or more visual properties and/or a display style for the wake user interface.
  • the wake user interface includes a user interface that corresponds to a restricted state of the computer system, such as a wake user interface user interface and/or a lock user interface user interface.
  • the wake user interface can be redisplayed as a coversheet user interface to block a currently displayed home screen or application user interface in response to a user input and then dismissed to reveal the last displayed home screen or application user interface in response to another user input.
  • the first user input includes a gesture that corresponds to a request to display a wake user interface selection user interface (e.g., user input 602 ( Figure 6A) causes the device 100 to display expanded face switcher user interface 606 ( Figure 6D)).
  • the first user input is a user input that corresponds to a request enter a customization mode for customizing the one or more backgrounds for the wake user interface user interface.
  • the first input is required to meet different sets of requirements in order to cause display of a customization user interface or selection user interface for the wake user interface, depending on what user interface is currently displayed at the time of the first input.
  • the computer system In response to detecting the request to change the wake user interface of the computer system, the computer system displays, via the display generation component, a first user interface (e.g., expanded face switcher user interface 606 ( Figure 6D)) for changing the wake user interface for the computer system (e.g., a wake user interface selection user interface).
  • a first user interface e.g., expanded face switcher user interface 606 ( Figure 6D)
  • the wake user interface selection user interface e.g., a wake user interface selection user interface
  • the representation 611 of a wake screen user interface is displayed with a portion of a representation of a related home screen user interface 620.
  • the home user interface includes user interface objects corresponding to multiple applications that can be selected to open different applications (e.g., as illustrated in Figures 6I-6J, the home screen user interface includes a plurality of application icons).
  • the first representation of the wake user interface corresponds to a first set of one or more wake user interface settings including a first wake user interface background.
  • the first representation of the home user interface corresponds to a first set of one or more home user interface settings including a first home user interface background (e.g., the first wake user interface background and the first home user interface background are a predefined pair that is to be used respectively for the wake user interface and the home user interface in accordance with user selection of the first representation of the wake user interface, and/or if the first wake user interface background is already the currently selected background for the wake user interface).
  • the representation of smiley face emoji wake screen user interface 611 is associated with the representation of smiley face emoji home screen user interface 620.
  • the computer system displays, concurrently with the first representation, a second representation (e.g., a preview) of the wake user interface (e.g., a wake user interface user interface, a lock user interface user interface) (and, optionally, a second representation (e.g., at least part of a preview) of the home user interface).
  • a second representation e.g., at least part of a preview
  • FIG. 6F at least a portion of representation 613 is displayed with the representation 611 of the wake screen user interface.
  • the second representation of the wake user interface corresponds to a second set of one or more wake user interface settings including a second wake user interface background that is different from the first wake user interface background.
  • the first user interface also includes a second representation of the home user interface that includes a second home user interface background to be displayed in the home user interface when the second wake user interface background is selected for the wake user interface (e.g., the second wake user interface background and the second home user interface background are a predefined pair that is to be used respectively for the wake user interface and the home user interface at a given time).
  • a second representation of the home user interface that includes a second home user interface background to be displayed in the home user interface when the second wake user interface background is selected for the wake user interface (e.g., the second wake user interface background and the second home user interface background are a predefined pair that is to be used respectively for the wake user interface and the home user interface at a given time).
  • the computer system detects (1010) a sequence of one or more inputs corresponding to selection of a respective representation of the wake user interface for the computer system from the first user interface. For example, the user selects a wake screen user interface from the representations displayed in the expanded face switcher user interface 606 ( Figure 6D) to set as a current wake screen user interface.
  • the computer system sets (1014) the wake user interface of the computer system based on the first set of one or more wake user interface settings associated with the first representation of the wake user interface, including using the first wake user interface background as a respective background for the wake user interface and set the home user interface of the computer system based on the first set of one or more home user interface settings, including using the first home user interface background as a respective background for the home user interface. For example, if the user selects representation 615 of the wake screen user interface ( Figure 6D), the selected wake screen having the dog photo is set as the current wake screen, and the corresponding home screen user interface is set as the current home screen user interface.
  • the computer system sets (1016) the wake user interface of the computer system based on the second set of one or more wake user interface settings associated with the second representation of the wake user interface, including using the second wake user interface background as the background for the wake user interface. For example, if the user selects representation 611 if the wake screen user interface, the emoji smiley face wake screen and corresponding emoji smiley face home screen are set as the current wake screen and home screen.
  • detecting the request to change the wake user interface of the computer system includes (1018) detecting a long press input (e.g., a touch input on a touch-screen display at a location that corresponds to an unoccupied background region of the first user interface, and that is maintained without substantial movement for at least a threshold amount of time).
  • a long press input e.g., a touch input on a touch-screen display at a location that corresponds to an unoccupied background region of the first user interface, and that is maintained without substantial movement for at least a threshold amount of time.
  • user input 602 is a long press input in the wake screen user interface 600 that is not detected over any of the one or more complications or the date and/or time indication.
  • the computer system before displaying the first user interface for changing the wake user interface, displays an authentication user interface in accordance with a determination that the computer system is in a restricted, or locked, mode, and the computer system displays the first user interface after receiving valid authentication information, as described with reference to Figures 6B-6C. Enabling the user to provide a long press input to enter a wake user interface selection mode reduces the number of inputs needed to change the wake user interface without displaying additional controls, and requiring valid authentication information to enter the wake user interface selection mode provides improved security and privacy by preventing unauthorized users from changing the wake user interface.
  • the first representation of the home user interface is displayed (1020) in response to detecting an end of the long press input.
  • the computer system in response to detecting the long press input meeting the time threshold, the computer system initially displays the first representation of the wake user interface and the second representation of the wake user interface, without displaying the first representation of the home user interface; and upon detecting the termination of the long press input (e.g., upon an end or liftoff of the long-press input), the computer system displays the first representation of the home user interface.
  • the computer system in response to detecting the long press input meeting the time threshold, the computer system initially displays the first representation of the wake user interface and the second representation of the wake user interface, and a hind of the first representation of the home user interface (e.g., peeking from behind the first representation of the wake user interface); and upon detecting the termination of the long press input (e.g., upon liftoff of the long-press input), the computer system displays the first representation of the home user interface expanding out from behind the first representation of the wake user interface. For example, optionally upon liftoff of user input 624 (Figure 6F), the device displays user interface 638 ( Figure 61).
  • Enabling the user to provide a long press input to enter a wake user interface selection mode, and displaying a preview of a corresponding home user interface with a preview of a respective candidate version of the wake user interface reduces the number of inputs needed to view the effect that potential changes may have on both the wake user interface and the home user interface.
  • the computer system while displaying the first user interface for changing the wake user interface for the computer system, the computer system detects (1022) a first user input corresponding to a request to rearrange an order of the first representation of the wake user interface and the second representation of the wake user interface in the first user interface. In response to detecting the first user input, the computer system enters a state in which ordinal positions of the first representation of the wake user interface and the second representation of the wake user interface are adjustable in accordance with one or more user inputs.
  • the first user input that corresponds to a request to rearrange the order of the representations of wake user interface in the first user interface includes a long press on a location corresponding to one of the representations of the wake user interface, after the long press input the ordinal positions of the representations of the wake user interface are adjustable in accordance with one or more drag inputs on one or more of the representations of the wake user interface (e.g., dragging one representation of the wake user interface from a left side to a right side of another representation of the wake user interface), as described with reference to Figure 6D.
  • the representations of the wake user interface are displayed in an order of automatic rotation (e.g., the wake user interface automatically and/or periodically update from one version of the wake user interface to another based on an order of their representations in the first user interface).
  • the order that different versions of the wake user interface is presented is based on the order of their representations in the first user interface.
  • the order that different version of the wake user interface is presented in response to the user's swipe through the different versions of the wake user interface e.g., horizontal, and/or arc swipes at the bottom edge region of the wake user interface
  • the order that different version of the wake user interface is presented in response to the user's swipe through the different versions of the wake user interface e.g., horizontal, and/or arc swipes at the bottom edge region of the wake user interface
  • Enabling the user to provide a long press input while in the wake user interface selection mode to rearrange different candidate versions of the wake user interface enables custom organization of the different wake user interface options to be performed without displaying additional controls.
  • the computer system detects (1024) a second user input corresponding to a request to remove the first representation of the wake user interface from the first user interface (e.g., the second user input is a swipe input in a first direction (e.g., a swipe up) that is detected at a location corresponding to the first representation of the wake user interface in the first user interface) (e.g., in some embodiments, the second user input includes a touch-hold input followed by an upward swipe input).
  • the computer system initiates a wake screen removal process for removing the first representation of the wake user interface from the first user interface.
  • the computer system in response to detecting the second user input, displays a user-selectable affordance (e.g., a deletion button and/or a deletion confirmation button) for deleting the first representation of the wake user interface, and in response to detecting a user input selecting the user-selectable affordance, the computer system removes the first representation of the wake user interface from the first user interface.
  • removing the first representation of the wake user interface removes the respective version of the wake user interface from the set of different versions of the wake user interface that is accessible through the automatic rotation that automatically cycles through the different versions of the wake user interface and/or through manually swiping through the different versions of the wake user interface when the wake user interface is displayed.
  • the device deletes a wake screen user interface in response to detecting user input 618 selecting delete icon 616 displayed under a respective representation 615 of a wake screen user interface in the expanded face switcher user interface 606).
  • Enabling the user to request to remove a candidate version of the wake user interface causes the device to automatically enable custom organization of and reduction of clutter in the wake user interface selection mode, while in some embodiments further reducing user mistakes by requiring the user to confirm the removal.
  • the computer system while displaying the first user interface concurrently including the first representation of the wake user interface, the first representation of the home user interface, and the second representation of the wake user interface, the computer system detects (1026) a second sequence of one or more user inputs corresponding to a request to navigate through one or more representations of the wake user interface in the first user interface (e.g., a sequence of swipe inputs in the first direction (e.g., horizontal direction, vertical direction, clockwise direction, or counterclockwise direction), such as user input 614 ( Figure 6D), a tap input on the edge of the first user interface (e.g., a tap on the right edge, or a tap on the left edge), such as user input 610 ( Figure 6D)).
  • a second sequence of one or more user inputs corresponding to a request to navigate through one or more representations of the wake user interface in the first user interface e.g., a sequence of swipe inputs in the first direction (e.g., horizontal direction, vertical direction, clockwise direction, or counterclock
  • the computer system In response to detecting the second sequence of one or more user inputs: in accordance with a determination that the second sequence of one or more user inputs includes a third user input corresponding to the request to navigate to the second representation of the wake user interface (e.g., a tap on the second representation while the second representation is partially or fully displayed in the first user interface, or a swipe that brings the second representation into the position of the first representation), the computer system displays the second representation of the wake user interface at an initial location of the first representation of the wake user interface in the first user interface (e.g., the location of the currently selected version of the wake user interface).
  • a third user input corresponding to the request to navigate to the second representation of the wake user interface
  • the computer system displays the second representation of the wake user interface at an initial location of the first representation of the wake user interface in the first user interface (e.g., the location of the currently selected version of the wake user interface).
  • the computer system displays the third representation of the wake user interface at the initial location of the first representation of the wake user interface in the first user interface.
  • the second sequence of user inputs cause the representations of the wake user interface in the first user interface to scroll (e.g., to the right and/or left) in accordance with the direction of the user inputs; and the representations of the wake user interface shift into and out of an initial location at which the representation of the currently selected wake user interface was displayed in the first user interface.
  • some earlier displayed representations may shift out of the display region of the display generation component, while other representations may shift into view of the display region of the display generation component.
  • the device in response to a user’s request to scroll left and/or right within expanded face switcher user interface 606 (Figure 6D), the device changes from displaying representation 615 of a wake screen user interface as the centered representation to displaying representation 611 of a wake screen user interface as the centered representation ( Figure 6F).
  • Enabling the user to provide different types of inputs e.g., swipes and/or edge taps
  • the wake user interface selection mode reduces the number of inputs needed to select a different wake user interface while providing the user with greater control over the precision with which to navigate.
  • the computer system in response to detecting the second sequence of one or more user inputs (1028): in accordance with a determination that the second sequence of one or more user inputs includes the third user input corresponding to the request to navigate to the second representation of the wake user interface (e.g., a tap on the second representation while the second representation is partially or fully displayed in the first user interface, or a swipe that brings the second representation into the position of the first representation), the computer system ceases display of the first representation of the home user interface, and displays a second representation of the home user interface concurrently with the second representation of the wake user interface.
  • the third user input corresponding to the request to navigate to the second representation of the wake user interface
  • the computer system in accordance with a determination that the second sequence of one or more user inputs includes the third user input corresponding to the request to navigate to the second representation of the wake user interface (e.g., a tap on the second representation while the second representation is partially or fully displayed in the first user interface, or a swipe that brings the second representation into the position of the first
  • second representation 611 of a wake screen user interface is displayed concurrently with representation 620 of the related home screen user interface, while the representation of the home screen user interface associated with representation 615 of a wake screen user interface ( Figure 6D) is no longer displayed.
  • the computer system in accordance with a determination that the second sequence of one or more user inputs includes the fourth user input corresponding to the request to navigate to a third representation of the wake user interface different from the first representation of the wake user interface and the second representation of the wake user interface (e.g., a tap on the third representation while the third representation is partially or fully displayed in the first user interface, or a swipe that brings the third representation into view on the first user interface and into the initial display position of the first representation), the computer system ceases display of the first representation of the home user interface, and displays a third representation of the home user interface concurrently with the third representation of the wake user interface.
  • a third representation of the wake user interface different from the first representation of the wake user interface and the second representation of the wake user interface
  • the second sequence of one or more user inputs includes a request to navigate to representations of a plurality of different wake user interfaces in the set of wake user interfaces (e.g., from a representation of a first wake user interface to a representation of a second wake user interface and/or to a representation of a third wake user interface) by one or more user inputs in a same direction (e.g., a plurality of swipe user inputs, or a single user input of different swipe amounts and/or duration, in a first direction (e.g., left to right or right to left)).
  • a same direction e.g., a plurality of swipe user inputs, or a single user input of different swipe amounts and/or duration
  • the second sequence of one or more user inputs includes a first user input in a first direction (e.g., changing from a first wake screen to a second wake screen) and a second user input in a second direction (e.g., changing from the second wake screen back to the first wake screen, and/or optionally changing from the first wake screen to a fourth wake screen in the set of wake screens).
  • the device navigates (e.g., scroll left and/or right) between representations of wake user interfaces in response to user inputs swiping in different directions.
  • only the respective representation of the currently selected version of the wake screen is displayed with a representation of a corresponding version of the home user interface (e.g., a hint of a corresponding version of the home screen).
  • the representation of the currently selected version of the wake user interface is the respective representation of the wake user interface that is displayed in a center portion of the first user interface.
  • the version of the wake user interface that has its representation displayed in the center portion of the first user interface is displayed as the wake user interface.
  • displaying a preview of a corresponding home user interface with a preview of just the currently selected candidate version of the wake user interface causes the device to automatically provide relevant feedback about the effect that potential changes may have on both the wake user interface and the home user interface without displaying too many additional controls.
  • the computer system displays (1030) (e.g., in response to detecting the second sequence of one or more user inputs and in accordance with a determination that an end of a list of representations of the wake user interface has been reached in the first user interface) a second user-selectable affordance for adding an additional representation of the wake user interface (e.g., an "add” button, a "+” button, and/or a "new” button) displayed at the end of the scrollable listing of representations of the wake user interface.
  • a second user-selectable affordance for adding an additional representation of the wake user interface e.g., an "add” button, a "+” button, and/or a “new” button
  • the computer system selects the second user-selectable affordance for adding the additional representation of the wake user interface and display a second user interface that includes one or more selectable options for customizing a set of one or more wake user interface settings for the wake user interface corresponding to the additional representation of the wake user interface.
  • the second user-selectable affordance is displayed in a predefined position in the first user interface for changing a wake user interface for the computer system (e.g., in a top right comer of the first user interface).
  • the device provides one or more user interfaces and/or selectable options for adding an additional version of the wake user interface from a distinct application. For example, the device provides access to the first user interface for changing a wake user interface, or adding a new wake user interface, from a photos application, a settings application and/or from a wake screen gallery that includes a plurality of automatically generated versions for the wake user interface. For example, in response to user input 608 (Figure 6D) directed to the “+” button, the user interface 652 ( Figure 6K) is displayed for the user to create a new wake screen user interface to add to the set of wake screen user interfaces.
  • While in the wake user interface selection mode, displaying a user-selectable affordance for adding another candidate version of the wake user interface reduces the number of inputs needed to customize the wake user interface, particularly if none of the listed options are of interest to the user.
  • the computer system in response to detecting the second sequence of one or more user inputs and in accordance with a determination that an end of a list of representations of the wake user interface has been reached in the first user interface (e.g., in accordance with a determination that the representation of the wake user interface that is currently displayed in the initial location of the first representation of the wake user interface is a last representation of the wake user interface in the list of representations of wake user interfaces), the computer system displays (1032) a third user interface (e.g., a wake screen gallery) that includes a plurality of representations of wake user interface corresponding to a plurality of automatically configured versions of the wake user interface.
  • a third user interface e.g., a wake screen gallery
  • the third user interface corresponds to user interface user interface 652 for creating a new wake screen user interface, as described with reference to Figure 6K.
  • displaying a gallery of wake user interfaces corresponding to automatically configured versions of the wake user interface causes the device to present multiple wake user interface options for the user to consider at the same time, thereby reducing the number of inputs needed to customize the wake user interface, particularly if none of the listed options are of interest to the user.
  • At least one representation of the wake user interface in the plurality of representations of wake user interface corresponds (1034) to an automatically generated version of the wake user interface that includes a third set of one or more wake user interface settings that are set by the computer system.
  • to the device provides one or more editing user interfaces and/or selectable options for modifying the one or more wake user interface settings that are set by the computer system.
  • the device displays selectable representations of one or more pre-generated versions of the wake user interface in a wake screen gallery and the device selects a pre-generated version of the wake user interface in response to a selection input directed to the respective representation of the pre-generated version of the wake user interface in the wake screen gallery, and then the device optionally modifies one or more of the wake user interface settings for the pre-generated version of the wake user interface in response to user inputs directed to one or more selectable options corresponding to those settings that have been displayed by the device.
  • the Featured Faces includes one or more wake screen user interfaces that are generated by the device 100. Displaying, in a gallery of automatically configured versions of the wake user interface, at least one version that is automatically system generated causes the device to automatically suggest an intelligently generated version of the wake user interface, thereby reducing the number of inputs needed to customize the wake user interface.
  • the computer system displays (1036) a selectable option to set, for a respective representation of the wake user interface (e.g., the currently selected representation of the wake user interface, or the representation that has the input focus), a restricted notification mode in which certain types of notifications are suppressed and/or delayed.
  • the restricted notification mode is selected from a plurality of focus modes, wherein each focus mode that defines when certain alerts are provided to the user based on a current set of circumstances, such as a current time of day, a current location of the user, a current status of the user (e.g., driving mode, sleeping mode), and/or do not disturb mode.
  • the user-selectable user interface element 609a (“Focus”) enables the user to set a restricted notification mode for the wake screen corresponding to representation 615 of a wake screen. While in the wake user interface selection mode, displaying for a respective candidate version of the wake user interface a user-selectable option to set a restricted notification mode for suppressing and/or delaying certain types of notifications improves privacy and security by limiting the amount of information that is automatically displayed and reduces an amount of time needed to perform a particular operation on the device by helping the user focus.
  • the computer system while displaying the first user interface, displays (1038) a selectable option to navigate to a third user interface (e.g., a wake screen gallery) that includes a plurality of representations of wake user interface corresponding to a plurality of automatically configured versions of the wake user interface.
  • a third user interface e.g., a wake screen gallery
  • the plurality of automatically configured versions of the wake user interface are generated by the computer system (e.g., based on parameters selected automatically by the computer system).
  • selection of the selectable option in the first user interface for navigating to the third user interface causes display of the first user interface to be replaced by the display of the third user interface (e.g., selection of option 609b ( Figure 6D) causes the device 100 to display the user interface 652 ( Figure 6K)).
  • Displaying, in the wake user interface selection mode, a user-selectable option to view a gallery of wake user interfaces corresponding to automatically configured versions of the wake user interface enables the user to consider multiple wake user interface options at the same time, thereby reducing the number of inputs needed to customize the wake user interface.
  • the third user interface that includes the plurality of representations of wake user interface corresponding to the plurality of automatically configured versions of the wake user interface includes (1040) one or more affordances for initiating corresponding wake user interface creation flows for creating new versions of the wake user interface.
  • the wake screen gallery includes one or more affordances for entering a multi-step wake screen creation flow (optionally multiple different affordances for entering different wake screen creation flows).
  • user interface 652 in Figure 6K enables the user to add and/or edit wake screen user interfaces to the set of wake screen user interfaces that device 100 rotates through periodically.
  • a user-selectable option for initiating a wake user interface creation flow or wizard helps the user through the process of creating a new wake user interface version, thereby reducing an amount of time needed to perform a particular operation on the device.
  • selection of the first representation of the wake user interface is based on (1042) a tap input in the sequence of one or more inputs that is detected on the first representation of the wake user interface
  • selection of the second representation of the wake user interface is based on a selection input (e.g., a tap input or other selection input) in the sequence of one or more inputs that is directed to (e.g., detected on or detected while attention is directed to) the second representation of the wake user interface.
  • the tap input detected on the representation that is displayed in the center region of the first user interface is recognized as a selection input that causes the version of the wake user interface corresponding to the selected representation to be displayed when the computer system exits the first user interface and returns to the wake user interface (e.g., a tap input directed to representation 615 ( Figure 6D) causes the wake screen user interface corresponding to representation 615 to be set as the current wake screen user interface).
  • the tap input detected on the representation that is displayed in a peripheral region of the first user interface is recognized as a navigation input that causes the selected representation to be moved to the central portion of the first user interface (e.g., tap input 610 on representation 611, Figure 6D), and, optionally, causes the version of the wake user interface corresponding to the selected representation to be displayed when the computer system exits the first user interface and returns to the wake user interface (e.g., optionally, another tap input is required on the representation when the representation is displayed in the central portion of the first user interface).
  • a navigation input that causes the selected representation to be moved to the central portion of the first user interface
  • the version of the wake user interface corresponding to the selected representation to be displayed when the computer system exits the first user interface and returns to the wake user interface (e.g., optionally, another tap input is required on the representation when the representation is displayed in the central portion of the first user interface).
  • the computer system while displaying the first user interface (e.g., including displaying a respective representation of the wake user interface as a currently selected representation of the wake user interface), the computer system detects (1044) a fifth user input corresponding to a request to edit a respective representation of the wake user interface (e.g., detecting a tap input on a "customize" button while the respective representation of the wake user interface is displayed in the center portion of the first user interface) from the first user interface; and in response to detecting the fifth user input corresponding to the request to edit the respective representation of the wake user interface, displays a first plurality of selectable options (e.g., in an overlay on the first user interface, or an editing user interface that replaces display of the first user interface) for changing a respective set of one or more wake user interface settings for a respective version of the wake user interface that corresponds to the respective representation of the wake user interface.
  • a fifth user input corresponding to a request to edit a respective representation of the wake user interface e.g., detecting a
  • the respective representation of the wake user interface is selected in accordance with the respective representation of the wake user interface being positioned in the center of the first user interface (and optionally displays other representations of the wake user interface on either side of the respective representation).
  • the user input is received on a user-selectable affordance (e.g., an edit button).
  • a user-selectable affordance e.g., an edit button.
  • user input 622 ( Figure 6F) on the Customize button causes device 100 display editing user interface 626 ( Figure 6G).
  • While in the wake user interface selection mode that presents a plurality of candidate wake user interface versions in response to a user input corresponding to a request to edit a respective candidate version, presenting a plurality of selectable options for changing one or more features and/or settings of the respective candidate version causes the device to automatically facilitate wake screen user interface editing, thereby reducing an amount of time needed to perform a particular operation on the device.
  • the computer system while displaying the first user interface, including concurrently displaying the first representation of the wake user interface and the first representation of the home user interface, the computer system detects (1046) a sixth user input corresponding to a request to view the first representation of the home user interface; and in response to detecting the sixth user input corresponding to the request to view the first representation of the home user interface, displays an expanded view of the first representation of the home user interface in the first user interface (including, e.g., moving the first representation of the home user interface into the center of the first user interface, increasing a size of the first representation of the home user interface, and/or moving the first representation of the home user interface from behind the first representation of the wake user interface, and optionally decreasing and/or moving the first representation of the wake user interface).
  • the first representation of the home user interface is initially displayed as at least partially occluded by the first representation of the wake user interface; and in response to the user input, the first representation of the home user interface is displayed without being occluded by the first representation of the wake user interface in the first user interface.
  • the device displays user interface 638 ( Figure 61).
  • the computer system while displaying the first user interface, including concurrently displaying the first representation of the wake user interface and the first representation of the home user interface, the computer system detects (1048) a seventh user input corresponding to a request for displaying a plurality of customization options (e.g., options for changing color scheme, options for changing font, and/or options for changing gradient).
  • a seventh user input corresponding to a request for displaying a plurality of customization options (e.g., options for changing color scheme, options for changing font, and/or options for changing gradient).
  • the computer system In response to detecting the seventh user input: in accordance with a determination that the first representation of the home user interface is positioned at a respective position (e.g., in the center or center region) of the first user interface, the computer system displays a first plurality of customization options for changing a first set of one or more home user interface settings for the first representation of the home user interface; and in accordance with a determination that the first representation of the wake user interface is positioned at the respective position of the first user interface, the computer system displays a second plurality of customization options for changing a first set of one or more wake user interface settings for the first representation of the wake user interface.
  • first plurality of customization options and the second plurality of customization options include at least one customization option (e.g., color, font, and/or gradient) that appears in both the first plurality and the second plurality of customization options.
  • While in the wake user interface selection mode displaying customization options for a respective wake user interface if the wake user interface is displayed at a particular position when a request for customization is received, or alternatively displaying customization options for the corresponding home user interface if the home user interface is at the particular position when the request for customization is received, causes the device to automatically enable customization for the user interface that has focus when the request for customization is received, thereby reducing an amount of time needed to perform a particular operation on the device.
  • the first plurality of customization options includes (1050) a first set of user-selectable options for configuring the first wake user interface background and the second plurality of customization options include a second set of user- selectable options for configuring the first home user interface background.
  • the respective sets of user-selectable option(s) for configuring the wake user interface background and/or the home user interface background includes one or more selectable photos, one or more selectable gradients, and/or one or more colors, that can be used as the background(s) of the wake user interface and/or home user interface.
  • the respective sets of user-selectable option(s) for configuring the wake user interface background and/or the home user interface background includes one or more selectable visual treatments, e.g., blur, transparency, and/or gradient that can be applied to a selected background image of the wake user interface and/or home user interface.
  • editing user interface 642 ( Figure 6J) includes an option for changing a color and/or gradient applied to the background, and includes an option for turning legibility blur on and/or off. Enabling the user to configure the background of a respective wake user interface and/or the corresponding home user interface while in the wake user interface selection mode reduces the number and extent of inputs needed to customize user interfaces on the device.
  • the computer system while displaying the first set of user-selectable options for configuring the first wake user interface background or the second set of user-selectable options for configuring the first home user interface background, the computer system detects (1052) an eighth user input that meets selection criteria (e.g., the eighth user input is a tap input on the respective user-selectable option, or the eighth user input is an air tap that is detected while a gaze input is on the respective user-selectable option).
  • selection criteria e.g., the eighth user input is a tap input on the respective user-selectable option, or the eighth user input is an air tap that is detected while a gaze input is on the respective user-selectable option.
  • the computer system in response to detecting the eighth user input: in accordance with a determination that a respective user-selectable option in the first set of user-selectable options is selected by the eighth user input, the computer system displays a first set of additional options associated with the respective user-selectable control function for configuring the first wake user interface; and in accordance with the determination that a respective user- selectable option in the second set of user-selectable options is selected by the eighth user input, the computer system displays a second set of additional options associated with the respective user-selectable control function for configuring the first home user interface.
  • the respective user-selectable option in the first set of user-selectable options for configuring the first wake user interface background and the respective user- selectable option in the second set of user-selectable options for configuring the first home user interface background include the same selectable option, and/or the first set of additional options and the second set of additional options include the same set of additional options.
  • the first set of additional options and/or the second set of additional functions include a color picker for selecting a color for the background and/or a photo picker for selecting a photo for the background.
  • the device 100 displays user interface element 630 ( Figure 6H) for changing a color (e.g., a font color and/or a background color).
  • a color e.g., a font color and/or a background color.
  • displaying additional options for modifying the particular feature causes the device to automatically enable additional levels of customization without displaying additional controls when not needed.
  • the first set of user-selectable options includes (1054) a first option for selecting a background (e.g., a solid color background, an image background, and/or a photo background) as the first wake user interface background for the wake user interface.
  • the second set of user-selectable options includes an option for selecting a background (e.g., a solid color background, an image background, and/or a photo background) as the first home user interface background for the home user interface.
  • option 628c Figure 6J
  • option 628c enables the user to view a photo picker (e.g., user interface 668, Figure 6N) to change a background of the home screen user interface. Enabling the user to configure the background of a respective wake user interface and/or the corresponding home user interface by choosing the background graphic reduces the number and extent of inputs needed to customize user interfaces on the device.
  • the second set of user-selectable options includes (1056) a second option for applying a respective visual effect (e.g., a blur, a transparency filter, a color filter, and/or a gradient filter) to the first wake user interface background.
  • the second set of user-selectable options includes an option for applying a respective visual effect (e.g., a blur, a transparency filter, a color filter, and/or a gradient filter) to the first home user interface background.
  • user interface 626 includes an option 628a for applying visual filter to the background of the wake screen user interface.
  • the computer system while displaying a first version of the wake user interface that corresponds to the first representation of the wake user interface, the computer system detects (1058) a ninth user input directed to a respective portion of the first version of the wake user interface.
  • the computer system displays the first version of the wake user interface in an editing view, wherein the first version of the wake user interface displayed in the editing view includes one or more selectable options for configuring one or more elements of the first version of the wake user interface (e.g., selectable options for configuring the date element, time element, the one or more user interface objects, and/or the background of the first version of the wake user interface).
  • the first version of the wake user interface displayed in the editing view includes one or more selectable options for configuring one or more elements of the first version of the wake user interface (e.g., selectable options for configuring the date element, time element, the one or more user interface objects, and/or the background of the first version of the wake user interface).
  • a user input 5028 is detected on a calendar complication, and in response to user input 5028, the device 100 displays editing user interface 565-3 (Figure 5AZ), including an option to modify the one or more complications of wake screen user interface 5026.
  • Displaying an editing view that enables the user to configure user interface objects in a wake user interface that are periodically updated with content from active applications makes more efficient use of the display area and reduces the number and extent of inputs needed for the user to view current relevant information about a state of the device.
  • the computer system while displaying a second version of the wake user interface that corresponds to the first representation of the wake user interface, the computer system detects (1060) a tenth user input directed to a respective portion of the second version of the wake user interface.
  • the computer system displays the second version of the wake user interface in an editing view, wherein the second version of the wake user interface displayed in the editing view includes a respective selectable option for adding one or more user interface objects that correspond to one or more applications and include respective content from the one or more applications and are updated periodically as information represented by the one or more user interface objects changes (e.g., the one or more user interface objects include one or more complications, widgets, and/or other similar user interface elements that correspond to different applications).
  • the second version of the wake user interface also includes one or more selectable options for configuring one or more elements of the second version of the wake user interface (e.g., selectable options for configuring the date element, time element, the one or more user interface objects, and/or the background of the first version of the wake user interface), for example reticles 568 and 569 ( Figure 5X).
  • the respective user-selectable affordance includes a plus symbol.
  • the respective user-selectable affordance is displayed in accordance with a determination that there are no user interface objects that correspond to respective applications and include respective content from the respective application that is updated periodically as information represented by the one or more user interface objects changes (e.g., a wake user interface that does not include complications is displayed with a "+" symbol to add complications, and a wake user interface that includes at least one complication is not displayed with a "+” symbol to add complications).
  • the respective selectable option for adding one or more user interface objects that correspond to one or more applications and include respective content from the one or more applications and are updated periodically as information represented by the one or more user interface objects changes is displayed in accordance with a determination that the second version of the wake user interface is of a first type (e.g., photo, color gradient, emoji, weather, nature, and/or astronomy type); and is not displayed if the second version of the wake user interface is of a second type (e.g., a portrait) different from the first type.
  • a first type e.g., photo, color gradient, emoji, weather, nature, and/or astronomy type
  • a second type e.g., a portrait
  • the device adds one or more complications, and overrides the default setting of not adding complications to a portrait mode background in response to detecting a user input at a predefined portion of the editing user interface (e.g., at the region above and/or below the time indication). Enabling the user to add, to a wake user interface, user interface objects that are periodically updated with content from active applications makes more efficient use of the display area and reduces the number and extent of inputs needed for the user to view current relevant information about a state of the device.
  • the first version of the wake user interface displayed (1062) in the editing view includes respective representations of one or more sets of recommended user interface objects that are selectable for inclusion in the first version of the wake user interface, wherein the recommended user interface objects correspond to respective applications and include respective content from the respective applications that is updated periodically as information represented by the recommended user interface objects changes.
  • each recommended user interface object is displayed with an indication of the respective application that contributed the user interface object in the editing view. For example, an application icon and/or badge is displayed with each complication to indicate which application is associated with the complication.
  • Figure 5AZ illustrates user interface object 570 that displays a plurality of complications, and indications of respective applications associated with each complication.
  • the computer system while displaying the first version of the wake user interface that corresponds to the first representation of the wake user interface in the editing view (or optionally, while displaying the second version of the wake user interface in the editing view), the computer system detects (1064) an eleventh user input directed to a respective portion of the first version of the wake user interface (or optionally, directed to a respective portion of the second version of the wake user interface).
  • the computer system In response to detecting the eleventh user input: in accordance with a determination that the eleventh user input is directed to a textual element of the first version of the wake user interface (e.g., the date element, the time and/or element), the computer system displays one or more selectable options for changing a font of the textual element displayed in the first version of the wake user interface, for example user interface element 586 ( Figure 5 AD) includes a plurality of style and/or color options to change a font of the time and/or date indication.
  • a textual element of the first version of the wake user interface e.g., the date element, the time and/or element
  • the one or more options to change a font includes an option to change a font (e.g., typeface), a font style (e.g., a font size, style, and/or weight), an option to change a font vibrancy, an option to change a color tint, and/or an option to change a luminance (e.g., light or dark).
  • the one or more options to change a font further control visual properties of complications, as described with reference to Figures 5AD-5AE.
  • changing a color of a font for the time element also causes the font of the date to be updated.
  • the device in response to detecting a user input that selects an option to change font properties for a first textual element, automatically changes font properties for a second textual element in accordance with the change to the font properties for the first textual element (e.g., a change to the time font causes a change to the date font).
  • the fonts are complimentary and/or the same font is applied to both textual elements (e.g., date and time). Enabling the user to configure the font of a textual element in a candidate wake user interface version using one or more displayed selectable options reduces the number and extent of inputs needed to customize user interfaces on the device.
  • the computer system while displaying the one or more selectable options for changing the font of the textual element displayed in the first version of the wake user interface, the computer system detects (1066) a twelfth user input selecting a respective one of the one or more selectable options for changing the font of the textual element.
  • the computer system changes the font of the textual element in a first manner in accordance with the selected respective one of the one or more selectable options for changing the font of the textual element; and changes the one or more visual properties of the one or more user interface objects that correspond to respective applications and include respective content from the respective application that is updated periodically as information represented by the one or more user interface objects changes that are displayed on the wake user interface, in the first version of the wake user interface.
  • the visual properties include a color scheme for the complications. For example, the complications are displayed in greyscale and/or as monochrome in accordance with the selection of a respective font color for the time and/or date element.
  • the complications are displayed with a color scheme that is automatically selected based on the background (e.g., wallpaper image) and/or based on the current settings for the font for the time and/or date element.
  • the same color scheme is used for the complications as the changed font color for the date and/or time element.
  • complimentary color schemes are used for the complications and the changed date and/or time elements. For example, in Figure 5AD, in response to user input 588, the font of the time and/or date indication of the wake screen user interface is updated to be displayed with “Style 5,” as illustrated in Figure 5AE.
  • the computer system displays (1068), in the first user interface, a respective user-selectable affordance (e.g., a "+" button in the upper right corner of the wake screen selector user interface, as illustrated in Figure 6D) for creating new versions of the wake user interface, wherein selection of the respective user-selectable affordance causes display of a third user interface that includes a set of selectable options for creating a new version of the wake user interface independent of the first version of the wake user interface.
  • the third user interface is a gallery user interface that includes options to create a new version of the wake user interface from different styles, display modes, and/or background types. While in the wake user interface selection mode, displaying a user-selectable affordance for adding another candidate version of the wake user interface reduces the number of inputs needed to customize the wake user interface, particularly if none of the listed options are of interest to the user.
  • the contacts, gestures, user interface objects and animations described above with reference to method 1000 optionally have one or more of the characteristics of the contacts, gestures, user interface objects and animations described herein with reference to other methods described herein (e.g., methods 900, 1100, 12000, 1300, 14000, 15000, 16000, and 17000). For brevity, these details are not repeated here.
  • FIGS 11 A-l 1H are flow diagrams illustrating method 1100 of for modifying a user interface in accordance with some embodiments.
  • Method 1100 is performed (1102) at a computer system (e.g., device 300, Figure 3, or portable multifunction device 80, Figure 1 A) that is in communication with a display generation component (e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like) (and optionally, the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as lidars, depth sensors, and/or distance sensors)
  • a display generation component e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a head-mounted display, or the like
  • method 1100 provides a method for customizing a wake screen user interface.
  • displaying a user interface for editing a respective editable object in a first version of the wake screen in response to a first type of input versus switching between different versions of the wake screen in response to a second type of input, reduces the number of inputs needed to make different types of configuration changes to the wake screen.
  • the computer system displays (1104), via the display generation component, a first user interface for configuring a wake user interface (e.g., an editing user interface 565 ( Figure 5X) for configuring a wake screen user interface, for configuring a home screen user interface associated with a wake screen user interface, and/or for configuring a lock screen user interface), wherein a respective version of the wake user interface includes a respective background and a respective plurality of editable user interface objects (e.g., a plurality of user selectable objects, such as complications, widgets, shortcuts and/or a plurality of user interface objects such as a time object, and/or a date object) overlaying the respective background, and wherein the first user interface displays at least a first representation of a first version of the wake user interface illustrating a first plurality of editable user interface objects overlaying a first background (e.g., the first user interface displays the first version of the wake user interface in an editing view).
  • Figure 5AL illustrates an editing user interface 565-2
  • the computer system While displaying the first user interface, the computer system detects (1106) a first input directed to the first user interface. For example, while displaying editing user interface 565-2 ( Figure 5AO), the device 100 detects a user input 5014-1.
  • the computer system displays (1108) a second user interface for editing a first user interface object of the first plurality of editable user interface objects (e.g., editing on the basis of the first user interface object as shown in the first version of the wake user interface), wherein the first user interface object is selected in accordance with a location of the first input (e.g., a user interface object located at or within a threshold range of a location of the tap input or double tap input is selected to be edited, and the second user interface provides various controls of editing one or more aspects of the selected user interface object
  • the device 100 displays option for editing visual properties (e.g., font properties) of the time indication, as illustrated in Figure 5AD.
  • option for editing visual properties e.g., font properties
  • the computer system updates (1110) the first user interface to replace display of the first representation of the first version of the wake user interface with display of a second representation of a second version of the wake user interface.
  • the second representation of the second version of the wake user interface includes a second plurality of editable user interface objects (e.g., widgets and/or complications, time object and/or date object) overlaying a second background that is different from the first background (e.g., wallpaper, image, and/or photo).
  • the second plurality of editable user interface objects is different from the first plurality of editable user interface objects (e.g., one or more objects in the second plurality of editable user interface objects have a different appearance and/or functionality than one or more corresponding objects in the first set of editable user interface objects).
  • Figure 5AO in response to user input 5014-1 the background and/or complications are changed from the wake screen user interface illustrated in Figure 5 AO to the wake screen user interface illustrated in Figure 5AT.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of a second version of the wake user interface includes (1112) changing one or more font properties of system generated text that is displayed in the wake user interface from a first set of font properties shown in the first representation of the first version of the wake user interface to a second set of font properties shown in the second representation of the second version of the wake user interface.
  • the system generated text includes text in a date object and/or time object in the wake user interface.
  • the system generated text further includes subject lines, and/or object names in system-generated objects that are displayed in the wake user interface (e.g., notification history, system prompts, and/or alerts).
  • the system generated text further includes text in complications or widgets included in the wake user interface.
  • the one or more font properties include one or more of: a typeface, a color, a size, and a weight of the font of the system generated text.
  • the time and/or date indication is updated to be displayed with a different font, font style, and/or font color (e.g., optionally in accordance with a determination that the user has not explicitly modified the time and/or date indication in the current and/or a previous editing session).
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1114) shifting the second representation of the second version of the wake user interface and moving system generated text on the wake user interface in accordance with the first input (e.g., moving the system generated text in the second representation of the second version of the wake user interface and/or moving the system generated text in the first representation of the first version of the wake user interface, in accordance with the first input that meets the second criteria).
  • the computer system displays a sliding visual effect over the first background to generate the second background.
  • the computer system moves the system generated text in a direction of the first input (e.g., a swipe input to the left moves the system generated text of the first representation of the first version of the wake user interface to the left to appear as if it is scrolling off the display area of the display generation component, while the system generated text of the second representation of the second version of the wake user interface slides onto the display generation component from right to left), as illustrated in Figures 5AO-5AT.
  • the system generated text is not moved. For example, in accordance with a determination that the user has edited font properties of the system generated text during the current editing session and/or in a past editing session, the system generated text is not moved when the background of the wake user interface is updated.
  • Switching between different versions of the wake screen in response to an input directed to the wake screen configuration user interface, by shifting the new version including system generated text onto the display in accordance with the input (e.g., by an amount controlled by the movement of the input) causes the device to automatically provide the user with a preview of the new version before the end of the input and optionally gives the user an opportunity to change their mind.
  • the system generated text includes (1116) one or more of a date indication, a time indication, and/or one or more user interface objects that correspond to a first plurality of applications and include respective content from the first plurality of applications and are updated periodically as information represented by the first plurality of user interface objects changes (e.g., complications, or widgets).
  • the wake screen user interface in Figure 5AO includes a date indication, a time indication, and complications.
  • the device When switching between different versions of the wake screen, including in the new version system generated text such as the date, the time, and/or one or more user interface objects that are periodically updated with content from active applications causes the device to automatically present current information that is likely to be of interest to the user in an efficient way in the display area, thereby reducing the number of inputs and amount of time that is needed for the user to view the information.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1118): in accordance with a determination that system generated text in the first version of the wake user interface meets editing criteria (e.g., criteria that are met in accordance with a determination that a user has edited the system generated text with a predetermined time period, during the current editing session, and/or in a previous editing session), shifting at least the second background of the second representation of the second version of the wake user interface in accordance with the first input, while maintaining display position of the system generated text in the first representation of the first version of the wake user interface (e.g., moving the background in the second representation of the second version of the wake user interface underneath the system generated text of the first representation of the first version of the wake user interface, in accordance with the first input that meets the second criteria) (e.g., the system generated text that has been edited becomes part of the second representation of the second version of the wake
  • the background of the wake screen user interface is updated without moving the date and/or time indications because the user has edited the date and/or time indications.
  • maintaining the display position of system generated text that the user has previously edited causes the device to automatically present current information that the user, through interacting with the system generated text, has already indicated is of interest to them, thereby reducing the number of inputs and amount of time that is needed for the user to view the information.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1120) replacing display of the first background with the second background.
  • the background in Figure 5AO is distinct from the background in Figure 5AT. Enabling a user to direct an input to a wake screen configuration user interface to switch between different versions of the wake screen, including to a version with a different background, reduces the number of inputs needed to make different types of configuration changes to the wake screen.
  • the first background includes (1122) at least a first portion of the first background (e.g., foreground portion, central portion, top portion, left portion, one or more foreground objects, or a main subject) and a second portion of the first background (e.g., background portion, peripheral portion, bottom portion, right portion, one or more far away objects, and/or one or more secondary subjects), and updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes changing at least a first display property (e.g., color scheme, blur radius, opacity, and/or luminance) of the first portion of the first background (and, optionally, maintaining an appearance of the second portion of the first background in the first representation of the first version of the wake user interface and/or changing at least a second display property (e.g., different from the first display property, or same as the first display property but by a different amount or manner of change) of
  • a first display property e.
  • the background in Figure 5AO includes a foreground portion that includes an individual (e.g., the subject) and a background color behind the individual.
  • the foreground portion does not change from the foreground in Figure 5AO, but the background color changes.
  • updating the first user interface to change a display mode of the wake user interface includes changing a color scheme, a level of blur, a level of luminance, and/or a level of opacity of one portion of the current background of the wake user interface in one manner and changing a color scheme, a level of blur, a level of luminance, and/or a level of opacity of another portion of the current background in a different manner and/or maintaining the appearance of other portions of the current background.
  • the updated background becomes the background of the second representation of the second version of the wake user interface (e.g., the new display mode for the wake user interface).
  • the computer system while displaying the first user interface, including the first representation of the first version of the wake user interface, the computer system detects (1124) a second input.
  • the computer system detects (1124) a second input.
  • the third criteria include a requirement that the second input is a pinch gesture, and/or a two-finger translation gesture, optionally a requirement that the second input is directed to the first background (e.g., region of the first representation of the first version of the wake user interface that is not occupied by a user interface object)
  • the computer system changes one or more spatial properties of the first background in the first representation of the first version of the wake user interface (e.g., changing a zoom level, dimensions, and/or center of the image of the first background in the first representation of the first version of the wake user interface).
  • an input meeting the third criteria can be directed to a second representation of a second version of the wake user interface and change the zoom level, center, and/or cropping style of the second background in the second representation of the second version of the wake user interface.
  • the respective background in a respective representation of a respective version of the wake user interface displayed in the first user interface is a photo
  • the second input causes the photo to be cropped, recentered, and/or zoomed in the respective representation in accordance with the second input.
  • a pinch gesture inwards e.g., the two contact points move toward each other
  • a pinch gesture outwards e.g., the two contact points move away from each other
  • a translation gesture e.g., a geometric center of the two contacts moving in the same direction
  • a user input 5004 e.g., an outward pinch gesture
  • user input 5006 e.g., an inward pinch gesture
  • the device 100 changes a size and/or crop of the background image (e.g., the subject of the portrait-style photo).
  • Enabling a user to direct a different type of input to a wake screen configuration user interface to resize, rescale, and/or reposition at least a portion of the background of the wake screen reduces the number of inputs needed to make different types of configuration changes to the wake screen without displaying additional controls.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1125): (optionally, in accordance with a determination that the first user interface is displayed in a first display mode (e.g., a photos mode) (e.g., a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode)) changing a first set of one or more photo visual effects displayed in the first representation of the first version of the wake user interface to a second set of one or more photo visual effects displayed in the second representation of the second version of the wake user interface (e.g., while maintaining the same basic image in the background).
  • a first display mode e.g., a photos mode
  • a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the photo visual effects are effects applied to a photograph that is used as the background of the wake user interface.
  • the first set of one or more photo visual effects includes original coloring and the second set of one or more visual effects includes duotone, studio color, studio black/white, sepia and/or other display effects that change a tone and/or color of the image in the first background.
  • the computer system switches the photo visual effect applied to the background image of the first representation of the first version of the wake user interface to generate the updated representations of the updated versions of the wake user interface one by one.
  • user input 5020 causes device 100 to change a visual effect, such by applying a filter and/or changing a background color, of the background from a first background of the wake screen user interface (Figure 5AV) to a second background of the wake screen user interface ( Figure 5AW).
  • the visual effect e.g., filter
  • the subject of the photo e.g., the individual
  • the background of the photo e.g., the individual
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1126): (optionally, in accordance with a determination that the first user interface is displayed in a second display mode (e.g., a portrait mode) (e.g., a second mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode)) replacing display of a first photo in the first representation of the first version of the wake user interface with display of a second photo, distinct from the first photo, in the second representation of the second version of the wake user interface.
  • a second display mode e.g., a portrait mode
  • a second mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the first photo is displayed as the first background and the second photo is displayed as the second background.
  • the first photo and the second photo include a main subject (e.g., an individual, a pet, and/or a landmark) and one or more secondary subjects (e.g., background objects and/or environment).
  • Figures 5BC-5BD illustrate changing a background image, including a foreground portion (e.g., the moon) and a background environment (e.g., the color and/or pattern behind the moon). Enabling a user to direct an input to a wake screen configuration user interface to switch between different versions of the wake screen, including to a version that shows a different photo or portrait, reduces the number of inputs needed to make different types of configuration changes to the wake screen.
  • the first photo includes a first main portion (e.g., a person, a pet, and/or a landmark) and one or more first peripheral portions (e.g., peripheral and/or background objects, and/or environment) and the second photo includes (1128) a second main portion (e.g., a person, a pet, and/or a landmark) and one or more second peripheral portions (e.g., peripheral and/or background objects, and/or environment), and wherein the first background includes the first main portion with the one or more first peripheral portions replaced with a first texture (e.g., a first color, a first pattern, and/or a first color gradient), and the second background includes the second main portion with the one or more second peripheral portions replaced with a second texture (e.g., a second color, a second pattern, and/or a second color gradient).
  • a first texture e.g., a first color, a first pattern, and/or a first color gradient
  • the second background includes the second main portion with the
  • the first texture and/or the second texture include a translucency gradient (e.g., increasing translucency from the center to the edge, from top to bottom, or vice versa).
  • the wake user interface displayed in the portrait mode includes a photo of an individual, wherein the individual is maintained, for example in the foreground, and background objects and/or colors in the original photo are replaced with a system-generated texture and/or color.
  • the system-generated texture and/or color has a translucency gradient. For example, the individual in the photo is displayed over a colored background that is not part of the original photo.
  • the system-generated texture and/or color is selected in accordance with one or more features (e.g., colors, color tones, or background objects) that are present in the original photo.
  • features e.g., colors, color tones, or background objects
  • the foreground e.g., the subject
  • the background having a color, texture and/or gradient is updated.
  • replacing the background of a respective photo or portrait with a visual texture causes the device to automatically give visual emphasis to objects in the foreground of the photo or portrait, thereby reducing an amount of time needed to choose between different versions of the wake screen by helping the user focus.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1130): (optionally, in accordance with a determination that the first user interface is displayed in a third display mode (e.g., an emoji mode) (e.g., a third mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode)) replacing display of a first pattern of one or more selected emojis displayed in the first representation of the first version of the wake user interface with display of a second pattern of the one or more selected emojis, distinct from the first pattern, in the second representation of the second version of the wake user interface.
  • a third display mode e.g., an emoji mode
  • a third mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the device provides an input region (e.g., an input region at the top of the emoji keyboard) that allows the user to select up to a predefined number of emojis to include in the respective wake user interface.
  • the device provides input slots for up to three (e.g., one, two, or three) distinct emojis in the input region where the emojis entered into the input region are to be displayed in a pattern on the wake user interface.
  • different versions of the wake user interface include different patterns (e.g., a swirl pattern, a grid pattern, and/or a line pattern) of the same emoji.
  • different versions of the wake user interface include different sets of emojis, as described with reference to Figure 6G. Enabling a user to direct an input to a wake screen configuration user interface to switch between different versions of the wake screen, including to a version that includes a different set or pattern of emojis, reduces the number of inputs needed to make different types of configuration changes to the wake screen.
  • updating the first user interface to replace display of the first representation of the first version of the wake user interface with display of the second representation of the second version of the wake user interface includes (1131): (optionally, in accordance with a determination that the first user interface is displayed in a fourth display mode (e.g., an astronomy mode) (e.g., a fourth mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode)) replacing display of a first type of celestial object displayed in the first representation of the first version of the wake user interface with display of a second type of celestial object, distinct from the first type of celestial object, in the second representation of the second version of the wake user interface.
  • a fourth display mode e.g., an astronomy mode
  • a fourth mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the first type and the second type of celestial objects are selected from the group consisting of: a moon, a planet, earth, and/or an orrery that includes a plurality of celestial objects and illustrates a relationship between the motion and/or location of the different celestial objects represented by the orrery.
  • Figures 5BA-5BD illustrate different phases of a moon as the background image of the wake screen user interface. Enabling a user to direct an input to a wake screen configuration user interface to switch between different versions of the wake screen, including to a version that shows a different celestial object or scene, reduces the number of inputs needed to make different types of configuration changes to the wake screen.
  • the first representation of the first version of the wake user interface is (1132) displayed in a motion mode, and displaying the first background (e.g., in the first representation of the first version of the wake user interface, and/or in the first version of the wake user interface) includes displaying an animated sequence of frames selected from a video.
  • the computer system also generates and displays additional interpolated frames to generate a motion visual effect, as described with reference to Figure 5AI.
  • Enabling a wake screen user interface to be displayed in a motion mode in which the wake screen includes a video clip that plays in the background of the wake screen causes the device to automatically enable the user to add a favorite video clip to the wake screen, thereby reducing the number of inputs and amount of time needed to navigate on the device to view the favorite video clip.
  • the computer system displays (1134), in the first user interface for configuring the wake user interface, a respective user interface object that indicates availability of one or more additional versions of the wake user interface, including the second version of the wake user interface.
  • the respective user interface object is a series of paging dots.
  • indication 574 ( Figure 5AJ) indicates additional views of the wake screen user interface are available.
  • the respective user interface object changes its appearance, e.g., highlights a different paging dot in a series of paging dots, in response to an input that meet the second criteria, to indicate that a next, different version of the wake user interface is now displayed in response to the input.
  • the device navigates between the one or more additional versions, including the second version, in response to user inputs that correspond to a user’s requests to navigate between the different versions of the wake user interface. For example, in response to one or more swipe inputs, or other navigation inputs directed to the different versions of the wake user interface, the device continues switching to additional versions of the wake user interface (e.g., 3 or more versions of the wake user interface are available to be viewed in the wake screen configuration user interface). Displaying, in a wake screen configuration user interface, an indication of whether additional versions of the wake user interface are available for previewing and optionally selection provides feedback about a state of the device.
  • the first user interface object includes (1136) system generated text (e.g., the first user interface object includes a time element, a date element, and/or system prompt, shown in the first representation of the first version of the wake user interface) and a respective user interface object (e.g., reticles, bounding boxes, and/or highlighting) is displayed at a respective location that corresponds to the first user interface object in the first user interface to indicate that the first user interface object is editable in the first user interface (e.g., in response to the first input that meets the first criteria and when the first input is directed to the respective location that corresponds to the first user interface object including system generated text).
  • system generated text e.g., the first user interface object includes a time element, a date element, and/or system prompt, shown in the first representation of the first version of the wake user interface
  • a respective user interface object e.g., reticles, bounding boxes, and/or highlighting
  • displaying the second user interface for editing the first user interface object includes displaying a user interface for configuring the system generated text included in the first user interface object.
  • a first reticle is displayed at a first location that corresponds to a first portion of the system generated text (e.g., a time indication) and a second reticle is displayed at a second location that corresponds to a second portion of the system generated text (e.g., a date indication).
  • only one reticle is displayed (e.g., the reticle for the time is displayed without displaying the reticle for the date).
  • the second user interface for configuring the system generated text includes one or more options for changing font properties, a color of the font, and/or a type of information that is displayed (e.g., day of the week, month, date, and/or year). Displaying, in a wake screen configuration user interface, an indication that system-generated text in the wake screen is editable provides feedback about a state of the device that reduces the number of inputs needed to edit system information.
  • the first background of the first representation of the first version of the wake user interface includes (1138) a first photo comprising a first main portion (e.g., a person, a pet, and/or a landmark) and one or more first peripheral portions (e.g., one or more background objects, secondary objects, and/or environment) (e.g., the first main portion is displayed at a first depth, the first peripheral portion is displayed at a second depth different from (e.g., larger than) the first depth, in the first representation of the first version of the wake user interface, as the first background of the first version of the wake user interface).
  • a first photo comprising a first main portion (e.g., a person, a pet, and/or a landmark) and one or more first peripheral portions (e.g., one or more background objects, secondary objects, and/or environment)
  • the first main portion is displayed at a first depth
  • the first peripheral portion is displayed at a second depth different from (e.g., larger than) the first depth, in
  • the first user interface object Prior to detecting the first input, the first user interface object is displayed behind the first main portion of the first photo in the first representation of the first portion of the wake user interface (e.g., the first user interface object is partially occluded by the first main portion of the first photo in the first representation of the first version of the wake user interface) (optionally, the first user interface object is displayed in front of other peripheral portions of the photo in the first representation of the first version of the wake user interface).
  • the first user interface object is displayed in front of the first main portion of the first photo in the first representation of the first portion of the wake user interface while displaying the second user interface for editing the first user interface object (optionally, the main portion of the first photo is displayed at a visual depth that is smaller than the other peripheral portions of the photo).
  • the individual e.g., the main portion of the photo
  • the time indication after detecting one or more user inputs.
  • an indication that system-generated text in a wake screen that includes a photo is editable, displaying the indication about the system-generated text behind a main portion of the photo until a user input selecting the system-generated text for editing is detected, and then displaying the system-generated text in front of the main portion of the photo in response, causes the device to automatically give visual emphasis to the main portion of the photo while providing feedback about a state of the device that reduces the number of inputs needed to edit system information, and then to automatically give visual emphasis to the portion of the wake screen with which the user has indicated intent to interact.
  • displaying the second user interface for editing the first user interface object of the plurality of editable user interface objects includes (1140): in accordance with a determination that the first user interface object includes system generated text (e.g., in accordance with a determination that the first user interface object is a time indication, a date indication, a set of complications, and/or a system generated prompt) (e.g., in accordance with a determination that the first input is directed to a respective object that is displayed with a reticle, a bounding box, or highlight that indicates the respective object is editable), displaying a first plurality of user-selectable color options (e.g., contrasting colors, complimentary colors, and/or matching colors) that are selected based on the first background of the first version of the wake user interface (e.g., based on the colors detected in the first background).
  • system generated text e.g., in accordance with a determination that the first user interface object is a time indication, a date indication, a set of complications, and/
  • matching and/or complimentary color options are selected based on colors detected in the first background of the first version of the wake user interface. In some embodiments, matching and/or complimentary color options are selected based on colors detected in the second background of the second version of the wake user interface, if an input meeting the first criteria is detected on objects containing system generated text in the second representation of the second version of the wake user interface. In some embodiments, contrasting color options are provided based on the colors in the background of the currently displayed representation of a version of the wake user interface.
  • the computer system in addition to displaying the color options, also provides a selectable option in the second user interface that, when selected, causes display of a full set of color options that is independent of the colors in the background of the currently displayed representation of a version of the wake user interface.
  • a selectable option in the second user interface that, when selected, causes display of a full set of color options that is independent of the colors in the background of the currently displayed representation of a version of the wake user interface.
  • user interface object 586 includes a plurality of selectable color options for the user to select to change a color of the text in the wake screen user interface.
  • displaying a palette of user- selectable color options for the system-generated text that are based on the background of the wake screen causes the device to automatically suggest color options that are intelligently generated so as to complement the wake screen, thereby reducing the number of inputs and amount of time needed for the user to select a text color.
  • displaying the second user interface for editing the first user interface object of the plurality of editable user interface objects includes (1142): in accordance with a determination that the first user interface object includes system generated text (e.g., in accordance with a determination that the first user interface object is a time indication, a date indication, a set of complications, and/or a system generated prompt) (e.g., in accordance with a determination that the first input is directed to a respective object that is displayed with a reticle, a bounding box, or highlight that indicates the respective object is editable), displaying respective user-selectable options for changing one or more visual properties (e.g., translucency, opacity, luminance, contrast, brightness, and/or saturation) of the system generated text.
  • visual properties e.g., translucency, opacity, luminance, contrast, brightness, and/or saturation
  • the computer system in response to the user selecting (e.g., using a tap input) one of the options for the changing the one or more visual properties to edit the corresponding visual property (e.g., to change translucency, brightness, and/or saturation), the computer system displays a value selector (e.g., a slider control, or a radial button control) for the user to select the desired value(s) for the corresponding visual property.
  • a value selector e.g., a slider control, or a radial button control
  • the user interface object 586 optionally includes a setting for changing a translucency, opacity, luminance, contract, brightness and/or saturation of the text in the wake screen user interface.
  • displaying a palette of user-selectable options for one or more visual properties of the system-generated text causes the device to automatically suggest text display options, thereby reducing the number of inputs and amount of time needed for the user to edit the appearance of the text.
  • the computer system displays (1144), in the first user interface, one or more respective selectable options (e.g., a set of selectors corresponding to different rotation frequencies) for configuring a frequency for switching the currently displayed photo in the first background, as described with reference to the “Smart Album” wake screen user interface in Figure 6K.
  • a photos mode e.g., a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, and/or emoji mode
  • the computer system displays (1144), in the first user interface, one or more respective selectable options (e.g., a set of selectors corresponding to different rotation frequencies) for configuring a frequency for switching the currently displayed photo in the first background, as described with reference to the “Smart Album” wake screen user interface in Figure 6K.
  • the currently displayed photo is part of a smart album or other collection of a plurality of photos.
  • the frequency is based on a predefined amount of time elapsing.
  • at least one selectable frequency is based on a current time (e.g., a time of day and/or a day of the week).
  • one or more selectable frequencies are based on one or more events of the computer system, such as an inactive period (e.g., after being locked) or in accordance with focus mode settings (e.g., a do not disturb (DND) mode, a sleep mode, or a work mode).
  • DND do not disturb
  • enabling a user to select a frequency with which to switch the currently displayed photo reduces the number of inputs and amount of time needed to customize the wake screen.
  • the computer system detects (1146) a respective plurality of user inputs that correspond to a request to add, to the first version of the wake user interface (e.g., by adding, to the first representation of the first version of the wake user interface displayed in the first user interface), one or more user interface objects that correspond to a plurality of applications and include respective content from the plurality of applications and that are updated periodically as information represented by the plurality of user interface objects changes (e.g., a request to add one or more complications corresponding to different applications and include content that is periodically updated as the information represented by the complications changes).
  • the computer system adds the one or more user interface objects to the first representation of the first version of the wake user interface in the first user interface.
  • a determination that the first background of the first version of the wake user interface includes a first photo e.g., in accordance with a determination that the first version of the wake user interface corresponds to a first display mode (e.g., a photos mode) (e.g., a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, emoji mode, and so on) in which the first background includes a currently displayed photo selected from a plurality of photos (e.g., photos from a selected album or photo set)), and that a main subject (e.g., a person, a pet, a landmark, and/or a central portion) of the first photo overlaps with at least one of the one or more user interface objects that are added to the first representation of the first version of the wake user
  • a first display mode e.g., a
  • the device 100 automatically adjusted a size of the individual of the portrait-style photo as to not overlap with the complications.
  • the device adjusts the size of the individual in response to a user’s input that corresponds to a request to resize the portion of the background corresponding to the individual, even if the individual will overlap the complications, as illustrated in Figures 5AJ- 5AL.
  • a wake screen that includes a photo in the background
  • resizing, rescaling, and/or repositioning the photo as needed so that the main subject of the photo and the one or more user interface objects do not overlap causes the device to automatically optimize the arrangement of various elements of the wake screen to reduce truncation or occlusion and make more efficient use of the display area.
  • the computer system in conjunction with changing the one or more spatial properties of the first photo in the first background (e.g., changing a zoom level, a cropping dimensions, a center, and/or a size of the first photo) such that the main subject of the first photo does not overlap with the one or more user interface objects in the first representation of the first version of the wake user interface, and in accordance with a determination that the main subject overlaps with system generated text in the first representation of the first version of the wake user interface, the computer system removes (1148) a visual effect applied to the main subject that adjusts a perceived depth of the main subject in the first photo (e.g., the first photo is initially displayed with a depth visual effect which places the main subject in front of the system generated text (e.g., before the complications were added), and the depth visual effect is removed if the main subject is moved and/or resized in the first photo to avoid overlapping with the newly added complications).
  • a visual effect applied to the main subject that adjusts a perceived depth of the main subject in the first photo
  • a perceived depth of the subject is automatically removed after the complications have been added.
  • removing a visual effect of depth in the photo and/or between the photo and other elements in the wake screen when resizing, rescaling, and/or repositioning the photo as needed so that the main subject of the photo and the one or more user interface objects do not overlap causes the device to automatically optimize the arrangement of various elements of the wake screen to reduce truncation or occlusion and make more efficient use of the display area.
  • changing the one or more spatial properties of the first photo in the first background is performed (1150) in accordance with a determination that a user has not modified the one or more spatial properties of the first photo.
  • the computer system in accordance with a determination that the user has specified a size, center, and/or dimensions of the first photo (e.g., the user manually cropped the first photo) in the current editing session, and/or in a previous editing session, the computer system forgoes changing the size, center, and/or dimensions of the first photo in the first background, even if the main subject would overlap with the one or more newly added complications. For example, after a user has manually edited the spatial properties of the first photo, the computer system forgoes automatically changing the spatial properties of the first photo (and/or the main subject) for other rules.
  • the computer system forgoes automatically changing the spatial properties of the first photo (and/or the main subject) for other rules.
  • a feature e.g., in a current editing session and/or a previous editing session
  • that feature is not automatically updated, or changed, by the device 100.
  • modifying one or more spatial properties of the photo e.g., by resizing, rescaling, and/or repositioning the photo
  • the main subject of the photo and the one or more user interface objects do not overlap
  • only if a user has not previously modified the spatial properties of the photo causes the device to automatically defer to previously-received user specification when determining how to arrange various elements of the wake screen to reduce truncation or occlusion and make more efficient use of the display area.
  • the computer system displays (1152), in the first user interface, a respective user- selectable option for specifying one or more rules for automatically selecting the plurality of photos (e.g., rather than manually selecting the photos using selection input(s) directed to one or more individual photos or sets of photos) to display in the first background of the first version of the wake user interface.
  • a photos mode e.g., a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the computer system displays (1152), in the first user interface, a respective user- selectable option for specifying one or more rules for automatically selecting the plurality of photos (e.g., rather than manually selecting the photos using selection input(s) directed to one or more individual photos or sets of photos) to display in the first background of the first version of the wake user interface.
  • the user selects rule(s) to include one or more types of photos (e.g., one or more individuals, pets, and/or albums) and/or selects rule(s) to exclude one or more types of photos (e.g., photos without people, photos of objects, photos of webpages, text, and/or screenshots).
  • the user selects rules that define a time period of photos (e.g., the creation date and/or most recent editing date) to display.
  • the user selects rule(s) that define an orientation of photos to be included and/or excluded (e.g., include portrait orientation photos and exclude landscape orientation photos). For example, in Figure 6K, a “Smart Album” is provided as an option under the Featured Faces.
  • a “Smart Album” enables the user to select certain types of photos to be included in the Smart Album wake screen user interface, such as pictures of a pet, pictures from a particular album, and/or pictures of an individual.
  • certain types of photos to be included in the Smart Album wake screen user interface such as pictures of a pet, pictures from a particular album, and/or pictures of an individual.
  • For a wake screen user interface that includes a photo in the background enabling a user to specify one or more rules for automatically selecting candidate photos to display in the background of the wake screen reduces the number of inputs and amount of time needed to select the candidate photos.
  • the computer system displays (1154) the currently displayed photo with a first aspect ratio in the first version of the wake user interface, wherein the first aspect ratio is distinct from a second aspect ratio corresponding to the first photo stored in a photo library of the computer system.
  • a photos mode e.g., a first mode of a plurality of modes such as photos mode, portrait mode, astronomy mode, weather mode, or emoji mode
  • the computer system displays (1154) the currently displayed photo with a first aspect ratio in the first version of the wake user interface, wherein the first aspect ratio is distinct from a second aspect ratio corresponding to the first photo stored in a photo library of the computer system.
  • a different aspect ratio is used for photos on the wake user interface than the photos stored and/or viewed in the photo library.
  • the aspect ratio of the first photo in the photo library is a square
  • the aspect ratio of the first photo displayed on the wake user interface is a rectangle such that it fits a size of the display.
  • an aspect ratio of the portrait-style photo is distinct from the stored photo (e.g., as viewed in a photos application).
  • displaying the photo in the wake screen with an aspect ratio that is different from an aspect ratio with which the photo is displayed in a different type of user interface of the computer system, such as in a photo library causes the device to automatically optimize the arrangement of various elements in the wake screen.
  • displaying the second user interface for editing the first user interface object of the first plurality of editable user interface objects includes (1156): in accordance with a determination that the location of the first input is a first location in the first user interface (e.g., the first user interface object is an object of a first type at the first location), displaying a first set of selectable options for editing the first user interface object; and in accordance with a determination that the location of the first input is a second location in the first user interface (e.g., the first user interface object is an object of a second type at the second location) different from the first location in the first user interface, displaying a second set of selectable options for editing the first user interface object, the second set of selectable options being different from the first set of selectable options.
  • the second user interface includes a respective editing panel that is specific to the user interface object that has been selected for editing by the location of the first input in the first user interface. If the location of the first input is the location of a time element or a date element, a first set of selectable options including options to edit the font and/or font color of the time element, and/or a format of the time element is displayed; and if the location of the first input is the location of a complication below the time element, a second set of selectable options including options to change the style of the complication, the information to be included in the complication, and/or the size and format of the complication.
  • the computer system while displaying the second user interface that includes the respective editing panel that is specific to the user interface object that has been selected by the first input, the computer system detects another input directed to a different location that corresponds to a second user interface object in the first user interface (e.g., some elements of the first user interface remain visible and selectable while the second user interface is displayed); and in response to detect the additional input, the computer system updates the second user interface to indicate selection of the second user interface object and deselection of the first user interface object, and displays an editing panel with a set of selectable options that is specific to the second user interface object (and ceases to display the editing panel with the set of selectable options that is specific to the first user interface object).
  • another input directed to a different location that corresponds to a second user interface object in the first user interface e.g., some elements of the first user interface remain visible and selectable while the second user interface is displayed
  • the computer system updates the second user interface to indicate selection of the second user interface object and deselection of the first user interface object, and displays
  • the user interface element 570 (e.g., as illustrated in Figure 5AC) for editing complications that are positioned below the indication of the time is distinct from the user interface element 5090 (e.g., illustrated in Figure 5AX4) for editing complications (e.g., textual complications) that are positioned above the indication of the time.
  • Enabling a user to direct inputs to objects at different locations in a wake screen configuration user interface to bring up different sets of selectable options enables the user to customize different objects in the wake screen and causes the device to automatically display a set of selectable options that is appropriate for the particular object that the user wants to edit.
  • the first user interface object is (1158) a user interface object that corresponds to a respective application, that includes respective content from the respective applications (e.g., the first user interface object is a complication and/or widget that corresponds to an application), and is updated periodically as information represented by the first user interface object changes
  • displaying the second user interface for editing the first user interface object of the first plurality of editable user interface objects includes: in accordance with a determination that the location of the first input is the first location in the first user interface (e.g., the first user interface object is a textual complication displayed above the time element), displaying the first set of selectable options (e.g., font, font size, and/or font color) including at least one selectable option for editing the first user interface object in a first format (e.g., textual format, and/or simplified format) (e.g., user interface element 5090 in Figure 5AX4 is displayed to edit the complications above the time indication); and in accordance with
  • user interface element 5082 illustrated in Figure 5 AX3 enables the user to modify a size of one or more complications that are positioned below the time indication.
  • Enabling a user to direct inputs to objects, which are periodically updated with content from associated active applications, at different locations in a wake screen configuration user interface to bring up different sets of selectable options enables the user to customize different objects in the wake screen and causes the device to automatically display a set of selectable options that is appropriate for the particular object that the user wants to edit.
  • displaying the second user interface for editing the first user interface object of the first plurality of editable user interface objects includes (1160): in accordance with a determination that the first user interface object includes system generated text (e.g., in accordance with a determination that the first user interface object is a time element, or a date element), displaying a third set of selectable options for editing a font and/or font color of the system generated text.
  • the third set of selectable options includes a set of vibrant materials that can be selected to use as the font color of the system generated text.
  • the third set of selectable options includes black or white font color depending on the visual properties (e.g., brightness, saturation level, and/or luminance) of the respective background currently used in the first user interface.
  • the third set of selectable options is also updated in accordance with the visual properties of the changed background.
  • the third set of selectable options include colors that are selected based on the colors that are automatically detected from the respective background that is currently used in the first user interface (e.g., colors that are similar in tone and tint as the colors in the background, and/or colors that have high contrast with the colors in the background).
  • the third set of selectable options include a respective option for displaying a color picker that allows the user to sample a color from the background media item.
  • the sampled color is added to the third set of selectable options.
  • the third set of selectable options e.g., font and color sheet
  • the third set of selectable options is removed from view to allow the user access to the entirety of the respective background to sample a color from it, and once the color is collected from the respective background, the third set of selectable options is redisplayed and now includes the newly collected color from the respective background.
  • user interface object 586 enables the user to modify a style of the text (e.g., a typeface or other font properties) of the indication of the time, and enables the user to modify a color of the text of the indication of the time.
  • a style of the text e.g., a typeface or other font properties
  • displaying a set of selectable options for the system-generated text causes the device to automatically display a set of selectable options that is appropriate for the particular object that the user wants to edit.
  • displaying the second user interface for editing the first user interface object of the first plurality of editable user interface objects includes: in accordance with a determination that the first user interface object is a user (1162) interface object that corresponds to a respective application, that includes respective content from the respective applications (e.g., the first user interface object is a complication and/or widget that corresponds to an application), and is updated periodically as information represented by the first user interface object changes (and optionally, further in accordance with a determination that the location of the first input and/or the location of the first user interface object is below the time element), displaying a fourth set of selectable options including one or more selectable options to edit a respective size of the first user interface object.
  • a plurality of options for modifying a size of the calendar complication is displayed in user interface element 5082, illustrated in Figure 5AX3.
  • a user interface object that is periodically updated with content from an associated active application displaying a set of selectable options for the user interface object that includes one or more size options for the user interface object causes the device to automatically display a set of selectable options that is appropriate for the particular object that the user wants to edit.
  • the computer system in response to detecting the first input directed to the first user interface in accordance with a determination that the first input meets the first criteria, displays (1162) respective affordances at locations corresponding to a subset of user interface objects of the first plurality of editable user interface objects, wherein the subset of user interface objects correspond to different applications and include respective content from the different applications and are updated periodically as information represented by the subset of user interface objects change, and wherein the respective affordances, when selected, remove corresponding user interface objects of the subset of user interface objects from the first user interface.
  • forgoing displaying respective affordances at locations corresponding to a subset of user interface objects of the first plurality of editable user interface objects (e.g., without displaying the editing user interface illustrated in Figures 5AX2-5AX3).
  • the one or more complications, including complication 5089-1 and 5089-2 are each displayed with a removal affordance (e.g., a minus symbol) that, when selected by the user (e.g., via user input 5086), causes the device to remove the complication associated with the selected removal affordance, as illustrated in Figure 5AX4.
  • a removal affordance e.g., a minus symbol
  • While displaying a user interface for editing one or more editable objects in a wake screen displaying, for a subset of objects that are periodically updated with content from associated active applications, corresponding affordances that are selectable to remove the corresponding object from the wake screen, reduces the number of inputs and amount of time needed to customize the wake screen.
  • the contacts, gestures, user interface objects and animations described above with reference to method 1100 optionally have one or more of the characteristics of the contacts, gestures, user interface objects and animations described herein with reference to other methods described herein (e.g., methods 900, 1000, 12000, 1300, 14000, 15000, 16000, and 17000). For brevity, these details are not repeated here.
  • Figures 12A-12E are flow diagrams illustrating method 12000 for displaying a representation of a plurality of notifications in different configurations, in accordance with some embodiments.
  • Method 12000 is performed at an electronic device (e.g., device 300, Figure 3, or portable multifunction device 100, Figure 1 A) with a display, a touch-sensitive surface, and one or more sensors to detect intensity of contacts with the touch- sensitive surface.
  • the display is a touch-screen display and the touch-sensitive surface is on or integrated with the display.
  • the display is separate from the touch-sensitive surface.
  • method 12000 is a method for displaying a representation of a plurality of notifications in different configurations, thereby providing the user with different configurations for displaying notifications (e.g., based on different circumstances), which provides additional control options without cluttering the user interface with additional displayed controls. Additionally, the available configurations are persistent, which reduces the number of user inputs needed to display notifications in a desired configuration (e.g., the user does not need to select the configuration every time a new notification comes in, or every time the device transitions to a wake state).
  • the method 12000 is performed at a computer system with a display component and one or more input devices.
  • the computer system displays (12002), via the display generation component, a first user interface (e.g., a wake screen user interface) that includes a plurality of notifications (e.g., notifications that have been recently received by the computer system).
  • a first user interface e.g., a wake screen user interface
  • the computer system displays (12004) a representation of the plurality of notifications in a first configuration in a first region of the first user interface.
  • the computer system displays (12006) the representation of the plurality of notifications in a second configuration in a second region of the first user interface that is smaller than the first region of the first user interface.
  • the device allows the user to select from a plurality of user-selectable modes, including the first mode and the second mode, that allow the user to control how notifications are displayed on the wake screen.
  • the first mode is a regular mode of the computer system (e.g., the first configuration is a default configuration and/or the first mode does not include any rules that affect notification delivery or display).
  • the second mode includes one or more rules that control notification delivery or display (e.g., while the second mode is active, some notifications are deferred or hidden by default).
  • the computer system While displaying the first user interface, the computer system detects (12008) a first user input (e.g., a tap, a long press, or a swipe) at a respective location on the first user interface corresponding to the representation of the plurality of notifications.
  • a first user input e.g., a tap, a long press, or a swipe
  • the computer system Scrolls (12012) notifications in the plurality of notifications in the first region of the first user interface (e.g., without expanding the display of the plurality of notifications in the first configuration and/or without expanding the first region of the first user interface) in accordance with the first user input; and in accordance with a determination that the first user input meets the first criteria and in accordance with a determination that the representation of the plurality of notifications is displayed with the second configuration (e.g., a condensed, minimize
  • the third region is the same as the second region (e.g., notifications are scrolled without changing a size of the second region). In some embodiments, the third region is larger than the second region (e.g., the second configuration is “expanded” such that more notifications of the plurality of notifications are visible, and/or more content from the notifications of the plurality of notifications is visible, when the representation of the plurality of notifications is displayed in the third region).
  • the computer system after scrolling the notifications in the plurality of notifications, the computer system detects (12016) the occurrence of a first event (e.g., timeout without user input, or the computer system entering a low power state). In some embodiments, in response to the first event, the computer system ceases to display the first user interface. In response to detecting the occurrence of the first event: in accordance with a determination that the computer system has the first mode for displaying notifications enabled, the computer system maintains display of the representation of the plurality of notifications in the first configuration in the first region; and in accordance with a determination that the computer system has the second mode for displaying notifications enabled, the computer system redisplays the representation of the plurality of notifications in the second configuration in the second region.
  • a first event e.g., timeout without user input, or the computer system entering a low power state.
  • the computer system in response to the first event: in response to detecting the occurrence of the first event: in accordance with a determination that the computer system has the first mode for displaying
  • the third region is different from the second region.
  • the third region is an expanded region of the second region, and after scrolling the notifications in the plurality of notifications in the third region, the notifications collapse back into the second region (e.g., after a threshold amount of time has passed without further scrolling).
  • the expanded configuration 7000-d is a temporary configuration (e.g., for scrolling notifications), and if the portable multifunction device 100 ceases to display the representation 7000 (e.g., in response to detecting a user input, or when the portable multifunction device transitions to a low power or sleep state), in response to a first event (e.g., that transitions the portable multifunction device 100 out of the low power or sleep state), the portable multifunction device 100 redisplays the representation 7000 in the second configuration 7000-b (e.g., and not in the expanded configuration 7000-d).
  • a first event e.g., that transitions the portable multifunction device 100 out of the low power or sleep state
  • Maintaining display of the representation of the plurality of notifications in the first configuration in the first region, in accordance with a determination that the computer system has the first mode for displaying notifications enabled, and redisplaying the representation of the plurality of notifications in the second configuration in the second region, in accordance with a determination that the computer system has the second mode for displaying notifications enabled, automatically redisplays the representation of the notifications in the second configuration (e.g., automatically collapses an expanded view of notifications back into a default view) without requiring further user input (e.g., a separate user input to redisplay the representation of the plurality of notifications in the second configuration in the second region).
  • the computer system after scrolling the notifications, ceases (12018) to display the first user interface (e.g., the computer transitions into a sleep state after a threshold amount of time passes, the computer system transitions into the sleep state in response to a user input, the computer system replaces display of the first user interface with another user interface (e.g., an application launch user interface, a user interface for adjusting one or more settings of the computer system, a camera user interface or other application user interface, a second user interface including one or more widgets)).
  • another user interface e.g., an application launch user interface, a user interface for adjusting one or more settings of the computer system, a camera user interface or other application user interface, a second user interface including one or more widgets.
  • the computer system redisplays the first user interface (e.g., in response to a request to wake the computer system).
  • the computer system displays the representation of the plurality of notifications in the first configuration in the first region of the first user interface.
  • the computer system displays the representation of the plurality of notifications in the second configuration in the second region of the first user interface.
  • the computer system ceases to display the first user interface when the computer system transitions to a low power state (e.g., a sleep state) or an off state, and the computer system redisplays the first user interface when the computer system transitions out of the low power state or off state (e.g., in response to a user input, in response to receiving/generating a new notification, or automatically based off settings of the computer system).
  • the first user interface is displayed immediately as and/or when the computer system transitions out of the low power state or off state (e.g., if the computer system was in a sleep state, the next time the computer system is woken, the first user interface is the first displayed user interface when the computer system is woken). For example, in Figures 7M-7P, even as the time advances from 9:41 to 12:00, the representation 7000 remains displayed in the second configuration 7000-b.
  • Redisplaying the first user interface including displaying the representation of the plurality of notifications in the first configuration in the first region of the first user interface in accordance with a determination that the computer system has the first mode for displaying notifications enabled, and displaying the representation of the plurality of notifications in the second configuration in the second region of the first user interface in accordance with a determination that the computer system has the second mode for displaying notifications enabled, reduces the number of inputs needed to display notifications in a desired configuration (e.g., the user does not need reselect the desired configuration for the plurality of notifications each time the first user interface is redisplayed).
  • the first user interface includes (12020) a first portion of a background image (e.g., a wallpaper for a wake screen), and the second region of the first user interface does not overlay the first portion of the background image.
  • the computer system scrolls the notifications in the plurality of notifications in the third region of the first user interface, in accordance with the first user input, and without overlaying the first portion of the background image. For example, in Figures 7H and 71, in response to the upward swipe gesture 7038, notifications are scrolled in the representation 7000 while maintaining the second configuration 7000-b, and without changing the size of the region in which the representation 7000 is displayed in.
  • Scrolling the notifications in the plurality of notifications in the third region of the first user interface, in accordance with the first user input, and without overlaying the first background image reduces the number of user inputs needed to display the representation of the plurality of notifications in the desired configuration (e.g., and without losing the ability to interact with, view content from, and scroll through, notifications in the plurality of notifications) (e.g., as the user does not need to perform additional user inputs to transition to a different configuration before scrolling the notifications).
  • the computer system displays (12022) text of a respective notification of the plurality of notifications with a first text size.
  • the computer system displays the text of the respective notification of the plurality of notifications with a second text size that is different from (e.g., smaller than) the first text size.
  • the notification 7002 could be displayed with a second text size in the second configuration 7000-b, and the notification 7002 could be displayed with a first text size (e.g., where the second text size is different from the first text size) when the notification 7002 is displayed in the first configuration 7000-a (e.g., as in Figure 7G).
  • a first text size e.g., where the second text size is different from the first text size
  • the computer system displays first text from a first notification of the plurality of notifications with the first text size and displays second text from a second notification of the plurality of notifications with the first text size.
  • the computer system displays the first text with the second text size that is different from the first text size and displays the second text with the second text size.
  • the computer system displays subsequent notifications (e.g., notifications received after, or notifications generated after, detecting the first user input) with the respective text size (e.g., based on which configuration the representation of the plurality of notifications is displayed in).
  • the notification 7002 and the notification 7004 could both be displayed with a first text size in the first configuration 7000-a, and the notification 7002 and the notification 7004 could both be displayed with a second text size in the second configuration 7000-b (e.g., in Figure 7H).
  • the computer system in accordance with the determination that the representation of the plurality of notifications is displayed in the first configuration, displays text of a subset of notifications of the plurality of notifications with the first text size, and in accordance with a determination that the representation of the plurality of notifications is displayed in the second configuration, the computer system displays the text of the subset of notifications of the plurality of notifications with the second text size that is different from the first text size.
  • the subset of notifications includes multiple notifications, but not all notifications, in the plurality of notifications.
  • the notification 7004 could be displayed with a different text size (e.g., a smaller text size) in the second configuration 7000- b, as compared to the same notification 7004 in the first configuration 7000-a (e.g., in Figure 7G).
  • the notification 7002 could be displayed with the same text size in both the second configuration 7000-b and the first configuration 7000-a (e.g., because the notification 7002 is the topmost notification in the stack of notifications in the second configuration 7000- b, reducing a text size of the notification 7002 would reduce visibility and/or readability).
  • the first text size and the second text size are selected based on characteristics of the first configuration and the second configuration, respectively.
  • the first configuration may take up a large amount of room on the display generation component, and the second configuration may be more compact and take up less room on the display generation component (e.g., is more compact and/or takes up less vertical space on the display generation component), relative to the first configuration.
  • the second text size is smaller than the first text size (e.g., is scaled to fit the size of the respective configuration).
  • the difference in size between the first text size and the second text size is proportional to the difference between the room occupied by the first configuration (on the display generation component) and the room occupied by the second configuration.
  • one or more notifications may have a different (e.g., smaller) text size in the second configuration 7000-b, as compared to the displayed text size in the first configuration 7000-a (e.g., in Figure 7G), because the second configuration 7000- b takes up a smaller amount of room on the display as compared to the first configuration 7000-a.
  • Displaying the text of the respective notification of the plurality of notifications with a second text size that is different from the first text size reduces the number of inputs needed to display notifications in a desired configuration (e.g., the user does not need to perform additional user inputs to adjust the text size of notification after displaying the representation of the plurality of notifications in a different configuration).
  • a respective notification of the plurality of notifications includes a first image.
  • the computer system displays (12024) the first image of the respective notification with a first image size.
  • the representation of the plurality of notifications is displayed in the second configuration, displays the first image of the respective notification with a second image size different from (e.g., smaller than) the first image size.
  • the notification 7002 could include a first image (e.g., an application icon for the application A) displayed with a second image size, and the first image is displayed with a first image size (e.g., where the second image size is different from the first image size) when the notification 7002 is displayed in the first configuration 7000-a (e.g., as in Figure 7G).
  • a first notification of the plurality of notifications includes a first image
  • a second notification of the plurality of notification includes a second image.
  • the computer system displays the first image with the first image size and displays the second image with the first image size.
  • the computer system displays the first image with the second image size and displays the second image with the second image size.
  • a subsequent notification e.g., a notification received or generated after detecting the first user input
  • the computer system displays the corresponding image with the respective image size (e.g., based on which configuration the representation of the plurality of notifications is displayed in).
  • the notification 7002 could include a first image (e.g., an application icon for the application A) and the notification 7004 could include a second image (e.g., an application icon for the application Z), and both the first image and the second image are displayed with a second image size in the second configuration 7000-b.
  • the first image and the second image could be displayed with a first image size (e.g., different from the second image size) when the notification 7002 and the notification 7004 are displayed in the first configuration 7000-a (e.g., in Figure 7G).
  • Displaying the first image of the respective notification with a second image size different from the first image size reduces the number of inputs needed to display notifications in a desired configuration (e.g., the user does not need to perform additional user inputs to adjust the image size of images in notifications after displaying the representation of the plurality of notifications in a different configuration).
  • the computer system detects (12026) a second user input that meets second criteria (e.g., the second user input is a swipe, a pinch, a de-pinch), wherein the second criteria are different than the first criteria.
  • the computer system expands the representation of the plurality of notifications. For example, in Figures 7S and 7T, in response to detecting the depinch gesture 7060, or the upward swipe gesture 7062, the portable multifunction device 100 displays the representation 7000 in an expanded configuration 7000-d (e.g., where the expanded configuration 7000-d is an expanded version of the second configuration 7000-b).
  • expanding the representation of the plurality of notifications includes displaying the plurality of notifications with the first configuration (e.g., the computer system transitions from displaying the plurality of notification in the second configuration, to displaying the plurality of notifications in the first configuration, in response to detecting the second user input).
  • the first configuration e.g., the computer system transitions from displaying the plurality of notification in the second configuration, to displaying the plurality of notifications in the first configuration, in response to detecting the second user input.
  • expanding the representation of the plurality of notifications includes displaying the plurality of notifications in a third configuration different than the first configuration and the second configuration (e.g., the first configuration is a regular or default configuration, the second configuration is a condensed, minimized, or reduced prominence configuration, and the third configuration is an intermediate configuration (e.g., the third configuration is less condensed, minimized, or has an increased prominence relative to the second configuration, but is more condensed, minimized, or has a reduced prominence relative to the first configuration)).
  • the first configuration is a regular or default configuration
  • the second configuration is a condensed, minimized, or reduced prominence configuration
  • the third configuration is an intermediate configuration (e.g., the third configuration is less condensed, minimized, or has an increased prominence relative to the second configuration, but is more condensed, minimized, or has a reduced prominence relative to the first configuration)).
  • the portable multifunction device 100 in response to detecting the depinch gesture 7060, or the upward swipe gesture 7062, displays the representation 7000 in an expanded configuration 7000-d, which is different from the first configuration 7000-a (e.g., as shown in Figure 7U) and the second configuration 7000-b (e.g., as shown in Figure 7S)
  • the representation of the plurality of notifications before detecting the second user input, is displayed with a default view of the second configuration, and expanding the representation of the plurality of notifications includes displaying the plurality of notification with an expanded view of the second configuration that is different from the default view of the second configuration.
  • the plurality of notifications remains expanded until collapsed (e.g., in response to a subsequent user input).
  • the plurality of notifications automatically collapse after a predetermined amount of time (e.g., 15 seconds, 30 seconds, 1 minute), and return to a default view of the second notification configuration.
  • the portable multifunction device 100 could display the representation 7000 in the expanded configuration 7000-d for a predetermined amount of time, after which the representation 7000 is displayed in the second configuration 7000-b. .Expanding the representation of the plurality of notifications in response to detecting the second user input, provides additional control options without cluttering the UI with additional displayed controls (e.g., a separate control for expanding the representation of the plurality of notifications.
  • displaying the representation of the plurality of notifications in the second configuration includes (12028) displaying the representation of the plurality of notifications as a stack of notifications; a first notification of the plurality of notifications is displayed at the top of the stack of notifications, and partially overlays other notifications of the plurality of notifications in the stack of notifications; and a first portion of a second notification of the plurality of notifications, different from the first notification, is visible in the stack of notifications.
  • the notifications in the plurality of notifications are stacked one on top of the other, and each a respective notification of the plurality of notifications overlays the notifications underneath it.
  • the first notification is the most recently received notification of the plurality of notifications.
  • At least some content for the second notification is visible in the stack of notifications.
  • some notifications in the plurality of notifications partially overlay other notifications while in the second configuration 7000-b, forming a “stack” of notifications.
  • Displaying the representation of the plurality of notifications as a stack of notifications, wherein a first notification displayed at the top of stack partially overlays other notifications in the stack, provides improved visual feedback to the user (e.g., improved visual feedback regarding the number of notifications in the plurality of notifications).
  • the second notification of the plurality of notifications is displayed (12030) at the bottom of the stack of notifications, and the first portion of the second notification of the plurality of notifications includes a count of notifications in the stack of notifications.
  • the second notification represents a group of notifications (e.g., those notifications that aren’t represented visually in the stack).
  • the count is a count of remaining notifications in the stack of notifications (e.g., notifications beyond those that are visible in the stack).
  • the notification counter 7024 indicates the number of notifications remaining in the stack of notifications (e.g., that are not currently displayed, and have not previously been scrolled off the display).
  • Displaying the first portion of the second notification, including a count of notifications in the stack of notifications provides improved visual feedback to the user (e.g., improved visual feedback regarding the number of notifications represented by the representation of the plurality of notifications and/or content associated with one or more notifications represented by the representation of the plurality of notifications).
  • the second notification of the plurality of notifications is displayed (12032) at the bottom of the stack of notifications
  • the first portion of the second notification of the plurality of notifications includes visual representations (e.g., application icons) of respective applications corresponding to (e.g., applications that generated) respective notifications in the stack of notifications.
  • the visible portion of the second notification includes a visual representation of each application associated with a notification in the stack of notifications.
  • the visible portion of the second notification includes up to a preset maximum number of visual representations (e.g., three application icons). For example, although there are eight applications that generated notifications that are included in the stack of notifications, the visible portion of the second notification includes only (the present maximum number of) three visual representations.
  • the notification count 7024 indicates that at least one notification that is not currently displayed (e.g., and has not been scrolled off the display) has been generated and/or is associated with the application A (“App A”).
  • Displaying the first portion of the second notifications, including visual representations of respective applications corresponding to respective notifications in the stack of notifications, provides improved visual feedback to the user (e.g., improved visual feedback regarding the specific applications and/or the number of applications associated with notifications in the stack of notifications).
  • the computer system in accordance with a determination that the computer system has the second mode (e.g., a Do Not Disturb mode or a focus mode, that causes at least some notifications to be suppressed in accordance with settings of the Do Not Disturb mode or focus mode) for displaying notifications enabled, displays (12034) a visual representation (e.g., an icon, a text label, or a combination of icon and text label) of the second mode for displaying notifications (e.g., a Do Not Disturb icon, an icon corresponding to the active focus mode, and/or a text label identifying the name of the second mode).
  • a visual representation e.g., an icon, a text label, or a combination of icon and text label
  • the computer system in accordance with a determination that the computer system has the first mode for displaying notifications enabled, displays the representation of the plurality of notifications in the first configuration, including displaying a visual representation (e.g., icon) of the first mode for displaying notifications.
  • the first mode for displaying notifications is a normal mode of the computer system (e.g., and notifications are not suppressed in accordance with settings of the normal mode of the computer system), and in accordance with a determination that the computer system has the first mode for displaying notifications enabled, the computer system displays the plurality of notifications in the first configuration without displaying a visual representation of the first mode for displaying notifications.
  • the second mode is a reduced notification mode (e.g., a Do Not disturb mode or focus mode).
  • a reduced notification mode can be associated with different contexts (e.g., a productivity mode, a social mode, a sleep mode, and/or an exercise mode), and can have a whitelist (e.g., or different whitelists, depending on the corresponding context) that lists users and/or applications from which notifications are allowed to “break through” the reduced notification mode.
  • a reduced notification mode causes the device to at least partially block notifications that are not whitelisted for that particular reduced notification mode.
  • a notification that is whitelisted for an active reduced notification mode will be displayed and the user will be notified of its arrival by a sound and/or a haptic alert.
  • a notification that is not whitelisted for the active reduced notification mode will be displayed without a sound or haptic alert (or will not be provided at all while the particular reduced notification mode is active).
  • a user can, for example, set the device to a productivity mode at work and not be distracted by social media (e.g., as shown in Figure 7B), or can set the device to a social mode at home to avoid being bothered by work emails (e.g., as shown in Figure 7C).
  • Figure 7B shows the representation 7000 includes a “Work” indicator 7022 corresponding to an active “Work” mode
  • Figure 7C shows a “Personal” indicator 7026 corresponding to an active “Personal” mode. Displaying a visual representation of the second mode for displaying notifications provides improved visual feedback to the user (e.g., improved visual feedback regarding the currently active mode for displaying notifications).
  • the computer system displays (12036) the representation of the plurality of notifications with the second configuration, detects occurrence of a second event at a first time.
  • the computer system displays a notification corresponding to the second event (e.g., in a fourth region different from the first region and the second region) separately from the representation of the plurality of notifications.
  • a threshold amount of time e.g., 30 seconds, 1 minute, 5 minutes, or 10 minutes
  • the computer system displays the notification corresponding to the second event with the representation of the plurality of notifications in the second configuration.
  • recent notifications are displayed separately (e.g., in a different region from) the representation of the plurality of notifications (e.g., in the first or second region), for increased visibility.
  • the notification is included in the plurality of notifications (e.g., collapses into or is coalesced with the representation of the plurality of notifications).
  • the notification 7050 has just been received or generated, and so the notification 7050 is displayed separately from the representation 7000.
  • the notification 7050 is displayed with the representation 7000.
  • Displaying a notification corresponding to the second event separately from the representation of the plurality of notifications, and displaying the notification corresponding to the second event with the representation of the plurality of notifications in the second configuration after a threshold amount of time has passed provides improved visual feedback to the user (e.g., improved visual feedback regarding whether the notification corresponding to the second event was recently received or generated).
  • the computer system detects (12038) a third user input at a location corresponding to the representation of the plurality of notifications.
  • the computer system transitions to displaying the representation of the plurality of notifications in a third configuration that is different from the first configuration and the second configuration.
  • the third configuration is the same as the first configuration.
  • the first configuration is a regular or default configuration
  • the second configuration is a condensed, minimized, or reduced prominence configuration
  • the third configuration is an intermediate configuration (e.g., the third configuration is less condensed, less minimized, or has increased prominence relative to the second configuration, but more condensed, more minimized, or reduced prominence relative to the first configuration).
  • the portable multifunction device in response to detecting depinch gestures, or upward swipe gestures, transitions to displaying the representation 7000 in a different configuration (e.g., transitions from the third configuration 7000-c to the second configuration 7000-b between Figures 7R and 7S, and transitions from the second configuration 7000-b to the first configuration 7000-a between Figures 7S and 7U).
  • a different configuration e.g., transitions from the third configuration 7000-c to the second configuration 7000-b between Figures 7R and 7S, and transitions from the second configuration 7000-b to the first configuration 7000-a between Figures 7S and 7U).
  • Transitioning to displaying the representation of the plurality of notifications in a third configuration different from the first configuration and second configuration, in accordance with a determination that the third user input meets third criteria, and in accordance with a determination that the representation of the plurality of notifications is displayed in the second configuration provides additional control options without cluttering the UI with additional displayed controls (e.g., additional displayed controls for transitioning and/or changing the configuration that the plurality of notifications is displayed in).
  • the computer system detects (12040) a fourth user input at a location corresponding to the representation of the plurality of notifications.
  • the computer system transitions to displaying the representation of the plurality of notifications in the second configuration.
  • the portable multifunction device in response to detecting pinch gestures, or downward swipe gestures, transitions to displaying the representation 7000 in a different configuration (e.g., transitions from the first configuration 7000-a to the second configuration 7000-b between Figures 7G and 7H, and transitions from the second configuration 7000-b to the third configuration 7000-c between Figures 7P and 7Q).
  • a different configuration e.g., transitions from the first configuration 7000-a to the second configuration 7000-b between Figures 7G and 7H, and transitions from the second configuration 7000-b to the third configuration 7000-c between Figures 7P and 7Q.
  • Transitioning to displaying the representation of the plurality of notifications in the second configuration, in accordance with a determination that the fourth user input meets fourth criteria, and in accordance with a determination that the representation of the plurality of notifications is displayed in the first configuration provides additional control options without cluttering the UI with additional displayed controls (e.g., additional displayed controls for transitioning and/or changing the configuration that the plurality of notifications is displayed in).
  • the first user interface before detecting the first user input, includes (12042) a first system user interface (e.g., a system user interface that should always be displayed in certain scenarios, such as a live session, media information and controls, an emergency notification, a time-sensitive notification, an urgent or emergency notification, and/or a contextually relevant system user interface such as a boarding pass) that is displayed separately from the plurality of notifications.
  • a first system user interface e.g., a system user interface that should always be displayed in certain scenarios, such as a live session, media information and controls, an emergency notification, a time-sensitive notification, an urgent or emergency notification, and/or a contextually relevant system user interface such as a boarding pass
  • the computer system maintains display of the first system user interface, separate from the plurality of notifications.
  • the first system user interface is displayed separate from the plurality of notifications regardless of which configuration the plurality of notifications is displayed with.
  • the first system user interface is displayed separate from the plurality of notifications regardless of which mode for displaying notifications is enabled for the computer system.
  • the session 7048 is displayed separate from the representation 7000 (e.g., regardless of which configuration the representation 7000 is displayed in).
  • Maintaining display of the first system user interface, separate from the plurality of notifications reduces the number of user inputs needed to display and/or interact with the first system user interface (e.g., there is a high probability the user will need to view and/or interact with the first system user interface, if the first system user interface is contextually relevant and/or time-sensitive, and the user does not need to perform additional user inputs to navigate to and/or interact with the first system user interface, when the first system user interface is displayed separate from the plurality of notifications).
  • the computer system detects (12044) a fifth user input (e.g., a tap, a swipe, or a long press) at a location corresponding to a respective notification of the plurality of notifications.
  • a fifth user input e.g., a tap, a swipe, or a long press
  • the computer system performs an operation associated with the respective notification without performing an operation associated with other concurrently displayed notifications (e.g., displaying an application associated with the respective notification or displaying one or more affordances (e.g., including an affordance for opening the notification, an affordance for dismissing the notifications, and/or an affordance for adjusting one or more notification settings for an application associated with the notification) for interacting with the respective notifications); and in accordance with a determination that the representation of the plurality of notifications is displayed in the second configuration, the computer system forgoes performing the operation associated with the respective notification (e.g., forgoing displaying the application associated with the respective notification and/or for
  • a user can only interact with the respective notification while the plurality of notifications is displayed in the first configuration, and cannot interact with the respective notification while the plurality of notifications is displayed with the second configuration (e.g., without first changing the configuration for the plurality of notifications).
  • the notification 7004 nor the notification 7006 are available for interaction (e.g., because, or while, the representation 7000 is displayed in the second configuration 7000-b).
  • Forgoing display of the one or more affordances for interacting with the respective notification, in accordance with a determination that the representation of the plurality of notifications is displayed in the second configuration, and displaying the one or more affordances for interacting with the respective notification in accordance with a determination that the representation of the plurality of notifications is displayed in the first configuration, provides additional control options without cluttering the UI with additional displayed controls (e.g., the second configuration is a minimalist configuration, and forgoing display of the one or more affordances for interacting with the respective notifications avoid cluttering the UI with the one or more affordances because the user has selected a minimalist configuration).
  • the computer system detects (12046) a sixth user input (e.g., a tap, a swipe, or a long press) at a location corresponding to a respective notification of the plurality of notifications.
  • a sixth user input e.g., a tap, a swipe, or a long press
  • the computer system performs an operation associated with the respective notification (e.g., displaying an application associated with the respective notification or displaying one or more affordances (e.g., including an affordance for opening the notification, an affordance for dismissing the notifications, and/or an affordance for adjusting one or more notification settings for an application associated with the notification) for interacting with the respective notification).
  • a user can interact with the respective notification regardless of which configuration the plurality of notifications is displayed with (e.g., the user can interact with the respective notification when the plurality of notifications is displayed in the first configuration, and when the plurality of notifications is displayed in the second configuration). For example, as described above with reference to Figure 7A, for example, where the user can interact (e.g., via the tap input 7012, the rightward swipe input 7014, or the leftward swipe input 7016) with notifications while the representation 7000 is in the first configuration 7000-a.
  • Displaying one or more affordances for interacting with the respective notification in response to detecting the sixth user input at a location corresponding to a respective notification of the plurality of notifications, provides additional control option without cluttering the UI with permanent display of certain controls (e.g., the one or more affordances for interacting with the respective notification).
  • the computer system detects (12048) a seventh user input (e.g., a tap, a swipe, or a long press) at a location corresponding to a respective notification of the plurality of notifications.
  • a seventh user input e.g., a tap, a swipe, or a long press
  • the computer system performs an operation associated with the respective notification without performing an operation associated with other concurrently displayed notifications (e.g., displaying an application associated with the respective notification or displaying one or more affordances (e.g., including an affordance for opening the notification, an affordance for dismissing the notifications, and/or an affordance for adjusting one or more notification settings for an application associated with the notification) for interacting with the respective notification); and in accordance with a determination that less than the threshold amount of the respective notification is visible while the representation of plurality of notifications is displayed in the second configuration, the computer system forgoes performing the operation associated with the respective
  • the notification 7004 (e.g., which is fully visible) is available for interaction (e.g., as shown in Figure 7K), but in some embodiments, the notification 7006 (e.g., for which the threshold amount is not visible) is not available for interaction.
  • Displaying one or more affordances for interacting with the respective notification in accordance with a determination that a threshold amount of the respective notification is visible while the representation of the plurality of notifications is displayed in the second configuration, and forgoing display of the one or more affordances for interacting with the respective notification in accordance with a determination that less than the threshold amount of the respective notification is visible while the representation of the plurality of notifications is displayed in the second configuration provides additional control options without cluttering the UI with permanently displayed controls (e.g., without permanently displaying the one or more affordances for interacting with the respective notification, particularly if the relevant notification is not sufficiently visible).
  • the computer system scrolls (12050) the notifications in the plurality of notifications in the third region of the first user interface, wherein scrolling the notifications includes displaying at least the threshold amount of the respective notification.
  • displaying the plurality of notifications in the second configuration includes displaying the plurality of notifications are displayed in a stack of notifications, and scrolling the notifications in the plurality of notifications includes displaying the respective notification at the top of the stack of notifications.
  • the computer system responds differently to the same input depending on how much of the respective notification (e.g., over which the input is received) is displayed.
  • the portable multifunction device 100 scrolls notifications (e.g., so that the threshold amount of the notification 7006 is visible, as shown by the notification 7006 being fully visible in Figure 7M).
  • Scrolling notifications in the plurality of notifications in the third region to display at least a threshold amount of a respective notification, in accordance with a determination that less than the threshold amount of the respective notification is visible while the plurality of notifications is displayed in the second configuration, automatically scrolls a notification to a position for user interaction in response to a user input at a location corresponding to the notification, without requiring further user input (e.g., a user input to first scroll the notifications in the plurality of notifications, before the user can request that the computer system display the one or more affordances for interacting with the notification).
  • the contacts, gestures, and user interface objects optionally have one or more of the characteristics of the contacts, gestures, and user interface objects, described herein with reference to other methods described herein (e.g., 900, 1000, 1100, 1300, 14000, 15000, 16000, and 17000). For brevity, these details are not repeated here.
  • Figures 13A-13G are flow diagrams illustrating method 1100 of displaying status information in a session region based on subscriptions in accordance with some embodiments.
  • Method 1300 is performed (1302) at a computer system (e.g., device 300, Figure 3, or portable multifunction device 80, Figure 1 A) that is in communication with a display generation component (e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a headmounted display, or the like) (and optionally, the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as lidars, depth sensors, and/or distance sensors)
  • a display generation component e.g., a hardware element, comprising one or more display devices, such as a display, a projector, a touch-screen display, a heads-up display, a headmounted display, or the like
  • the computer system is further in communication with one or more input devices, one or more cameras, and/or one or more 3D sensing and/or determination devices, such as lid
  • the computer system detects (1304) one or more inputs (e.g., inputs directed to the first application and inputs directed to the second application; and/or inputs directed to a subscription interface that lists both events from the first application and events from the second application) to subscribe to updates from a first application for a first event, and to subscribe to updates from a second application for a second event (e.g., optionally, the first application is different from the second application, and/or the first event is different from the second event).
  • user input 808 ( Figure 8A) and user input 812 (Figure 8C) correspond to requests to subscribe to a food delivery event
  • user inputs 826 and 828 correspond to requests to subscribe to two sports events.
  • the first event and the second event are not subscribed to at the same time. In some embodiments, the first event and the second event do not overlap in time. In some embodiments, the first event and the second event are both subscribed to during a first period of time, but the first event and the second event do not overlap in time. In some embodiments, the first event and the second event overlap in time for some portions of the two events, but they do not start and/or end at the same time. In some embodiments, the first and second applications are the same application and the first event is distinct from the second event. In some embodiments, the first application and second application are distinct applications.
  • the first event is a sporting event (e.g., with updates for starting time, score updates, interruptions, intermissions, and/or final score).
  • the first event is a delivery event (e.g., for food, for online purchases, and/or for express mail) (e.g., with updates for pickup, enroute, arrival, and/or expected delays).
  • the first event is a ride share event (e.g., with updates for estimated arrival time, arrival, pickup, intermediate stops, and/or drop off).
  • additional events e.g., that are related to the first event
  • the computer system automatically subscribes another event corresponding to a second game for the first sport team.
  • the computer system displays (1306) a first user interface (e.g., a user interface that corresponds to a restricted state of the computer system, such as a wake screen user interface and/or a lock screen user interface).
  • the first user interface includes a first region at a first location in the first user interface (e.g., a region directly below the time indication of the wake screen user interface, or a region in the bottom portion of the wake screen user interface).
  • Displaying the first user interface includes: in accordance with a determination that the first event is active (e.g., ongoing and providing updates, and/or has not ended) and that the second event is not active (e.g., not providing updates, and/or has not started), displaying a first representation of the first event in the first region of the first user interface, and updating first information contained in the first representation of the first event in accordance with updates received from the first application for the first event (e.g., updating the information is made substantially in real-time of receipt of the updates for the first event).
  • the computer system while displaying the first user interface, in accordance with a determination that the first event is no longer active (e.g., the event has ended) or the user has unsubscribed the first event, the computer system ceases to display the first representation of the first event in the first user interface.
  • session 816-1 is displayed in a session region of wake screen user interface 800 (Figure 8D) while the food delivery session is ongoing, and is no longer displayed (Figure 8J) after the event has concluded (e.g., the food is delivered).
  • Displaying the first user interface includes: in accordance with a determination that the second event is active (e.g., ongoing and providing updates, and/or has not ended) and that the first event is not active (not providing updates, and/or has not started), displaying a second representation of the second event in the first region of the first user interface, and updating second information contained in the second representation of the second event in accordance with updates received from the second application for the second event (e.g., updating the information is made substantially in real-time of receipt of the updates for the second event).
  • a third subscribed event is active concurrently with the second event, and a third representation of the third event is displayed in a second region of the first user interface concurrently with display of the second representation displayed in the first region of the first user interface.
  • the first region of the first user interface does not include a user interface object when there is no subscribed events, or when no subscribed events are currently active.
  • session 830-1 for a first sports game e.g., Golden State vs Chicago
  • the second sports game to which the user subscribed e.g., via user input 828, Figure 8L
  • a session is not displayed for the second sports game in user interface 800.
  • a third event (e.g., session 864-1) is active concurrently with the first event (e.g., session 830-5).
  • session 864-1 is displayed in a second portion of the session region concurrently with session 830-5 that is displayed in a first portion of the session region.
  • the first user interface is (1308) a wake screen user interface (e.g., the user interface that is initially displayed when the display generation component transitions from a power saving mode (e.g., a display off state, or a dimmed always on state) to a normal mode in response to an event (e.g., arrival of a notification, a press input on the power button or touch screen, and/or a change in an orientation of the display generation component)); the first representation of the first event is displayed in the first region while the first event is active; and the second representation of the second event is displayed in the first region of the wake screen user interface while the second event is active.
  • a wake screen user interface e.g., the user interface that is initially displayed when the display generation component transitions from a power saving mode (e.g., a display off state, or a dimmed always on state) to a normal mode in response to an event (e.g., arrival of a notification, a press input on the power button or touch screen, and/or a change in
  • session 830-1 for the game is displayed in the session region of wake screen user interface 800 ( Figure 8N).
  • both the first and the second representations are displayed concurrently (e.g., with the representation of the event that started earlier displayed in the first region, and the representation of the event that started later displayed below the first region).
  • a representation of a respective event is only displayed while the respective event is active and automatically disappears in accordance with a determination that the respective event has ended (e.g., delivery arrived, game is over, or workout is over).
  • the representation of the event that started earlier and ended earlier is removed from the first region, and then the representation of the event that started later and ended later is moved to the first region once it has been vacated by the representation of the earlier event.
  • Displaying updates for active application events in a wake screen user interface enables the user to view different types of status information for the computer system quickly, prior to navigation to another user interface such as a home user interface, and optionally without requiring user authentication, thereby reducing an amount of time needed to view the status information.
  • the computer system while the first event is active (1310): at a first time, the computer system displays the wake user interface with the first representation of the first event in the first region of the wake user interface; at a second time after the first time, the computer system ceases display of the wake user interface (e.g., and optionally, ceases to display the first representation of the first event) in response to detecting that a first condition is met (e.g., in response to the display generation component transitioning into a power saving mode after prolonged inactivity and/or a press input on the power button, or in response to navigation to another user interface (e.g., home user interface and/or widget screen user interface)); and at a third time after the second time, in response to detecting that a second condition is met, the computer system redisplays the wake user interface with the first representation of the first event in the first region of the wake user interface.
  • a first condition e.g., in response to the display generation component transitioning into a power saving mode after prolonged inactivity and/or a press
  • the respective representation of a respective subscribed event is persistently displayed on the wake user interface, as long as the respective event is still active and receiving updates, even if the wake user interface has been dismissed one or more times (e.g., display is turned off or dimmed, and/or other user interfaces has replaced display of the wake user interface) while the respective event is active.
  • session 830-2 is displayed for the sports event, before detecting user input 856 (Figure 8X) for putting device 100 into a low-power state. After the device has entered the low-power state, and comes out of the low-power state, session 830-4 ( Figure 8Y) for the sports event continues to be displayed in the session region of the wake screen user interface 800.
  • Displaying, in a wake screen user interface, updates for active application events, including when the wake screen user interface is redisplayed after being dismissed or turned off, enables the user to view different types of status information for the computer system quickly, thereby reducing an amount of time needed to view the status information.
  • the computer system while the first event is active (1312): at a fourth time, the computer system displays the first user interface with the first representation of the first event in the first region of the first user interface, wherein the first user interface does not include notifications; and at a fifth time later than the fourth time, the computer system redisplays one or more notifications (e.g., displays notification history including one or more previously saved notifications, and/or displaying newly received and/or unread notifications) in the first user interface (e.g., the wake screen user interface, the lock screen user interface, or a blurred and/or dimmed version of the wake user interface or lock user interface) in response to a third condition being met (e.g., arrival of new notifications, and/or a user input that corresponds to a request to display the notification history), and maintains display of the first representation of the first event in the first user interface (e.g., in the first region of the first user interface, or in a second region of the first user interface different from the first region of the first user interface
  • session 830-1 is displayed in the session region of the wake screen user interface without notifications
  • a plurality of notifications e.g., notifications 806-7 and 806-8) are displayed concurrently with session 830-2.
  • notifications are displayed in a distinct region of the first user interface (e.g., below the first region).
  • the first region that displays the ongoing event is a dedicated region that is distinct from the region(s) that display notifications.
  • notifications are scrollable and may shift underneath the first region that displays the first representation of the first event.
  • the first representation of the first event scrolls with the notifications in a direction of a swipe input when the swipe input (e.g., an upward swipe) is detected on the notifications.
  • the notifications when notifications are dismissed, cleared, or saved, the notifications cease to be displayed in the first user interface, while the first representation of the first event is persistently displayed in the first user interface as long as the first event is still active, as described with reference to Figures 8P-8Q.
  • the computer system while the notifications are displayed in the first user interface, in accordance with a determination that the first event has ended, the computer system ceases to display the first representation of the first event and continues to display the notifications on the first user interface. Displaying updates for active application events in a user interface that is also occasionally updated to include one or more notifications enables the user to view different types of status information for the computer system together in the same user interface, thereby reducing an amount of time needed to view the status information while making more efficient use of the display area.
  • the computer system while the first event is active (1314): at a sixth time, the computer system displays the first user interface with the first representation of the first event in the first region of the wake user interface; and at a seventh time after the sixth time: the computer system replaces display of the first user interface with display of a second user interface that includes a plurality of application icons that, when selected, cause display of corresponding applications (e.g., the second user interface is one of a home screen user interface, an application launch user interface, and/or a widget screen).
  • the second user interface is one of a home screen user interface, an application launch user interface, and/or a widget screen.
  • the computer system in response to detecting that a fourth condition is met (e.g., in response to detecting a upward swipe input from a bottom edge of the display region of the display generation component, a press on a home button, or a rightward swipe from the left edge of the display region), replaces display of the first representation of the first event in the first region of the first user interface with display of a third representation of the first event (e.g., a reduced representation as compared to the first representation) (e.g., a bubble, or pill shaped user interface object that includes less information than the first representation of the first event) in a second region of the second user interface (e.g., in one of the upper left corner, upper right corner, and/or a screen cutout region).
  • a third representation of the first event e.g., a reduced representation as compared to the first representation
  • a bubble, or pill shaped user interface object e.g., a bubble, or pill shaped user interface object that includes less information than the first representation of the
  • session 830-3 displayed on the wake screen user interface 800 is shifted to a session region 838-1 in Figure 8R while displaying the home screen user interface 805.
  • the respective representation of a respective subscribed event is transformed into a reduced representation that is persistently displayed on the display, as long as the respective event is still active and receiving updates, even if the first user interface has been dismissed one or more times and replaced with display of another user interface, such as the home screen user interface, the widget screen user interface, and/or a notification center user interface. Displaying updates for active application events in a home user interface reduces the number of inputs and the amount of time needed to view current status information for aspects of the computer system even without otherwise displaying user interfaces of the active applications.
  • detecting the one or more inputs to subscribe to updates from the first application for the first event includes (1316): while displaying a first notification corresponding to the first application (e.g., while the first notification is displayed on the first user interface) (e.g., optionally, the first notification is regarding a first update from the first event), detecting a first set of inputs directed to the first notification, wherein the first set of inputs meet respective criteria for subscribing to updates from the first application for the first event.
  • the first set of inputs include one or more inputs from: an input causing display of a selectable option for subscribing to the first event, an input selecting the selectable option for subscribing to the first event, and/or an input confirming subscription to the first event.
  • user inputs 810 and 812 cause the device 100 to subscribe to the food delivery event from notification 804. Enabling a user to subscribe to updates from a respective active application by directing one or more inputs to a notification corresponding to the respective application reduces the number of inputs needed to pin and view status information for the computer system.
  • detecting a first set of inputs directed to the first notification includes (1318) detecting selection of a first affordance displayed with the first notification.
  • a first affordance displayed with the first notification For example, in Figure 8 A, user input 808 selects the “Subscribe” option 802 from notification 804.
  • the first affordance is a button displayed on the first notification (e.g., automatically based on a determination that the notification is a type of notification associated with repeated future updates).
  • the first affordance is displayed next to the first notification (e.g., the first notification shifts in response to a user input (e.g., a touch-hold input on the first notification, or a sideways swipe on the first notification) and the first affordance is revealed where the first notification is shifted away from).
  • the first affordance is displayed in a menu of affordances (e.g., for dismissing the first notification, for opening an application associated with the first notification, and/or for storing the first notification in a notification history). Enabling a user to subscribe to updates from a respective active application by selecting an affordance for doing so that is displayed with a notification corresponding to the respective application reduces the number of inputs needed to pin and view status information for the computer system.
  • detecting (1320) the one or more inputs to subscribe to updates from the second application for the second event includes: while displaying one or more search results (e.g., search results including content from the computer system, and/or content from outside of the computer system) corresponding to a search input (e.g., one or more search keywords), including a first search result that corresponds to the second application (e.g., the first search result includes at least one of an application icon for the second application, content from the second application (e.g., text message from a messaging application), and/or a widget corresponding to the second application), detecting a second set of inputs directed to the first search result, wherein the second set of inputs meet respective criteria for subscribing to updates from the second application for the second event.
  • search results e.g., search results including content from the computer system, and/or content from outside of the computer system
  • a search input e.g., one or more search keywords
  • the first search result includes at least one of an application icon for the second application
  • the second set of inputs includes one or more inputs selected from: an input causing display of a selectable option for subscribing to the second event, an input selecting the selectable option for subscribing to the second event, and/or an input confirming subscription to the second event.
  • the first event of the first application can also be subscribed from a search result that corresponds to the first application in a manner analogous to those described above with respect to the second event of the second application. For example, as described with reference to Figure 8V-8X, the user subscribes to a workout and/or health application from a search result.
  • Enabling a user to subscribe to updates from a respective active application by directing one or more inputs to a search result that corresponds to the respective application reduces the number of inputs and amount of time needed to locate the respective application and pin and view corresponding status information using the computer system.
  • the computer system while displaying a respective user interface of a third application (e.g., same as the first application, same as the second application, different from the first application, and/or different from the second application), the respective user interface including a respective affordance for subscribing to updates from the third application for a third event, the computer system detects (1322) selection of the respective affordance for subscribing to updates from the third application for the third event.
  • a third application e.g., same as the first application, same as the second application, different from the first application, and/or different from the second application
  • the respective user interface including a respective affordance for subscribing to updates from the third application for a third event
  • the computer system detects (1322) selection of the respective affordance for subscribing to updates from the third application for the third event.
  • the computer system displays a third representation of the third event in the first region of the first user interface (e.g., when the first event and the second event are not active), and updates information contained in the third representation of the third event in accordance with updates received from the third application for the third event.
  • the first event of the first application and/or the second event of the second application can also be subscribed respectively from the first application and/or the second application in a manner analogous to those described above with respect to the third event of the third application.
  • the device subscribes to one or more sports events from user interface 803 ( Figure 8L) for the sports application in response to user inputs 826 and 828.
  • Enabling a user to subscribe to updates from a respective active application by selecting an affordance for doing so that is displayed within a user interface of the respective application reduces the number of inputs and amount of time needed to pin and view status information for the computer system.
  • the computer system in accordance with a determination that a user of the computer system has enabled an option for automatic subscription, automatically subscribes (1324) to updates from a fourth application for a fourth event in response to detecting that a fifth condition (e.g., a new event corresponding to the option for automatic subscription has been created) has been met.
  • the user selects, for a particular application, to subscribe to all events, or a subset of events, for the application; and after the selection, the computer system automatically subscribes to any new events that are created for the application without requiring further user inputs.
  • the user selects to subscribe to a subset of events from a sports application, the subset of events corresponding to a first team participating in the event, and/or the user selects to subscribe to a subset of a certain type of events (e.g., basketball games, but not football games, or sports games that occur at a selected location (or set of locations) or include a selected team (or set of teams) but not sports games that do not occur at a selected location and/or do not include a selected team), and when a new game event for the first team and/or when a new instance of the certain type of events becomes available (e.g., not yet active) in the particular application, the computer system automatically subscribes to the new game event and/or the new instance of the certain type of events without requiring
  • the user selects to subscribe to all events for a particular application, for example, a plurality of (e.g., most or all) rideshare requests and/or a plurality of (e.g., most or all) food deliveries, are automatically subscribed to such that upon initiation of a new rideshare and/or food delivery event, the status of the new event is updated in the first region of the first user interface.
  • the device enables and/or disables automatic subscription for an event in accordance with a state of the toggle option 880 and in response to a user input changing the toggle state of the toggle option 880 (e.g., for future rides in the rideshare application). Enabling a user to configure the computer system to automatically subscribe to updates from a respective active application reduces the number of inputs and amount of time needed to pin and view status information for the computer system.
  • the computer system in accordance with a determination that past user behavior meets one or more subscription criteria, automatically subscribes (1326) to updates from a fifth application for a fifth event.
  • the user does not always need to actively select to subscribe to events from an application.
  • the computer system determines that the user has elected to subscribe to a threshold number of events (of a certain type, or from a certain application) and automatically subscribes the user, without additional user input, to future events that satisfy similarity criteria to the events that the user has previously subscribed. For example, the computer system determines that the user tends to follow sports events for a first sports team, and automatically subscribes the user to future events for the first sports team.
  • the computer system learns from user feedback for automatic subscribed events and determines whether to continue to automatically subscribe to similar events. For example, as described with reference to Figure 8AE, in some embodiments, the computer system automatically subscribes the user to particular events based on the user’s past behavior (e.g., past subscriptions). Automatically subscribing to updates from a respective active application based on past user behavior causes the device to automatically pin and view status information for the computer system that is likely to be of interest to the user without requiring additional user input.
  • past behavior e.g., past subscriptions
  • the first application is (1328) a rideshare application and the first event is an instance of a respective ride requested in the rideshare application.
  • the first information contained in the first representation of the first event that is updated in accordance with updates received from the first application for the first event includes location information of the respective ride requested in the rideshare application.
  • the first information includes a distance and/or other indication of location of the hailed ride, optionally displayed in a map.
  • the first information includes an approximate time until the hailed ride arrives.
  • the first information includes information about a drop off location (e.g., a distance, time and/or route to a drop off location), that is displayed while the user is riding in the hailed ride.
  • the second application is a rideshare application and the second event is an instance of a respective ride requested in the rideshare application; and the second information contained in the second representation of the second event that is updated in accordance with updates received from the second application for the second event includes location information of the respective ride requested in the rideshare application.
  • Figures 8AC-8AF illustrate an example of initiating a session for a rideshare application, wherein session 884-1 for the rideshare application is displayed in the wake screen user interface 800.
  • the first application is (1330) a delivery application (e.g., a food or package delivery application) and the first event is an instance of a respective delivery requested in the delivery application.
  • the first information contained in the first representation of the first event that is updated in accordance with updates received from the first application for the first event includes delivery information of the respective delivery requested in the delivery application.
  • the first information includes a distance and/or other indication of time of arrival of the requested delivery, optionally displayed in a map.
  • the second application is a delivery (e.g., food or package delivery) application and the second event is an instance of a respective delivery requested in the food delivery application; and the second information contained in the second representation of the second event that is updated in accordance with updates received from the second application for the second event includes delivery information of the respective delivery requested in the delivery application.
  • the user subscribes to a food delivery event. Enabling a user to subscribe to updates from an active delivery application, which are displayed at a consistent location within another user interface such as a wake screen user interface, reduces the number of inputs and amount of time needed to view current, potentially time-sensitive status information for the computer system while making more efficient use of the display area.
  • the second application is (1332) a sports application (e.g., an application associated with a particular sport, a video application that includes sports game videos, and/or a news application that includes sports game news) and the second event is an instance of a game reported by the sports application.
  • the second information contained in the second representation of the second event that is updated in accordance with updates received from the second application for the second event includes score information for the instance of the game.
  • the second information includes a time indicator (e.g., time remaining in a quarter or a half), in the game.
  • the score information includes updated scores for each team participating in the instance of the game.
  • the first application is a sports application and the first event is an instance of a game reported by the sports application; and the first information contained in the first representation of the first event that is updated in accordance with updates received from the first application for the first event includes score information for the instance of the game.
  • session 830-1 ( Figure 8N) is a session for a basketball game associated with a sports application.
  • the second application is (1334) a workout application and the second event is an instance of a workout logged by the workout application.
  • the second information contained in the second representation of the second event that is updated in accordance with updates received from the second application for the second event includes activity information for the instance of the workout.
  • the activity information includes a pace of an activity (e.g., a running and/or walking pace).
  • the activity information includes an indication of a length of time of the activity.
  • the activity information includes an indication of distance covered in the activity (e.g., mileage).
  • the activity information includes an indication of a location of the activity (e.g., a path taken during the activity).
  • the first application is a workout application and the first event is an instance of a workout logged by the workout application; and the first information contained in the first representation of the first event that is updated in accordance with updates received from the first application for the first event includes activity information for the instance of the workout.
  • session 864-1 ( Figure 8Z) is a session for a workout associated with a workout and/or health application.
  • the computer system while displaying the first user interface (e.g., while displaying the first user interface after the initial display of the first/second representation of the first/second event): in accordance with a determination that the first representation of the first event is currently displayed in the first region of the first user interface (e.g., the first representation of the first event has been updated one or more times based on updates received from the first application for the first event): in accordance with a determination that the first event is still active, the computer system maintains (1336) display of the first representation of the first event in the first region of the first user interface (and optionally, continuing to update the first representation based on future updates received from the first application for the first event); and in accordance with a determination that the first event is no longer active (e.g., after the last update has been received and represented in the first representation of the first event), ceases display of the first representation of the first event in the first region of the first user interface.
  • the basketball game associated with session 830 is still active, and the session remains displayed in the wake screen user interface 800, and in Figure 8AK, the basketball game is over and the session 830 is no longer displayed.
  • the computer system in accordance with a determination that the second representation of the second event is currently displayed in the first region of the first user interface: the computer system, in accordance with a determination that the second event is still active, maintains display of the second representation of the second event in the first region of the first user interface; and the computer system, in accordance with a determination that the second event is no longer active, ceases display of the second representation of the second event in the first region of the first user interface.
  • the computer system automatically, without user input, ceases display of the representation of an event after the event ends.
  • a final status is provided before ceasing display of the representation of the event.
  • the computer system displays “delivered” for a delivery event and maintains display of the final status for a predetermined amount of time (e.g., 30 seconds, 1 minute, or 5 minutes) before automatically removing the status and the representation of the event, without requiring the user to request to clear the status and the representation of the event from the first user interface.
  • a status region that is updated with status information for an active event
  • ceasing to display the status region when the event is no longer active causes the device to automatically reduce the number of displayed elements when no longer needed.
  • the computer system while displaying the first user interface (e.g., displaying the first user interface after the initial display of the first/second representation of the first/second event, optionally after navigating to another user interface and/or turning the display off and on again): in accordance with a determination that the first event is inactive (e.g., has ended and/or no longer receiving updates) and a determination that the first representation of the first event was last displayed (e.g., the first event ended at a time after the first user interface was last displayed) or is currently displayed (e.g., the first event ended at a time while the first user interface is displayed) in the first region of the first user interface: in accordance with a determination that a sixth condition is not met (e.g., the sixth condition requires that the first representation of the first event is displayed at least once after the first user interface is dismissed and redisplayed after the first event ended), the computer system displays (1338) the first representation of the first event in the first region of the first user interface, the first representation
  • the computer system in accordance with a determination that the second representation of the second event is currently displayed or was last displayed in the first region of the first user interface, and a determination that the second event is inactive: the computer system, in accordance with a determination that the sixth condition is not met, displays the second representation of the second event in the first region of the first user interface, the second representation of the second event including the second information that has been updated in accordance with a second final update received from the second application for the second event; and the computer system, in accordance with a determination that the sixth condition is met, forgoes displaying the second representation of the second event in the first region of the first user interface.
  • the sixth condition is satisfied in accordance with a determination that the first user interface (e.g., the wake screen user interface and/or the lock screen user interface) is displayed and dismissed at least once after the first/second event is no longer active (e.g., the user has displayed and then dismissed the wake screen after the first/second event ends such that the device makes the final update for the first/second event on the wake screen user interface (e.g., a location to which an item was delivered or a final score for a sports event) visible to the user for at least once.
  • the first user interface e.g., the wake screen user interface and/or the lock screen user interface
  • session 816-6 indicates “Your food has been delivered” and session 816-6 is optionally maintained until the device 100 has entered a low-power state, such that when the device 100 leaves the low-power state, the wake screen user interface 800 no longer displays the final status of session 816-6 ( Figure 8J).
  • a status region was updated with status information for an active event
  • maintaining display of the status region after the event is no longer active until a particular condition (e.g., based on a user viewing the final status of the active event) is met causes the device to automatically ensure that the user is provided with an opportunity to view the final status information and reduce the number of displayed elements when no longer needed.

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Un système informatique affiche une première version d'une première interface utilisateur qui comprend des premiers objets d'interface utilisateur affichés simultanément avec un premier arrière-plan. Les premiers objets d'interface utilisateur comprennent un contenu provenant d'une première pluralité d'applications. En réponse à la détection d'une première entrée, lorsque la première entrée comprend un mouvement dans une première direction, le système remplace l'affichage de la première version de la première interface utilisateur par une seconde interface utilisateur qui comprend des représentations d'une seconde pluralité d'applications. Lorsque la première entrée comprend un second mouvement dans une seconde direction, le système remplace l'affichage de la première version de la première interface utilisateur par l'affichage d'une seconde version de la première interface utilisateur qui comprend des seconds objets d'interface utilisateur affichés avec un second arrière-plan. Les seconds objets d'interface utilisateur correspondent à une troisième pluralité d'applications et comprennent un contenu provenant de la troisième pluralité d'applications.
PCT/US2023/021376 2022-05-10 2023-05-08 Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application WO2023219959A1 (fr)

Applications Claiming Priority (6)

Application Number Priority Date Filing Date Title
US202263340388P 2022-05-10 2022-05-10
US63/340,388 2022-05-10
US202263349128P 2022-06-05 2022-06-05
US63/349,128 2022-06-05
US17/949,128 2022-09-20
US17/949,128 US11868601B2 (en) 2022-05-10 2022-09-20 Devices, methods, and graphical user interfaces for providing notifications and application information

Publications (1)

Publication Number Publication Date
WO2023219959A1 true WO2023219959A1 (fr) 2023-11-16

Family

ID=86688816

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2023/021376 WO2023219959A1 (fr) 2022-05-10 2023-05-08 Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application

Country Status (1)

Country Link
WO (1) WO2023219959A1 (fr)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180335920A1 (en) * 2017-05-16 2018-11-22 Apple Inc. Devices, Methods, and Graphical User Interfaces for Accessing Notifications

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180335920A1 (en) * 2017-05-16 2018-11-22 Apple Inc. Devices, Methods, and Graphical User Interfaces for Accessing Notifications

Similar Documents

Publication Publication Date Title
AU2022203957B2 (en) Context-specific user interfaces
EP4283437A1 (fr) Interfaces utilisateur spécifiques au contexte
KR102503135B1 (ko) 시간과 관련된 사용자 인터페이스들
EP4276596A1 (fr) Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application
WO2023219959A1 (fr) Dispositifs, procédés et interfaces utilisateur graphiques pour fournir des notifications et des informations d'application

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23728514

Country of ref document: EP

Kind code of ref document: A1