CN114840280A - Display method and electronic equipment - Google Patents

Display method and electronic equipment Download PDF

Info

Publication number
CN114840280A
CN114840280A CN202110134456.7A CN202110134456A CN114840280A CN 114840280 A CN114840280 A CN 114840280A CN 202110134456 A CN202110134456 A CN 202110134456A CN 114840280 A CN114840280 A CN 114840280A
Authority
CN
China
Prior art keywords
electronic device
display area
application
display
user interface
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110134456.7A
Other languages
Chinese (zh)
Inventor
高凌云
史雪菲
赵博
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huawei Technologies Co Ltd
Original Assignee
Huawei Technologies Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huawei Technologies Co Ltd filed Critical Huawei Technologies Co Ltd
Priority to CN202110134456.7A priority Critical patent/CN114840280A/en
Priority to PCT/CN2022/070132 priority patent/WO2022161119A1/en
Publication of CN114840280A publication Critical patent/CN114840280A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures

Abstract

The embodiment of the application provides a display method and electronic equipment, wherein the method is applied to the electronic equipment comprising a first display area and a second display area, and the method comprises the following steps: receiving a first operation; in response to the first operation, the electronic device determines that the first display area is associated with the first application; when the electronic equipment is in a first physical state, displaying a user interface of a first application through a first display area; receiving a second operation; responding to the second operation, and displaying a user interface of the second application through the first display area by the electronic equipment; in response to the physical state of the electronic device changing from the first physical state to the second physical state, the electronic device displaying a user interface of a second application through a second display area; in response to the physical state of the electronic device changing from the second physical state to the first physical state, the electronic device displays a user interface of the first application through the first display area. The application switching method and the application switching device can enable the user to quickly and conveniently switch the application displayed by the electronic equipment.

Description

Display method and electronic equipment
Technical Field
The present application relates to the field of electronic devices, and in particular, to a display method and an electronic device.
Background
In the process of using an electronic device such as a smart phone or a tablet computer, a user often needs to frequently switch an application program (application for short) displayed by the electronic device. For example, when a user is entertaining using a gaming application, it is often necessary to switch to a social application to reply to the message. At present, a user can close or hide an application currently displayed on an electronic device through functions such as three-key navigation and gesture navigation, and then open other applications to be used so that the electronic device displays user interfaces of the other applications. Such operations are cumbersome and have a slow response speed. Or, the user can also make the electronic device display the user interfaces of a plurality of applications through the functions of split screen and floating window, but the user interface of each application is small, the display effect is not good, and the user is inconvenient to use.
Disclosure of Invention
The embodiment of the application discloses a display method and electronic equipment, which can rapidly and conveniently switch displayed applications, and the display effect of each application is better.
In a first aspect, an embodiment of the present application provides a display method applied to an electronic device, where the electronic device includes a first display area and a second display area, and the method includes: receiving a first user operation; in response to the first user operation, the electronic device determines that the first display area is associated with a first application; when the electronic equipment is in a first physical state, displaying a user interface of the first application through the first display area; receiving a second user operation; responding to the second user operation, the electronic equipment displays a user interface of a second application through the first display area, wherein the first application and the second application are different; when the user interface of the second application is displayed through the first display area, responding to a third user operation, the physical state of the electronic equipment is changed from the first physical state to a second physical state, and the user interface of the second application is displayed through the second display area by the electronic equipment; when the user interface of the second application is displayed through the second display area, the electronic device displays the user interface of the first application through the first display area in response to the physical state of the electronic device being changed from the second physical state to the first physical state.
In this embodiment, after the electronic device determines that the first display area is associated with the first application, when the physical state of the electronic device is changed to the first physical state, the electronic device may display a user interface of the first application through the first display area. That is to say, the user can rapidly switch the applications displayed by the electronic device by changing the physical state of the electronic device, without exiting or hiding the currently displayed applications and reopening the applications that the user wants to view many times, which is greatly convenient for the user to use. In addition, the display areas of the electronic equipment for displaying the user interface in different physical states are different, the existing layout of the application interface cannot be changed, and the display effect is better.
In one possible implementation, the first display area includes at least a portion of the second display area; and/or the second display area comprises at least part of the first display area.
Illustratively, the first display area and the second display area belong to the same display screen of the electronic device. The first display area and the second display area are overlapped display areas, or the first display area is a partial display area of the second display area, or the second display area is a partial display area of the first display area.
In one possible implementation manner, the electronic device is a foldable electronic device; when the electronic equipment is in an unfolded state, the first display area and the second display area are in the same plane; when the electronic device is in a folded state, the light-emitting surface of the first display area and the light-emitting surface of the second display area are opposite to each other.
In some embodiments, the electronic device includes a first display screen, the first display area is at least a partial display area of the first display screen, and the second display area is at least a partial display area of the first display screen.
In some embodiments, the electronic device includes a first display screen and a second display screen, the first display area is a display area of the first display screen, and the second display area is a display area of the second display screen.
In some embodiments, the first physical state and the second physical state are both folded states, and the third user operation is an operation of the user to turn over the electronic device.
In a possible implementation manner, the electronic device is a foldable electronic device, the electronic device includes a first display screen, the second display area is a full-screen display area of the first display screen, the first display area is a partial display area of the first display screen, the first physical state is a folded state, and the second physical state is an unfolded state.
In a possible implementation manner, the electronic device is a foldable electronic device, and when the electronic device is in an unfolded state, a light emitting surface of the first display area and a light emitting surface of the second display area are opposite to each other; the first physical state is the unfolded state, the second physical state is the folded state, and the third user operation is an operation of the user to change the electronic device from the unfolded state to the folded state.
In a possible implementation manner, the electronic device is a foldable electronic device, and when the electronic device is in an unfolded state, a light emitting surface of the first display area and a light emitting surface of the second display area are opposite to each other; the first physical state and the second physical state are both the extended states, and the third user operation is an operation of the user to flip the electronic device.
In the embodiment of the application, the forms of the first display area and the second display area are various, the display method can be applied to foldable electronic equipment in various forms, the application scenes are wide, and the usability is high.
In a possible implementation manner, the method further includes: when the user interface of the first application is displayed through the first display area, the electronic equipment displays the user interface of the first application through the second display area in response to the physical state of the electronic equipment being changed from the first physical state to the second physical state; receiving a fourth user operation when the electronic device displays the user interface of the first application through the second display area; and responding to the fourth user operation, the electronic equipment displays a user interface of a third application through the second display area, wherein the third application is different from the first application.
In this embodiment, when the physical state of the electronic device is changed from the first physical state to the second physical state, the application displayed when the electronic device is in the first physical state may be continuously displayed through the second display area. Therefore, the user can continue to use the application displayed when the electronic equipment is in the first physical state, the user does not need to manually open the application again, the data of the application cannot be lost, and the user can use the electronic equipment more conveniently.
In one possible implementation manner, when the user interface of the second application is displayed through the second display area, in response to the physical state of the electronic device being changed from the second physical state to the first physical state, after the user interface of the first application is displayed through the first display area by the electronic device, the method further includes: and in response to the physical state of the electronic equipment being changed from the first physical state to the second physical state, the electronic equipment displays a user interface of the second application through the second display area.
In this embodiment, when the physical state of the electronic device is changed from the first physical state to the second physical state, the application displayed when the electronic device is in the second physical state last time may be displayed through the second display area. Therefore, the user can continue to use the application displayed when the electronic equipment is in the second physical state last time, the user does not need to manually open the application again, the data of the application cannot be lost, and the user can use the electronic equipment more conveniently.
In a possible implementation manner, the method further includes: receiving a fifth user operation; in response to the fifth user operation, the electronic device determines that the second display area is associated with a fourth application; when the user interface of the first application is displayed through the first display area, the electronic device displays the user interface of the fourth application through the second display area in response to the physical state of the electronic device being changed from the first physical state to the second physical state, wherein the first application and the fourth application are different.
In the embodiment of the application, a user can set different display areas and different application associations. If the user needs to frequently use the first application and the fourth application, the user can directly convert the physical state of the electronic equipment so as to quickly switch the application displayed by the electronic equipment without multiple manual operations, and the use of the user is greatly facilitated.
In a possible implementation manner, when the receiving of the first user operation is performed, the electronic device is in the first physical state, and displays a user interface of the first application through the first display area; the first user operation is a user operation for displaying a display area of a user interface of the first application.
In the embodiment of the application, a user can directly set the first display area to be associated with the first application when the electronic equipment displays the user interface of the first application through the first display area, and the user does not need to set a specific interface, so that the user can use the electronic equipment more conveniently.
In a second aspect, an embodiment of the present application provides a display method for a Graphical User Interface (GUI), which is applied to an electronic device, where the electronic device includes a first display area and a second display area, and the display method is provided in any one of the first aspect and an implementation manner of the first aspect.
In a third aspect, an embodiment of the present application provides an electronic device, which includes a first display area, a second display area, one or more memories, and one or more processors; the one or more memories are configured to store a computer program, the one or more processors are configured to invoke the computer program, and the computer program includes instructions that, when executed by the one or more processors, cause the electronic device to perform the display method provided in any one of the implementations of the first aspect and the first aspect.
In a fourth aspect, the present application provides a computer storage medium, which includes a computer program including instructions that, when executed on a processor, implement the display method provided in any one of the first aspect and the implementation manner of the first aspect.
In a fifth aspect, the present application provides a computer program product, which when run on an electronic device, causes the electronic device to execute the display method provided in any one of the implementation manners of the first aspect and the first aspect.
In a sixth aspect, embodiments of the present application provide a chip, where the chip includes at least one processor and an interface circuit, and optionally, the chip further includes a memory; the memory, the interface circuit and the at least one processor are interconnected by a circuit, and the at least one memory stores a computer program; the computer program implements the display method provided by any one of the implementation manners of the first aspect and the first aspect when executed by the at least one processor.
It is to be understood that the electronic device provided by the third aspect, the computer storage medium provided by the fourth aspect, the computer program product provided by the fifth aspect, and the chip provided by the sixth aspect are all configured to execute the first aspect and the display method provided by any implementation manner of the first aspect. Therefore, the advantageous effects achieved by the method can refer to the advantageous effects of the display method provided in the first aspect, and are not described herein again.
Drawings
The drawings used in the embodiments of the present application are described below.
Fig. 1 is a schematic hardware structure diagram of an electronic device according to an embodiment of the present disclosure;
fig. 2 is a schematic diagram of a software architecture of an electronic device according to an embodiment of the present application;
3-5 are schematic diagrams of physical states of some electronic devices provided by embodiments of the present application;
6-17 are schematic diagrams of some of the human-machine interactions provided by embodiments of the present application;
18-22 are schematic diagrams of some user interface embodiments provided by embodiments of the present application;
23-26 are flow diagrams of hardware-driven interaction within some electronic devices provided by embodiments of the present application;
fig. 27 is a flowchart illustrating a display method according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be described in detail and clearly with reference to the accompanying drawings. The terminology used in the description of the embodiments herein is for the purpose of describing particular embodiments herein only and is not intended to be limiting of the application.
The application provides a display method which can be applied to electronic equipment, and the electronic equipment can comprise a first display area and a second display area. The physical state of the electronic equipment changes, the display area of the electronic equipment changes, and the application of the display changes. For example, a first display area is associated with a first application and a second display area is associated with a second application. Therefore, when the physical state of the electronic device is in the first physical state, the electronic device can display the user interface of the first application through the first display area. When the physical state of the electronic device is in the second physical state, the electronic device may display a user interface of the second application through the second display area. Therefore, the user can quickly and conveniently switch the application displayed by the electronic equipment, and the use is more convenient.
In some embodiments, the electronic device is configured with a foldable display screen (which may be referred to as a folding screen). The electronic device may be referred to as a foldable electronic device. The folding of the folding screen may also be referred to as the folding of the electronic device, and the physical state of the folding screen may also be referred to as the physical state of the electronic device. For simplicity of description, the following embodiments will be described with the foldable electronic device simply referred to as an electronic device.
In this application, an electronic device may run at least one application. The electronic device may or may not display the user interface of the running application. When the electronic device displays a user interface of an application, a user operation (e.g., a click operation on a display screen) acting on the electronic device may be detected, and in response to the operation, the electronic device may perform a corresponding task based on the application. That is, the user may operate the application by operating the user interface of the application. Such applications can provide users with more functionality and more resources required. The electronic device may run the application through any available resource, such as a network resource like bandwidth, a system resource like a Central Processing Unit (CPU).
And although other applications run by the electronic device are not displayed by the electronic device, the electronic device can display notification messages, frame popping messages and other prompting messages of the applications. For example, when the social application is running but not displayed in the electronic device, if the electronic device receives a message sent by another user based on the social application, the message may be pushed to the user in the form of a notification message. That is, although the user cannot operate the user interface of the application, the user may receive the prompt information of the application, so as to obtain the related process of the application through the prompt information. Such applications may provide less functionality and require less resources to the user. For example, when the electronic device displays a user interface of a social application, a user may input text or voice through the user interface of the social application, whereas the user cannot input text or voice without displaying the user interface. The electronic device may run the other applications described above through partially available resources.
The electronic device according to the embodiment of the present application may be, but not limited to, a mobile phone, a tablet Computer, a Personal Digital Assistant (PDA), a handheld Computer, a wearable electronic device (e.g., a smart watch, a smart bracelet), an Augmented Reality (AR) device (e.g., AR glasses), a terminal device such as a Virtual Reality (VR) device (e.g., VR glasses), a smart home device such as a smart television, or other devices such as a desktop, a laptop, a notebook, an Ultra-mobile Personal Computer (UMPC), and a netbook. The embodiment of the present application does not limit the specific type of the electronic device.
Next, the structure of an exemplary electronic device provided in the embodiment of the present application is described.
Referring to fig. 1, fig. 1 shows a schematic structural diagram of an electronic device 100.
As shown in fig. 1, the electronic device 100 may include a processor 110, an external memory interface 120, an internal memory 121, a Universal Serial Bus (USB) interface 130, a charging management module 140, a power management module 141, a battery 142, an antenna 1, an antenna 2, a mobile communication module 150, a wireless communication module 160, an audio module 170, a speaker 170A, a receiver 170B, a microphone 170C, an earphone interface 170D, a sensor module 180, a button 190, a motor 191, an indicator 192, a camera 193, a display screen 194, a Subscriber Identity Module (SIM) card interface 195, and the like. The sensor module 180 may include a pressure sensor 180A, a gyroscope sensor 180B, an air pressure sensor 180C, a magnetic sensor 180D, an acceleration sensor 180E, a distance sensor 180F, a proximity light sensor 180G, a fingerprint sensor 180H, a temperature sensor 180J, a touch sensor 180K, an ambient light sensor 180L, an angle sensor 180M, and the like.
It is to be understood that the illustrated structure of the embodiment of the present invention does not specifically limit the electronic device 100. In other embodiments of the present application, electronic device 100 may include more or fewer components than shown, or some components may be combined, some components may be split, or a different arrangement of components. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 110 may include one or more processing units, such as: the processor 110 may include an Application Processor (AP), a modem processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a video codec, a Digital Signal Processor (DSP), a baseband processor, and/or a neural-Network Processing Unit (NPU), etc. The different processing units may be separate devices or may be integrated into one or more processors.
The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in processor 110 for storing instructions and data. In some embodiments, the memory in the processor 110 is a cache memory. The memory may hold instructions or data that have just been used or recycled by the processor 110. If the processor 110 needs to reuse the instruction or data, it can be called directly from the memory. Avoiding repeated accesses reduces the latency of the processor 110, thereby increasing the efficiency of the system.
The charging management module 140 is configured to receive charging input from a charger. The charger may be a wireless charger or a wired charger. In some wired charging embodiments, the charging management module 140 may receive charging input from a wired charger via the USB interface 130. In some wireless charging embodiments, the charging management module 140 may receive a wireless charging input through a wireless charging coil of the electronic device 100. The charging management module 140 may also supply power to the electronic device through the power management module 141 while charging the battery 142.
The power management module 141 is used to connect the battery 142, the charging management module 140 and the processor 110. The power management module 141 receives input from the battery 142 and/or the charge management module 140, and supplies power to the processor 110, the internal memory 121, the display 194, the camera 193, the wireless communication module 160, and the like. The power management module 141 may also be used to monitor parameters such as battery capacity, battery cycle count, battery state of health (leakage, impedance), etc. In other embodiments, the power management module 141 may be disposed in the processor 110. In other embodiments, the power management module 141 and the charging management module 140 may be disposed in the same device.
The wireless communication function of the electronic device 100 may be implemented by the antenna 1, the antenna 2, the mobile communication module 150, the wireless communication module 160, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 100 may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 150 may provide a solution including 2G/3G/4G/5G wireless communication applied to the electronic device 100. The mobile communication module 150 may include at least one filter, a switch, a power amplifier, a Low Noise Amplifier (LNA), and the like. The mobile communication module 150 may receive the electromagnetic wave from the antenna 1, filter, amplify, etc. the received electromagnetic wave, and transmit the electromagnetic wave to the modem processor for demodulation. The mobile communication module 150 may also amplify the signal modulated by the modem processor, and convert the signal into electromagnetic wave through the antenna 1 to radiate the electromagnetic wave. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be disposed in the processor 110. In some embodiments, at least some of the functional modules of the mobile communication module 150 may be disposed in the same device as at least some of the modules of the processor 110.
The wireless communication module 160 may provide a solution for wireless communication applied to the electronic device 100, including Wireless Local Area Networks (WLANs) (e.g., wireless fidelity (Wi-Fi) networks), bluetooth (bluetooth, BT), Global Navigation Satellite System (GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared (IR), and the like. The wireless communication module 160 may be one or more devices integrating at least one communication processing module. The wireless communication module 160 receives electromagnetic waves via the antenna 2, performs frequency modulation and filtering processing on electromagnetic wave signals, and transmits the processed signals to the processor 110. The wireless communication module 160 may also receive a signal to be transmitted from the processor 110, perform frequency modulation and amplification on the signal, and convert the signal into electromagnetic waves through the antenna 2 to radiate the electromagnetic waves.
In some embodiments, antenna 1 of electronic device 100 is coupled to mobile communication module 150 and antenna 2 is coupled to wireless communication module 160 so that electronic device 100 can communicate with networks and other devices through wireless communication techniques. The wireless communication technology may include global system for mobile communications (GSM), General Packet Radio Service (GPRS), code division multiple access (code division multiple access, CDMA), Wideband Code Division Multiple Access (WCDMA), time-division code division multiple access (time-division code division multiple access, TD-SCDMA), Long Term Evolution (LTE), LTE, BT, GNSS, WLAN, NFC, FM, and/or IR technologies, etc. The GNSS may include a Global Positioning System (GPS), a global navigation satellite system (GLONASS), a beidou navigation satellite system (BDS), a quasi-zenith satellite system (QZSS), and/or a Satellite Based Augmentation System (SBAS).
The electronic device 100 implements display functions via the GPU, the display screen 194, and the application processor. The GPU is a microprocessor for image processing, and is connected to the display screen 194 and an application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. The processor 110 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 194 is used to display images, video, and the like. The display screen 194 includes a display panel. The display panel may adopt a Liquid Crystal Display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (active-matrix organic light-emitting diode, AMOLED), a flexible light-emitting diode (FLED), a miniature, a Micro-oeld, a quantum dot light-emitting diode (QLED), and the like. In some embodiments, the electronic device 100 may include 1 or N display screens 194, with N being a positive integer greater than 1.
In this application, the display screen 194 may include at least one display area, and the display screen 194 includes a first display area and a second display area.
When the electronic device 100 is in the first physical state, the first display area of the display screen 194 is used for displaying a Graphical User Interface (GUI), which may also be referred to as a user interface. At this time, the first display area is in a bright screen state, and the second display area may be in a dark screen state. When the electronic device 100 is in the second physical state, the second display area of the display screen 194 is used for displaying the user interface. At this time, the second display area is in a bright screen state, and the first display area may be in a screen-off state. The second display area when the electronic device 100 is in the first physical state and the first display area when the electronic device 100 is in the second physical state may also be in a screen-locked state or a screen-bright state, which is not limited in this application.
For convenience of description, in the following embodiments, the first display area is turned on and the second display area is turned off when the electronic device 100 is in the first physical state, and the first display area is turned off and the second display area is turned on when the electronic device 100 is in the second physical state.
In some embodiments, the display screen 194 of the electronic device 100 is a folding screen, and the physical state of the electronic device 100 may include an unfolded state, a folded state, and a folded state. When the electronic device 100 is in the unfolded state, the light emitting surface of the first display area is opposite to the light emitting surface of the second display area, and at this time, the first display area may be in the flattened state, for a specific example, see fig. 3 below. When the electronic device 100 is in the folded state and the folded state, the first display region may be folded, for a specific example, see fig. 4 to 5 below. Illustratively, the first physical state is an unfolded state, the second physical state is a folded state, and the first display area and the second display area are the first display screen 200 and the second display screen 300 shown in fig. 3-5, respectively.
The electronic device 100 may implement a shooting function through the ISP, the camera 193, the video codec, the GPU, the display 194, the application processor, and the like. In some embodiments, the electronic device 100 may unlock a face, access an application lock, and the like through the face information acquired by the shooting function.
The ISP is used to process the data fed back by the camera 193. For example, when a photo is taken, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing and converting into an image visible to naked eyes. The ISP can also carry out algorithm optimization on the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in camera 193.
The camera 193 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image to the photosensitive element. The photosensitive element may be a Charge Coupled Device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor. The light sensing element converts the optical signal into an electrical signal, which is then passed to the ISP where it is converted into a digital image signal. And the ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into image signal in standard RGB, YUV and other formats. In some embodiments, electronic device 100 may include 1 or N cameras 193, N being a positive integer greater than 1.
In some embodiments, the display screen 194 is a folding screen, and the display screen 194 includes a first display area and a second display area. The electronic device 100 may include N cameras 193, and the N cameras 193 are disposed on the first display region and the second display region, respectively. The electronic apparatus 100 can determine whether the electronic apparatus 100 is flipped or not by the detection signals of the N cameras 193.
Illustratively, the first display area may be in a flattened state and opposite the user's face when the electronic device 100 is in the unfolded state. At this time, the camera 193 on the first display area may acquire the face information of the user, and the camera 193 on the second display area may not acquire the face information of the user. The electronic device 100 may display the user interface of the application a through the first display area. The electronic device 100 may receive a user's flipping operation for changing the relative positions of the first display area and the second display area, i.e., for opposing the second display area to the user's face. When the camera 193 in the second display area can acquire the face information of the user and the camera 193 in the first display area cannot acquire the face information of the user, the electronic device 100 may determine that the electronic device 100 is turned over. At this time, the second display area is opposite to the user face, and the electronic device 100 may display the user interface of the application B through the second display area.
The external memory interface 120 may be used to connect an external memory card, such as a Micro SD card, to extend the memory capability of the electronic device 100. The external memory card communicates with the processor 110 through the external memory interface 120 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
The internal memory 121 may be used to store computer-executable program code, which includes instructions. The internal memory 121 may include a program storage area and a data storage area. The storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required by at least one function, and the like. The storage data area may store data (such as audio data, phone book, etc.) created during use of the electronic device 100, and the like. In addition, the internal memory 121 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (UFS), and the like. The processor 110 executes various functional applications of the electronic device 100 and data processing by executing instructions stored in the internal memory 121 and/or instructions stored in a memory provided in the processor.
The electronic device 100 may implement audio functions via the audio module 170, the speaker 170A, the receiver 170B, the microphone 170C, the headphone interface 170D, and the application processor. Such as music playing, recording, etc.
The audio module 170 is used to convert digital audio information into an analog audio signal output and also to convert an analog audio input into a digital audio signal. The audio module 170 may also be used to encode and decode audio signals. In some embodiments, the audio module 170 may be disposed in the processor 110, or some functional modules of the audio module 170 may be disposed in the processor 110.
The speaker 170A, also called a "horn", is used to convert the audio electrical signal into an acoustic signal. The electronic apparatus 100 can listen to music through the speaker 170A or listen to a handsfree call.
The receiver 170B, also called "earpiece", is used to convert the electrical audio signal into a sound signal. When the electronic apparatus 100 receives a call or voice information, it can receive voice by placing the receiver 170B close to the ear of the person.
The microphone 170C, also referred to as a "microphone," is used to convert sound signals into electrical signals. When making a call or transmitting voice information, the user can input a voice signal to the microphone 170C by speaking the user's mouth near the microphone 170C. The electronic device 100 may be provided with at least one microphone 170C. In other embodiments, the electronic device 100 may be provided with two microphones 170C to achieve a noise reduction function in addition to collecting sound signals. In other embodiments, the electronic device 100 may further include three, four or more microphones 170C to collect sound signals, reduce noise, identify sound sources, perform directional recording, and so on.
The headphone interface 170D is used to connect a wired headphone. The headset interface 170D may be the USB interface 130, or may be a 3.5mm open mobile electronic device platform (OMTP) standard interface, a cellular telecommunications industry association (cellular telecommunications industry association of the USA, CTIA) standard interface.
The pressure sensor 180A is used for sensing a pressure signal, and converting the pressure signal into an electrical signal. In some embodiments, the pressure sensor 180A may be disposed on the display screen 194. The pressure sensor 180A can be of a wide variety, such as a resistive pressure sensor, an inductive pressure sensor, a capacitive pressure sensor, and the like. The capacitive pressure sensor may be a sensor comprising at least two parallel plates having an electrically conductive material. When a force acts on the pressure sensor 180A, the capacitance between the electrodes changes. The electronic device 100 determines the strength of the pressure from the change in capacitance. When a touch operation is applied to the display screen 194, the electronic device 100 may obtain a corresponding touch operation strength according to the detection signal of the pressure sensor 180A. The electronic device 100 may also calculate a position of the touch area (simply referred to as a touch position) on the display screen 194 where the touch operation is applied according to the detection signal of the pressure sensor 180A. The electronic device 100 may further calculate the shape of the touch area according to the detection signal of the pressure sensor 180A.
In some embodiments, the touch operations that are applied to the same touch position but have different touch operation intensities may correspond to different operation instructions. For example: and when the touch operation with the touch operation intensity smaller than the first pressure threshold value acts on the short message application icon, executing an instruction for viewing the short message. And when the touch operation with the touch operation intensity larger than or equal to the first pressure threshold value acts on the short message application icon, executing an instruction of newly building the short message.
The gyro sensor 180B may detect the magnitude of angular velocity of the electronic device 100 in various directions (typically three axes, i.e., x, y, and z axes), which may be used to determine the motion attitude of the electronic device 100. Alternatively, the gyro sensor 180B may be provided on a circuit board of the electronic apparatus 100. The electronic apparatus 100 may determine whether the bending angle of the electronic apparatus 100 is changed or whether the electronic apparatus 100 is flipped according to the detection signal of the gyro sensor 180B. In some embodiments, the gyro sensor 180B may also be used for photographing anti-shake. Illustratively, when the shutter is pressed, the gyro sensor 180B detects a shake angle of the electronic device 100, calculates a distance to be compensated for by the lens module according to the shake angle, and allows the lens to counteract the shake of the electronic device 100 through a reverse movement, thereby achieving anti-shake. The gyroscope sensor 180B may also be used for navigation, somatosensory gaming scenes.
The acceleration sensor 180E can detect the magnitude of acceleration of the electronic apparatus 100 in various directions (generally, three axes). The magnitude and direction of gravity can be detected when the electronic device 100 is stationary. The method can also be used for recognizing the posture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and other applications. Alternatively, the acceleration sensor 180E may be provided on a circuit board of the electronic apparatus 100. The electronic apparatus 100 may determine whether the bending angle of the electronic apparatus 100 is changed or whether the electronic apparatus 100 is turned over according to the detection signal of the acceleration sensor 180E.
In some embodiments, the electronic device 100 may include multiple acceleration sensors 180E and/or multiple gyroscope sensors 180B. The display screen 194 is a folding screen, and the display screen 194 includes a first display area and a second display area. When the electronic device 100 is in the unfolded state, the first display region may be in a flat state, and when the electronic device 100 is in the bent state and the folded state, the first display region may be bent. The first display area can be divided into two display areas when being bent, and planes of the two display areas are intersected. The plurality of acceleration sensors 180E and/or the plurality of gyro sensors 180B may be disposed on the circuit boards on the two display area sides, respectively. Accordingly, the electronic apparatus 100 may determine whether the bending angle of the electronic apparatus 100 is changed or whether the electronic apparatus 100 is flipped according to the detection signals of the plurality of acceleration sensors 180E and/or the plurality of gyro sensors 180B.
Illustratively, when the electronic device 100 is in the unfolded state, the first display area faces upward and displays the user interface of the application a, and the second display area faces downward and displays the screen. The electronic device 100 may receive a user's flipping operation for changing the relative positions of the first display area and the second display area, i.e., the first display area facing downward and the second display area facing upward. The electronic apparatus 100 may determine that the electronic apparatus 100 is flipped according to the detection signals of the acceleration sensor 180E and/or the gyro sensor 180B. In response to the flipping operation, the electronic device 100 may display the user interface of the application B through the second display region, while the first display region faces downward and turns the screen.
Illustratively, when the electronic device 100 is in the unfolded state, the first display area faces upward and displays the user interface of the application a, and the second display area faces downward and displays the screen. The electronic apparatus 100 may receive a folding operation of a user for folding the electronic apparatus 100. The electronic apparatus 100 may determine that the bending angle of the electronic apparatus 100 becomes smaller and is in the folded state from the detection signal of the acceleration sensor 180E and/or the gyro sensor 180B. When the electronic device 100 is in the folded state, the second display region is bent and can be divided into two display regions, and planes of the two display regions intersect. In response to the folding operation, the electronic apparatus 100 may display the user interface of the application B through the display area facing upward of the two display areas, and the display area facing downward of the two display areas may be turned off. Among them, the electronic apparatus 100 may determine the display area facing upward in the two display areas according to the detection signal of the acceleration sensor 180E and/or the gyro sensor 180B.
The fingerprint sensor 180H is used to collect a fingerprint. The electronic device 100 can utilize the collected fingerprint characteristics to unlock the fingerprint, access the application lock, photograph the fingerprint, answer an incoming call with the fingerprint, and so on.
The touch sensor 180K is also called a "touch device". In some embodiments, the touch sensor 180K may be disposed on the display screen 194, and the touch sensor 180K and the display screen 194 form a touch screen, also referred to as a "touch screen". The touch sensor 180K is used to detect a touch operation applied thereto or nearby. The touch sensor may communicate the detected touch operation to the application processor to determine the location of the touch operation on the display screen 194, the shape of the touch area, and thus the type of touch event. The electronic device 100 may provide visual output related to touch operations via the display screen 194. In other embodiments, the touch sensor 180K may be disposed on a surface of the electronic device 100, different from the position of the display screen 194.
The angle sensor 180M may take the angle information and convert it to a usable electrical signal output. In some embodiments, the angle sensor 180M may be disposed in the display screen 194 for detecting a bending angle of the electronic device 100. The processor 110 may determine whether the physical state of the electronic device 100 (e.g., the unfolded state, the folded state, or the folded state) and the physical state of the electronic device 100 are changed according to the detection signal of the angle sensor 180M.
The application is not limited to a particular type of sensor for detecting a physical condition.
The keys 190 include a power-on key, a volume key, and the like. The keys 190 may be mechanical keys. Or may be touch keys. The electronic apparatus 100 may receive a key input, and generate a key signal input related to user setting and function control of the electronic apparatus 100.
The motor 191 may generate a vibration cue. The motor 191 may be used for incoming call vibration cues, as well as for touch vibration feedback. For example, touch operations applied to different applications (e.g., photographing, audio playing, etc.) may correspond to different vibration feedback effects. The motor 191 may also respond to different vibration feedback effects for touch operations applied to different areas of the display screen 194. Different application scenes (such as time reminding, receiving information, alarm clock, game and the like) can also correspond to different vibration feedback effects. The touch vibration feedback effect may also support customization.
Indicator 192 may be an indicator light that may be used to indicate a state of charge, a change in charge, or a message, missed call, notification, etc.
The SIM card interface 195 is used to connect a SIM card. The SIM card can be brought into and out of contact with the electronic apparatus 100 by being inserted into the SIM card interface 195 or being pulled out of the SIM card interface 195. The electronic device 100 may support 1 or N SIM card interfaces, N being a positive integer greater than 1. The SIM card interface 195 may support a Nano SIM card, a Micro SIM card, a SIM card, etc. The same SIM card interface 195 can be inserted with multiple cards at the same time. The types of the plurality of cards may be the same or different. The SIM card interface 195 may also be compatible with different types of SIM cards. The SIM card interface 195 may also be compatible with external memory cards. The electronic device 100 interacts with the network through the SIM card to implement functions such as communication and data communication. In some embodiments, the electronic device 100 employs esims, namely: an embedded SIM card. The eSIM card can be embedded in the electronic device 100 and cannot be separated from the electronic device 100.
In the present application, the pressure sensor 180A and/or the touch sensor 180K may be disposed in the display screen 194. When the display screen 194 displays a user interface of an application, the pressure sensor 180A and/or the touch sensor 180K may detect a user operation applied to the user interface by a user. In response to the user operation, the electronic device 100 may perform a corresponding task based on the application. For example, if the user clicks on the head portrait of a friend in the social application, the electronic device 100 may display personal information published by the friend through the social application on the display screen 194.
The software system of the electronic device 100 may employ a layered architecture, an event-driven architecture, a micro-core architecture, a micro-service architecture, or a cloud architecture. For example, the software system of the hierarchical architecture may be an Android (Android) system, or a Huawei Mobile Services (HMS) system. The embodiment of the present application takes an Android system with a layered architecture as an example, and exemplarily illustrates a software structure of the electronic device 100.
Fig. 2 is a block diagram of a software configuration of the electronic apparatus 100 according to the embodiment of the present invention.
The layered architecture divides the software into several layers, each layer having a clear role and division of labor. The layers communicate with each other through a software interface. In some embodiments, the Android system is divided into four layers, an application layer, an application framework layer, an Android runtime (Android runtime) and system library, and a kernel layer from top to bottom.
The application layer may include a series of application packages.
As shown in fig. 2, the application packages may include social applications, gaming applications, video applications, payment applications, cameras, galleries, maps, notes, electronic books, and the like.
The application framework layer provides an Application Programming Interface (API) and a programming framework for the application program of the application layer. The application framework layer includes a number of predefined functions.
As shown in FIG. 2, the application framework layers may include a window manager, content provider, view system, phone manager, resource manager, notification manager, and the like.
The window manager is used for managing window programs. The window manager can obtain the size of the display screen, judge whether a status bar exists, lock the screen, intercept the screen and the like.
The content provider is used to store and retrieve data and make it accessible to applications. Such data may include video, images, audio, calls made and received, browsing history and bookmarks, phone books, etc.
The view system includes visual controls such as controls to display text, controls to display pictures, and the like. The view system may be used to build applications. The display interface may be composed of one or more views. For example, a display interface of an application, including a display interface of a notification message of an application, may include a view displaying text and a view displaying pictures.
The phone manager is used to provide communication functions of the electronic device 100. Such as management of call status (including on, off, etc.).
The resource manager provides various resources for the application, such as localized strings, icons, pictures, layout files, video files, and the like.
The notification manager enables the application to display notification information in the status bar, can be used to convey notification-type messages, can disappear automatically after a short dwell, and does not require user interaction. Such as a notification manager used to inform download completion, message alerts, etc. The notification manager may also be a notification that appears in the form of a chart or scroll bar text at the top status bar of the system, such as a notification of a background running application, or a notification that appears on the screen in the form of a dialog window. For example, prompting text information in the status bar, sounding a prompt tone, vibrating the electronic device, flashing an indicator light, etc.
The Android Runtime comprises a core library and a virtual machine. The Android runtime is responsible for scheduling and managing an Android system.
The core library comprises two parts: one part is a function which needs to be called by java language, and the other part is a core library of android.
The application layer and the application framework layer run in a virtual machine. And executing java files of the application program layer and the application program framework layer into a binary file by the virtual machine. The virtual machine is used for performing the functions of object life cycle management, stack management, thread management, safety and exception management, garbage collection and the like.
The system library may include a plurality of functional modules. For example: surface managers (surface managers), Media Libraries (Media Libraries), three-dimensional graphics processing Libraries (e.g., OpenGL ES), 2D graphics engines (e.g., SGL), and the like.
The surface manager is used to manage the display subsystem and provide fusion of 2D and 3D layers for multiple applications.
The media library supports a variety of commonly used audio, video format playback and recording, and still image files, among others. The media library may support a variety of audio-video encoding formats, such as MPEG4, h.264, MP3, AAC, AMR, JPG, PNG, and the like.
The three-dimensional graphic processing library is used for realizing three-dimensional graphic drawing, image rendering, synthesis, layer processing and the like.
The 2D graphics engine is a drawing engine for 2D drawing.
The kernel layer is a layer between hardware and software. The inner core layer at least comprises a display driver, a camera driver, an audio driver and a sensor driver. Among them, the sensor drive may be used to drive a plurality of sensors in the control hardware, such as the pressure sensor 180A, the gyro sensor 180B, the acceleration sensor 180E, the touch sensor 180K, the angle sensor 180M, and the like shown in fig. 1.
The following describes exemplary workflow of the software and hardware of the electronic device 100 in conjunction with a display scenario. Here, a case where the sensor for detecting the physical state is the gyro sensor 180B will be described as an example.
Assume that the first display area is associated with application a. Illustratively, when the electronic device 100 is in the unfolded state and the first display area faces upward, the electronic device 100 runs the application a and displays the user interface of the application a through the first display area. The electronic device 100 may receive a user operation, and the gyro sensor 180B reports a detection signal to the kernel layer. The core layer may process the detection signal into an input event (e.g., the electronic device 100 is in the unfolded state, the first display area is facing down and the second display area is facing up, or a relative position of the first display area and the second display area changes), which is stored in the core layer. The application framework layer obtains an input event from the kernel layer and determines that the physical state of the electronic device 100 changes according to the input event. Assuming that the second display area is associated with application B, for example, the electronic device 100 is in the unfolded state, and the second display area faces upward, the electronic device 100 runs application B and displays the user interface of application B through the second display area. Therefore, in response to the change of the physical state of the electronic device 100, the second application may call the interface of the application framework layer, launch the second application, and further launch the display driver by calling the kernel layer, and display the user interface of the second application through the second display area.
Next, the physical state of the electronic apparatus equipped with the folding screen will be exemplarily described.
Referring to fig. 3, fig. 3 is a schematic view illustrating an unfolded state of an electronic device according to an embodiment of the present disclosure. Fig. 3 (a) shows a schematic view of a viewing angle of the electronic device. Fig. 3 (B) shows a schematic view of another viewing angle of the electronic device.
As shown in fig. 3 (a), the electronic device 100 may include a first display 200 and a second display 300. The electronic device may be in an unfolded state, a folded state or a bent state. When the electronic device is in the unfolded state, the first display screen 200 is in the flattened state, as shown in fig. 3 (a), the bending angle α of the first display screen 200 is about 180 degrees, but not limited thereto, the bending angle of the first display screen 200 may also be greater than or equal to 170 degrees and less than or equal to 180 degrees, and the specific value of the bending angle of the first display screen 200 in the unfolded state is not limited in the present application. When the electronic device is in the unfolded state, the light emitting surface of the first display screen 200 is opposite to the light emitting surface of the second display screen 300.
The first display screen 200 may include three display areas, i.e., a first display area 201, a second display area 202, and a third display area 203. The third display region 203 is bendable and located on a bending portion of the electronic device. The two ends of the bending part of the electronic device are respectively connected with the first display area 201 and the second display area 202. The bending angle of the first display 200 can also be understood as: an angle α between a plane where the first display region 201 is located and a plane where the second display region 202 is located, that is, an angle α between both ends of the bending portion shown in fig. 3 (a). As shown in fig. 3 (a), when the first display screen 200 is in the flat state, the first display area 201 and the second display area 202 are equivalent to being in the same plane.
The second display screen 300 may include three display regions, i.e., a fourth display region 301, a fifth display region 302, and a sixth display region 303. When the electronic device is in the unfolded state, the first display region 201 is opposite to the fourth display region 301, the second display region 202 is opposite to the sixth display region 303, and the third display region 203 is opposite to the fifth display region 302.
As shown in fig. 3 (B), the fifth display area 302 of the second display 300 is bendable and located at a bending portion of the electronic device. The two ends of the bending part of the electronic device are also respectively connected with the fourth display area 301 and the sixth display area 303. It can be understood that the bending angle α of the first display panel 200 is also: 360 degrees-the bending angle α 'of the second display 300, i.e., α is 360- α'. The bending angle of the second display 300 can also be understood as: the angle between the plane of the fourth display area 301 and the plane of the sixth display area 303.
As shown in fig. 3 (B), when the second display screen 200 is in the flat state, the fourth display area 301 and the sixth display area 303 are equivalently in the same plane.
In some embodiments, the electronic device may include at least one camera. As shown in fig. 3 (B), the electronic apparatus may include a camera 3031 (including a camera 3031A and a camera 3031B) disposed at an upper portion of the sixth display region 303, and a camera 3032 (including a camera 3032A and a camera 3032B) disposed at a lower portion of the sixth display region 303. Optionally, the electronic device may acquire face information of the user through the at least one camera, and perform face verification or determine whether a physical state of the electronic device changes according to the acquired face information. Optionally, the electronic device restarts the corresponding application and displays a user interface of the application when the face verification passes.
Without being limited to the above-listed cameras, in a specific implementation, the electronic device may also include only the camera 3031 disposed at the upper portion of the sixth display region 303. Alternatively, the electronic apparatus may include only the camera 3032 provided in the lower portion of the sixth display region 303. Alternatively, the electronic device may include a camera provided in the fourth display region 301 and/or the fifth display region 302. Alternatively, the electronic device may also include a camera disposed on the first display 200, which is not limited in this embodiment.
In some embodiments, the electronic device in the unfolded state shown in fig. 3 may be partially folded (in this case, an included angle between two ends of a bending portion of the electronic device may be changed from α to β), so as to obtain the electronic device in the bent state, as shown in fig. 4 below.
Referring to fig. 4, fig. 4 is a schematic view illustrating a bending state of an electronic device according to an embodiment of the present disclosure. Fig. 4 (a) shows a schematic view of an electronic device from one perspective. Fig. 4 (B) is a schematic view showing another view angle of the electronic apparatus.
As shown in fig. 4 (a), when the electronic device is in the bent state, the first display 200 is bent, and the plane of the first display area 201 intersects with the plane of the second display area 202, as shown in fig. 4 (a), the bending angle β of the first display 200 may be about 120 degrees. Without limitation, the bending angle of the first display 200 may be greater than 0 degree and smaller than 180 degrees, such as but not limited to 60 degrees, 90 degrees, 100 degrees, 120 degrees, and the like, and the specific value of the bending angle of the first display 200 in the bent state is not limited in the present application. The description of fig. 4 (B) is similar to that of fig. 4 (a), and is not repeated.
In some embodiments, the electronic device in the unfolded state shown in fig. 3 may be folded (in this case, the included angle between two ends of the bending portion of the electronic device may be changed from α to γ), so as to obtain the electronic device in the folded state. Alternatively, the electronic device in the folded state shown in fig. 4 may be folded (in this case, the angle between both ends of the folded portion of the electronic device may be converted from β to γ) to obtain the folded electronic device, as shown in fig. 5.
As shown in fig. 5 (a), when the electronic apparatus is in the folded state, the first display 200 is bent or folded, and as shown in fig. 5 (a), the bending angle γ of the first display 200 may be about 0 degree. Without limitation, the bending angle of the first display 200 may be greater than or equal to 0 degree and less than or equal to 20 degrees, and the specific value of the bending angle of the first display 200 in the folded state is not limited in the present application.
It can be understood that when the electronic device is in the folded state, the bent state, and the unfolded state, the bending angle of the display screen 194 may be different, but the specific value of the bending angle of the display screen 194 is not limited.
As shown in fig. 5 (a) and 5 (B), when the electronic device is in a folded state, the first display region 201 and the second display region 202 are opposite, the fourth display region 301 and the sixth display region 303 are opposite, and the third display region 203 and the fifth display region 302 are opposite.
In some embodiments, the first display screen 200 and the second display screen 300 may be different display areas on the same flexible folding screen (flexible screen for short). Alternatively, the first display area 201, the second display area 202, the third display area 203, the fourth display area 301, the fifth display area 302 and the sixth display area 303 may be different areas on the one flexible screen, all for displaying a user interface.
In some embodiments, the first display 200 may be a flexible screen of the electronic device. Optionally, the first display area 201, the second display area 202, and the third display area 203 are different areas on the flexible screen, and are all used for displaying a user interface. The second display screen 300 may be another flexible screen of the electronic device. Optionally, the fourth display area 301, the fifth display area 302, and the sixth display area 303 are different areas on the second display screen 300, and are all used for displaying a user interface.
In some embodiments, the first display screen 200 may be a display screen formed by splicing a rigid screen and a flexible screen, a chain, or other connecting components. For example, the first display screen 200 may be formed by splicing two rigid screens and one flexible screen. The first display area 201 and the second display area 202 may be areas on the two rigid screens, respectively, and the third display area 203 may be an area on the one flexible screen, both for displaying a user interface. Alternatively, the first display 200 may be formed by splicing two rigid panels and a chain for connecting the two rigid panels. The first display area 201 and the second display area 202 may be areas on the two rigid screens, respectively, for displaying a user interface. The third display area 203 is the chain used to connect the two rigid screens as described above. The second display screen 300 may also be a display screen formed by splicing a rigid screen, a flexible screen, a chain and other connecting components, and the specific description is similar to that of the first display screen 200 and is not repeated.
For convenience of description, the electronic device is in a folded state when the bending angle of the electronic device is smaller than a first angle threshold, is in an unfolded state when the bending angle of the electronic device is larger than a second angle threshold, and is in the folded state when the bending angle of the electronic device is larger than or equal to the first angle threshold and smaller than or equal to the second angle threshold. Illustratively, the first angle threshold is 30 degrees and the second angle threshold is 160 degrees.
As shown in fig. 3 to 5, when the electronic device is in the unfolded state, if the first display 200 faces upward, the user interface is displayed through the first display 200, and if the second display 300 faces upward, the user interface is displayed through the second display 300. When the electronic device is in the folded state, if the fourth display area 301 faces upward, the user interface is displayed through the fourth display area 301 (and optionally the fifth display area 302). If the sixth display area 303 is facing upwards, the user interface is displayed through the sixth display area 303 (and optionally the fifth display area 302).
The following describes an application scenario related to the embodiment of the present application and a human-computer interaction diagram in the scenario. In fig. 6 to 17, the structure of the electronic device 100 is described by taking the structure shown in fig. 3 to 5 as an example.
Referring to fig. 6, fig. 6 illustrates a schematic diagram of human-computer interaction.
As shown in fig. 6 (a), the electronic device 100 is in the unfolded state, and the first display 200 faces upward, and the electronic device may display the user interface of the first application through the first display 200. The electronic device may receive a user operation when displaying a setting interface or a desktop, and in response to the user operation, the electronic device sets the first display screen 200 to be associated with the first application. Without being limited thereto, the electronic device may also receive a user operation (for example, clicking a lock control) acting on the first display screen 200 when the user interface of the first application is displayed through the first display screen 200, and in response to the user operation, the electronic device sets the first display screen 200 and the first application association, which is not limited in this application.
As shown in fig. 6 (a), the electronic apparatus may detect a folding operation of a user, and the folding operation may reduce an angle between both ends of the bending portion. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display a user interface through the sixth display area 303 of the second display screen 300. Without being limited thereto, the electronic device may also display the user interface together through the fifth display area 302 and the sixth display area 303. Assuming that there is no application associated with the sixth display area 303, after the electronic device has set the first display screen 200 and the first application to be associated, and the first display screen is changed from the unfolded state to the folded state for the first time, the sixth display area 303 may still display the user interface of the application displayed by the electronic device in the unfolded state, that is, the user interface of the first application, and at this time, the electronic device may be in the state shown in (B) of fig. 6 below.
As shown in fig. 6 (B), the electronic device is in the folded state, and the sixth display area 303 faces upward, and the electronic device can continue to display the user interface of the first application displayed in the unfolded state through the sixth display area 303. The electronic device may receive a user operation to open a second application, such as returning to a desktop via a three-key navigation function, and click an icon of the second application on the desktop, wherein the first application and the second application are different. In response to the user operation, the electronic device may cancel displaying the user interface of the first application and display the user interface of the second application through the sixth display area 303, at which time the electronic device may be in the state shown in fig. 6 (C) below.
As shown in fig. 6 (C), the electronic device is in a folded state, and the sixth display area 303 faces upward, the electronic device may display the user interface of the second application through the sixth display area 303. The electronic device can detect the unfolding operation of a user, and the unfolding operation can enlarge the included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being greater than the second angle threshold, the electronic device may display, through the first display screen 200, a user interface of an application associated with the first display screen 200, that is, a user interface of the first application, where the electronic device may be in a state shown in fig. 6 (a).
In some embodiments, after the electronic device 100 sets the first display screen 200 and the first application association, a user operation may be received while a user interface of the first application is displayed through the first display screen 200. The user operation is used for opening a third application, for example, entering a multi-task interface through a gesture navigation function, and clicking a thumbnail of the third application in the multi-task interface, wherein the first application and the third application are different. In response to the user operation, the electronic device may display a user interface of the third application through the first display screen 200. At this time, the physical state of the electronic device may be transformed into a folded state, and the user interface of the third application is displayed through the second display screen 300. A specific example is shown in fig. 7 below.
Referring to fig. 7, fig. 7 is a schematic diagram illustrating still another human-computer interaction.
Fig. 7 (a) is identical to fig. 6 (a), and the description thereof is omitted. In the state shown in fig. 7 (a), after the electronic device sets the first display screen 200 and the first application association, the electronic device may receive a user operation for opening the third application. In response to the user operation, the electronic apparatus may cancel displaying the user interface of the first application and display the user interface of the third application through the first display screen 200, at which time the electronic apparatus may be in the state shown in fig. 7 (B) below.
As shown in fig. 7 (B), the electronic apparatus is in the unfolded state, and the first display 200 is facing up, the electronic apparatus displays the user interface of the third application through the first display 200. The electronic device can detect the folding operation of the user, and the folding operation can reduce the included angle between the two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display a user interface through the sixth display area 303 of the second display screen 300. Without being limited thereto, the electronic device may also display the user interface together through the fifth display area 302 and the sixth display area 303. Assuming that there is no application associated with the sixth display area 303, after the electronic device sets the first display screen 200 to be associated with the first application and is transformed from the unfolded state to the folded state for the first time, the sixth display area 303 may still display the user interface of the application displayed by the electronic device in the unfolded state, that is, the application interface of the third application, and at this time, the electronic device may be in the state shown in (C) of fig. 7.
As shown in fig. 7 (C), the electronic device is in the folded state, and the sixth display area 303 faces upward, and the electronic device can continue to display the user interface of the third application displayed in the unfolded state through the sixth display area 303. The electronic device can detect the unfolding operation of a user, and the unfolding operation can enlarge the included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being greater than the second angle threshold, the electronic device may display, through the first display screen 200, a user interface of an application associated with the first display screen 200, that is, a user interface of the first application, where the electronic device may be in a state shown in fig. 7 (a).
It is to be understood that, if the electronic device is not provided with the first application and the first display 200, for example, the first display 200 and the first application are not provided in the state shown in fig. 6 (a), the electronic device is changed from the other physical state to the unfolded state, for example, in response to the included angle between the two ends of the bending portion being greater than the second angle threshold, the electronic device may display the user interface of the application displayed by the electronic device in the other physical state, which is not limited to this, and may also be a desktop or a screen locking interface of the electronic device.
In some embodiments, the state change of the electronic device is, in order: fig. 6 (a), fig. 6 (B), fig. 6 (C), and fig. 6 (a). At this time, the electronic device is in the unfolded state, and the first display 200 faces upward, and the user interface of the first application is displayed through the first display 200. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display, through the sixth display area 303, the user interface of the application displayed by the electronic device in the folded state last time, that is, the user interface of the second application, where the electronic device may be in the state shown in (C) of fig. 6.
In some embodiments, the state change of the electronic device is, in order: fig. 7 (a), fig. 7 (B), fig. 7 (C), and fig. 7 (a). At this time, the electronic device is in the unfolded state, and the first display 200 faces upward, and the user interface of the first application is displayed through the first display 200. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display, through the sixth display area 303, the user interface of the application displayed by the electronic device in the folded state last time, that is, the user interface of the third application, where the electronic device may be in the state shown in (C) of fig. 7.
In some embodiments, the electronic device may receive a user operation while displaying the settings interface or desktop, in response to which the electronic device sets the sixth display area 303 to be associated with a fourth application, wherein the first application and the fourth application are different. Without being limited thereto, the electronic device may receive a user operation (for example, clicking a lock control) acting on the sixth display area 303 when the user interface of the fourth application is displayed through the sixth display area 303, and in response to the user operation, the electronic device sets the sixth display area 303 and the fourth application to be associated, which is not limited in this application. In this case, when the electronic device is in the folded state and the sixth display area 303 is facing upward, the user interface of the fourth application may be displayed through the sixth display area 303 (and optionally the fifth display area 302).
For example, when the electronic device is in the unfolded state, the electronic device may detect a folding operation of a user, and the folding operation may reduce an included angle between two ends of the bending portion. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display, through the sixth display area 303, a user interface of an application associated with the sixth display area 303, that is, a user interface of a fourth application.
It should be noted that the user interface of the application displayed in the folded state by the electronic device last time is a case after the first display screen 200 and the first application are associated with each other.
Not limited to the above list, in a specific implementation, the electronic device may set the sixth display area 303 to be associated with the fourth application. In response to the physical state of the electronic device being changed from the folded state to the unfolded state, the electronic device may display a user interface of an application displayed in the unfolded state last time by the electronic device through the first display screen 200. If the application displayed by the electronic device in the folded state last time does not exist, that is, after the electronic device sets the second display screen 300 and the fourth application to be associated, the electronic device is changed from the folded state to the unfolded state for the first time, the electronic device may continue to display the user interface of the application displayed by the electronic device in the folded state through the sixth display area 303. In response to the physical state of the electronic device changing from the unfolded state to the folded state, the electronic device may display a user interface of an application associated with the sixth display region 303, i.e., a user interface of a fourth application, through the sixth display region 303.
Fig. 6 to 7 illustrate an example in which the sixth display region 303 faces upward when the electronic apparatus is in a folded state. If the fourth display area 301 faces upward when the electronic device is in the folded state, the electronic device may display the user interface through the fourth display area 301, but is not limited thereto, and the electronic device may also display the user interface through the fourth display area 301 and the fifth display area 302 together. For example, the electronic device may detect the display area facing upward through the acceleration sensor 180E and/or the gyro sensor 180B.
In some embodiments, the electronic device has set the first display 200 to associate with the video application and also has set the second display 300 to associate with the social application, examples of which may be found in fig. 6-7 above. When the electronic device is in the unfolded state and the first display screen 200 faces upward, the electronic device may display the user interface 110 of the video application through the first display screen 200, as shown in fig. 8 (a). When the electronic device is in the unfolded state and the second display screen 300 is facing upward, the electronic device may display the user interface 120 of the social application through the second display screen 300, as shown in fig. 8 (B).
As shown in fig. 8 (a), the electronic apparatus may detect a user's flip operation, which may change the relative positions of the first display 200 and the second display 300. In response to the first display screen 200 facing downward and the second display screen 300 facing upward, the electronic device may display the user interface of the application associated with the second display screen 300, i.e., the user interface 120 of the social application, through the second display screen 300, at which time the electronic device may be in the state shown in fig. 8 (B).
As shown in (B) of fig. 8, the electronic apparatus may detect a user's flip operation, which may change the relative positions of the first display 200 and the second display 300. In response to the second display screen 300 facing downward and the first display screen 200 facing upward, the electronic device may display the user interface of the application associated with the first display screen 200, i.e., the user interface 110 of the video application, through the first display screen 200, at which time the electronic device may be in the state shown in (a) of fig. 8.
In some embodiments, the electronic device has set the first display 200 to associate with the video application and also has set the second display 300 to associate with the social application, examples of which may be found in fig. 6-7 above. And the electronic device determines that the application associated with the fourth display area 301 and the application associated with the sixth display area 303 are the same as the application associated with the second display screen 300. I.e. the applications associated with the fourth display area 301 and the sixth display area 303 are also social applications. The electronic device may display a user interface of the social application when the physical state of the electronic device is transformed into the collapsed state, a specific example of which is shown in fig. 9 below.
Referring to fig. 9, fig. 9 illustrates yet another human-computer interaction diagram.
As shown in fig. 9 (a), the electronic device is in an unfolded state, and the first display screen 200 is facing up, the electronic device displays the user interface 110 of the video application through the first display screen 200. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display the user interface through the sixth display area 303 or the fourth display area 301, where the electronic device may be in the state shown in fig. 9 (B) or fig. 9 (C).
As shown in fig. 9 (B), the electronic device is in a folded state, and the sixth display region 303 faces upward, and the electronic device displays, through the sixth display region 303, a user interface of an application associated with the sixth display region 303, which is also a user interface of an application associated with the second display screen 300, that is, the user interface 120 of the social application. Without being limited thereto, the electronic device may also display the user interface 120 of the social application together through the sixth display area 303 and the fifth display area 302.
As shown in (C) of fig. 9, the electronic device is in a folded state, and the fourth display area 301 faces upward, the electronic device displays a user interface of an application associated with the fourth display area 301 through the fourth display area 301, and also a user interface of an application associated with the second display screen 300, that is, the user interface 120 of the social application. Without being limited thereto, the electronic device may also display the user interface 120 of the social application together through the fourth display area 301 and the fifth display area 302.
In the state shown in (B) of fig. 9, the electronic apparatus may detect a flip operation by the user, which may change the relative positions of the fourth display region 301 and the sixth display region 303. In response to the sixth display area 303 facing downward and the fourth display area 301 facing upward, the electronic device may display the user interface 120 of the social application through the fourth display area 301, and at this time, the electronic device may be in a state shown in (C) of fig. 9. Similarly, in the state shown in fig. 9 (C), the electronic apparatus may also detect a flip operation by the user, which may change the relative positions of the fourth display region 301 and the sixth display region 303. In response to the fourth display area 301 facing downward and the sixth display area 303 facing upward, the electronic device may display the user interface 120 of the social application through the sixth display area 303, at which time the electronic device may be in the state shown in fig. 9 (B).
In some embodiments, the electronic device has set the first display 200 to be associated with a video application and also has set the second display 300 to be associated with a social application. And, the electronic device has also set the fourth display area 301 to be associated with a gallery and the sixth display area 303 to be associated with a camera, examples of the setting modes can be seen in fig. 6-7 above. That is, the application associated with the fourth display area 301, the application associated with the sixth display area 303, and the application associated with the second display screen 300 are different from each other. The electronic device may determine the application of the display according to the display area facing upward when the physical state of the electronic device is changed to the folded state, as shown in fig. 10 below.
Referring to fig. 10, fig. 10 illustrates yet another human-computer interaction diagram. Fig. 10 (a) is identical to fig. 9 (a), and the description thereof is omitted.
As shown in fig. 10 (B), the electronic apparatus is in a folded state, and the sixth display region 303 faces upward, the electronic apparatus displays the user interface of the application associated with the sixth display region 303, i.e., the user interface 130 of the camera, through the sixth display region 303. Without being limited thereto, the electronic device may also display the user interface 130 of the camera together through the sixth display area 303 and the fifth display area 302.
As shown in (C) of fig. 10, the electronic device is in a folded state, and the fourth display area 301 is facing upward, the electronic device displays the user interface of the application associated with the fourth display area 301, i.e., the user interface 140 of the gallery, through the fourth display area 301. Without being limited thereto, the electronic device may also display the user interface 140 of the gallery together through the fourth display area 301 and the fifth display area 302.
In the state shown in (B) of fig. 10, the electronic apparatus may detect a flip operation by the user, which may change the relative positions of the fourth display region 301 and the sixth display region 303. In response to the sixth display area 303 facing downward and the fourth display area 301 facing upward, the electronic device may display the user interface 140 of the gallery through the fourth display area 301, at which time the electronic device may be in the state shown in fig. 10 (C). Similarly, in the state shown in fig. 10 (C), the electronic apparatus may also detect a flip operation by the user, which may change the relative positions of the fourth display region 301 and the sixth display region 303. In response to the fourth display area 301 facing downward and the sixth display area 303 facing upward, the electronic device may display the user interface 130 of the camera through the sixth display area 303, at which time the electronic device may be in the state shown in fig. 10 (B).
Not limited to the above list, in a specific implementation, the application associated with the fourth display area 301 may be the same as the application associated with the second display screen 300, for example, a social application, and the application associated with the sixth display area 303 may be different from the application associated with the second display screen 300, for example, a camera. Alternatively, the application associated with the sixth display area 303 may be the same as the application associated with the second display 300, and the application associated with the fourth display area 301 may be different from the application associated with the second display 300. Alternatively, the application associated with the fourth display area 301 is the same as the application associated with the sixth display area 303 but different from the application associated with the second display screen 300.
In some embodiments, the electronic device has set the first display screen 200 to associate with the video application. When the electronic device is in the bent state, the user interface of the application may be displayed on the first display 200 in a split manner, as shown in fig. 11 to 12 below.
Referring to fig. 11, fig. 11 illustrates yet another human-computer interaction diagram. Wherein, when the electronic device is in the bending state, the two applications displayed in the split screen mode are different from the application associated with the first display screen 200.
As shown in fig. 11 (a), the electronic device is in a bent state, and the first display 200 is facing upward, the electronic device displays the user interface 150 of the short message through the first display area 201 and a part of the third display area 203, and displays the user interface 160 of the electronic book through the second display area 202 and a part of the third display area 203. The electronic device can detect the unfolding operation of a user, and the unfolding operation can enlarge the included angle between two ends of the bending part. In response to that the included angle between the two ends of the bending portion is greater than the second angle threshold and the first display screen 200 faces upward, the electronic device may display the user interface of the application associated with the first display screen 200, that is, the user interface 110 of the video application, through the first display screen 200, and at this time, the electronic device may be in the state shown in fig. 11 (B). In some embodiments, in response to the angle between the two ends of the bending portion being greater than the second angle threshold and the second display screen 300 facing upward, the electronic device may display a user interface of an application associated with the second display screen 300 through the second display screen 300, and at this time, the electronic device may be in the state shown in fig. 8 (B).
In some embodiments, the electronic device may also detect a folding operation of the user, which may reduce an angle between two ends of the bending portion. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold and the fourth display area 301 facing upward, the electronic device may display the user interface of the application associated with the fourth display area 301 through the fourth display area 301, and at this time, the electronic device may be in the state shown in fig. 9 (C) or fig. 10 (C). Alternatively, in response to the angle between the two ends of the bending portion being smaller than the first angle threshold and the sixth display area 303 facing upward, the electronic device may display the user interface of the application associated with the sixth display area 303 through the sixth display area 303, and at this time, the electronic device may be in the state shown in fig. 9 (B) or fig. 10 (B).
As shown in fig. 11 (B), the electronic device is in the unfolded state, and the first display screen 200 faces upward, and the electronic device displays the user interface 110 of the video application through the first display screen 200. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being greater than or equal to the first angle threshold and being less than or equal to the second angle threshold, the electronic device may display the user interface of the application through the first display screen 200 in a split-screen manner, where the electronic device may be in the state shown in fig. 11 (a) or fig. 12 (a) below.
Referring to fig. 12, fig. 12 illustrates yet another human-computer interaction diagram. One of the two applications displayed in a split screen mode when the electronic device is in the bent state is an application associated with the first display screen 200, namely, a video application.
As shown in fig. 12 (a), the electronic device is in a bent state with the first display screen 200 facing upward, the electronic device displays the user interface 150 for short messages through the first display area 201 and a portion of the third display area 203, and displays the user interface 110 for a video application through the second display area 202 and a portion of the third display area 203. At this time, the description of the change of the physical state when the electronic device detects the user operation is similar to that in fig. 11 (a), and is not repeated. Fig. 12 (B) is similar to fig. 11 (B), and will not be described again.
In a specific implementation, if the electronic device is not provided with an application associated with the second display 300, the electronic device may also display, through the second display 300, a user interface displayed when the electronic device was in the unfolded state or the folded state last time in response to the physical state of the electronic device being changed from the folded state to the unfolded state or the folded state. Illustratively, in response to the angle between the two ends of the bending portion being greater than the second angle threshold and the second display screen 300 being upward, the electronic device may display, through the second display screen 300, the user interface that the electronic device has recently displayed through the second display screen 300. In response to that the included angle between the two ends of the bending portion is smaller than the first angle threshold and the fourth display area 301 faces upward, the electronic device may display, through the fourth display area 301, the user interface that the electronic device has recently displayed through the fourth display area 301. In response to that the included angle between the two ends of the bending portion is smaller than the first angle threshold and the sixth display area 303 faces upward, the electronic device may display, through the sixth display area 303, the user interface that the electronic device has recently displayed through the sixth display area 303.
Not limited to the above-mentioned cases, in a specific implementation, in response to the physical state of the electronic device being changed from the unfolded state to the folded state, the electronic device may also continue to display the user interface of the application displayed when the electronic device is in the unfolded state through the first display 200. Alternatively, the electronic device may also display, through the first display screen 200, a user interface displayed when the electronic device was last in the bent state. Alternatively, the electronic device may also display a user interface of an application associated with the first display 200 through the first display 200, which is not limited in this application.
In some embodiments, the physical state of the electronic device may also be changed from the folded state to the folded state, and the electronic device may display the user interface of the application through the first display 200 in a split manner, in which case the electronic device may be in the state shown in fig. 11 (a) or fig. 12 (a). Not limited thereto, the electronic apparatus may also continue to display the user interface of the application displayed when the electronic apparatus is in the folded state through the first display screen 200. Alternatively, the electronic device may also display, through the first display screen 200, a user interface displayed when the electronic device was last in the bent state. Alternatively, the electronic device may also display a user interface of an application associated with the first display screen 200 through the first display screen 200, which is not limited in this application.
In some embodiments, when the electronic device is in the unfolded state and the first display screen 200 faces upward and is in the landscape state, the electronic device may display the user interface 110 of the video application through the first display screen 200, as shown in fig. 13 (a). When the electronic device is in the unfolded state and the first display 200 faces upward and is in the portrait state, the electronic device may display the user interface 160 of the electronic book through the first display 200, as shown in fig. 13 (B).
As shown in fig. 13 (a), the electronic apparatus may detect a rotation operation of the user, which may change the physical state of the electronic apparatus from the landscape screen state to the portrait screen state. In response to the physical state of the electronic device being changed from the landscape state to the portrait state, the electronic device may display the user interface 160 of the electronic book through the first display 200, at which time the electronic device may be in the state shown in (B) of fig. 13.
As shown in fig. 13 (B), the electronic apparatus may detect a rotation operation by the user, which may change the physical state of the electronic apparatus from the portrait screen state to the landscape screen state. In response to the physical state of the electronic device being changed from the portrait screen state to the landscape screen state, the electronic device may display the user interface 110 of the video application through the first display screen 200, at which time the electronic device may be in the state shown in (a) of fig. 13.
Not limited to the above list, in a specific implementation, when the electronic device is in the unfolded state and the second display 300 faces upward and is in the landscape state, the electronic device may display the user interface of the application a through the second display 300. When the electronic device is in the unfolded state and the second display screen 300 faces upward and is in the vertical screen state, the electronic device may display the user interface of the application B through the second display screen 300, which is similar to that shown in fig. 13 and will not be described again.
For example, the electronic device may detect that the electronic device is in the landscape state or the portrait state through the acceleration sensor 180E and/or the gyro sensor 180B.
In a possible implementation manner, the second display screen 300 may include only the fourth display area 301, and optionally, may also include the fourth display area 301 and the fifth display area 302. Alternatively, the second display 300 may include only the sixth display area 303, and optionally, may include the sixth display area 303 and the fifth display area 302. For example, the second display 300 is a rigid screen. When the physical state of the electronic device changes, a display area for displaying the user interface may change, and a displayed application may also change, as shown in fig. 14 as a specific example.
Referring to fig. 14, fig. 14 illustrates yet another human-computer interaction diagram. Fig. 14 illustrates an example in which the second display 300 includes only the sixth display area 303.
As shown in fig. 14 (a), the electronic device is in an unfolded state, and the first display screen 200 faces upward, and the electronic device displays the user interface 110 of the video application through the first display screen 200. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display, through the sixth display area 303, the user interface of the application associated with the sixth display area 303 (i.e., the second display screen 300), that is, the user interface 120 of the social application, where the electronic device may be in the state shown in fig. 14 (B). The electronic device may also detect a flipping operation by the user, which may change the relative positions of the first display 200 and the second display 300. In response to the first display screen 200 facing downward and the second display screen 300 facing upward, the electronic device may display the user interface of the application associated with the sixth display area 303, i.e., the user interface 120 of the social application, through the sixth display area 303, at which time the electronic device may be in the state shown in (C) of fig. 14.
As shown in fig. 14 (B), the electronic apparatus is in a folded state, and the user interface 120 of the social application is displayed through the sixth display area 303. The electronic device can detect the unfolding operation of a user, and the unfolding operation can enlarge the included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being greater than the second angle threshold and the first display screen 200 facing upward, the electronic device may display the user interface of the application associated with the first display screen 200, that is, the user interface 110 of the video application, through the first display screen 200, and at this time, the electronic device may be in the state shown in fig. 14 (a). Or, in response to that the included angle between the two ends of the bending portion is greater than the second angle threshold and the second display screen 300 faces upward, the electronic device may continue to display the user interface of the application associated with the sixth display area 303, that is, the user interface 120 of the social application, through the sixth display area 303, at this time, the electronic device may be in the state shown in (C) of fig. 14.
As shown in fig. 14 (C), the electronic device is in the unfolded state, and the second display screen 300 is facing up, the electronic device displays the user interface 120 of the social application through the sixth display area 303. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may continue to display the user interface of the application associated with the sixth display area 303, that is, the user interface 120 of the social application, through the sixth display area 303, at this time, the electronic device may be in the state shown in fig. 14 (B). The electronic device may also detect a flipping operation by the user, which may change the relative positions of the first display 200 and the second display 300. In response to the second display screen 300 facing downward and the first display screen 200 facing upward, the electronic device may display the user interface of the application associated with the first display screen 200, i.e., the user interface 110 of the video application, through the first display screen 200, at which time the electronic device may be in the state shown in (a) of fig. 14.
In a possible implementation manner, the first display screen 200 may include only the first display area 201, and optionally, may also include the first display area 201 and the third display area 203. Alternatively, the first display screen 200 may include only the second display area 202, and optionally, may include the second display area 202 and the third display area 203. For example, the first display screen 200 is a rigid screen. When the physical state of the electronic device changes, a display area for displaying the user interface may change, and a displayed application may also change, as shown in fig. 15 as a specific example.
Referring to fig. 15, fig. 15 is a schematic diagram illustrating still another human-computer interaction. Fig. 15 illustrates an example in which the first display 200 includes only the first display area 201.
As shown in fig. 15 (a), the electronic device is in an unfolded state, and the first display screen 200 faces upward, the electronic device displays the user interface 110 of the video application through the first display area 201. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display, through the sixth display area 303, the user interface of the application associated with the sixth display area 303, that is, the user interface 120 of the social application, where the electronic device may be in the state shown in fig. 15 (B). The electronic device may also detect a flipping operation by the user, which may change the relative positions of the first display 200 and the second display 300. In response to the first display screen 200 facing downward and the second display screen 300 facing upward, the electronic device may display the user interface of the application associated with the sixth display area 303, i.e., the user interface 120 of the social application, through the sixth display area 303, at which time the electronic device may be in the state shown in (C) of fig. 15. Fig. 15 (B) and (C) are similar to fig. 14 (B) and (C), and are not described again.
Without being limited to the example of fig. 15, in a specific implementation, the second display screen 300 may also include a fourth display area 301 and a sixth display area 303, optionally, and a fifth display area 302, which may be specifically referred to the description of the second display screen 300 in fig. 3 to 5 above. Then in (C) of fig. 15 above, the electronic device is in the unfolded state and the second display 300 is facing upward, and the electronic device can display the user interface of the application associated with the second display 300 through the second display 300. In this case, the schematic diagram of the electronic device when the physical state changes is similar to that in fig. 8-12 and 15, and is not repeated.
In a possible implementation manner, the electronic device may only include the second display 300, and the structure of the second display 300 may be as described above with reference to the second display 300 in fig. 3 to 5. The electronic device may have set the second display screen 300 to be associated with a social application, and the electronic device has also set the fourth display area 301 to be associated with a gallery and the sixth display area 303 to be associated with a camera, examples of which may be found in fig. 6-7 above. When the physical state of the electronic device changes, a display area for displaying the user interface may also change, and a displayed application may also change, as shown in fig. 16 as a specific example.
Referring to fig. 16, fig. 16 illustrates yet another human-computer interaction diagram.
As shown in fig. 16 (a), the electronic device is in the unfolded state, and the second display screen 300 is facing upward, and the electronic device displays the user interface 120 of the social application through the second display screen 300. The electronic device can detect a folding operation of a user, and the folding operation can reduce an included angle between two ends of the bending part. In response to the angle between the two ends of the bending portion being smaller than the first angle threshold, the electronic device may display the user interface through the sixth display area 303 or the fourth display area 301, where the electronic device may be in the state shown in fig. 16 (B) or fig. 16 (C). Fig. 16 (B) and (C) are the same as fig. 10 (B) and (C), and are not described again.
Not limited to the above list, in a specific implementation, the application associated with the fourth display area 301 may be the same as the application associated with the second display screen 300, for example, a social application, and the application associated with the sixth display area 303 may be different from the application associated with the second display screen 300, for example, a camera. Alternatively, the application associated with the sixth display area 303 may be the same as the application associated with the second display 300, and the application associated with the fourth display area 301 may be different from the application associated with the second display 300. Alternatively, the application associated with the fourth display area 301 is the same as the application associated with the sixth display area 303 but different from the application associated with the second display screen 300. Alternatively, the application associated with the fourth display area 301, the application associated with the sixth display area 303, and the application associated with the second display screen 300 are the same, for example, social applications.
In one possible implementation, the electronic device 100 is not foldable, i.e., the display screen 194 of the electronic device 100 is a flat panel screen. The display screen 194 includes a first display screen 200 and a second display screen 300, and a light emitting surface of the first display screen 200 is opposite to a light emitting surface of the second display screen 300. Optionally, the first display screen 200 and the second display screen 300 are each a rigid screen. Optionally, the first display screen 200 and the second display screen 300 are integrally formed flexible screens, and the first display screen 200 and the second display screen 300 are different display areas on the one flexible screen. An example of the structure of the electronic device 100 is shown in fig. 17 below. When the physical state of the electronic device changes, a display area for displaying the user interface may change, and a displayed application may also change, as shown in fig. 17 as a specific example.
Referring to fig. 17, fig. 17 illustrates yet another human-computer interaction diagram.
As shown in fig. 17 (a), the first display screen 200 of the electronic device faces upward, and the electronic device displays the user interface 110 of the video application through the first display screen 200. The electronic device may detect a user's flipping operation, which may change the relative positions of the first display 200 and the second display 300. In response to the first display screen 200 facing downward and the second display screen 300 facing upward, the electronic device may display the user interface of the application associated with the second display screen 300, i.e., the user interface 120 of the social application, through the second display screen 300, at which time the electronic device may be in the state shown in (B) of fig. 17.
As shown in fig. 17 (B), the second display screen 300 of the electronic device faces upward, and the electronic device displays the user interface 120 of the social application through the second display screen 300. The electronic device may detect a user's flipping operation, which may change the relative positions of the first display 200 and the second display 300. In response to the second display screen 300 facing downward and the first display screen 200 facing upward, the electronic device may display the user interface of the application associated with the first display screen 200, i.e., the user interface 110 of the video application, through the first display screen 200, at which time the electronic device may be in the state shown in (a) of fig. 17.
Not limited to the above-mentioned cases, in a specific implementation, when the electronic device is in the unfolded state and the first display 200 is facing the user, the electronic device may display the user interface through the first display 200. At this time, the camera 193 on the first display screen 200 may acquire the face information of the user, and the camera 193 on the second display screen 300 may not acquire the face information of the user. When the electronic device is in the unfolded state and the second display 300 is facing the user, the electronic device can display the user interface through the second display 300. At this time, the camera 193 on the second display screen 300 may acquire the face information of the user, and the camera 193 on the first display screen 200 may not acquire the face information of the user.
Similarly, when the electronic device is in a folded state and the sixth display area 303 of the second display screen 300 is facing the user, the electronic device may display the user interface through the sixth display area 303 (and optionally the fifth display area 302). At this time, the camera 193 in the sixth display area 303 can acquire the face information of the user, and the camera 193 in the fourth display area 301 cannot acquire the face information of the user. When the electronic device is in a folded state and the fourth display area 301 of the second display screen 300 is facing the user, the electronic device may display the user interface through the fourth display area 301 (and optionally the fifth display area 302). At this time, the camera 193 in the fourth display area 301 may acquire the face information of the user, and the camera 193 in the sixth display area 303 may not acquire the face information of the user.
In a specific implementation, when the electronic device is in the unfolded state, both the first display 200 and the second display 300 may be lit, where the first display 200 displays a user interface of an application associated with the first display 200, and the second display 300 displays a user interface of an application associated with the second display 300. Similarly, when the electronic device is in the folded state, both the fourth display area 301 and the sixth display area 303 (and optionally the fifth display area 302) may be lit, wherein the fourth display area 301 displays a user interface of an application associated with the fourth display area 301, the sixth display area 303 displays a user interface of an application associated with the sixth display area 303, and optionally the fourth display area 301 or the sixth display area 303 displays the user interface together with the fifth display area 302.
It is understood that after the electronic device sets the first display screen 200 and the first application to be associated, the first application may be closed (e.g., the electronic device is powered off, or a user operation for closing the first application is received, etc.). When the physical state of the electronic device (at this time, the electronic device is turned on) is changed from another physical state (for example, a folded state) to an unfolded state, the electronic device may automatically open the first application and display a user interface of the first application through the first display screen 200. That is to say, the user can open the application quickly by changing the physical state of the electronic device, so that the operation steps of starting the application by the user are saved, and the electronic device is more convenient and quicker. Other examples of display area and application association are similar to the above examples and are not described in detail.
Not limited to the above-mentioned example, after the electronic device sets the first display screen 200 and the first application association, the electronic device may receive a user operation while displaying a setting interface or a desktop, and in response to the user operation, the electronic device cancels the first display screen 200 and the first application association. Without being limited thereto, the electronic device may also receive a user operation (e.g., clicking a lock control) acting on the first display screen 200 while displaying the user interface of the first application through the first display screen 200, and in response to the user operation, the electronic device cancels the association of the first display screen 200 and the first application. Alternatively, when the first application is closed (e.g., the electronic device is powered off, or a user operation for closing the first application is received, etc.), the electronic device cancels the association of the first display screen 200 and the first application. This is not a limitation of the present application. The description of the electronic device canceling the association of other display areas and applications is similar to the above process, and is not repeated.
In this embodiment, the association relationship between the display area and the application may be preset by the system, for example, the second display 300 of the electronic device displays the user interface of the payment application by default. The association relationship between the display area and the application may also be determined in real time by the electronic device, for example, the electronic device determines that the application displayed last time through the first display screen 200 is associated with the first display screen 200. The association relationship between the display area and the application may also be customized in response to a user operation, and specific examples may be found in fig. 18 to 22 below.
The following embodiment is described by taking a foldable screen configured for an electronic device as an example of the foldable screen shown in fig. 3 to 5, and examples of the change of the physical state of the electronic device at this time can be seen in the embodiments shown in fig. 6 to 13.
Referring to fig. 18, fig. 18 illustrates a schematic diagram of an embodiment of a user interface. User interface 180 may include settings interface 151, guide examples 152, and toggle options 153. Wherein:
the settings interface 151 may include a first title 1511, a video application option 1512, a social application option 1513, a game application option 1514. The first title 1511 of the setting interface 151 includes text information: "set associated application", which indicates that the setting interface 151 is used to set the association relationship between the display area and the application.
The setting interface 151 may be used to set an association relationship between any one of the applications and the display area, for example, an association relationship between a video application, a social application, a game application, and the display area may be set. Video application options 1512 may include a first option 1512A and a second option 1512B. First option 1512A may be for a user to set whether the video application and second display screen 300 are associated, including textual information: the "outer screen display in the folded state", wherein the outer screen is the second display screen 300. The electronic device may detect a user operation (e.g., a click or slide operation) on the first option 1512A, and in response to the operation, the electronic device may determine that the video application is associated with the second display screen 300 or cancel the association of the video application with the second display screen 300. The first option 1512A of the user interface 180 characterizes that the video application and the second display screen 300 are not associated.
The second option 1512B may be used for the user to set whether the video application is associated with the first display screen 200, including text information: the "display of the inner screen in the expanded state" is the first display screen 200. The electronic device may detect a user operation (e.g., a click or slide operation) on the second option 1512B, and in response to the operation, the electronic device may determine that the video application is associated with the first display screen 200 or cancel the association of the video application with the first display screen 200. The second option 1512B of the user interface 180 characterizes the video application as being associated with the first display screen 200.
Similarly, the social application options 1513 may also include a third option 1513A and a fourth option 1513B, and the game application may also include a fifth option 1514A and a sixth option. Third option 1513A and fifth option 1514A are similar to the description of first option 1512A, and fourth option 1513B and sixth option are similar to the description of second option 1512B, and are not described again.
The guide instance 152 may include a second title 1521 and a picture instance 1522. The second title 1521 of the guide example 152 includes text information: "in-screen display in the expanded state" is used to indicate that the scene shown in the picture example 1522 is: the electronic device is in the unfolded state and displays the user interface of the application through the first display screen 200. In the case where the video application is associated with the first display screen 200, the electronic device may display the user interface of the video application through the first display screen 200 in response to the physical state of the electronic device transitioning from the other physical state to the expanded state, at which point the electronic device may be in the state shown in the guide example 152.
Switching options 153 may include a first example option 153A and a second example option 153B. A first example option 153A of toggle options 153 is a selected state, indicating that boot example 152 is a first interface of a boot example displayed by the electronic device. The electronic device may receive a sliding operation (e.g., sliding from right to left) by the user acting on the guide example 152, the toggle option 153, or a blank area of the user interface 180, and in response to the sliding operation, the electronic device may toggle to display a second interface of the guide example, i.e., the picture example 1542 of the guide example 154 shown in fig. 19. When the electronic device displays the second interface of the guidance example, the second example option 153B is in a selected state, as shown in fig. 19.
Unlike the guidance example 154 shown in fig. 19 and the guidance example 153 shown in fig. 18, the guidance example 154 may include a third title 1541 and a picture example 1542. The third title 1541 includes text information: "outer screen display in folded state" is used to indicate that the scene shown in the picture example 1542 is: the electronic device is in a folded state, displaying an example of a user interface of an application through the second display screen 300. In the case where the social application is associated with the second display 300, in response to the physical state of the electronic device changing from the other physical state to the collapsed state, the electronic device may display the user interface of the social application through the sixth display area 303 (and optionally the fifth display area 302) of the second display 300, at which point the electronic device may be the state shown in the guide example 154. Without being limited thereto, the electronic device may also display the user interface of the social application through the fourth display area 301 (and optionally the fifth display area 302) of the second display screen 300. The rest of fig. 19 is identical to fig. 18 and will not be described again.
It will be appreciated that in order to ensure the display effect and usage effect of each application, a display area is typically associated with only one application. Therefore, in the scenario of setting up the associated application shown in fig. 18-19, at most two options are open, which are: for setting up the association of application a with the first display screen 200 and for setting up the association of application B with the second display screen 300. In the embodiments shown in fig. 18-19, applications that may be set include video applications, social applications, and gaming applications. Only one of first option 1512A, third option 1513A, and fifth option 1514A may be in an on state, i.e., at most one application associated with second display 300, and only one of second option 1512B, fourth option 1513B, and sixth option may be in an on state, i.e., at most one application associated with first display 200.
Moreover, in a case that one application is associated with the first display 200, if the user sets other applications to be associated with the first display 200, the electronic device cancels the association relationship between the original application and the first display 200, and sets other applications to be associated with the first display 200. For example, in the embodiment shown in FIGS. 18-19, second option 1512B for the video application is an open state, indicating that the video application has been associated with first display screen 200. At this time, if the electronic device detects a user operation (e.g., a click operation) on the fourth option 1513B, in response to the user operation, the electronic device sets the second option 1512B to the off state and sets the fourth option 1513B to the on state. That is, in response to the user operation, the electronic device cancels the association relationship of the video application and the first display screen 200, and sets the social application to be associated with the first display screen 200. The second display 300 is similar to the first display 200 and will not be described in detail.
Referring to fig. 20, fig. 20 illustrates a schematic diagram of yet another embodiment of a user interface. The user interface 170 may include a status bar 171 and an application icon list 172. Wherein:
the status bar 171 may include the name of the accessed mobile network, WI-FI icon, signal strength, and current remaining power. The accessed mobile network is a fifth generation mobile communication technology (5G) network with a signal format number of 4 (i.e. the signal strength is the best).
The application icon list 172 may include, for example, a set icon 1721, a calculator icon 1722, a music icon 1723, a gallery icon 1724, a dialing icon 1725, a contact icon 1726, an internet icon 1727, a text message icon 1728, a camera icon 1729, and other application icons, which are not limited in this embodiment. The icon of any application can be used for responding to the operation of the user, such as a touch operation, so that the electronic equipment starts the application corresponding to the icon. The icon of any one of the applications may be used to cause the electronic device to display an editing interface (e.g., editing interface 173 of the gallery) of the application in response to a user operation, such as a long press operation.
Illustratively, the electronic device may detect a user operation (e.g., a long press operation) by the user on the icon 1724 of the gallery, and in response to the user operation, the electronic device may display the editing interface 173 of the gallery. The gallery editing interface 173 may include, for example, a removal option 173A, an association option 173B, an edit option 173C, a more option 173D, and the like, and may further include other options, which are not limited in this embodiment. The electronic device may detect a user operation (e.g., a click operation) acting on the association option 173B, and in response to the user operation, the electronic device may set the first display screen 200 to be associated with the gallery application or cancel the association relationship between the first display screen 200 and the gallery application.
It is to be understood that the user interface 170 shown in fig. 20 is a user interface displayed through the first display screen 200 when the electronic device is in the unfolded state. The association option 173B displayed by the electronic device at this time is used for the user to set the association relationship of the gallery application and the first display screen 200. However, when the electronic device is in the folded state and the user interface 170 is displayed through the second display 300, the electronic device receives a user operation (e.g., a click operation) to the associated option 173B. In response to the user operation, the electronic device may set the second display screen 300 to be associated with the gallery application or cancel the association relationship between the second display screen 300 and the gallery application. The electronic device may also determine the application associated with the fourth display area 301 and the application associated with the sixth display area 303 in the manner described above. That is, the association option of the application is used for the user to set the association relationship between the display area currently used for display by the electronic device and the application.
Referring to FIG. 21, FIG. 21 illustrates a schematic diagram of yet another embodiment of a user interface. The user interface 210 may include a thumbnail 181 for a video application, a thumbnail 182 for a social application, a thumbnail 183 for music, and a close option 184. The user interface 210 may be a multitasking interface accessed by a user through a gesture navigation function (for example, the user slides up and stops from the bottom edge of the screen), or may be a multitasking interface accessed by a user operation (for example, a click operation) on a multitasking option of a triple-key navigation function or a hover navigation function, which is not limited in this embodiment of the present application.
The thumbnail image 181 of the video application displays an application name 181A (i.e., video application), and an association option 181B. The electronic device may detect a user operation (e.g., a click operation) acting on the association option 181B, and in response to the user operation, the electronic device may determine that the first display screen 200 is associated with the video application or cancel the association relationship of the first display screen 200 and the video application. The video application association option 181B shown in the user interface 210 characterizes that the video application has been associated with the first display screen 200. At this time, if the user clicks the association option 181B, the electronic device cancels the association relationship between the first display screen 200 and the video application, and the association option 181B is displayed as the state shown by the association option 182A of the social application.
The thumbnail 182 of the social application displays an associated option 182A. The association option 182A of the social application shown in the user interface 210 characterizes that the social application is not associated with the first display screen 200. If the electronic device detects a user operation (e.g., a click operation) on the association option 182A at this time, the electronic device may set the first display screen 200 social application association in response to the user operation. The thumbnail images 183 of music have application names 183A (i.e., music) displayed thereon.
The thumbnail 182 of the social application and the thumbnail 183 of the music display only partial content, and the thumbnail 181 of the video application 181 displays full content. The electronic device may detect a slide left and right operation applied to the user interface 210, and in response to the slide operation, the electronic device may switch the positions of the thumbnails of the plurality of applications. For example, when the user slides from right to left, the electronic device may place a thumbnail 181 of the video application in the position of a thumbnail 182 of the social application and a thumbnail 183 of the music in the position of the thumbnail 181 of the video application. If the applications run by the electronic device include only social applications, video applications, and music, the electronic device may place thumbnail 182 of the social application in the position of thumbnail 183 of the music. If the electronic device also runs other applications, such as a gaming application, the thumbnail of the gaming application may be placed in the position of the thumbnail 183 of the music, at which point the thumbnail 182 of the social application is not visible.
The close option 184 may be used to close all applications running on the electronic device, and when the electronic device detects a user operation (e.g., a click operation) acting on the close option 184, the electronic device may close all applications running on the electronic device and display a desktop of the electronic device in response to the user operation. Without being limited thereto, in response to the user operation, the electronic device may also close a part of the running applications, displaying only the user interface of one application (e.g., the user interface of the video application displaying the full thumbnail in the user interface 210).
The association option 182A of the social application and the association option 181B of the video application shown in fig. 21 are similar to the association option 173B shown in fig. 20, and are used for setting the association relationship between the display area and the application currently used for displaying by the electronic device. For example, when the electronic device is in a folded state and the user interface is displayed through the sixth area 303, the electronic device receives a user operation (e.g., a click operation) that acts on the association option 181B of the video application. The electronic device may determine that the sixth area 303 is associated with the video application or cancel the association of the sixth area 303 with the video application in response to the user operation.
Turning to fig. 22, fig. 22 illustrates a schematic diagram of yet another embodiment of a user interface.
As shown in fig. 22, the electronic device is in the unfolded state, and the user interface 110 of the video application is displayed through the first display screen 200. The user interface 110 may include an association option 110A. The electronic device may detect a user operation (e.g., a click operation) acting on the association option 110A, and in response to the user operation, the electronic device may determine that the first display screen 200 is associated with the video application or cancel the association relationship of the first display screen 200 and the video application. The association option 110A shown in the user interface 110 characterizes that the video application has been associated with the first display screen 200. At this time, if the user clicks the association option 110A, the electronic device cancels the association relationship between the first display screen 200 and the video application, and the association option 110A is displayed in the state shown in the association option 182A of the social application in fig. 21.
Similarly, the electronic device is in a folded state when the user interface of the video application is displayed through the second display screen 300. The user interface of the video application may also include an association option. The electronic device may also receive a user action (e.g., a click action) on the associated option. In response to the user operation, the electronic device may determine that the second display screen 300 is associated with the video application or cancel the association of the second display screen 300 with the video application. The electronic device may also determine the application associated with the fourth display area 301 and the application associated with the sixth display area 303 in the manner described above.
It will be appreciated that in order to ensure the display effect and usage effect of each application, a display area is typically associated with only one application. Illustratively, in the scenario of setting the associated application shown in fig. 21, at most one association option is turned on, and this association option is used to associate the setting application C with the first display screen 200. In the embodiment shown in fig. 21, applications that can be set include a video application, a social application, and music. Only one of the video application association option 181B, the social application association option 182A, and the music association option may be in an on state. Moreover, in a case that one application is associated with the first display 200, if the user sets other applications to be associated with the first display 200, the electronic device cancels the association relationship between the original application and the first display 200, and sets other applications to be associated with the first display 200. For example, in the embodiment shown in fig. 21, the association option 181B of the video application is in an on state, indicating that the video application has been associated with the first display screen 200. At this time, if the electronic device detects a user operation (e.g., a click operation) on the association option 182A of the social application, in response to the user operation, the electronic device sets the association option 181B of the video application to the closed state and sets the association option 182A of the social application to the open state. That is, in response to the user operation, the electronic device cancels the association relationship of the video application and the first display screen 200, and sets the social application to be associated with the first display screen 200. The second display 300 is similar to the first display 200 and will not be described in detail.
It will be appreciated that the manner in which the first display screen 200 is set and associated with the application may be the same as the manner in which the second display screen 300 is set and associated with the application. For example, in the embodiment shown in fig. 18-19, the electronic device may receive a click operation on first option 1512A and fourth option 1513B. In response to the clicking operation, the electronic device may determine that the first display screen 200 is associated with a social application and determine that the second display screen 300 is associated with a video application. The manner in which the first display screen 200 and the application associations are set, and the manner in which the second display screen 300 and the application associations are set, may also be different. For example, in the embodiment shown in fig. 21, the electronic device sets the first display screen 200 to be associated with a video application in response to a user operation. And the association relationship between the second display screen 300 and the application may be determined in real time by the electronic device, that is, the application associated with the second display screen 300 is: the electronic device has last passed the application displayed by the second display 300. The embodiment of the present application does not limit the specific manner and the specific time for setting different display areas and application associations.
Next, the cooperation relationship of the components in the electronic device 100 in the scenarios shown in fig. 6 to 7 will be exemplarily described, and specifically, as shown in fig. 23 to 26 below. The following embodiment will be described by taking as an example the angle sensor 180M detecting the bending angle of the display screen 194.
Referring to fig. 23, fig. 23 illustrates a cooperation relationship of the components in the electronic device 100 in the scenario illustrated in fig. 6.
In the process of the electronic apparatus 100 changing from the state shown in fig. 6 (a) to the state shown in fig. 6 (B), the cooperation relationship is specifically as follows:
1. the electronic device 100 is in the unfolded state, and the user interface of the first application is displayed on the first display screen 200.
2. The angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
3. The processor 110 determines that the bending angle of the first display screen 200 is smaller than the first angle threshold, and determines the display area as the sixth display area 303 of the second display screen 300.
4. The processor 110 determines that there is no application associated with the sixth display area 303 of the second display screen 300.
5. The processor 110 instructs the sixth display area 303 of the second display screen 300 to continue displaying the user interface of the application displayed by the electronic device 100 in the unfolded state, i.e. the user interface of the first application.
6. The sixth display area 303 of the second display screen 300 displays the user interface of the first application.
In the process of the electronic apparatus 100 changing from the state shown in fig. 6 (B) to the state shown in fig. 6 (C), the cooperation relationship is specifically as follows:
7. the pressure sensor 180A detects a touch operation applied to the sixth display area 303 of the second display screen 300, the touch operation being used to open the second application.
8. The pressure sensor 180A reports the touch event to the processor 110.
9. The processor 110 determines the displayed application as the second application according to the touch operation event.
10. The processor 110 instructs the sixth display area 303 of the second display screen 300 to display the user interface of the second application.
11. The sixth display area 303 of the second display screen 300 displays the user interface of the second application.
The order of 3 and 4 is not limited, and may be performed simultaneously.
Referring to fig. 24, fig. 24 illustrates still another cooperation relationship of the components in the electronic device 100 in the scenario illustrated in fig. 6.
In the process of the electronic apparatus 100 changing from the state shown in fig. 6 (C) to the state shown in fig. 6 (a), the cooperation relationship is specifically as follows:
12. the angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
13. The processor 110 determines that the bending angle of the electronic device 100 is greater than the second angle threshold, and determines that the display area is the first display screen 200.
14. The processor 110 determines that the first display screen 200 is associated with a first application. The association relationship may be determined by the processor 110 itself or set in response to a user operation, for example, see fig. 18 to 22.
15. The processor 110 instructs the first display screen 200 to display the user interface of the application associated with the first display screen 300, i.e. the user interface of the first application.
16. The first display screen 200 displays a user interface of a first application.
In the process of the electronic apparatus 100 changing from the state shown in fig. 6 (a) to the state shown in fig. 6 (C), the cooperation relationship is specifically as follows:
17. the angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
18. The processor 110 determines that the bending angle of the electronic device 100 is smaller than the first angle threshold, and determines that the display area is the sixth display area 303 of the second display screen 300.
19. The processor 110 determines that the application displayed by the electronic device 100 in the folded state last time is the second application, i.e. the application displayed by the second display screen 300 last time is the second application.
20. The processor 110 instructs the sixth display area 303 of the second display screen 300 to display the user interface of the application displayed by the electronic device 100 in the folded state last time, i.e. the user interface of the second application.
21. The sixth display area 303 of the second display screen 300 displays the user interface of the second application.
The order of 13 and 14 is not limited, and may be performed simultaneously. The order of 18 and 19 is not limited, and may be performed simultaneously.
Next, a cooperative relationship of the components in the electronic device 100 in the scenario shown in fig. 7 will be described.
Referring to fig. 25, fig. 25 illustrates a cooperation relationship of the components in the electronic device 100 in the scenario illustrated in fig. 7.
In the process of the electronic apparatus 100 changing from the state shown in fig. 7 (a) to the state shown in fig. 7 (B), the cooperation relationship is specifically as follows:
1. the electronic device 100 is in the unfolded state, and the user interface of the first application is displayed on the first display screen 200.
2. The pressure sensor 180A detects a touch operation applied to the first display screen 200, the touch operation being used to open the third application.
3. The pressure sensor 180A reports the touch event to the processor 110.
4. The processor 110 determines the displayed application as a third application according to the event of the touch operation.
5. The processor 110 instructs the first display screen 200 to display the user interface of the third application.
6. The first display screen 200 displays a user interface of the third application.
In the process of the electronic apparatus 100 changing from the state shown in fig. 7 (B) to the state shown in fig. 7 (C), the cooperation relationship is specifically as follows:
7. the angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
8. The processor 110 determines that the bending angle of the electronic device 100 is smaller than the first angle threshold, and determines that the display area is the sixth display area 303 of the second display screen 300.
9. The processor 110 determines that there is no application associated with the sixth display area 303 of the second display screen 300.
10. The processor 110 instructs the sixth display area 303 of the second display screen 300 to continue displaying the user interface of the application displayed by the electronic device 100 in the unfolded state, i.e. the user interface of the third application.
11. The sixth display area 303 of the second display screen 300 displays the user interface of the third application.
The order of the above 8 and 9 is not limited, and may be performed simultaneously.
Referring to fig. 26, fig. 26 illustrates still another collaboration of the components of the electronic device 100 in the scenario illustrated in fig. 7.
In the process of the electronic apparatus 100 changing from the state shown in fig. 7 (C) to the state shown in fig. 7 (a), the cooperation relationship is specifically as follows:
12. the angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
13. The processor 110 determines that the bending angle of the electronic device 100 is greater than the second angle threshold, and determines that the display area is the first display screen 200.
14. The processor 110 determines that the first display screen 200 is associated with a first application. The association relationship may be determined by the processor 110 itself or set in response to a user operation, for example, see fig. 18 to 22.
15. The processor 110 instructs the first display screen 200 to display the user interface of the application associated with the first display screen 300, i.e. the user interface of the first application.
16. The first display screen 200 displays a user interface of a first application.
In the process of the electronic apparatus 100 changing from the state shown in fig. 7 (a) to the state shown in fig. 7 (C), the cooperation relationship is specifically as follows:
17. the angle sensor 180M detects the bending angle of the electronic device 100 and reports the bending angle to the processor 100.
18. The processor 110 determines that the bending angle of the electronic device 100 is smaller than the first angle threshold, and determines that the display area is the sixth display area 303 of the second display screen 300.
19. The processor 110 determines that the application displayed by the electronic device 100 in the folded state last time is the third application, i.e., the application displayed by the second display screen 300 last time is the third application.
20. The processor 110 instructs the sixth display area 303 of the second display screen 300 to display the user interface of the application displayed by the electronic device 100 in the folded state last time, i.e. the user interface of the third application.
21. The sixth display area 303 of the second display screen 300 displays the user interface of the third application.
The order of 13 and 14 is not limited, and may be performed simultaneously. The order of 18 and 19 is not limited, and may be performed simultaneously.
Based on some embodiments shown in fig. 1-26, the display method provided by the present application is described below.
Referring to fig. 27, fig. 27 is a display method according to an embodiment of the present disclosure. The method may be applied to the electronic device 100 shown in fig. 1. The method may also be applied to the electronic device 100 shown in fig. 2. The method includes, but is not limited to, the steps of:
s101: a first user operation is received.
S102: in response to the first user operation, the electronic device determines that the first display area is associated with the first application.
In particular, examples of the electronic device determining the display area and application association in response to user operation may be found in the embodiments illustrated in fig. 18-22 above.
S103: when the electronic equipment is in the first physical state, the user interface of the first application is displayed through the first display area.
S104: and receiving a second user operation.
S105: in response to the second user operation, the electronic device displays a user interface of the second application through the first display area.
S106: when the user interface of the second application is displayed through the first display area, the physical state of the electronic equipment is changed from the first physical state to the second physical state in response to a third user operation, and the user interface of the second application is displayed through the second display area by the electronic equipment.
S107: when the user interface of the second application is displayed through the second display area, the electronic device displays the user interface of the first application through the first display area in response to the physical state of the electronic device changing from the second physical state to the first physical state.
For an example of the physical state change process of the electronic device shown in fig. 27, see fig. 7 above.
Specifically, the electronic device includes a first display area and a second display area, the first display area displays the user interface when the electronic device is in the first physical state, and the second display area displays the user interface when the electronic device is in the second physical state.
In some embodiments, the first display region comprises at least a portion of the second display region; and/or the second display area comprises at least part of the first display area. Illustratively, the first display area and the second display area belong to the same display screen of the electronic device, such as the second display screen 300 shown in fig. 3-5 above. The first display region includes a fourth display region 301 and a fifth display region 302 shown in fig. 3 to 5, the second display region includes a sixth display region 303 and a fifth display region 302 shown in fig. 3 to 5, and the first display region and the second display region have an overlapping display region, i.e., the fifth display region 302.
In some embodiments, the electronic device is a foldable electronic device. When the electronic equipment is in the unfolding state, the first display area and the second display area are in the same plane. When the electronic device is in a folded state, the light-emitting surface of the first display area and the light-emitting surface of the second display area are opposite to each other. Illustratively, the first display area and the second display area belong to the same display screen of the electronic device, such as the second display screen 300 shown in fig. 3-5 above. The first display area and the second display area are respectively the fourth display area 301 and the sixth display area 303 shown in fig. 3 to 5. In the above case, the first physical state and the second physical state may both be folded states, and the third user operation is an operation of the user to turn over the electronic device. For example, in the first physical state, the first display area faces upwards, and the second display area faces downwards; in the second physical state, the second display area faces upwards, and the first display area faces downwards. Or the first display area is over against the user in the first physical state, and the second display area is over against the user in the second physical state.
Optionally, the electronic device includes a first display screen, the first display area is at least a partial display area of the first display screen, and the second display area is at least a partial display area of the first display screen. For example, the first display screen may be a flexible folding screen, and the first display area and the second display area are display areas on the flexible folding screen.
Optionally, the electronic device includes a first display screen and a second display screen, the first display area is a display area of the first display screen, and the second display area is a display area of the second display screen. For example, the first display screen may be a display screen formed by splicing two rigid screens and connecting assemblies such as a flexible screen and a chain, the first display area is a display area on one rigid screen, and the second display area is a display area on the other rigid screen.
In some embodiments, the electronic device is a foldable electronic device, the electronic device includes a first display screen, the second display area is a full-screen display area of the first display screen, the first display area is a partial display area of the first display screen, the first physical state is a folded state, and the second physical state is an unfolded state. Illustratively, the first display screen is the second display screen 300 shown in fig. 3-5 above, and the second display area includes the fourth display area 301, the fifth display area 302, and the sixth display area 303 shown in fig. 3-5 above. The first display area is the fourth display area 301 or the sixth display area 303, and optionally, the first display area may further include a fifth display area 302.
An example of the structure and physical state change process of the electronic device in the above case can be seen in fig. 16 above.
In some embodiments, the electronic device is a foldable electronic device, and when the electronic device is in the unfolded state, the light emitting surface of the first display area and the light emitting surface of the second display area are opposite to each other. Illustratively, the first display area and the second display area are the first display screen 200 and the second display screen 300 shown in fig. 3-5. Alternatively, the first display area is at least one display area on the first display screen 200, such as the first display area 201. Alternatively, the second display area is at least one display area on the second display screen 300, such as the fourth display area 301. Optionally, the first physical state is an unfolded state, the second physical state is a folded state, and the third user operation is an operation of the user to convert the electronic device from the unfolded state to the folded state. Optionally, the first physical state and the second physical state are both unfolded states, and the third user operation is an operation of the user to flip the electronic device. For example, in the first physical state, the first display area faces upwards, and the second display area faces downwards; in the second physical state, the second display area faces upwards, and the first display area faces downwards. Or the first display area is over against the user in the first physical state, and the second display area is over against the user in the second physical state. In this case, examples of the structure and physical state change process of the electronic device can be seen in fig. 3 to 15 above.
In some embodiments, the electronic device is not foldable, and the light emitting surface of the first display area and the light emitting surface of the second display area are opposite to each other. The third user operation is an operation in which the user turns the electronic device. For example, in the first physical state, the first display area faces upwards, and the second display area faces downwards; in the second physical state, the second display area faces upwards, and the first display area faces downwards. Or the first display area is over against the user in the first physical state, and the second display area is over against the user in the second physical state. In this case, an example of the structure and physical state change process of the electronic device can be seen in fig. 17 above.
In some embodiments, the method may further comprise: when the user interface of the first application is displayed through the first display area, in response to the physical state of the electronic equipment being changed from the first physical state to the second physical state, the electronic equipment displays the user interface of the first application through the second display area; receiving a fourth user operation when the electronic equipment displays the user interface of the first application through the second display area; in response to the fourth user operation, the electronic device displays a user interface of a third application through the second display area, the third application being different from the first application. In this case, an example of the physical state change process of the electronic device can be seen in fig. 6 above.
In some embodiments, in response to the physical state of the electronic device changing from the first physical state to the second physical state, the electronic device may display, via the second display area, a user interface of an application that was displayed by the electronic device most recently in the second physical state. After S107, the method may further include: in response to the physical state of the electronic device changing from the first physical state to the second physical state, the electronic device displays a user interface of a second application through the second display area.
In some embodiments, the electronic device may also customize the second display area and the fourth application association in response to a user operation, the fourth application being different from the first application, and specific examples may refer to the embodiments shown in fig. 18-22 above. Illustratively, the method may further comprise: receiving a fifth user operation; in response to a fifth user operation, the electronic device determines that the second display area is associated with a fourth application; when the user interface of the first application is displayed through the first display area, the electronic device displays the user interface of the fourth application through the second display area in response to the physical state of the electronic device being changed from the first physical state to the second physical state.
In a specific implementation, the first display area and the first application association may also be preset by the electronic device, and the manner of determining the display area and the application association is not limited in the present application.
The implementation of the association relationship is not limited in the present application. Alternatively, the electronic device may store the association of the display area and the application through a data set. Alternatively, the data set may be temporarily stored when the electronic device determines that an association exists, thereby reducing unnecessary storage overhead. Illustratively, the data set may include a set of data including an identification of the first display region, an identification of an application associated with the first display region, and optionally an identification of the first physical state. Alternatively, the data set may include two sets of data, where one set of data is the above-mentioned set of data, and the other set of data includes an identifier of the second display area and an identifier of an application associated with the second display area, and optionally, may further include an identifier of the second physical state. The electronic device may determine an association relationship between the display area and the application according to the data set, so that the associated application is displayed through the corresponding display area in different physical states.
In some embodiments, the electronic device stores the association of the first display region, the second display region, and the application by a data set that includes two sets of data. After the electronic device determines that the first display area is associated with the first application, one set of data in the data set includes an identification of the first display area and an identification of the first application, and another set of data may be empty. After the electronic device determines that the first display area is associated with the first application, when the physical state is changed to the second physical state for the first time, the electronic device can continue to display the user interface of the application displayed when the electronic device is in the first physical state through the second display area. Assume that the application displayed by the electronic device through the second display area last time before the physical state of the subsequent electronic device is changed from the second physical state to the first physical state is the second application. The electronic device can determine that another set of data in the data set includes an identification of the second display area and an identification of the second application. Thus, upon a subsequent transition of the physical state of the electronic device to the first physical state, the electronic device may determine, from the data set, to display the user interface of the first application through the first display region. When the physical state of the electronic device is changed to a second physical state, the electronic device may determine, from the data set, to display a user interface of a second application through a second display area.
In some embodiments, the association relationship between the display area and the application may change, and the electronic device may update the application identification stored in the data set. For example, assume that the data set originally includes an identification of the second display region and an identification of application a. The electronic device receives a user operation acting on the setting interface, the user operation is used for setting the second display area to be associated with the application B, and the electronic device can update another group of data in the data set, wherein the updated another group of data comprises the identifier of the second display area and the identifier of the application B. Or before the physical state of the electronic device is changed from the second physical state to the first physical state, if the application that the electronic device has last displayed through the second display area is application C, the electronic device may update another set of data in the data set, where the updated another set of data includes the identifier of the second display area and the identifier of application C.
In some embodiments, in the case where the electronic device determines that an application is associated with a display area, if the electronic device runs the application, the electronic device may reserve all or most of the processes of the application without reclaiming network resources, system resources, and the like required to run the application even if the application is not displayed by the electronic device. For example, the electronic device determines that a first display region is associated with a first application, and the electronic device runs the first application and a second application. When the electronic device is in the second physical state and the user interface of the second application is displayed through the second display area, the electronic device does not recycle the resources required by running the first application. Therefore, when the physical state of the electronic device is changed to the first physical state, the electronic device may display, through the first display area: and the electronic equipment displays the user interface of the first application through the first display area when being in the first physical state last time. Therefore, the situations of user data loss and the like caused by the fact that the electronic equipment closes any application process of the first application are avoided, the speed of opening the first application is increased, and the use feeling of a user is improved.
In the method illustrated in fig. 27, after the electronic device determines that the first display area is associated with the first application, the electronic device may display the user interface of the first application through the first display area whenever the physical state of the electronic device is transformed into the first physical state. That is to say, the user can switch the displayed application quickly by changing the physical state of the electronic device, and does not need to quit or hide the currently displayed application for many times and reopen the application that the user wants to view, which is greatly convenient for the user to use. In addition, the display areas of the electronic equipment for displaying the user interface in different physical states are different, the existing layout of the application interface cannot be changed, the display effect is better, and the use is more convenient.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions. When the above computer program instructions are loaded and executed on a computer, the processes or functions according to the present application are produced in whole or in part. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, digital subscriber line) or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that includes one or more of the available media. The usable medium may be a magnetic medium (e.g., a floppy disk, a hard disk, a magnetic tape), an optical medium (e.g., a Digital Versatile Disk (DVD)), a semiconductor medium (e.g., a Solid State Disk (SSD)), or the like.
In short, the above description is only an example of the technical solution of the present invention, and is not intended to limit the scope of the present invention. Any modifications, equivalents, improvements and the like made in accordance with the disclosure of the present invention are intended to be included within the scope of the present invention.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present invention without departing from the spirit and scope of the invention. Thus, if such modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (15)

1. A display method is applied to an electronic device, wherein the electronic device comprises a first display area and a second display area, and the method comprises the following steps:
receiving a first user operation;
in response to the first user operation, the electronic device determining that the first display area is associated with a first application;
when the electronic equipment is in a first physical state, displaying a user interface of the first application through the first display area;
receiving a second user operation;
in response to the second user operation, the electronic equipment displays a user interface of a second application through the first display area, wherein the first application and the second application are different;
when the user interface of the second application is displayed through the first display area, responding to a third user operation, the physical state of the electronic equipment is changed from the first physical state to a second physical state, and the user interface of the second application is displayed through the second display area by the electronic equipment;
when the user interface of the second application is displayed through the second display area, the electronic device displays the user interface of the first application through the first display area in response to the physical state of the electronic device being changed from the second physical state to the first physical state.
2. The method of claim 1, wherein the first display area comprises at least a portion of the second display area; and/or the second display region comprises at least part of the first display region.
3. The method of claim 1, wherein the electronic device is a foldable electronic device; when the electronic equipment is in an unfolded state, the first display area and the second display area are in the same plane; when the electronic device is in a folded state, the light-emitting surface of the first display area and the light-emitting surface of the second display area are opposite to each other.
4. The method of claim 3, wherein the electronic device includes a first display screen, the first display area being at least a portion of a display area of the first display screen, the second display area being at least a portion of a display area of the first display screen.
5. The method of claim 3, wherein the electronic device comprises a first display screen and a second display screen, the first display area being a display area of the first display screen, the second display area being a display area of the second display screen.
6. The method of claim 1, 2 or 4, wherein the electronic device is a foldable electronic device, the electronic device comprising a first display screen, the second display area being a full screen display area of the first display screen, the first display area being a partial display area of the first display screen, the first physical state being a folded state, the second physical state being an unfolded state.
7. The method of claim 3, wherein the first physical state and the second physical state are both folded states, the third user operation being an operation of the user to flip the electronic device.
8. The method of claim 1, wherein the electronic device is a foldable electronic device, and when the electronic device is in an unfolded state, a light emitting surface of the first display area and a light emitting surface of the second display area are opposite to each other; the first physical state is the unfolding state, the second physical state is the folding state, and the third user operation is the operation that the user converts the electronic equipment from the unfolding state to the folding state.
9. The method of claim 1, wherein the electronic device is a foldable electronic device, and when the electronic device is in an unfolded state, a light emitting surface of the first display area and a light emitting surface of the second display area are opposite to each other; the first physical state and the second physical state are both the unfolded states, and the third user operation is an operation of the user to flip the electronic device.
10. The method of any one of claims 1-9, further comprising:
when the user interface of the first application is displayed through the first display area, in response to the physical state of the electronic device being changed from the first physical state to the second physical state, the electronic device displays the user interface of the first application through the second display area;
receiving a fourth user operation when the electronic equipment displays the user interface of the first application through the second display area;
in response to the fourth user operation, the electronic device displays a user interface of a third application through the second display area, the third application being different from the first application.
11. The method of any of claims 1-9, wherein in response to the physical state of the electronic device changing from the second physical state to the first physical state while the user interface of the second application is displayed by the second display area, the method further comprises, after the electronic device displays the user interface of the first application through the first display area:
in response to the physical state of the electronic device changing from the first physical state to the second physical state, the electronic device displays a user interface of the second application through the second display area.
12. The method of any one of claims 1-9, further comprising:
receiving a fifth user operation;
in response to the fifth user operation, the electronic device determining that the second display area is associated with a fourth application;
when the user interface of the first application is displayed through the first display area, in response to the physical state of the electronic device being changed from the first physical state to the second physical state, the electronic device displays the user interface of the fourth application through the second display area, wherein the first application and the fourth application are different.
13. The method of any of claims 1-9, wherein upon receiving the first user operation, the electronic device is in the first physical state and displays a user interface of the first application via the first display area; the first user operation is a user operation for displaying a display area of a user interface of the first application.
14. An electronic device, comprising a first display area, a second display area, one or more memories, one or more processors; the one or more memories are for storing a computer program, the one or more processors are for invoking the computer program, the computer program comprising instructions that, when executed by the one or more processors, cause the electronic device to perform the method of any of claims 1-13.
15. A computer storage medium comprising a computer program comprising instructions which, when run on a processor, implement the method of any one of claims 1 to 13.
CN202110134456.7A 2021-01-30 2021-01-30 Display method and electronic equipment Pending CN114840280A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202110134456.7A CN114840280A (en) 2021-01-30 2021-01-30 Display method and electronic equipment
PCT/CN2022/070132 WO2022161119A1 (en) 2021-01-30 2022-01-04 Display method and electronic device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110134456.7A CN114840280A (en) 2021-01-30 2021-01-30 Display method and electronic equipment

Publications (1)

Publication Number Publication Date
CN114840280A true CN114840280A (en) 2022-08-02

Family

ID=82561336

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110134456.7A Pending CN114840280A (en) 2021-01-30 2021-01-30 Display method and electronic equipment

Country Status (2)

Country Link
CN (1) CN114840280A (en)
WO (1) WO2022161119A1 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024034810A1 (en) * 2022-08-08 2024-02-15 삼성전자주식회사 Electronic device for processing input for multiple windows
WO2024043440A1 (en) * 2022-08-24 2024-02-29 삼성전자주식회사 Method and device for switching, between plurality of displays, display on which screen of application is displayed

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109889630A (en) * 2019-01-11 2019-06-14 华为技术有限公司 Display methods and relevant apparatus
CN110602273A (en) * 2019-08-05 2019-12-20 华为技术有限公司 Message display method and electronic equipment
CN111078091A (en) * 2019-11-29 2020-04-28 华为技术有限公司 Split screen display processing method and device and electronic equipment
CN111182137A (en) * 2019-12-19 2020-05-19 华为技术有限公司 Display method of electronic equipment with flexible screen and electronic equipment

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109766053B (en) * 2019-01-15 2020-12-22 Oppo广东移动通信有限公司 User interface display method, device, terminal and storage medium
CN111124561B (en) * 2019-11-08 2021-02-12 华为技术有限公司 Display method applied to electronic equipment with folding screen and electronic equipment
KR20210132491A (en) * 2020-04-27 2021-11-04 삼성전자주식회사 Method for controling a display and electronic device thereof

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109889630A (en) * 2019-01-11 2019-06-14 华为技术有限公司 Display methods and relevant apparatus
CN110602273A (en) * 2019-08-05 2019-12-20 华为技术有限公司 Message display method and electronic equipment
CN111078091A (en) * 2019-11-29 2020-04-28 华为技术有限公司 Split screen display processing method and device and electronic equipment
CN111182137A (en) * 2019-12-19 2020-05-19 华为技术有限公司 Display method of electronic equipment with flexible screen and electronic equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024034810A1 (en) * 2022-08-08 2024-02-15 삼성전자주식회사 Electronic device for processing input for multiple windows
WO2024043440A1 (en) * 2022-08-24 2024-02-29 삼성전자주식회사 Method and device for switching, between plurality of displays, display on which screen of application is displayed

Also Published As

Publication number Publication date
WO2022161119A1 (en) 2022-08-04

Similar Documents

Publication Publication Date Title
CN112217923B (en) Display method of flexible screen and terminal
CN113794800B (en) Voice control method and electronic equipment
CN110381282B (en) Video call display method applied to electronic equipment and related device
CN112714901B (en) Display control method of system navigation bar, graphical user interface and electronic equipment
CN110119296B (en) Method for switching parent page and child page and related device
CN111176506A (en) Screen display method and electronic equipment
CN111669459B (en) Keyboard display method, electronic device and computer readable storage medium
CN111078091A (en) Split screen display processing method and device and electronic equipment
CN112231025A (en) UI component display method and electronic equipment
CN115866121A (en) Application interface interaction method, electronic device and computer-readable storage medium
CN113994317A (en) User interface layout method and electronic equipment
CN111602108B (en) Application icon display method and terminal
CN113961157B (en) Display interaction system, display method and equipment
CN114125130B (en) Method for controlling communication service state, terminal device and readable storage medium
CN114115769A (en) Display method and electronic equipment
CN112068907A (en) Interface display method and electronic equipment
CN114281439A (en) Screen splitting method and device and electronic equipment
CN113746961A (en) Display control method, electronic device, and computer-readable storage medium
CN114089932A (en) Multi-screen display method and device, terminal equipment and storage medium
WO2022161119A1 (en) Display method and electronic device
CN114115770A (en) Display control method and related device
CN113641271A (en) Application window management method, terminal device and computer readable storage medium
CN112449101A (en) Shooting method and electronic equipment
CN113438366B (en) Information notification interaction method, electronic device and storage medium
CN113448658A (en) Screen capture processing method, graphical user interface and terminal

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination