WO2014079821A1 - Adaptable input - Google Patents
Adaptable input Download PDFInfo
- Publication number
- WO2014079821A1 WO2014079821A1 PCT/EP2013/074121 EP2013074121W WO2014079821A1 WO 2014079821 A1 WO2014079821 A1 WO 2014079821A1 EP 2013074121 W EP2013074121 W EP 2013074121W WO 2014079821 A1 WO2014079821 A1 WO 2014079821A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- input
- user interface
- mobile communications
- communications terminal
- controller
- Prior art date
Links
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M1/00—Substation equipment, e.g. for use by subscribers
- H04M1/72—Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
- H04M1/724—User interfaces specially adapted for cordless or mobile telephones
- H04M1/72403—User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
- G06F3/04166—Details of scanning methods, e.g. sampling time, grouping of sub areas or time sharing with display driving
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
- G06F3/0418—Control or interface arrangements specially adapted for digitisers for error correction or compensation, e.g. based on parallax, calibration or alignment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
Definitions
- This application relates to a method, a mobile communications terminal and a computer program product for adapting the input, and in particular to a method, a mobile communications terminal and a computer program product for adapting input according to changes in a user interface presentation.
- Mobile communications terminals are becoming more and more advanced running many different sorts of applications. Some applications are arranged to adapt their display view according to the orientation of the mobile communications terminal. This leads to a number of issues.
- the change of the user interface style or orientation is effected just as a user makes an input, the change may be too quick for the user to realize that his input is incorrect. It may be so that the mobile communications terminal receives an input and interprets it for a changed user interface while the use meant the input to be for the original (or unchanged user interface style).
- Another example is when the mobile communications terminal is currently under a heavy processing load and user interface style change is not executed in time. A user may then perform an input in a style that has not changed yet, but the mobile communications terminal will interpret the input as being for the changed user interface style.
- a mobile communications terminal comprising a memory, a display arranged to display a representation of a first user interface state, and a controller, wherein said controller is configured to detect a user interface changing event, change from said first user interface state to a second user interface state cause said display to display a representation of said second user interface state, receive an input, interpret said input based on a time period to said event, and process said input according to said interpretation.
- a mobile communications terminal comprising a memory, a display arranged to display a representation of a first user interface state, and a controller, said method comprising detecting a user interface changing event, changing from said first user interface state to a second user interface state, causing said display to display a representation of said second user interface state, receiving an input, interpreting said input based on a time period to said event, and processing said input according to said interpretation.
- teachings provided herein may beneficially be used in touch-based user interfaces where graphical objects may be rearranged according to a user interface state.
- the teachings provided herein also find use in user interfaces which utilize so-called softkeys, as the labels for the softkeys may be re-arranged or otherwise changed depending on the user interface state.
- a change in the graphical representation of the user interface may not be perceived by a user even for longer periods of time and other measures may need to be taken such as using audible notifications.
- the manner taught herein may also be used for such systems.
- One example of such a system is a computer using a graphic user interface, possibly windows based, where a cursor is operated to manipulate the various graphical objects.
- FIG. 2 is a schematic view of the components of a mobile communications terminal according to the teachings herein;
- Figure 3 is a schematic view of a computer-readable medium according to the teachings herein;
- Figures 4A and 4B each shows a mobile communications terminal displaying a user interface state according to one embodiment of the teachings of this application;
- Figure 5 shows a flowchart for a method according to one embodiment of the teachings of this application.
- Figures 6A and 6B each shows a mobile communications terminal displaying a user interface state according to one embodiment of the teachings of this application;
- Figure 7 shows a time graph according to one embodiment of the teachings of this application.
- Figure 8 shows a flowchart for a method according to one embodiment of the teachings of this application.
- Figures 9A and 9B each shows a mobile communications terminal displaying a user interface state according to one embodiment of the teachings of this application.
- FIGs 1 A and IB generally show a mobile communications terminal 100 according to an embodiment herein.
- a mobile communications terminal 100 are an internet tablet, a mobile telephone, a smart phone and a personal digital assistant.
- Two embodiments will be exemplified and described as being a mobile telephone, such as a smartphone, in figure 1 Aand an internet tablet in figure IB.
- a smartphone 100 comprises a housing 110 in which a display
- the display 120 is arranged.
- the display 120 is a touch display.
- the display 120 is a non-touch display.
- the smartphone 100 comprises two keys 130a, 130b. In this embodiment there are two keys 130, but any number of keys is possible and depends on the design of the smartphone 100.
- the smartphone 100 is configured to display and operate a virtual key 135 on the touch display 120. It should be noted that the number of virtual keys 135 is dependent on the design of the smartphone 100 and an application that is executed on the smartphone 100.
- an internet tablet 100 comprises a housing 110 in which a display 120 is arranged.
- the display 120 is a touch display.
- the display 120 is a non-touch display.
- the internet tablet 100 may comprise one or more keys 130. In this embodiment there is one key 130, but any number of keys is possible and depends on the design of the internet tablet 100.
- the internet tablet 100 is configured to display and operate one or more virtual keys 135 on the touch display 120. It should be noted that the number of virtual keys 135 is dependent on the design of the internet tablet 100 and an application that is executed on the internet tablet 100.
- FIG 2 shows a schematic view of the general structure of a mobile communications terminal according to figures 1A and IB.
- the mobile communications terminal 200 comprises a controller 210 which is responsible for the overall operation of the mobile communications terminal 200 and is preferably implemented by any commercially available CPU ("Central Processing Unit"), DSP ("Digital Signal Processor") or any other electronic programmable logic device.
- the controller 210 may be implemented using instructions that enable hardware functionality, for example, by using executable computer program instructions in a general-purpose or special-purpose processor that may be stored on a computer readable storage medium (disk, memory etc) 240 to be executed by such a processor.
- the controller 210 is configured to read instructions from the memory 240 and execute these instructions to control the operation of the mobile communications terminal 200.
- the memory 240 may be
- the mobile communications terminal 200 further comprises one or more applications 250.
- the applications are sets of instructions that when executed by the controller 210 control the operation of the mobile communications terminal 200.
- the memory 240 is used for various purposes by the controller 210, one of them being for storing application data and program instructions 250 for various software modules in the mobile communications terminal 200.
- the software modules include a real-time operating system, drivers for a man-machine interface 220, an application handler as well as various applications 250.
- the applications 250 can include a messaging application such as electronic mail, a browsing application, a media player application, as well as various other applications 250, such as applications for voice calling, video calling, document reading and/or document editing, an instant messaging application, a calendar application, a control panel application, one or more video games, a notepad application, etc.
- the mobile communications terminal 200 further comprises a user interface 220, which in the mobile communications terminal 100 of figure of 1A or IB, is comprised of the display 120 and the keypad 130.
- the user interface (UI) 220 also includes one or more hardware controllers, which together with the user interface drivers cooperate with the display 120, keypad 130, as well as various other I/O devices such as sound system, LED indicator, etc.
- the user may operate the mobile communications terminal 200 through the man-machine interface thus formed.
- the mobile communications terminal 200 further comprises a radio frequency interface 230, which is adapted to allow the mobile communications terminal to communicate with other devices through a radio frequency band through the use of different radio frequency technologies.
- the mobile communications terminal 200 may further comprise a wired interface, which is adapted to allow the mobile communications terminal to communicate with other devices through the use of different network technologies. Examples of such technologies are USB, Ethernet, and Local Area Network, TCP/IP (Transport Control Protocol/Internet Protocol) to name a few.
- a wired interface which is adapted to allow the mobile communications terminal to communicate with other devices through the use of different network technologies. Examples of such technologies are USB, Ethernet, and Local Area Network, TCP/IP (Transport Control Protocol/Internet Protocol) to name a few.
- the mobile communications terminal 200 may further comprise an orientation sensing device 235, such as an accelero meter, a gyro or other means capable of detecting movement or a change in orientation.
- an orientation sensing device 235 such as an accelero meter, a gyro or other means capable of detecting movement or a change in orientation.
- Such devices are common in the field of mobile communications terminals and a skilled person would know which device to select depending on the overall design of the mobile communications terminal 200.
- the controller 210 is configured to operably execute applications 250 such as the web browsing or email application through the RF interface 230 using software stored in the memory 240 which software includes various modules, protocol stacks, drivers, etc. to provide communication services (such as transport, network and connectivity) for the RF interface 230 and optionally a Bluetooth interface and/or an IrDA interface for local connectivity.
- applications 250 such as the web browsing or email application through the RF interface 230 using software stored in the memory 240 which software includes various modules, protocol stacks, drivers, etc. to provide communication services (such as transport, network and connectivity) for the RF interface 230 and optionally a Bluetooth interface and/or an IrDA interface for local connectivity.
- FIG. 3 shows a schematic view of a computer-readable medium 32 as described in the above.
- the computer-readable medium 32 is in this embodiment housed in a memory stick 30, such as a Universal Serial Bus (USB) stick.
- the USB stick 30 comprises a housing 33 having an interface, such as a connector 34, and a memory chip 32.
- the memory chip 32 is a flash memory, that is, a non-volatile data storage that can be electrically erased and re- programmed.
- the memory chip 32 is programmed with instructions 31 that when loaded (possibly via the interface 34) into a controller, such as a processor, executes a method or procedure according to the embodiments disclosed above.
- the USB stick is arranged to be connected to and read by a reading device, such as a mobile communications terminal 100, 200 according to figures 1 and 2, for loading the instructions into the controller (referenced 210 in figure 2).
- a computer-readable medium can also be other mediums such as compact discs, digital video discs, hard drives or other memory technologies commonly used.
- the instructions can also be downloaded from the computer-readable medium via a wireless interface to be loaded into the controller. References to 'computer-readable storage medium', 'computer program product', 'tangibly embodied computer program' etc. or a 'controller', 'computer', 'processor' etc.
- references to computer program, instructions, code etc. should be understood to encompass software for a programmable processor or firmware such as, for example, the programmable content of a hardware device whether instructions for a processor, or configuration settings for a fixed- function device, gate array or programmable logic device etc.
- Figure 4A shows a mobile communications terminal 400, such as the smart phone 100 of figure 1A.
- the mobile communications terminal 400 in figure 4A is operating in a first user interface state and displaying a graphic representation 440 of the first user interface state on a display 420.
- the display 420 is a touch display, but it should be noted that the functionality exemplified is also applicable to a display arranged to display labels for so-called softkeys (referenced 130 in figure 1A).
- a user interface state is, in this context, defined as an arrangement of icons, virtual buttons and other graphical objects that a user may manipulate, and the
- the user interface state is displayed as a graphical representation comprising graphical representations of the icons, virtual buttons and other manipulative objects. It should be clear to a skilled person what is meant by this term.
- buttons 441-444 displayed and four text fields 445-448.
- the virtual buttons 441-444 are grouped with each a text field 445-448.
- the graphical representation of the user interface is displayed in a portrait orientation in figure 4A.
- a touch area 450 is indicated with a circle roughly representing the area touched by for example a user's finger.
- the mobile communications terminal 400 is configured to detect an event that causes a change from the first user interface state to a second user interface state. Such events may be change in position or orientation of the mobile communications terminal 400, an application prompting a displayed prompt or window, an application changing its operational status and thereby the displayed user interface, an application receiving a communication (call, message) from a remote source (network), or a progressive changing event such as downloading a web page and displaying it while downloading it. Such an event may cause the web page to be displayed differently depending on which objects that has been downloaded. For example, the arrangement of virtual buttons, for example representing links, may vary if an image that is supposed to be displayed between the two virtual buttons has been downloaded fully or not.
- the action of detecting an event may include the actual reception of the event.
- the detecting of an event may also include the processing of an event to load the event and any resulting user interface notification into a stack for displaying representations of user interfaces.
- the controller (not shown in figure 4A, but referenced 210 in figure 2) detects that such an event has occurred, the controller causes the display 420 to display the graphical representation of the second user interface state.
- the event is a re-orientation of the mobile communications terminal 400 and the graphical representation of the user interface of the user interface state will be adapted accordingly changing from a portrait to landscape mode.
- the resulting (second) graphical representation of the user interface is shown in figure 4B.
- the arrangement of the virtual buttons 441-444 and the text fields 445-448 has changed although their interrelated arrangement is the same.
- the virtual buttons 441-444 are arranged along a long edge of the display in figure 4 A, but along a short edge of the display in figure 4B.
- touch area 450 overlaps one virtual button 441 in figure 4A, but a second virtual button 444 in figure 4B.
- the input would be interpreted as being aimed at the second virtual button 444, whereas the user may not have been aware of the upcoming graphical representation of the user interface change.
- One situation where this is possible is when the user is holding his mobile communications terminal 400 slightly angled and accidentally tilts the mobile communications terminal just a little bit causing the mobile communications terminal to determine that it should change the user interface state. The user may, in such a situation, not have time to react to the sudden change of graphical representation of the user interface and not be able to abort his input action before the input is registered by the mobile communications terminal 400.
- communications terminal 400 is in a second user interface state, but a graphical representation of the user interface of a first user interface state is displayed. This causes all inputs in this time interval to be wrongly interpreted by prior art mobile
- Figure 5 shows a flow chart of how a method of how a mobile communications terminal can overcome the problems listed above, and also the problems listed in the background section and other where in this description.
- the mobile communications terminal detects 510 a user interface state changing event and causes the display to display 520 a graphical representation of the user interface of a second user interface state.
- the mobile communications terminal further receives an input 530 and determines how to interpret 540 the input based on a time from the event. The input is then processed 550 according to the interpretation.
- the controller detects 510 a user interface changing event when a browser application (not shown, but referenced 250 in figure 2) receives a HTML5 server-sent event for a chat invitation.
- the controller changes from said first user interface state to a second user interface state by causing a JavaScript callback to construct a new absolutely positioned ⁇ div> DOM (Document Object Model) node as pop-up notification box and also causes said display (120) to display 520 a representation of said second user interface state, whereby the browser application (250) renders the modified DOM on the display (120).
- the controller also receives an input 530 by the browser application receiving a touch input event and registering the event in an event buffer.
- the controller interprets 540 said input based on a time period to said event, whereby the browser application reads the touch input event from the event buffer and interprets it, and the controller process 550 said input according to said interpretation. Based on browser application's interpretation, the touch input can either be applied to the changed DOM which is the pop-up notification box or the previous state of the DOM which is whatever that lies beneath the notification box.
- the controller is configured to interpret the input so that if the input is received within a first time period from the detection of the event, the input is interpreted as having been made in the first user interface state. This enables the mobile communications terminal to avoid mistaken input made by the user.
- the controller is configured to interpret the input so that if the input is received within a first time period from the detection of the event, the input is ignored. This enables the mobile communications terminal to avoid mistaken input made by the user and allows the user to input the same input again at a later stage when it is clear to the user which graphical representation of the user interface is in effect. In one embodiment the input is ignored by being deleted from an input buffer.
- the input is ignored by cancelling any activities that have been initiated by the input thereby nullifying the input.
- actions in certain applications should not be ignored irrespective of when they are being input.
- One example of such an application is a gaming application.
- the manner of interpreting the input may be based on which application is receiving the input.
- the first time period is 0.2 seconds, 0.4 seconds, 0.8 seconds or 1.6 seconds or any value in between.
- the first time period is set by a user. In one embodiment the first time period is based on statistics on or of heuristics of user input.
- the controller is configured to interpret an input by postponing the input until after the application or action causing the user interface changing event has finished execution. This allows for a user to focus on one application or action at a time.
- the controller is configured to display a prompt requesting the user to confirm that an input should be ignored and only ignore the input if the user gives his acceptance.
- FIG. 6A shows an example of a mobile communications terminal 600 displaying a graphical representation of the user interface 640 of a phone book application. Displayed are text fields 645-648 indicating the name of contacts along with virtual buttons 641-644 for initiating communication, for example via a chat application, with the corresponding contact.
- the user is aiming to initiate a chat communication with a contact "Nicola” by touching the corresponding virtual button 643. This is indicated with a touch area 650. Substantially simultaneously an incoming call is received.
- This is a user interface state changing event as the mobile communications terminal 600 is (commonly) configured to either change the graphical representation of the user interface completely or display a prompting window indicating that a call is incoming and from whom.
- Figure 6B shows an example of such a prompt 660.
- the user does not have time to react before touching the display 620 and the touch area 650 overlaps a virtual button 649 for accepting the incoming call.
- the controller is configured to interpret the input so that if the input is received within a second time period before the detection of the event, the input is interpreted as having been made in the first user interface state.
- An alternative to enabling the controller to accommodate for periods of high processor load during which the controller may be slow in updating or displaying the graphical representation of the second user interface state is to configure the controller to interpret the input so that if the input is received within a second time period before the detection of the event, the input is ignored.
- the second time period is 0.2 seconds, 0.4 seconds, 0.8
- the second time period is relative and depends on the number of events in an input buffer.
- the second time period may include the input events stored in an input buffer and which have not yet been processed. In such an
- the second time period is set by a user. In one embodiment the second time period is based on statistics on or of heuristics of user input.
- a further alternative to enabling the controller to accommodate for periods of high processor load during which the controller may be slow in updating or displaying the graphical representation of the second user interface state is to configure the controller to determine whether the graphical representation of the second user interface state is displayed or not. If the graphical representation of the second user interface state is not displayed, the controller is configured to interpret the input as having been made in the first user interface state.
- Figure 7 shows a schematic view of a time graph of events and the corresponding interpretation of inputs according to one embodiment of the teachings herein.
- the user interface changing event is detected. Any input received within a time period T2 before the event is detected will either be ignored or interpreted as having been made in the first user interface state, as has been disclosed above. In one embodiment an event received within the time period T2 will be ignored by being cancelled. This ensures that any activities that have been initiated by the input is cancelled thereby nullifying the input.
- the controller detects that the graphical representation of the second user interface state is displayed. Any input received before this, but after the event is detected (T2 ⁇ T ⁇ T3) is interpreted as having been made in the first user interface state.
- T3 is typically about 20 ms, but can be as long as a few seconds depending on the current load of the controller or processor.
- an input received within a first time period Tl will be interpreted as having been made in the first user interface state or ignored, depending on the embodiment chosen by the designer of the mobile communications terminal, by the designer of an application or by the user of the mobile communications terminal.
- the first time period Tl starts at TO.
- the first time period Tl starts at T3. In the situations when T3 is small it will have little effect if Tl starts at TOor if Tl starts at T3, but if T3 is long, it may even be longer than Tl in which case Tl will not have any effect on the input system.
- the mobile communications terminal For progressively changing graphical representation of the user interfaces it may be difficult or impossible to determine which user interface state the mobile communications terminal is operating in at any time, and especially for determining which state an input refers to. For example, when downloading a webpage, the graphical representation of the user interface may change gradually. In such situations it may be best to ignore any input received while downloading the web page or between updates of the graphical representation, or utilize long time periods - indicating a stable view of the graphical representation of the user interface.
- Figure 8 shows a flow chart of how a method of how a mobile communications terminal can overcome the problems listed above, and also the problems listed in the
- the mobile communications terminal detects 810 a user interface state changing event and causes the display to display 820 a graphical representation of the user interface of a second user interface state.
- the mobile communications terminal further receives an input 830, determines if the graphical representation of the second user interface state is being displayed and determines how to interpret 840 the input based on a time from the event and whether the graphical representation of the second user interface state is being displayed.
- the input is then processed 850 according to the interpretation.
- FIGs 9Aand 9B illustrate another problem situation.
- a mobile communications terminal 900 is lying flat on a surface, for example a bedside table.
- An application for example an alarm is activated and corresponding graphic representation of the first user interface state is displayed having two virtual buttons 941, 942.
- One button 941 being for snoozing an alarm function and the other button 942 for cancelling the alarm.
- the graphic representation 940 will be displayed in a portrait orientation. A user tilts the mobile communications terminal 900 to better reach the button and touches the snooze button 941.
- the position detecting device senses a change in orientation and causes a user interface change, see figure 9B, where by the graphic representation 940 is displayed in a landscape orientation.
- the user since the user is already in the middle of his action of touching the button, he does not have time to notice that the orientation is changing and his touch input mistakenly targets the cancel button 942.
- the immediate result is that the alarm is stopped, regardless of which button is pressed and the user will thus not be aware that the wrong input has been received by the mobile communications terminal 900 until it is too late. In this example, the user will most likely oversleep as the alarm has been cancelled instead of snoozed, which the user is totally unaware of.
- a mobile communications terminal that is configured to use the manners taught herein will thus provide a different result for a specific input in a specific user interface mode depending on the timing of the specific input.
- a mobile communications terminal utilizing the teachings herein would be identified by showing that an input is treated differently depending on how close to the user interface changing event the input is received.
- One benefit of the teachings herein is that a user will be made aware of what actions or consequences will be effected as the user makes an input. Another benefit of the teachings herein lies in that the mobile communications terminal will be enabled to adapt to high processor loads and ensure that the input received is timed correctly to the display of a changed user interface state.
- Another benefit of the teachings herein lies in that user input will be interpreted according to the user's assumed whishes instead of being erroneously interpreted based on a mismatch between an internal user interface state and a display of another user interface state's graphical representation.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
- Mobile Radio Communication Systems (AREA)
- Telephonic Communication Services (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
RU2015124168A RU2626095C2 (en) | 2012-11-23 | 2013-11-19 | Adaptive input |
JP2015543404A JP6034508B2 (en) | 2012-11-23 | 2013-11-19 | Adaptive input |
US14/440,913 US9804706B2 (en) | 2012-11-23 | 2013-11-19 | Systems and methods of adapting input based on a change in user interface presentation |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP12193958.1 | 2012-11-23 | ||
EP12193958.1A EP2735956B3 (en) | 2012-11-23 | 2012-11-23 | Adaptable Input |
US201261738561P | 2012-12-18 | 2012-12-18 | |
US61/738,561 | 2012-12-18 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014079821A1 true WO2014079821A1 (en) | 2014-05-30 |
Family
ID=47221232
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2013/074121 WO2014079821A1 (en) | 2012-11-23 | 2013-11-19 | Adaptable input |
Country Status (6)
Country | Link |
---|---|
US (1) | US9804706B2 (en) |
EP (1) | EP2735956B3 (en) |
JP (1) | JP6034508B2 (en) |
PT (1) | PT2735956E (en) |
RU (1) | RU2626095C2 (en) |
WO (1) | WO2014079821A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3628157A1 (en) | 2018-09-28 | 2020-04-01 | Basf Se | Method of controlling insecticide resistant insects and virus transmission to plants |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10359918B2 (en) * | 2015-12-31 | 2019-07-23 | Spotify Ab | System and method for preventing unintended user interface input |
US10725627B2 (en) * | 2016-07-15 | 2020-07-28 | International Business Machines Corporation | Managing inputs to a user interface with system latency |
US20180024723A1 (en) * | 2016-07-19 | 2018-01-25 | Red Hat, Inc. | Synchronizing user input with a user interface |
US10678404B2 (en) * | 2017-12-15 | 2020-06-09 | International Business Machines Corporation | Operation of a data processing system during graphical user interface transitions |
EP4093000A3 (en) * | 2022-09-21 | 2023-03-29 | Riesenhuber, Thomas | Method and apparatus for reducing network traffic |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010029506A1 (en) * | 2008-09-12 | 2010-03-18 | Koninklijke Philips Electronics N.V. | Navigating in graphical user interface on handheld devices |
US20110109563A1 (en) * | 2009-11-11 | 2011-05-12 | Tang-Hui Liu | Electronic device with dynamically adjusted touch area |
US20110296307A1 (en) * | 2009-02-17 | 2011-12-01 | Satoshi Inami | Object selecting apparatus, object selecting program, integrated circuit used for the object selecting apparatus, and object selecting method |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP1076871A1 (en) * | 1998-05-15 | 2001-02-21 | Unicast Communications Corporation | A technique for implementing browser-initiated network-distributed advertising and for interstitially displaying an advertisement |
JP2002123455A (en) * | 2000-10-17 | 2002-04-26 | Toshiba Tec Corp | Print system |
US7218313B2 (en) * | 2003-10-31 | 2007-05-15 | Zeetoo, Inc. | Human interface system |
US20090174679A1 (en) * | 2008-01-04 | 2009-07-09 | Wayne Carl Westerman | Selective Rejection of Touch Contacts in an Edge Region of a Touch Surface |
US20090193357A1 (en) * | 2008-01-26 | 2009-07-30 | Panico Michael W | Method and System to Prevent Unintended Graphical Cursor Input |
US8633901B2 (en) * | 2009-01-30 | 2014-01-21 | Blackberry Limited | Handheld electronic device having a touchscreen and a method of using a touchscreen of a handheld electronic device |
JP2010224976A (en) * | 2009-03-24 | 2010-10-07 | Kyocera Mita Corp | Display controller and image forming device |
JP5726754B2 (en) * | 2009-11-25 | 2015-06-03 | レノボ・イノベーションズ・リミテッド(香港) | Portable information terminal, input control method, and program |
JP5556423B2 (en) * | 2010-01-29 | 2014-07-23 | ブラザー工業株式会社 | Input device and input control program |
US8661120B2 (en) * | 2010-09-21 | 2014-02-25 | Amazon Technologies, Inc. | Methods and systems for dynamically managing requests for computing capacity |
-
2012
- 2012-11-23 PT PT121939581T patent/PT2735956E/en unknown
- 2012-11-23 EP EP12193958.1A patent/EP2735956B3/en active Active
-
2013
- 2013-11-19 WO PCT/EP2013/074121 patent/WO2014079821A1/en active Application Filing
- 2013-11-19 RU RU2015124168A patent/RU2626095C2/en active
- 2013-11-19 US US14/440,913 patent/US9804706B2/en active Active
- 2013-11-19 JP JP2015543404A patent/JP6034508B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010029506A1 (en) * | 2008-09-12 | 2010-03-18 | Koninklijke Philips Electronics N.V. | Navigating in graphical user interface on handheld devices |
US20110296307A1 (en) * | 2009-02-17 | 2011-12-01 | Satoshi Inami | Object selecting apparatus, object selecting program, integrated circuit used for the object selecting apparatus, and object selecting method |
US20110109563A1 (en) * | 2009-11-11 | 2011-05-12 | Tang-Hui Liu | Electronic device with dynamically adjusted touch area |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3628157A1 (en) | 2018-09-28 | 2020-04-01 | Basf Se | Method of controlling insecticide resistant insects and virus transmission to plants |
WO2020064408A1 (en) | 2018-09-28 | 2020-04-02 | Basf Se | Method of controlling insecticide resistant insects and virus transmission to plants |
Also Published As
Publication number | Publication date |
---|---|
PT2735956E (en) | 2015-10-20 |
US9804706B2 (en) | 2017-10-31 |
EP2735956B3 (en) | 2021-03-24 |
US20150301683A1 (en) | 2015-10-22 |
RU2015124168A (en) | 2017-01-10 |
JP2016506554A (en) | 2016-03-03 |
EP2735956A1 (en) | 2014-05-28 |
JP6034508B2 (en) | 2016-11-30 |
EP2735956B1 (en) | 2015-08-26 |
RU2626095C2 (en) | 2017-07-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9965158B2 (en) | Touch screen hover input handling | |
US9804706B2 (en) | Systems and methods of adapting input based on a change in user interface presentation | |
US9268481B2 (en) | User arrangement of objects on home screen of mobile device, method and storage medium thereof | |
EP3901756B1 (en) | Electronic device including touch sensitive display and method for operating the same | |
US9342235B2 (en) | Device, method, and storage medium storing program | |
US9357396B2 (en) | Terminal device | |
US9619139B2 (en) | Device, method, and storage medium storing program | |
KR102116840B1 (en) | Unlocking a device by performing gestures on an unlock image | |
US9524091B2 (en) | Device, method, and storage medium storing program | |
US9158399B2 (en) | Unlock method and mobile device using the same | |
US20100107067A1 (en) | Input on touch based user interfaces | |
US20130055160A1 (en) | Device, method, and storage medium storing program | |
US20130265250A1 (en) | Device, method and storage medium storing program | |
US9785324B2 (en) | Device, method, and storage medium storing program | |
US20130162574A1 (en) | Device, method, and storage medium storing program | |
US9690391B2 (en) | Keyboard and touch screen gesture system | |
US10146401B2 (en) | Electronic device, control method, and control program | |
US10372286B2 (en) | Method for controlling notification and electronic device thereof | |
KR20150001095A (en) | Method for processing character input and apparatus for the same | |
KR101354841B1 (en) | Electronic Device With Touch Screen And Input Data Processing Method Thereof | |
US20150042589A1 (en) | Method and electronic device for wireless connection | |
US20190260864A1 (en) | Screen Locking Method, Terminal, and Screen Locking Apparatus | |
JP6405024B1 (en) | Electronic device, control method, and control program | |
US20130139062A1 (en) | Audio Indicator of Position Within a User Interface | |
JP5969320B2 (en) | Mobile terminal device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13792380 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14440913 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2015543404 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2015124168 Country of ref document: RU Kind code of ref document: A |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13792380 Country of ref document: EP Kind code of ref document: A1 |