EP3695293A1 - Method for providing a haptic feedback message to an operator of a touch-sensitive display device - Google Patents
Method for providing a haptic feedback message to an operator of a touch-sensitive display deviceInfo
- Publication number
- EP3695293A1 EP3695293A1 EP18773992.5A EP18773992A EP3695293A1 EP 3695293 A1 EP3695293 A1 EP 3695293A1 EP 18773992 A EP18773992 A EP 18773992A EP 3695293 A1 EP3695293 A1 EP 3695293A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- touch
- display device
- haptic feedback
- button
- sensitive display
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000000034 method Methods 0.000 title claims abstract description 21
- 230000006870 function Effects 0.000 claims description 16
- 238000004458 analytical method Methods 0.000 claims description 12
- 238000001514 detection method Methods 0.000 claims description 7
- 238000004590 computer program Methods 0.000 claims description 4
- 239000003086 colorant Substances 0.000 claims description 2
- 238000010191 image analysis Methods 0.000 description 8
- 238000012545 processing Methods 0.000 description 6
- 238000012360 testing method Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 5
- 238000013528 artificial neural network Methods 0.000 description 4
- 238000012546 transfer Methods 0.000 description 4
- 238000013461 design Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000004891 communication Methods 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 238000011156 evaluation Methods 0.000 description 2
- 230000012447 hatching Effects 0.000 description 2
- 238000000926 separation method Methods 0.000 description 2
- 230000001960 triggered effect Effects 0.000 description 2
- 230000033228 biological regulation Effects 0.000 description 1
- 238000009530 blood pressure measurement Methods 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000012905 input function Methods 0.000 description 1
- 238000004573 interface analysis Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000011017 operating method Methods 0.000 description 1
- 239000000523 sample Substances 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/016—Input arrangements with force or tactile feedback as computer generated output to the user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/014—Force feedback applied to GUI
Definitions
- Touchscreen known in which a pressure measurement for detecting a touch of the touch screen is carried out at a certain position, depending on a touch and depending on a one touch of a position associated with the element of a graphical user interface haptic feedback is issued to an operator.
- a haptic feedback may be dependent on the pressure that is detected when touching the user interface at the location of the element shown in the graphical user interface.
- Feedback to an operator of a touch-sensitive display device has the advantage that a haptic feedback can also be made for a user interface shown in the touch-sensitive display with buttons shown in the user interface, if in addition to image information itself no information for the design of the
- a haptic feedback should be given to a user even in such a situation when he touches the touch-sensitive display device in the area of a button, so he a
- touch-sensitive display device are also performed, for example, in the event that a user does not have the display device in its field of view during operation. In the event that he is the
- touch-sensitive display with haptic feedback but be known a position to which a haptic feedback should be given, otherwise a targeted haptic feedback is not possible only when a button touch.
- a button is understood to mean, on the one hand, a surface whose contact, similar to a pushbutton, performs a circuit.
- a button but also an area for more, different designed controls, such as sliders or rotary elements.
- the haptic feedback must also be made as soon as possible after a detected touch. If the haptic feedback occurs only one or possibly several seconds later, then a haptic
- the touch-sensitive display device so that a timely haptic feedback can be done.
- the display device is able to determine automatically only from the transmitted image data possible positions of buttons in a display of a user interface to be displayed.
- Display of the touch-sensitive display is not provided externally, e.g. if only image data of a user interface from an external
- Calculator be transferred or if the position of the button on the touch-sensitive display device by a user or by an application has been changed.
- haptic feedback it may be possible, for example, for haptic feedback to be given even in the event that a computing device that provides a user interface is not aware that a touch-sensitive display device for a user device is being used
- Computing device provided user interface at a touch of buttons shown and upon detection of a touch on one of the buttons specifically gives a haptic feedback.
- touch-sensitive display with haptic feedback from a further computing device are omitted, since all for the haptic
- buttons in user interfaces represented graphically in a display as a rule differ significantly from a background in the display in terms of their color, brightness and / or texture, since even with a selection without haptic feedback, a user has a
- buttons should recognize easily. These differences in parameters of the display are exploited for the image analysis, namely by regular structures, such as bounding lines and resulting shapes, especially taking into account a minimum or maximum size or area, for determining a position of such buttons by image analysis be used. Thus, a fast haptic
- Feedback are also made possible for a user interface, if a layout of the user interface and in particular positions of the buttons were previously known to the touch-sensitive display device.
- touch-sensitive display device to the arithmetic unit to return from which the image data of the user interface has been provided.
- This can be assigned to the user interface from the coordinate determine and thus assign a function or a button and then trigger the assigned function.
- Computing device are processed.
- Interface advantageous, as well as a connection of an external device, such as a smartphone, to a touch-sensitive device.
- Display device is enabled.
- buttons it is advantageous for the stored forms of buttons to include polygons, ellipses or shapes with predefined textures or colors, since in this way a large number of possible buttons can be detected.
- a shape can be understood to be the outer geometric shape of a button.
- different shapes can also be understood to mean a same shape in a different color design.
- a property of the haptic feedback is selected depending on a property of the form identified as a button, which corresponds to the display area of the surface of the
- Buttons also provide different haptic feedback, for example, different frequencies of vibration or a different number of vibrations. This not only has a user the
- Computing device such a form specification is transmitted together with the image data. Rather, the touch-sensitive display device can determine such a regulation itself.
- touch-sensitive display device has been touched several times in the area of this shape.
- a successful operation of a button causes the design of the user interface to change, either by displaying a confirmation of an operation or by displaying a new operating menu.
- a user touches the screen at one point and then changes the screen content, completely or partially, it can be deduced that a
- an analysis of transmitted image data only takes place if the image data remain unchanged for a predetermined period of time.
- the transmitted image data are data of a user interface which as a rule remain unchanged for at least 10 seconds, at least for the purpose of recording an operation by a user for a predetermined period of time.
- animations such as the appearance of a Control window by flying in or zooming as well as a representation of moving images from a representation of a user interface for analysis on possible buttons to distinguish a haptic feedback at a touch of these buttons to distinguish.
- Such a suppression of a haptic feedback during a change of an image representation is not disadvantageous since, as a rule, for example, during an animation carried out a user input usually should not be made.
- 1 shows a display device with a touch-sensitive
- Figure 2 shows an example of an assignment in the touch-sensitive
- FIG. 3 shows an example of a method for providing a haptic
- Display devices are used.
- the use for a driver information device in a motor vehicle is advantageous because the driver should concentrate on the driving situation in a vehicle and thus he should direct his view of a traffic environment of the vehicle, which he controls.
- a haptic feedback By being able to give him an advantageous haptic feedback, he can leave the view on the road, while he can perform a particular blind operation of a user interface with various buttons.
- a mobile telephone in particular a so-called smartphone
- various protocols with a driver information device installed in the vehicle.
- the image of the user interface is generated and preferably rendered by the mobile telephone or the portable computer device, the image data thus generated being sent to the computer
- Driver information system in the vehicle are transmitted and displayed by this. Additional information about a specific image content itself, such as a position of buttons in a
- the driver information system can decide from the transmitted image data at which positions of the transmitted image there is a button for which a haptic feedback is to be given in the vehicle when touched at a corresponding point on a touch-sensitive display device of the driver information device , Therefore, the present invention is explained below using the example of a driver information device in a vehicle.
- FIG. 1 shows a touch-sensitive display device 10 of FIG.
- Display device 10 has a display unit 11, which has a
- Display area 12 has.
- the display unit 11 is embodied, for example, as a liquid crystal display or as an OLED display.
- the Display surface 12 is formed touch-sensitive, so that a coordinate of a touch of the display surface 12 is determined and from a
- Processing unit 13 can be evaluated and processed further.
- An image content displayed in the display area 12 is provided by a computing unit 14.
- the arithmetic unit 14 prepares for display in the display area 12 image data, which are transmitted in the embodiment shown here via an interface 15 from an external computing device 16 via a data link 17.
- the computing device 16 For a communication via the data connection 17, the computing device 16 also has a
- the data connection 17 is
- a near-field communication running such as
- Computer 16 which is embodied for example as a portable computer, in particular a mobile telephone or a so-called smartphone, determines the image data of an image displayed in the display device 12.
- the arithmetic unit 19 determines the computing device 16
- Image data of a user interface, with which a control of functions of the computing device 16 is to take place For this the data of the
- FIG. 1 shows an example of a representation of a user interface.
- a first icon 21, a second icon 22, and a third icon 23 are shown in an upper half of the display area 12.
- Processing unit 13 detected and transferred to the arithmetic unit 14.
- the arithmetic unit 14 transmits the coordinate value via the data connection 17 In particular, this coordinate, together with a coordinate with respect to an extension of the total, in the
- Pass display surface 12 shown user interface For example, a reference to a top left corner of the user interface shown is given.
- the arithmetic unit 19 can thus a position of the detected
- the arithmetic unit 19 can determine that a touch has occurred within the first symbol 21.
- Software now runs in the arithmetic unit 19, according to which a touch of the first symbol 21 is assigned to a specific function of the computing device 16, e.g. a start of a route route navigation.
- the computing device 16 now performs this function.
- this changes a display surface 12 to be displayed in the display surface.
- the second and third symbols 22, 23 may be different
- the arithmetic unit 19 determines the coordinate that no function is stored for this position and thus the image representation in the display area 12 remains unchanged and also none Function of the computing device 16 is executed.
- a control element 30 can be controlled and moved within the surface of the slider 29 by a user by detecting a touch of the display surface 12 and a subsequent continuous contact and movement of a contact point, for example a finger on the display surface 12 in the direction of arrow 31, the graphic element 30 is also displaced in the direction of arrow 31.
- Vibration unit 40 which is controlled by the processing unit 13 and the computing unit 14.
- a haptic feedback takes place shortly after a detected touch, in the form of a so-called "click", so that a short-term haptic feedback gives a user the feeling of having pressed an actually tactile button
- Vibration and / or the number of clicks can be specified and generate depending on the selection, if necessary, a tactile different haptic
- a coordinate table could be provided to which
- the processing unit 13 or the vibration unit 40 are made available.
- an additional arithmetic unit 41 is provided which is based on this
- Coordinate information performs a control of the vibration unit 40, so that as soon as possible after detection of a touch of the
- Display surface 12 can be given a haptic feedback to a user.
- Display surface 12 can be given a haptic feedback to a user.
- User interface for example, as shown in the figure 1, is such information of the arithmetic unit 14 and thus also the additional processing unit 41 initially not available. Therefore, in the event that one
- Calculation device 16 is transmitted, an analysis method by means of a processing unit 42 is started, which analyzes the transmitted image data. Based on the analysis of the image data, areas and thus coordinates are determined at which a haptic feedback is to take place. Such an analysis occurs when the transmitted image data remains unchanged for a predetermined period of time, for example, two to five seconds. Thus, it is assumed that the computing device 16 does not transmit, for example, a movie or an animation of a user interface, but a provides constant image representation of a user interface.
- the corresponding determined coordinate information for a button are stored after the analysis and used by the arithmetic unit 41 for controlling the vibration unit 40 until the arithmetic unit 14 determines that the image information transmitted by the computing device 16 has changed and thus also haptic feedback if necessary must be adjusted.
- buttons in a representation of a user interface differ in a systematic manner from their surroundings. For example, they can be separated from the environment by a border. Furthermore, they may also differ in brightness and / or color of presentation from a background. As a rule, a delimitation of a button from the background is effected by a straight or at least partially rectilinear or smooth line, e.g. a bow.
- edges For example, in the illustration according to FIG. 1, the first symbol 21 is delimited by a frame 43 from a background of the user interface.
- Gradient displays for a background gradient are hidden in the background and are ignored. Furthermore, it is advantageous to provide a minimum dimension for a length of an edge, so that, for example, text outputs, such as the text representation 27, 28 in FIG
- a touch of the optical separation line 25 can be assigned a haptic feedback, in particular due to their length.
- Slider 29 recognized as those areas where a haptic feedback is to be given.
- a minimum width for a respective frame as a driving rule for a haptic
- a user may thus at least feel the edges of buttons shown in the display.
- no haptic feedback is output because it does not limit a button.
- buttons are differentiated from backgrounds on the basis of an image texture, such as a pattern, with which the background and / or the buttons are deposited.
- Such a demarcated area is initially referred to as a shape which occupies a certain area of the surface of the display surface 12 with its surface
- the arithmetic unit 14 now compares the size, the shape and / or the dimensions of the detected shape with forms or values stored in a memory 44.
- a minimum size is advantageously provided for detection as a button, which corresponds to an operation by a finger or by a probe element.
- Stored in the memory 44 are predetermined examples of shapes for buttons, for example polygons, in particular rectangles and ellipses, in particular circles, which are usually used by designers of control surfaces for a function control.
- a polygon or a rectangle should also be understood to mean polygons or rectangles in which the corners are rounded off.
- a haptic feedback is to take place at a detected touch at the respective coordinate of the display area 12.
- Appropriate Coordinate information is transmitted from the arithmetic unit 14 to the additional arithmetic unit 41.
- the color of the respective shape in particular the filling color of the form, can also be determined and taken into account for the recognition of a button.
- FIG. 2 schematically illustrates the display surface 12 in a representation 50 in which, when a first surface 51 of the first symbol 21 contacts, a second surface 52 of the second symbol 22, a third surface 53 of the third symbol 23, a fourth surface 54 of the button 26 and a fifth surface 55 of the slider 29, a haptic feedback is issued. If the user interface is actuated outside the surfaces 51, 52, 53, 54, 55, for example at the position 56 designated by X, no haptic feedback occurs. In one embodiment, it is also possible for the haptic feedback, for example, according to a vibration frequency or a number of
- Ad interface 12 takes place.
- a first haptic feedback takes place in accordance with one in the figure
- a second haptic feedback represented by a second hatching of the fourth surface 54 and upon actuation of the slider 29, a third haptic feedback, represented by a third hatching
- a distinction can also be made as to whether an actuation occurs within the area of the slider element 30 or of the remaining area of the slider 29, in which case a fourth haptic feedback occurs, represented by a hatch of a sixth Area 58.
- buttons in the form of a learning system in particular in the form of a so-called deep-learning system or a neural network is formed.
- Learning can be done, for example, by the fact that a form that is not initially detected as a button will in the future be a button is captured and stored.
- the elliptically shaped button 26 is initially not stored in the memory 44.
- only rectangular buttons are stored, for example according to the symbols 21, 22, 23 and the slider 29.
- the arithmetic unit 14 determines a closed shape during its analysis, but determines that this shape is not defined as a button.
- the touch coordinate is forwarded by the arithmetic unit 14 to the computing device 16. Since a touch is interpreted within the elliptical area 26 but as a touch of a button, as a result of the present operation, the computing device 16 changes the image representation and, for example, displays another operating menu. This is detected by the arithmetic unit 14. It detects this change in the image data as a result of touching the user interface as a touch
- the arithmetic unit 14 assigns the function of a button for future operations of an elliptical area according to the area 26 and stores this information in the memory 44.
- a trained neural network which is equipped with a plurality of different, in particular different, possible computing devices, e.g.
- the memory 44 may be locally associated with the computing unit 14, for example in a vehicle. Furthermore, however, it is also possible that the memory 44 is connected to the arithmetic unit 14 via a data network and is available to a large number of arithmetic units. Thus, a learning algorithm can also be designed as a cloud-based neural network.
- FIG. 3 shows an exemplary method sequence. Starting from a start step 60, for example, from a connection of a mobile
- Computing device to a driver information device in the vehicle the process is started.
- image transfer step 61 image information is transmitted to a computing unit of a touch-sensitive display device.
- a subsequent first test step 62 it is checked whether the image data has changed within a predetermined period of, for example, two seconds. If this is the case, then that becomes
- Image transfer step 61 branches back. If this is not the case, i. If the image data remain constant, at least in terms of content, an analysis step 63 is continued in which the transmitted image data are analyzed to determine at which points edges or closed forms are located.
- Color or texture determines which elements in the image representation are shapes, in particular closed surfaces, to which a switching function is assigned.
- Image data transfer step 61 branches back. If this is not the case, a branch is made to a third test step 66. In the third test step 66 it is checked whether a touch of the touch-sensitive display device has taken place at a coordinate at which a haptic feedback is to take place. If this is not the case, the system branches back to the second test step 65. If this is the case, a further branch is made to a haptic output step 67. From the haptic output step 67 is branched back to the second test step 65. The method is continued as long as the computing device 16 outputs information via a user interface to the touch-sensitive display device.
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102017218120.6A DE102017218120A1 (en) | 2017-10-11 | 2017-10-11 | A method of providing haptic feedback to an operator of a touch-sensitive display |
PCT/EP2018/075417 WO2019072512A1 (en) | 2017-10-11 | 2018-09-20 | Method for providing a haptic feedback message to an operator of a touch-sensitive display device |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3695293A1 true EP3695293A1 (en) | 2020-08-19 |
Family
ID=63683176
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP18773992.5A Withdrawn EP3695293A1 (en) | 2017-10-11 | 2018-09-20 | Method for providing a haptic feedback message to an operator of a touch-sensitive display device |
Country Status (5)
Country | Link |
---|---|
US (1) | US11061562B2 (en) |
EP (1) | EP3695293A1 (en) |
CN (1) | CN111201501A (en) |
DE (1) | DE102017218120A1 (en) |
WO (1) | WO2019072512A1 (en) |
Families Citing this family (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10560563B1 (en) * | 2019-06-25 | 2020-02-11 | Bouton Sms Inc. | Haptic device |
US11820215B2 (en) | 2019-11-04 | 2023-11-21 | Hyundai Mobis Co., Ltd. | System and method for controlling display of vehicle |
CN115702409A (en) * | 2020-05-29 | 2023-02-14 | 三菱电机株式会社 | Display device |
EP3936971A1 (en) * | 2020-07-08 | 2022-01-12 | Leica Microsystems CMS GmbH | Auxiliary appliance for a user interface device |
FR3115132B1 (en) * | 2020-10-13 | 2023-06-30 | Continental Automotive Gmbh | Method and device for generating a haptic signal |
WO2023126645A1 (en) | 2021-12-27 | 2023-07-06 | Bosch Car Multimedia Portugal S.A | Interface and alert device with haptic and thermal feedback for autonomous vehicle |
WO2023126660A1 (en) | 2021-12-29 | 2023-07-06 | Bosch Car Multimedia Portugal S.A | Haptic and thermal feedback touchpad device, system and method thereof for an automotive setting |
WO2023126664A1 (en) | 2021-12-29 | 2023-07-06 | Bosch Car Multimedia Portugal S.A | System and method for providing a web browser online user interface with haptic feedback for an automotive setting |
WO2023126662A1 (en) | 2021-12-29 | 2023-07-06 | Bosch Car Multimedia Portugal S.A | Armrest keyboard with haptic and thermal feedback |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP4860625B2 (en) * | 2004-10-08 | 2012-01-25 | イマージョン コーポレーション | Haptic feedback for simulating buttons and scrolling motion on touch input devices |
US7382357B2 (en) * | 2005-04-25 | 2008-06-03 | Avago Technologies Ecbu Ip Pte Ltd | User interface incorporating emulated hard keys |
US7509588B2 (en) * | 2005-12-30 | 2009-03-24 | Apple Inc. | Portable electronic device with interface reconfiguration mode |
JP2008033739A (en) | 2006-07-31 | 2008-02-14 | Sony Corp | Touch screen interaction method and apparatus based on tactile force feedback and pressure measurement |
US10313505B2 (en) * | 2006-09-06 | 2019-06-04 | Apple Inc. | Portable multifunction device, method, and graphical user interface for configuring and displaying widgets |
US9619143B2 (en) * | 2008-01-06 | 2017-04-11 | Apple Inc. | Device, method, and graphical user interface for viewing application launch icons |
CA2731042A1 (en) * | 2008-07-23 | 2010-01-28 | Research In Motion Limited | Portable electronic device and method of controlling same |
JP2011048686A (en) * | 2009-08-27 | 2011-03-10 | Kyocera Corp | Input apparatus |
DE102009048832A1 (en) | 2009-10-09 | 2011-04-14 | Continental Automotive Gmbh | Multi-media system has central arithmetic unit, display unit and operating arrangement, where display unit is provided with unit that is formed to evaluate operation of input fields |
JP5635274B2 (en) * | 2010-01-27 | 2014-12-03 | 京セラ株式会社 | Tactile sensation presentation apparatus and tactile sensation presentation method |
US10788976B2 (en) * | 2010-04-07 | 2020-09-29 | Apple Inc. | Device, method, and graphical user interface for managing folders with multiple pages |
EP2839366A4 (en) * | 2012-04-18 | 2016-05-11 | Nokia Technologies Oy | A display apparatus with haptic feedback |
US9501147B2 (en) * | 2013-12-29 | 2016-11-22 | Immersion Corporation | Haptic device incorporating stretch characteristics |
US9648062B2 (en) * | 2014-06-12 | 2017-05-09 | Apple Inc. | Systems and methods for multitasking on an electronic device with a touch-sensitive display |
KR102354769B1 (en) * | 2014-10-29 | 2022-01-25 | 삼성전자주식회사 | Terminal apparatus and method for controlling thereof |
US9916008B2 (en) | 2015-01-12 | 2018-03-13 | International Business Machines Corporation | Microfluidics three-dimensional touch screen display |
US20170090571A1 (en) * | 2015-09-29 | 2017-03-30 | General Electric Company | System and method for displaying and interacting with ultrasound images via a touchscreen |
-
2017
- 2017-10-11 DE DE102017218120.6A patent/DE102017218120A1/en active Pending
-
2018
- 2018-09-20 US US16/651,504 patent/US11061562B2/en active Active
- 2018-09-20 EP EP18773992.5A patent/EP3695293A1/en not_active Withdrawn
- 2018-09-20 WO PCT/EP2018/075417 patent/WO2019072512A1/en unknown
- 2018-09-20 CN CN201880065758.XA patent/CN111201501A/en active Pending
Also Published As
Publication number | Publication date |
---|---|
DE102017218120A1 (en) | 2019-04-11 |
WO2019072512A1 (en) | 2019-04-18 |
CN111201501A (en) | 2020-05-26 |
US11061562B2 (en) | 2021-07-13 |
US20200264772A1 (en) | 2020-08-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2019072512A1 (en) | Method for providing a haptic feedback message to an operator of a touch-sensitive display device | |
EP3097468B1 (en) | User interface and method for adapting a view of a display unit | |
DE102012219119A1 (en) | Intelligent windowing in a graphical user interface | |
DE102018221024A1 (en) | DEVICE AND METHOD FOR OPERATING A STEERING WHEEL BASED ON A TOUCH CONTROL | |
WO2017211817A1 (en) | Operator control comprising eye tracker unit and method for calibrating an eye tracker unit for an operator control | |
EP2876526B1 (en) | Device for gesture recognition and method for recognition of gestures | |
EP3486805A1 (en) | Method for operating an electronic device or an application and corresponding device | |
DE102012020607B4 (en) | A motor vehicle with a gesture control device and method for controlling a selection element | |
DE112013006066T5 (en) | Emulate the pressure sensitivity on multi-touch units | |
DE102013011689A1 (en) | Method and device for processing touch signals of a touchscreen | |
DE102015116477A1 (en) | Data processing method and electronic device | |
DE102012224365A1 (en) | Fast cursor localization | |
EP2653335A2 (en) | Method and apparatus for displaying a hand of an operator of an operating element of a vehicle | |
DE102012009429A1 (en) | Method for remote configuration of electric car, involves displaying visually perceptible connection between two graphic elements of two display modes, where one of modes corresponds to actual control arrangement in vehicle | |
EP3570164B1 (en) | Method and system for generating a mobile application from a desktop application | |
EP2924551A1 (en) | Method and device for providing a graphical user interface in a vehicle | |
DE102017117393A1 (en) | A vehicle display device and method for controlling the vehicle display device | |
EP3418861A1 (en) | Method for operating a display device, and a motor vehicle | |
DE102014226553A1 (en) | Method for operating an input device, input device, motor vehicle | |
WO2014067774A1 (en) | Method and device for operating an input device | |
WO2017140569A1 (en) | Motor vehicle operator control apparatus and method for operating an operator control apparatus in order to cause an interaction between a virtual presentation plane and a hand | |
DE102016115479A1 (en) | User interface for electronic device, input processing method and electronic device | |
DE102014224599A1 (en) | Method for operating an input device, input device | |
DE10313019A1 (en) | Single handed operation of a mobile terminal by gesture recognition, whereby movement of the terminal relative to the user is measured and used to control the positioning of input selection means | |
EP2555097A1 (en) | Method and device for determining a section of an image and triggering imaging using a single touch-based gesture |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: UNKNOWN |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE INTERNATIONAL PUBLICATION HAS BEEN MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
17P | Request for examination filed |
Effective date: 20200511 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAV | Request for validation of the european patent (deleted) | ||
DAX | Request for extension of the european patent (deleted) | ||
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20220623 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20231121 |