EP3049909A1 - User interface and method for assisting a user with the operation of an operating unit - Google Patents
User interface and method for assisting a user with the operation of an operating unitInfo
- Publication number
- EP3049909A1 EP3049909A1 EP13770487.0A EP13770487A EP3049909A1 EP 3049909 A1 EP3049909 A1 EP 3049909A1 EP 13770487 A EP13770487 A EP 13770487A EP 3049909 A1 EP3049909 A1 EP 3049909A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- operating unit
- user
- predefined
- button
- sound
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
- 238000000034 method Methods 0.000 title claims abstract description 26
- 230000004044 response Effects 0.000 claims abstract description 8
- 238000012545 processing Methods 0.000 claims description 6
- 230000003287 optical effect Effects 0.000 claims description 4
- 230000007704 transition Effects 0.000 claims description 2
- 238000001514 detection method Methods 0.000 claims 1
- 230000005236 sound signal Effects 0.000 abstract description 5
- 230000006870 function Effects 0.000 description 18
- 230000008447 perception Effects 0.000 description 4
- 230000008569 process Effects 0.000 description 3
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000004888 barrier function Effects 0.000 description 1
- 238000002485 combustion reaction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000033001 locomotion Effects 0.000 description 1
- 238000001556 precipitation Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Arrangement of adaptations of instruments
-
- B60K35/10—
-
- B60K35/26—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/044—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- B60K2360/113—
-
- B60K2360/115—
-
- B60K2360/141—
-
- B60K2360/1438—
-
- B60K2360/1442—
-
- B60K2360/146—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04108—Touchless 2D- digitiser, i.e. digitiser detecting the X/Y position of the input means, finger or stylus, also when it does not touch, but is proximate to the digitiser's interaction surface without distance measurement in the Z direction
Definitions
- the present invention relates to a user interface and a method for assisting a user in the operation of a touch-sensitive
- the present invention relates to the assistance of a driver of a means of locomotion in the operation of a permanently arranged in the means of transport control unit during the perception of his driving task.
- HMIs human-machine interfaces
- rotary / push button, pushbutton and other mechanical elements are also known as touch-sensitive surfaces, which include, for example, capacitive sensors.
- systems are known in which a user gesture executed in front of a screen is recognized without requiring contact with the operating unit.
- Such an operating step is referred to as “hovering.”
- the gesture recognition via optical sensors (cameras) and alternatively or additionally via capacitive sensors is detected and assigned to a predetermined button.
- the operating system “Windows 8" (registered trademark) and referred to “Air View” (registered trademark) of the Samsung Galaxy S4 (registered trademark).
- WO 2009/062677 A2 shows a multimodal user interface of an infotainment system for inputting and presenting information in which gestures are used in conjunction with voice commands for calling predefined functions.
- Speech recognition system used to detect voice commands.
- the inventive method is used to assist a user in the operation of a touch-sensitive control unit, which may be configured as built in a means of transport HMI.
- a presence of an input means in a predefined first area with respect to the operating unit is detected.
- the input means may be, for example, a stylus, a user's finger, another body part of the user, etc.
- the predefined first area may be in contact with the operation unit ("touch operation").
- control unit can be used to detect the various inputs capacitive sensors, camera-based
- Sensors or similar include.
- an audio output is generated by means of which the presence is acknowledged to the user.
- the audio output comprises a first sound character, which is context-specifically assigned to a first button displayed on the operating unit.
- the first button may represent, for example, a predefined range of functions which the first sound character symbolizes in a suitable manner.
- the first sound signal can also be understood as an "acoustic icon" (also called “Earcon”).
- the function associated with the button determines the shape of the tone so that when listening to the tone, the user associates a particular function of a plurality of functions as nearest. In this way, reaching the first predefined area may cause an audible indication of the function associated with the first button.
- the button is pressed (for example, by a touch input, a gesture, etc.), one of the buttons is assigned
- Performed function which according to the invention already recognized by the user due to the audio output. Accordingly, leaving the predefined first area can be accompanied by the same or an alternative sound in the context of an audio output.
- the present invention supports the user in the operation of the operating unit according to the invention, in particular during the perception of a driving task, during which the driver has a limited
- the method according to the invention further comprises detecting a presence of an input means in a predefined second area with respect to the operating unit.
- the process is acknowledged to the user by a predefined second audio output.
- This includes a second tone that is assigned context-specific to a second button displayed on the control unit.
- the second button stands for a different function than the first button, with the second sound character in conjunction with the second button, the above applies accordingly.
- the user may conclude a function associated with the second button and substantially preclude performing the function associated with the first button. This improves the support of the user in the operation of the user interface according to the invention during the performance of his driving task.
- the predefined first area and the predefined second area may be delimited by an interface lying substantially parallel to the surface of the operating unit, which is spaced apart in the direction perpendicular to the surface of the operating unit.
- a first parallel interface may be defined by the surface of the operating unit itself or a Höver range may be defined by a between the surface of the operating unit and the first interface.
- Vertical boundary surfaces of the first and the second area can, for example, with the boundaries of a respective illustrated
- the illustrated button may be bounded by the border of a vertical projection of the first or second area on the surface of the operating unit.
- the first or the second region may be at a distance of -1 mm to + 3 mm from the surface of the operating unit.
- a negative distance range stands for a touch operation, while a positive distance range stands for a Höver range. This typically extends to a distance of 10 mm to 15 cm from the surface of the control unit.
- the method according to the invention can further recognize a transfer of the input means from the first area into a predefined third area with respect to the operating unit.
- a predefined third audio output which comprises a third sound character, which is context-specifically assigned to a third button displayed on the operating unit.
- the third button stands for a third function, for the user through the third Sound symbol is symbolized.
- Driving task be supported acoustically, so that his attention can remain essentially on the traffic.
- the audio outputs may comprise a sample or a plurality of mutually related samples, which may be varied, in particular with repeated output.
- a typical example is a clicking sound, which can be changed in frequency depending on the operating step. For example, when reaching a Höver range, a clicking sound of a first, higher frequency can be reproduced, while when changing from a first Höver range to a second Höver range, a clicking sound of a second, lower frequency is reproduced.
- the clicking sounds are an example of a poorly contextual association between the button and the sound. More specifically, for example, a weather-related button can be acknowledged by a thunder sound as a sound signal
- Rain or wind noise can be used. In this way, a current weather situation can be reproduced, provided that this can be symbolized with one of the sounds aptly.
- Another example is the announcement of a telephone function by playing a DTMF tone sequence or a tone as a tone.
- a navigation function can be announced by a sonar noise.
- the aforementioned assignments represent particularly intuitively recognizable relationships for the orientation of the user.
- the inventive method comprises acknowledging the detected presence by a predefined alternative optical representation of a on the
- Control unit button In this case, an increased noise level can be partially compensated by the alternative optical representation draws the attention of the user to the control unit and thus only weakly compared to the noise level perceived audio output can be better recognized.
- the associations between the respective audio output and a respective button can be defined by a user. This can be, for example be made such that the user can assign in a configuration menu a whole list of buttons respective sound.
- an additional sound signal eg self-designed or purchased
- a user interface which in particular can be permanently installed in a motor vehicle.
- Such user interfaces are also known as human-machine interfaces (HMI).
- HMI human-machine interfaces
- the motor vehicle may be, for example, a road vehicle (car, truck), which is driven electrically, via an internal combustion engine or hybrid.
- User interface includes a touch-sensitive control unit, a
- Recognition device for recognizing a user input and a
- the touch-sensitive operating unit may comprise, for example, a display which forms a touch screen together with a touch-sensitive element.
- the recognition device may comprise the touch-sensitive element and a camera system for detecting hover gestures.
- Processing means may comprise a processor which is arranged within the HMI and adapted, for example, to process navigation commands.
- the user interface is also configured to perform a method as described in connection with the first aspect of the invention.
- the operating unit may preferably be integrated centrally in the dashboard of a motor vehicle or arranged as an instrument cluster behind the steering device of the vehicle.
- the recognition device may comprise a camera system and / or a capacitive sensor in order to be able to recognize and assign hover gestures and touch gestures.
- the user interface may comprise a storage means in which reference signals representing predefined gestures are stored. The references may be read by the processing means and compared with inputs recognized by the recognizer. This increases the variety of usable operating steps, so that on average fewer consecutive operating steps have to be carried out.
- a vehicle is proposed which comprises a user interface according to the second aspect of the invention.
- the user interface may be arranged as a compact unit or as a distributed system within the vehicle. In this way, existing hardware can be used, whereby the present invention essentially in the form of
- Figure 1 is a schematic overview of components of an embodiment of a user interface according to the invention.
- FIG. 2 shows an illustration of a possible operating step of an exemplary embodiment of a user interface according to the invention
- FIG. 3 shows a representation of an alternative operating step of an exemplary embodiment of a user interface according to the invention.
- FIG. 4 is a flow chart illustrating the steps of an embodiment of a method according to the invention.
- FIG. 1 shows a means of transport 8, in which a screen of an HMI as
- Operating unit 1 is embedded in the dashboard.
- a camera 6 is provided in addition to (not shown) proximity sensor via a light barrier and capacitive technologies as a recognition device in the windshield.
- the camera 6 is connected to an electronic control unit 7 as a processing device for data processing. Further, the electronic control unit 7 is connected to a speaker 9 for generating an audio output and to the operating unit 1, on the screen of which a first button 10 and a second button 20 are shown.
- the buttons 10, 20 are respective, cuboidal areas 1 1, 21, shown in dashed lines, in which a presence of an input means as a touch input or as Hovern (depending on a respective distance a to the surface of the operating unit 1) is detected.
- 21 are oriented by a first, parallel to the surface of the control unit 1 oriented surface 12 and 22, a second, parallel to the surface of the control unit 1 oriented and remote from the surface second surface 13 and 23 and by four perpendicular to the surface of the control unit 1 oriented Areas 14 or 24, 15 or 25, 16 or 26 and 17 or 27 defined such that the buttons 10, 20 vertical projections of the areas 1 1. 21 represent on the control unit 1 and are limited by these.
- Figure 2 shows a possible user interface, which can be displayed on the control unit 1.
- a map section is displayed, which corresponds to a part of a route calculated by a navigation system.
- a second button 20 is displayed in an upper area, over which a currently playing title, its artist and the album on which the title is contained are reproduced.
- a third button 30 is shown in which the weather in Braunschweig is displayed in the form of an icon in conjunction with a degree Celsius indication and a current precipitation.
- the hand of a user 2 is located in front of the first button 10.
- a first predefined audio output 3 which includes a sounder signal as Earcon, the presence of the hand is acknowledged, and the user in the course given an orientation that he with a a navigation function associated button 10 is operated.
- Figure 3 shows the view shown in connection with Figure 2, in which, however, an alternative gesture is performed by the hand of the user 2.
- the hand changes along an arrow P from a first area 1 1 in front of the first button 10 in a third area in front of the third button 30.
- the detected crossing is by a predefined second audio output 4, which includes a discreet, high-frequency clinking, acknowledged.
- the second audio output 4 corresponds to the content of the third
- a high-frequency soft chime can cause a corresponding association with the user 2.
- FIG. 4 shows a flowchart, illustrating steps of an embodiment of a method according to the invention. The method begins in step 100 with the
- Assigning a first sound character to a displayable on the operating unit first button by a user This can be done in a configuration menu of a user interface according to the invention. In this case, individual or all buttons of the operating system of the user interface by a
- predefined (but expandable) pool of sound marks are added to selected sound symbols. After the assignment, the presence of a
- Input means are detected in a predefined first area with respect to the operating unit (for example as a touch or hover gesture).
- the user is acknowledged by a predefined first audio output, the audio output comprising a first sound character which is context-specifically assigned to a first button displayed on the control unit.
- the audio output comprising a first sound character which is context-specifically assigned to a first button displayed on the control unit.
- Button associated function a process that is as simple as possible associated with the first tone by the user. Further, in step 400, the presence of the input means in a predefined second (other) range with respect to
- Operating unit e.g., on or above a second button
- step 500 acknowledge to the user the detected presence through a predefined second (other) audio output.
- the second audio output comprises a second sound character, which represents an acoustic symbol for a function associated with the second button.
- the input means enters from a first area into a predefined third (again different) area with respect to the operating unit.
- step 700 the user is acknowledged by a predefined third audio output, the third audio output comprising a third sound character corresponding to a third one displayed on the control unit
- Button is assigned context-specific. In other words, the third one
Abstract
Description
Claims
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/EP2013/070214 WO2015043652A1 (en) | 2013-09-27 | 2013-09-27 | User interface and method for assisting a user with the operation of an operating unit |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3049909A1 true EP3049909A1 (en) | 2016-08-03 |
Family
ID=49261561
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP13770487.0A Ceased EP3049909A1 (en) | 2013-09-27 | 2013-09-27 | User interface and method for assisting a user with the operation of an operating unit |
Country Status (5)
Country | Link |
---|---|
US (1) | US10248382B2 (en) |
EP (1) | EP3049909A1 (en) |
KR (1) | KR101805328B1 (en) |
CN (1) | CN105683903A (en) |
WO (1) | WO2015043652A1 (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9563329B1 (en) * | 2015-09-15 | 2017-02-07 | Thunder Power Hong Kong Ltd. | Interchangeable display of information panels on a dashboard |
GB2572614B (en) * | 2018-04-05 | 2021-12-08 | Steris Solutions Ltd | Handset for controlling a support device or a movable surface |
CN114397996A (en) * | 2021-12-29 | 2022-04-26 | 杭州灵伴科技有限公司 | Interactive prompting method, head-mounted display device and computer readable medium |
Family Cites Families (22)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7760187B2 (en) * | 2004-07-30 | 2010-07-20 | Apple Inc. | Visual expander |
US7257536B1 (en) * | 1999-11-23 | 2007-08-14 | Radiant Systems, Inc. | Audio request interaction system |
US20080129520A1 (en) | 2006-12-01 | 2008-06-05 | Apple Computer, Inc. | Electronic device with enhanced audio feedback |
DE102007039450A1 (en) * | 2007-08-21 | 2009-02-26 | Siemens Ag | Touch-sensitive screen i.e. touch screen, for use in infotainment system for motor vehicle, has tactile feedback devices indicating position of push buttons and/or type and function of instruction by acoustic and haptic feedbacks |
DE102008051757A1 (en) | 2007-11-12 | 2009-05-14 | Volkswagen Ag | Multimodal user interface of a driver assistance system for entering and presenting information |
US9170649B2 (en) | 2007-12-28 | 2015-10-27 | Nokia Technologies Oy | Audio and tactile feedback based on visual environment |
US20090225043A1 (en) | 2008-03-05 | 2009-09-10 | Plantronics, Inc. | Touch Feedback With Hover |
US20100250071A1 (en) * | 2008-03-28 | 2010-09-30 | Denso International America, Inc. | Dual function touch switch with haptic feedback |
KR20120027334A (en) * | 2009-05-13 | 2012-03-21 | 코닌클리즈케 필립스 일렉트로닉스 엔.브이. | Audio feedback and dependency on light functionality and setting |
DE102009032069A1 (en) * | 2009-07-07 | 2011-01-13 | Volkswagen Aktiengesellschaft | Method and device for providing a user interface in a vehicle |
DE102009036369A1 (en) | 2009-08-06 | 2011-02-10 | Volkswagen Ag | Method for operating an operating device and operating device in a vehicle |
TWI423112B (en) * | 2009-12-09 | 2014-01-11 | Ind Tech Res Inst | Portable virtual human-machine interaction device and method therewith |
US10705794B2 (en) * | 2010-01-18 | 2020-07-07 | Apple Inc. | Automatically adapting user interfaces for hands-free interaction |
US8863256B1 (en) * | 2011-01-14 | 2014-10-14 | Cisco Technology, Inc. | System and method for enabling secure transactions using flexible identity management in a vehicular environment |
US9285944B1 (en) * | 2011-04-22 | 2016-03-15 | Angel A. Penilla | Methods and systems for defining custom vehicle user interface configurations and cloud services for managing applications for the user interface and learned setting functions |
WO2013029257A1 (en) * | 2011-08-31 | 2013-03-07 | Ooros Automotive Co., Ltd. | Vehicle's interactive system |
KR101852821B1 (en) * | 2011-09-08 | 2018-04-27 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
WO2013187137A1 (en) * | 2012-06-15 | 2013-12-19 | 株式会社ニコン | Electronic device |
KR101943320B1 (en) * | 2012-09-21 | 2019-04-17 | 엘지전자 주식회사 | Mobile terminal and method for controlling the same |
KR101990037B1 (en) * | 2012-11-13 | 2019-06-18 | 엘지전자 주식회사 | Mobile terminal and control method thereof |
US8818716B1 (en) * | 2013-03-15 | 2014-08-26 | Honda Motor Co., Ltd. | System and method for gesture-based point of interest search |
US9300779B2 (en) * | 2013-03-15 | 2016-03-29 | Blackberry Limited | Stateful integration of a vehicle information system user interface with mobile device operations |
-
2013
- 2013-09-27 US US15/025,526 patent/US10248382B2/en active Active
- 2013-09-27 KR KR1020167010261A patent/KR101805328B1/en active IP Right Grant
- 2013-09-27 CN CN201380080494.2A patent/CN105683903A/en active Pending
- 2013-09-27 EP EP13770487.0A patent/EP3049909A1/en not_active Ceased
- 2013-09-27 WO PCT/EP2013/070214 patent/WO2015043652A1/en active Application Filing
Non-Patent Citations (2)
Title |
---|
None * |
See also references of WO2015043652A1 * |
Also Published As
Publication number | Publication date |
---|---|
US20160239261A1 (en) | 2016-08-18 |
CN105683903A (en) | 2016-06-15 |
KR20160057474A (en) | 2016-05-23 |
US10248382B2 (en) | 2019-04-02 |
WO2015043652A1 (en) | 2015-04-02 |
KR101805328B1 (en) | 2017-12-07 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2223046B1 (en) | Multimode user interface of a driver assistance system for inputting and presentation of information | |
EP2451672B1 (en) | Method and device for providing a user interface in a vehicle | |
EP1853465B1 (en) | Method and device for voice controlling a device or system in a motor vehicle | |
DE102014200993A1 (en) | User interface and method for adapting a view on a display unit | |
WO2019007811A1 (en) | User interface for a means of transportation, and means of transportation containing a user interface | |
EP2883738B1 (en) | Method and assembly for controlling functions of a motor vehicle | |
DE102013014887B4 (en) | Motor vehicle operating device with low-distraction input mode | |
DE102017201799A1 (en) | User interface, means of transport and method of user discrimination | |
EP3049909A1 (en) | User interface and method for assisting a user with the operation of an operating unit | |
EP3049911B1 (en) | User interface and method for assisting a user in the operation of an operator control unit | |
DE102009018590B4 (en) | Motor vehicle with an operating device and associated method | |
WO2014117932A1 (en) | System-initiated help function for each journey and/or for each user for operating an apparatus associated with a vehicle | |
DE102013016196B4 (en) | Motor vehicle operation using combined input modalities | |
DE102015221304A1 (en) | Method and device for improving the recognition accuracy in the handwritten input of alphanumeric characters and gestures | |
DE102008033441B4 (en) | Method for operating an operating system for a vehicle and operating system for a vehicle | |
DE102011116122A1 (en) | Method for providing an operating device in a vehicle and operating device | |
DE102014202833A1 (en) | User interface and method for switching from a first user interface operating mode to a 3D gesture mode | |
DE102019129396A1 (en) | Graphical user interface, means of transportation and method for operating a graphical user interface for a means of transportation | |
EP3049910B1 (en) | User interface and method for assisting a user when operating an operating unit | |
DE102011015693A1 (en) | Method for providing graphical user interface (GUI) for operating navigation system in vehicle, involves selecting voice modes of GUI by control keys whose positions are independent or dependent on graphical objects of GUI | |
DE102014018423B4 (en) | Motor vehicle with a status display that can be interpreted by peripheral vision and a corresponding method | |
DE102016218270A1 (en) | Method for operating a motor vehicle operating device with speech recognizer, operating device and motor vehicle | |
DE102017200769A1 (en) | Method for operating an input device of a motor vehicle and input device for a motor vehicle | |
DE102012021220A1 (en) | Operating arrangement for detection of gestures in motor vehicle, has gesture detection sensor for detecting gestures and for passing on gesture signals, and processing unit for processing gesture signals and for outputting result signals | |
DE102012024867A1 (en) | Method for operating operation device of motor vehicle e.g. passenger car, involves transmitting feedback signal to audio system for outputting feedback signal by operating device for detecting operation by sensor arrangement |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20160428 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
DAX | Request for extension of the european patent (deleted) | ||
17Q | First examination report despatched |
Effective date: 20190523 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R003 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN REFUSED |
|
18R | Application refused |
Effective date: 20201010 |