US20180018074A1 - Method and device for displaying information arranged in lists - Google Patents
Method and device for displaying information arranged in lists Download PDFInfo
- Publication number
- US20180018074A1 US20180018074A1 US15/716,109 US201715716109A US2018018074A1 US 20180018074 A1 US20180018074 A1 US 20180018074A1 US 201715716109 A US201715716109 A US 201715716109A US 2018018074 A1 US2018018074 A1 US 2018018074A1
- Authority
- US
- United States
- Prior art keywords
- list
- operator action
- displayed
- entries
- movement
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title claims abstract description 25
- 230000009471 action Effects 0.000 claims abstract description 79
- 230000001133 acceleration Effects 0.000 claims description 14
- 230000008859 change Effects 0.000 claims description 14
- 230000015654 memory Effects 0.000 claims description 7
- 230000004044 response Effects 0.000 claims 2
- 239000010408 film Substances 0.000 description 17
- 230000007704 transition Effects 0.000 description 10
- 238000004378 air conditioning Methods 0.000 description 6
- 238000001514 detection method Methods 0.000 description 6
- 230000003068 static effect Effects 0.000 description 6
- 238000013459 approach Methods 0.000 description 4
- 239000005262 ferroelectric liquid crystals (FLCs) Substances 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000003213 activating effect Effects 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000010168 coupling process Methods 0.000 description 2
- 238000005859 coupling reaction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 239000004988 Nematic liquid crystal Substances 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001105 regulatory effect Effects 0.000 description 1
- 238000011895 specific detection Methods 0.000 description 1
- 239000010409 thin film Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K35/00—Arrangement of adaptations of instruments
-
- B60K35/10—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60K—ARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
- B60K37/00—Dashboards
- B60K37/04—Arrangement of fittings on dashboard
- B60K37/06—Arrangement of fittings on dashboard of controls, e.g. controls knobs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
-
- B60K2350/1024—
-
- B60K2350/1028—
-
- B60K2360/143—
-
- B60K2360/1438—
Definitions
- the present invention relates to a method for displaying information, in which a portion of a list is displayed on a display area, which includes a plurality of list entries.
- the displayed portion of the list is changed by a first operator action of a user, the first operator action running consecutively through the entries of the list.
- the present invention furthermore relates to a device for displaying information including a display device having a display area, a memory for storing a list, which includes a plurality of list entries, of which a portion is displayable on the display area, and an operating device, which is operable by a user, in order to change the displayed portion of the list by a first operator action, the first operator action running consecutively through the entries of the list.
- Mobile devices in consumer electronics such as e.g. devices for playing back audio and/or video files have memories, in which very large amounts of data may be stored.
- Mobile music players for example, are able to store several thousands of musical pieces.
- the mobile devices there arises the problem of how the user is able to access these large amounts of data in a simple manner.
- the mobile devices often have a display and an operating device.
- the operating device may include a touch-sensitive surface of the display or may be an operating element that is separate from the display.
- the data sets stored in the device may be stored at least in part as a list, of which a portion may be displayed on the display. With the aid of the operating device, the user is able to scroll through the list in order to reach the desired list entry.
- the display of information in a vehicle also entails the problem that very large data sets must be visualized in the vehicle.
- Modern vehicles include for example a plurality of driver assistance systems, the information of which must be displayed in the vehicle.
- vehicles often include a navigation system.
- Such a navigation system is able to display digital geographic road maps including a route and possibly various additional information.
- modern vehicles often include communication and multimedia applications, including a mobile telephone interface and devices for playing back music and voice. For these applications it must also be possible to display information in the vehicle.
- the display of information in a vehicle and the operation of the manifold devices of the vehicle entail very specific requirements. In the vehicle, it is, among others, the driver who takes in the information and operates the respective device.
- the information should thus be represented in the vehicle in such a way that while taking in the information the driver is not distracted from driving.
- the represented information should therefore be ascertainable intuitively and quickly by the driver such that he has to take his eyes off the driving action only very briefly when taking in the information.
- the operation of the vehicle devices should be as simple and intuitive as possible such that the driver is able to operate the devices even while driving. If the operation is supported or guided by a display, then the display should occur in such a way that the driver has to view the display only very briefly in order to carry out the operation.
- the representation of information in a vehicle often occurs in that a portion of the total information is displayed on the display area in the vehicle and that the user is able to change the portion.
- the total information may be a list having a plurality of list entries for example. A portion of the entries of the entire list is displayed.
- the user is able to scroll so as to display other list entries.
- the user may perform for example an operator movement of a certain extent.
- the user may furthermore actuate an operating element for a certain operating duration.
- the quantity or the measure of the change of the displayed information i.e. in this case the number of items, by which the displayed list entries change, depends in this case on the extent of the operator movement or on the operating duration.
- Example embodiments of the present invention provide a method and a device of the type mentioned at the outset, by which the displayed portion of the list may be changed in a manner that is simple and intuitive for the user such that the operator action required for this purpose may be performed as quickly as possible. Furthermore, as little attention as possible is to be exacted from the user.
- a list should be understood as a data set that is divisible into a plurality of data units.
- the data units then correspond to the list entries.
- the data units may be brought into a certain sequence such that they then represent a list.
- the user is able to run through the entries of the list consecutively.
- a portion of the list i.e. multiple list entries, may be represented on the display area on the basis of a graphic object.
- the list may include a plurality of audio or video files for example. Furthermore, it may pertain to entries of a telephone or address book, entries of a menu of a hierarchical menu structure for operating a device or other data units, which may be integrated in a list and represented abstractly on the basis of a graphic object.
- the list may be scanned much more quickly in order to reach a desired list entry since it is not necessary to run through the entire list entry by entry before reaching the desired entry, as the second operator action rather makes it possible to skip multiple entries in order to reach the desired entry more quickly.
- the first operator action may include steering a movement of a list entry in a first direction and the second operator action may include steering a movement of a list entry in a second direction, which differs from the first direction.
- the second direction runs in particular substantially perpendicular to the first direction.
- a list entry may be moved by the operator action in the horizontal direction in order to scroll through the individual list entries.
- a vertical movement of the list entry has the result that multiple list entries are skipped so as to reach a list entry that was not yet displayed when the respective list entry was moved in the vertical direction.
- any other directional combinations are possible for the two operator actions as long as the two directions for the operator actions differ.
- the entries of the list may be divided into classes.
- the second operator action takes one to a list entry of another class.
- the classes in particular also form a list, the individual classes including sorted list entries.
- the second operator action preferably takes one to the first list entry of the next or previous class.
- An example of such a class division is an alphabetical sorting of list entries.
- a class in this case includes all list entries beginning with a particular letter.
- the first operator action is in particular used to scroll through the list.
- the second operator action may be used to abbreviate the scrolling by skipping multiple list entries.
- At least one operator action may include a movement on touch-sensitive surface of an input device. Many other operator actions are possible, however, which substantially depend on the type of input device. If the input device allows for so-called gesture detection, then the operator action may include a gesture performed by the hand of a user. If the input device is a mechanical operating element, then the operator action is to actuate this operating element, e.g. to turn it or to press on it for a certain time.
- the touch-sensitive surface of the input device may be provided on the display area.
- the entries of the portion of the list may be displayed in a one-dimensional line on the display area.
- the first operator action in this case may include a movement on the touch-sensitive surface, which runs substantially in the direction of the one-dimensional line.
- the second operator action may include a movement on the touch-sensitive surface, which runs substantially in a direction perpendicular to the one-dimensional line.
- the movements may run exactly in the direction of the one-dimensional line or perpendicular to it. They may also run in an angular range of e.g. ⁇ 5 degrees or ⁇ 10 degrees around these directions.
- the second operator action may furthermore be a movement, which starts with a touch of the touch-sensitive surface at a position, at which an entry of the portion of the list is displayed.
- At least one operator action may include a movement of the display area in space, and the acceleration of the display area in space may be detected.
- the user moves the device, which includes the display area, e.g. in a particular direction.
- This movement includes an acceleration of the display area, which is detected.
- the direction in which a list is to be run through is derivable from the direction of movement.
- the movement in a different direction may be interpreted as a second operator action, in which multiple list entries are skipped.
- the displayed list entries may be represented as arranged on a virtual ring or ring section represented in perspective.
- the first operator action rotates the displayed list entries on the virtual ring or ring section in carousel-like fashion.
- Rotating should be understood as rotating the virtual ring or ring section. Since the three-dimensional virtual ring on the display is represented in a two-dimensional perspective view, the objects are actually shifted with respect to their position on the display.
- the user is able to actuate the operating device by a second operator action, by which multiple list entries are skipped in order to arrive at a list entry that is not contained in the displayed portion of list entries that were displayed while performing the second operator action.
- the device may be able to implement the method steps described above partially or entirely.
- the operating device of the display device may include a user interface device and an input device.
- the user interface device As a function of an input via the input device, the user interface device is able to produce graphic data, which control the display device such that a particular display is represented on the display area.
- the user interface device may include a computing unit, which produces graphic data for intermediate images for the animation. It is also possible, however, for these intermediate images to be already stored in advance in the memory.
- the first or second operator action which is implemented by the operating device of the device into a change of the display on the display area, includes in particular the operator actions described above with reference to the method.
- the input device may include a touch-sensitive surface. This surface may be part of a so-called touch pad. Furthermore, the surface may be provided on the display area, i.e. the display device is equipped with a touch screen.
- the input device may be a device for detecting and evaluating a gesture of a part of the body of a user, which is performed in front of the display area.
- the input device may include in particular a receiving device, to which a signal from the body part of the user may be transmitted capacitively when the body part is in proximity of the receiving device. The position of the body part may be detected via this capacitive coupling. From the change of this position over time, a gesture of the user may be inferred.
- the input device may include an infrared light source or a receiver for reflected infrared light for detecting the gesture of the body part of the user. In this case as well, the position of the body part and its change over time are detected and interpreted as a gesture.
- an approach sensor may be provided, which detects the operating intention of a user on the basis of the approach of a part of the body of the user.
- the device may include an acceleration sensor, which is able to detect the acceleration of the display area in space.
- the operating device is able to interpret the data detection of the acceleration sensor so as to assign it to the first or to the second operator action.
- FIG. 1 schematically shows display device according to an example embodiment of the present invention and the coupling of this display device to the electronic system of a motor vehicle.
- FIG. 2 shows a display of information produced by the method according to an example embodiment of the present invention in a motor vehicle, in which multiple objects are represented.
- FIG. 3 shows a display of information of an activated object produced by the method according to an example embodiment of the present invention.
- FIG. 4 shows a display of information of another object produced by the method according to an example embodiment of the present invention.
- FIG. 5 shows the time characteristic of the rotational angle of an object when rotating on the virtual ring.
- FIGS. 6 through 8 show the change of the displayed list entries in a second operator action.
- An exemplary embodiment described below relates to a display device and a method for displaying information in a vehicle, in particular a motor vehicle.
- Example embodiments of the present invention may be utilized in the same manner in a mobile or stationary consumer electronics device, in particular a device for playing back audio and video files.
- the display device includes a display 1 for the graphical representation of information on a display area.
- Display 1 may be a matrix display, e.g. an LCD (liquid crystal display), especially a color display using TFT (thin-film transistor) technology.
- the display may be a so-called twisted nematic-liquid crystal display (TN-LCD), a super twisted nematic (STN) display, a double-layer STN, an FLC (ferroelectric liquid crystal) display or an SSFLC (surface stabilized ferroelectric liquid crystal) display.
- Display 1 has an associated back-lighting (not shown), which may be provided by one or more light-emitting diodes.
- Display 1 is freely programmable, i.e. any desired graphic data may be produced, which are represented on display 1 .
- display 1 is mounted in an area of the vehicle that is clearly visible at least for the driver. If the operation of the devices of the vehicle is coupled directly to the position of display 1 such that the user must bring e.g. his hand or his finger at least near display 1 in order to make inputs, display 1 is positioned in such a way that the driver of the vehicle may reach it readily with his hand or his finger. Display 1 may be accommodated in the center console of the vehicle for example.
- Display 1 is connected to a user interface device 2 , which is able to produce graphic data for graphic objects 6 representable on display 1 . Furthermore, user interface device 2 is connected to an input device 4 , via which the user is able to control devices of the vehicle, the information of which is displayed on display 1 . User interface device 2 and input device 4 form an operating device, which may be actuated by a user and which implements the input of a user in graphic data for representation on display 1 .
- User interface device 2 is furthermore connected to a memory 15 .
- data for a list as well as the graphic objects 6 associated with the individual list entries may be stored in this memory.
- Input device 4 may be for example a device for detecting and evaluating a gesture of a part of a user's body.
- the hand of the user may perform the gesture in front of display 1 for example.
- the three-dimensional position of the hand is detected in a specific detection area in front of display 1 , without it being necessary to touch display 1 .
- the admissible detection area depends on the placement of display 1 in the motor vehicle. The area should be selected such that the presence of the hand of a user in this detection area may be associated unequivocally with an operative control of input device 4 .
- the boundary of the detection area may be e.g. 40 cm to 10 cm in front of display 1 .
- Input device 4 detects the position and the movement of the hand of the user in the detection area. In the process, various gestures performed by the hand are recognized and interpreted as inputs.
- Input device 4 may include e.g. infrared light sources and infrared light receivers, which detect the infrared light reflected by the hand. Details of such an input device are described in German Patent Document No. 100 58 244 which is incorporated herein by reference. Further input devices, which may be used in conjunction with the display device, are described in the following publications: German Patent Document No. 103 05 341 and German Patent Document No. 10 2004 048 956.
- the position of the hand and its change over time may also be detected by an optical system.
- a light-emitting diode emits e.g. square-wave, amplitude-modulated light. This light is reflected by the object to be detected, i.e. the hand, and after the reflection reaches a photodiode.
- Another light-emitting diode likewise emits square-wave, amplitude-modulated light to the photodiode, which light is phase-shifted by 180° however.
- the two light signals are superposed and cancel each other out if they have exactly the same amplitude.
- the light emission of the second diode is regulated via a control loop such that the total received signal again adds up to zero. If the position of the object changes, this also causes a change in the light component which arrives at the photodiode from the first light-emitting diode via the reflection on the object. This brings about a correction of the intensity of the second light-emitting diode through the control loop.
- the control signal is therefore a measure for the reflection of the light, which is emitted by the first diode, on the object. In this manner, it is possible to derive from the control signal a signal that is characteristic for the position of the object.
- the input device may be a touch-sensitive film, which is provided on display 1 .
- the film makes it possible to detect the position at which the user touches display 1 situated behind the film.
- the film may be arranged e.g. as a resistive touch film, a capacitive touch film or piezoelectric film.
- the film may be arranged so as to measure a heat flow emanating e.g. from the finger of a user.
- Various inputs may be obtained from the development of the touch of the film over time. In the simplest case, for example, a touch of the film at a specific position may be allocated to a graphic object shown on display 1 . Furthermore, sliding movements of the finger over the film may be interpreted. In particular, in this manner the user is able to define a line on display 1 , in that he touches the film at one point, slides on the film toward another point and removes the finger from the film at the other point.
- a separate operating element may be used as input device.
- the separate operating element is a mechanical operating element.
- a rotary switch may be provided, by which objects shown on the display are controllable and are selectable by pressing the rotary switch.
- an angle of rotation may also be input directly by the rotary switch, as will be explained later.
- separate press switches may be disposed around the rotary switch, the arrangement of display fields on the display, which are associated with the press switches, corresponding at least schematically to the arrangement of the press switches.
- the informational system may include a multifunction operating device as described in European Patent Document No. 1 212 208 for example.
- User interface device 2 is further coupled to a system clock 8 and if applicable to a vehicle bus 7 .
- User interface device 2 may be connected e.g. to driver assistance systems of the vehicle via vehicle bus 7 .
- User interface device 2 receives data from these driver assistance systems via vehicle bus 7 , and processes these data so that they are displayed graphically to the driver or the vehicle occupants via display 1 .
- user interface device 2 produces graphic data for objects displayable on display 1 , which, inter alia, graphically represent the information of the driver assistance systems.
- User interface device 2 is furthermore connected via vehicle bus 7 to various information, communication and entertainment devices of the vehicle. The varied information from these devices of the vehicle is processed in user interface device 2 and converted into graphic data for a graphical representation.
- the user interface device includes a computing unit 3 , which accesses system clock 8 in producing intermediate images.
- the graphic data, which user interface device 2 produces for display 1 relate to a portion of a list. Multiple entries from this list are represented as objects arranged on a virtual ring section 5 shown in perspective.
- the perspective representation shows virtual ring section 5 as though the observer were looking at an angle from above onto ring section 5 .
- An object 6 in the foreground is shown as the largest in the perspective view. Further back, additional objects 6 are shown, partially hidden by objects 6 situated in front of them.
- One or more devices of the vehicle are assigned to objects 6 or to the associated list entries. Objects 6 shown in FIG. 2 either give only an indication of the associated vehicle device or additionally already provide informational content of this device.
- object 6 represented in the foreground refers to the control of the air conditioning of the interior of the vehicle.
- the object adjacent to the right and behind this object 6 is assigned to the music playback device of the vehicle, and the object adjacent to the left of object 6 associated with the air-conditioning control is assigned to the navigation system of the vehicle.
- An object situated further back relates to the settings of the display device itself.
- buttons for controlling the graphic data produced by user interface device 2 for display 1 are buttons for controlling the graphic data produced by user interface device 2 for display 1 .
- a first button 9 the user is able to return in any display state to the initial state, in which objects 6 are displayed as situated on the virtual ring section 5 shown in perspective.
- buttons 10 and 11 objects 6 on virtual ring section 5 may be rotated in carousel-like fashion by a position clockwise or counterclockwise, as will be explained later in detail.
- Buttons 12 and 13 may be used to actuate functions within an object 6
- the user is able to access e.g. directly the settings menu for the graphics.
- any of the objects 6 shown on virtual ring section 5 may be selected and activated using input device 4 .
- the input device includes a touch-sensitive film, the user can tap e.g. using his finger on the object 6 to be activated.
- input device 4 detects e.g. the position of the finger or the hand of the user in front of display 1 , it suffices for the user to bring his finger near the position of display 1 , at which the object 6 to be activated is represented. If an object 6 has been activated with the aid of input device 4 , then in a smooth transition, virtual ring section 5 with the objects 6 situated on it is faded out and only activated object 6 is still represented.
- activated object 6 displaces the other objects 6 of ring section 5 and is enlarged continuously until only activated object 6 is still displayed.
- the graphic data for this smooth transition are calculated by the user interface device 2 .
- user interface device 2 uses system time 8 in order to ensure that the animation is reproduced smoothly without jerky movements.
- FIG. 3 for example shows a view of display 1 , which is shown when activating the object 6 associated with the air conditioning.
- the representation of object 6 for the air conditioning shown in FIG. 2 is magnified. Buttons 9 through 14 are preserved in the process. If the user actuates button 9 from this view shown in FIG. 3 , then the representation of object 6 for the air conditioning is continuously scaled down, and the other objects 6 of virtual ring section 5 appear again, until the view shown in FIG. 2 is displayed.
- object 6 for the air conditioning is similarly scaled down and the other objects 6 of virtual ring section 5 are displayed again as shown in FIG. 2 . Subsequently, however, virtual ring section 5 is rotated clockwise or counterclockwise in carousel-like fashion to the next position, whereupon the next object 6 now positioned in the foreground is activated and in a smooth transition is shown enlarged and as displacing the other objects 6 , until only this next object 6 is displayed.
- FIG. 4 shows object 6 associated with the settings of the graphics system itself. When activating this object 6 , various settings may be adjusted for the graphic display using input device 4 .
- the list entries, to which graphic objects 6 belong, are part of a hierarchical structure, as described above, since the activation of a list entry or a graphic object 6 results in the display of more detailed information. Furthermore, the activation of a graphic object 6 or a list entry may result e.g. in the playback of an audio file. In this case too, the playback of the audio file corresponds to a lower level of the hierarchy. When displaying such a hierarchically organized data set, the transition between two static representations on display 1 is of particular importance.
- the display device is installed in a motor vehicle, it is very important that the driver of the motor vehicle in particular is able to detect the information represented on display 1 quickly, simply and intuitively, even if he directs his eyes onto display 1 only for brief intervals. For this purpose it is particularly important that the viewer is at all times able to orient himself well within the hierarchical structure. It has been shown that abrupt changes of displayed images are disadvantageous for this purpose. It has been shown in particular that smooth transitions between two static representations, in which intermediate images are displayed, which emphasize the transition between the two static representations, facilitate orientation within the structure as well as the grasp of the informational content.
- intermediate images visualize the transition between two static representations.
- a linear transition from one static representation to the next static representation is not suitable for use in a motor vehicle.
- intermediate images are advantageous, which illustrate a movement of objects 6 that is at least partially slowed down.
- the parameters for the choice of the slowdown and the duration of the transition must also be adapted for use in the vehicle because conventional animations known from the field of computers are usually not usable.
- the starting point of the rotation is the arrangement of objects 6 on virtual ring section 5 shown in FIG. 2 .
- Objects 6 are thus represented on a ring-shaped one-dimensional line.
- the user By a first operator action via input device 4 , the user now specifies the total rotational angle ⁇ Rotation , by which objects 6 are to be rotated on the ring as a whole, of which only a ring section 5 is shown.
- the user enters a line on display 1 using input device 4 for example. If input device 4 includes a touch-sensitive film, the user may run his finger over the film in order to specify this line.
- This line for entering the first operator action runs in particular substantially on the line on which objects 6 are represented, or parallel to this line.
- the line is entered in this manner. If the line is not oriented in the direction of the width of display 1 , but rather runs aslant, preferably only the length of the horizontal component of the entered line oriented in parallel to the width of display 1 is used, as long as the deviation from the horizontal does not exceed 45°.
- the total rotational angle ⁇ Rotation is calculated from the ratio of the length of this horizontal component of the line to the total width of display 1 . In fact, in the perspective view on display 1 , objects 6 are shifted only in the plane of display 1 . Rotational angle ⁇ is thus transformed accordingly by user interface device 2 when producing the graphic data.
- the total rotational angle may also be input via the rotational angle of the rotary encoder.
- the shown portion changes as the virtual ring is rotated.
- objects 6 disappear on the one side of ring section 5 , while new objects 6 are displayed on the other side of ring section 5 , which belong to list entries that were originally not displayed.
- the first operator action may be used to run through the entries of the list consecutively, the entries of the list being shown on the basis of graphic objects 6 .
- a computing unit 3 of user interface device 2 produces an animation, which includes the graphic data for successive intermediate images, which represent successive rotational angles ⁇ for objects 6 .
- the transition from the initial state having an initial angle of the objects to a final state, in which the objects have been rotated by total angle ⁇ Rotation occurs by a decelerated movement, which is calculated by the following steps:
- t is the system time, which computing unit 3 retrieves from system clock 8 .
- the system time is zero.
- FIG. 5 shows the time characteristic of rotational angle ⁇ for the intermediate images if an object 6 is rotated by a total rotational angle ⁇ Rotation of 30° in 2.5 seconds. It can be seen that the rotational speed is initially relatively fast and is then slowed down.
- the total rotational angle ⁇ Rotation is additionally adapted to a rotational angle, which corresponds to a carousel-like rotation of objects 6 on virtual ring section 5 , at which in the final state an object 6 appears in the foreground, as shown in FIG. 2 .
- the rotational angle input by the user via input device 4 is thus rounded up or down to these discrete rotational angles.
- this object 6 is automatically activated after the expiration of a time interval. Following the expiration of the time interval, this object 6 is thus magnified, while at the same time the virtual ring section 5 with the other objects 6 vanishes until only the activated object 6 is still displayed.
- the data of the speedometer of the motor vehicle may be transmitted to user interface device 2 and thus computing unit 3 via vehicle bus 7 . These data are used by computing unit 3 to ascertain the duration of the rotation. The higher the speed of the motor vehicle, the longer is the duration of the total rotation.
- a second operator action may be performed in the method described herein or the device described herein.
- the user performs an operator movement that differs from the operator movement of the first operator action.
- a vertical operator movement upward or downward is performed.
- the user may move the finger upward or downward on the touch-sensitive surface.
- the second operator movement in particular starts with a touch of the touch-sensitive surface at a position at which an object 6 , i.e. an entry of the portion of the list, is displayed.
- an object 6 i.e. an entry of the portion of the list
- the second operator action has the consequence that user interface device 2 changes the graphic data in such a way that multiple list entries are skipped in the display of the portion of the list so as to reach a list entry, which, or the associated graphic object 6 of which, is not contained in the displayed portion of the list entries that were displayed while performing the second operator action.
- a graphic object 6 is thus displayed in the foreground, which was previously not displayed.
- the total ring with the list entries was transported directly into a rotated state, in which another portion of the list is shown, without it having been necessary to run through the individual list entries consecutively.
- the starting point is the display of a portion of a list on the display area of display 1 , as it is represented in FIG. 6 .
- This informational display substantially corresponds to the informational display represented in FIG. 2 of the list entries on the basis of graphic objects 6 .
- three list entries G 22 , G 23 , and G 24 are shown by way of example.
- the entries of this list are divided into classes, the individual classes in turn forming a list.
- this may be an alphabetic sorting of concepts associated with the list entries for example.
- list entries G 22 , G 23 and G 24 are the entries no. 22 through 24 beginning with the letter G.
- the list entries having other initial letters are sorted accordingly.
- the user may perform a movement upward in the direction of arrow A or downward in the direction of arrow B.
- An operator action upward in the direction of arrow A has the consequence that the displayed portion of the list entries is changed immediately such that multiple list entries are skipped and the first list entry of the next initial letter is reached, i.e. as shown in FIG. 7 , list entry F 1 .
- the displayed portion of the list may change such that the last list entry or the last entries beginning with the letter G is or are not included in the displayed portion.
- the second operator action by contrast includes an operator movement downward in the direction of arrow B
- the first list entry beginning with the next lower initial letter is reached directly, i.e. as shown in FIG. 8 , list entry H 1 .
- a portion of the list is thus displayed, which begins with the list entry H 1 and thereupon shows additional list entries beginning with the letter H.
- User interface device 2 may furthermore be coupled to an acceleration sensor 16 , which, alternatively or additionally, may be used as an input device.
- Acceleration sensor 16 is able to detect an acceleration of the device, which in this case is arranged in particular as a mobile device, as well as the direction of acceleration.
- a movement of the device, including display area 1 , in space corresponds in this case to the previously explained movement on the touch-sensitive surface of display area 1 .
- a movement to the left makes it possible to scroll through the list in one direction for example, while a movement to the right makes it possible to scroll through the list in the other direction.
- An upward movement results in multiple list entries being skipped forward in the list, while a movement downward results in multiple list entries being skipped backward, as was explained above.
Abstract
Description
- The present application is a continuation application of U.S. patent application Ser. No. 13/125,175, filed on Jul. 15, 2011, as a national stage application of PCT/EP2009/060249, filed on Aug. 6, 2009, which claims priority to German Patent Application No. 10 2008 052 485.9, filed on Oct. 21, 2008, each of which is incorporated herein in its entirety by reference thereto.
- The present invention relates to a method for displaying information, in which a portion of a list is displayed on a display area, which includes a plurality of list entries. The displayed portion of the list is changed by a first operator action of a user, the first operator action running consecutively through the entries of the list. The present invention furthermore relates to a device for displaying information including a display device having a display area, a memory for storing a list, which includes a plurality of list entries, of which a portion is displayable on the display area, and an operating device, which is operable by a user, in order to change the displayed portion of the list by a first operator action, the first operator action running consecutively through the entries of the list.
- Mobile devices in consumer electronics such as e.g. devices for playing back audio and/or video files have memories, in which very large amounts of data may be stored. Mobile music players, for example, are able to store several thousands of musical pieces. In such mobile devices there arises the problem of how the user is able to access these large amounts of data in a simple manner. For navigation, the mobile devices often have a display and an operating device. The operating device may include a touch-sensitive surface of the display or may be an operating element that is separate from the display. The data sets stored in the device may be stored at least in part as a list, of which a portion may be displayed on the display. With the aid of the operating device, the user is able to scroll through the list in order to reach the desired list entry. If the list contains very many entries, however, there arises the problem that it takes very long to reach an entry that is arranged in the list very far away from the displayed list entries. For this reason it has been proposed to group the list entries hierarchically. This entails the disadvantage, however, that the user must navigate through a deeply branched hierarchical menu structure in order to arrive at a desired list entry.
- Furthermore, the display of information in a vehicle also entails the problem that very large data sets must be visualized in the vehicle. Modern vehicles include for example a plurality of driver assistance systems, the information of which must be displayed in the vehicle. Furthermore, vehicles often include a navigation system. Such a navigation system is able to display digital geographic road maps including a route and possibly various additional information. Finally, modern vehicles often include communication and multimedia applications, including a mobile telephone interface and devices for playing back music and voice. For these applications it must also be possible to display information in the vehicle. The display of information in a vehicle and the operation of the manifold devices of the vehicle entail very specific requirements. In the vehicle, it is, among others, the driver who takes in the information and operates the respective device. The information should thus be represented in the vehicle in such a way that while taking in the information the driver is not distracted from driving. The represented information should therefore be ascertainable intuitively and quickly by the driver such that he has to take his eyes off the driving action only very briefly when taking in the information. Likewise, the operation of the vehicle devices should be as simple and intuitive as possible such that the driver is able to operate the devices even while driving. If the operation is supported or guided by a display, then the display should occur in such a way that the driver has to view the display only very briefly in order to carry out the operation.
- The representation of information in a vehicle often occurs in that a portion of the total information is displayed on the display area in the vehicle and that the user is able to change the portion. The total information may be a list having a plurality of list entries for example. A portion of the entries of the entire list is displayed. By an operator action, the user is able to scroll so as to display other list entries. In the operator action, the user may perform for example an operator movement of a certain extent. The user may furthermore actuate an operating element for a certain operating duration. The quantity or the measure of the change of the displayed information, i.e. in this case the number of items, by which the displayed list entries change, depends in this case on the extent of the operator movement or on the operating duration.
- If a very long list having many entries is to be accessed in the vehicle, there exists the problem that the operator action for scrolling to very distant list entries requires much time and attention on the part of the user. In the application in a vehicle, however, this is disadvantageous, as explained above.
- Example embodiments of the present invention provide a method and a device of the type mentioned at the outset, by which the displayed portion of the list may be changed in a manner that is simple and intuitive for the user such that the operator action required for this purpose may be performed as quickly as possible. Furthermore, as little attention as possible is to be exacted from the user.
- In the method according to example embodiments of the present invention, by a second operator action, multiple list entries are skipped so as to reach a list entry that is not contained in the displayed portion of the list entries that were displayed while performing the second operator action.
- A list should be understood as a data set that is divisible into a plurality of data units. The data units then correspond to the list entries. The data units may be brought into a certain sequence such that they then represent a list. By the first operator action, the user is able to run through the entries of the list consecutively. For this purpose, a portion of the list, i.e. multiple list entries, may be represented on the display area on the basis of a graphic object.
- The list may include a plurality of audio or video files for example. Furthermore, it may pertain to entries of a telephone or address book, entries of a menu of a hierarchical menu structure for operating a device or other data units, which may be integrated in a list and represented abstractly on the basis of a graphic object.
- Using the second operator action, which the method described herein makes possible, the list may be scanned much more quickly in order to reach a desired list entry since it is not necessary to run through the entire list entry by entry before reaching the desired entry, as the second operator action rather makes it possible to skip multiple entries in order to reach the desired entry more quickly.
- The first operator action may include steering a movement of a list entry in a first direction and the second operator action may include steering a movement of a list entry in a second direction, which differs from the first direction. The second direction runs in particular substantially perpendicular to the first direction. For example, a list entry may be moved by the operator action in the horizontal direction in order to scroll through the individual list entries. In this case, a vertical movement of the list entry has the result that multiple list entries are skipped so as to reach a list entry that was not yet displayed when the respective list entry was moved in the vertical direction. Of course, any other directional combinations are possible for the two operator actions as long as the two directions for the operator actions differ.
- The entries of the list may be divided into classes. In this case, the second operator action takes one to a list entry of another class. The classes in particular also form a list, the individual classes including sorted list entries. In this case, the second operator action preferably takes one to the first list entry of the next or previous class. An example of such a class division is an alphabetical sorting of list entries. A class in this case includes all list entries beginning with a particular letter.
- The first operator action is in particular used to scroll through the list. The second operator action may be used to abbreviate the scrolling by skipping multiple list entries.
- At least one operator action may include a movement on touch-sensitive surface of an input device. Many other operator actions are possible, however, which substantially depend on the type of input device. If the input device allows for so-called gesture detection, then the operator action may include a gesture performed by the hand of a user. If the input device is a mechanical operating element, then the operator action is to actuate this operating element, e.g. to turn it or to press on it for a certain time.
- The touch-sensitive surface of the input device may be provided on the display area. The entries of the portion of the list may be displayed in a one-dimensional line on the display area. The first operator action in this case may include a movement on the touch-sensitive surface, which runs substantially in the direction of the one-dimensional line.
- The second operator action may include a movement on the touch-sensitive surface, which runs substantially in a direction perpendicular to the one-dimensional line. For this purpose, it is not necessary for the movements to run exactly in the direction of the one-dimensional line or perpendicular to it. They may also run in an angular range of e.g. ±5 degrees or ±10 degrees around these directions.
- If the touch-sensitive surface of the input device is provided on the display area, then the second operator action may furthermore be a movement, which starts with a touch of the touch-sensitive surface at a position, at which an entry of the portion of the list is displayed.
- At least one operator action may include a movement of the display area in space, and the acceleration of the display area in space may be detected. In this case, the user moves the device, which includes the display area, e.g. in a particular direction. This movement includes an acceleration of the display area, which is detected. The direction in which a list is to be run through is derivable from the direction of movement. The movement in a different direction may be interpreted as a second operator action, in which multiple list entries are skipped.
- The displayed list entries may be represented as arranged on a virtual ring or ring section represented in perspective. In this case, the first operator action rotates the displayed list entries on the virtual ring or ring section in carousel-like fashion.
- Rotating should be understood as rotating the virtual ring or ring section. Since the three-dimensional virtual ring on the display is represented in a two-dimensional perspective view, the objects are actually shifted with respect to their position on the display.
- In a device according to example embodiments of the present invention for displaying information, the user is able to actuate the operating device by a second operator action, by which multiple list entries are skipped in order to arrive at a list entry that is not contained in the displayed portion of list entries that were displayed while performing the second operator action.
- The device may be able to implement the method steps described above partially or entirely.
- The operating device of the display device may include a user interface device and an input device. As a function of an input via the input device, the user interface device is able to produce graphic data, which control the display device such that a particular display is represented on the display area. For changing the graphic data, for example for producing an animation, the user interface device may include a computing unit, which produces graphic data for intermediate images for the animation. It is also possible, however, for these intermediate images to be already stored in advance in the memory.
- The first or second operator action, which is implemented by the operating device of the device into a change of the display on the display area, includes in particular the operator actions described above with reference to the method. For this purpose, the input device may include a touch-sensitive surface. This surface may be part of a so-called touch pad. Furthermore, the surface may be provided on the display area, i.e. the display device is equipped with a touch screen. Moreover, the input device may be a device for detecting and evaluating a gesture of a part of the body of a user, which is performed in front of the display area. For this purpose, the input device may include in particular a receiving device, to which a signal from the body part of the user may be transmitted capacitively when the body part is in proximity of the receiving device. The position of the body part may be detected via this capacitive coupling. From the change of this position over time, a gesture of the user may be inferred.
- Furthermore, the input device may include an infrared light source or a receiver for reflected infrared light for detecting the gesture of the body part of the user. In this case as well, the position of the body part and its change over time are detected and interpreted as a gesture.
- Furthermore, an approach sensor may be provided, which detects the operating intention of a user on the basis of the approach of a part of the body of the user.
- Furthermore, the device may include an acceleration sensor, which is able to detect the acceleration of the display area in space. The operating device is able to interpret the data detection of the acceleration sensor so as to assign it to the first or to the second operator action.
- Example embodiments of the present invention are explained in more detail below with reference to the Figures.
-
FIG. 1 schematically shows display device according to an example embodiment of the present invention and the coupling of this display device to the electronic system of a motor vehicle. -
FIG. 2 shows a display of information produced by the method according to an example embodiment of the present invention in a motor vehicle, in which multiple objects are represented. -
FIG. 3 shows a display of information of an activated object produced by the method according to an example embodiment of the present invention. -
FIG. 4 shows a display of information of another object produced by the method according to an example embodiment of the present invention. -
FIG. 5 shows the time characteristic of the rotational angle of an object when rotating on the virtual ring. -
FIGS. 6 through 8 show the change of the displayed list entries in a second operator action. - An exemplary embodiment described below relates to a display device and a method for displaying information in a vehicle, in particular a motor vehicle. Example embodiments of the present invention, however, may be utilized in the same manner in a mobile or stationary consumer electronics device, in particular a device for playing back audio and video files.
- The display device includes a
display 1 for the graphical representation of information on a display area.Display 1 may be a matrix display, e.g. an LCD (liquid crystal display), especially a color display using TFT (thin-film transistor) technology. Furthermore, the display may be a so-called twisted nematic-liquid crystal display (TN-LCD), a super twisted nematic (STN) display, a double-layer STN, an FLC (ferroelectric liquid crystal) display or an SSFLC (surface stabilized ferroelectric liquid crystal) display.Display 1 has an associated back-lighting (not shown), which may be provided by one or more light-emitting diodes.Display 1 is freely programmable, i.e. any desired graphic data may be produced, which are represented ondisplay 1. - In particular,
display 1 is mounted in an area of the vehicle that is clearly visible at least for the driver. If the operation of the devices of the vehicle is coupled directly to the position ofdisplay 1 such that the user must bring e.g. his hand or his finger at least neardisplay 1 in order to make inputs,display 1 is positioned in such a way that the driver of the vehicle may reach it readily with his hand or his finger.Display 1 may be accommodated in the center console of the vehicle for example. -
Display 1 is connected to auser interface device 2, which is able to produce graphic data forgraphic objects 6 representable ondisplay 1. Furthermore,user interface device 2 is connected to an input device 4, via which the user is able to control devices of the vehicle, the information of which is displayed ondisplay 1.User interface device 2 and input device 4 form an operating device, which may be actuated by a user and which implements the input of a user in graphic data for representation ondisplay 1. -
User interface device 2 is furthermore connected to amemory 15. In particular data for a list as well as thegraphic objects 6 associated with the individual list entries may be stored in this memory. - Input device 4 may be for example a device for detecting and evaluating a gesture of a part of a user's body. The hand of the user may perform the gesture in front of
display 1 for example. For this purpose, the three-dimensional position of the hand is detected in a specific detection area in front ofdisplay 1, without it being necessary to touchdisplay 1. The admissible detection area depends on the placement ofdisplay 1 in the motor vehicle. The area should be selected such that the presence of the hand of a user in this detection area may be associated unequivocally with an operative control of input device 4. The boundary of the detection area may be e.g. 40 cm to 10 cm in front ofdisplay 1. If the hand of the user is brought up closer than this threshold value to display 1, this is detected by input device 4 or by a separate approach sensor, and the approach is interpreted as an operating intention. For instance, this may result in a change in the manner in which the objects represented ondisplay 1 are rendered. Input device 4 detects the position and the movement of the hand of the user in the detection area. In the process, various gestures performed by the hand are recognized and interpreted as inputs. - Input device 4 may include e.g. infrared light sources and infrared light receivers, which detect the infrared light reflected by the hand. Details of such an input device are described in German Patent Document No. 100 58 244 which is incorporated herein by reference. Further input devices, which may be used in conjunction with the display device, are described in the following publications: German Patent Document No. 103 05 341 and German Patent Document No. 10 2004 048 956.
- Furthermore, the position of the hand and its change over time may also be detected by an optical system. In this system, a light-emitting diode emits e.g. square-wave, amplitude-modulated light. This light is reflected by the object to be detected, i.e. the hand, and after the reflection reaches a photodiode. Another light-emitting diode likewise emits square-wave, amplitude-modulated light to the photodiode, which light is phase-shifted by 180° however. At the photodiode, the two light signals are superposed and cancel each other out if they have exactly the same amplitude. If the signals do not cancel each other out at the photodiode, the light emission of the second diode is regulated via a control loop such that the total received signal again adds up to zero. If the position of the object changes, this also causes a change in the light component which arrives at the photodiode from the first light-emitting diode via the reflection on the object. This brings about a correction of the intensity of the second light-emitting diode through the control loop. The control signal is therefore a measure for the reflection of the light, which is emitted by the first diode, on the object. In this manner, it is possible to derive from the control signal a signal that is characteristic for the position of the object.
- In addition, the input device may be a touch-sensitive film, which is provided on
display 1. The film makes it possible to detect the position at which the user touchesdisplay 1 situated behind the film. The film may be arranged e.g. as a resistive touch film, a capacitive touch film or piezoelectric film. Furthermore, the film may be arranged so as to measure a heat flow emanating e.g. from the finger of a user. Various inputs may be obtained from the development of the touch of the film over time. In the simplest case, for example, a touch of the film at a specific position may be allocated to a graphic object shown ondisplay 1. Furthermore, sliding movements of the finger over the film may be interpreted. In particular, in this manner the user is able to define a line ondisplay 1, in that he touches the film at one point, slides on the film toward another point and removes the finger from the film at the other point. - Finally, a separate operating element may be used as input device. In particular, the separate operating element is a mechanical operating element. For instance, a rotary switch may be provided, by which objects shown on the display are controllable and are selectable by pressing the rotary switch. Furthermore, an angle of rotation may also be input directly by the rotary switch, as will be explained later. In addition, separate press switches may be disposed around the rotary switch, the arrangement of display fields on the display, which are associated with the press switches, corresponding at least schematically to the arrangement of the press switches. The informational system may include a multifunction operating device as described in European Patent Document No. 1 212 208 for example.
-
User interface device 2 is further coupled to asystem clock 8 and if applicable to avehicle bus 7.User interface device 2 may be connected e.g. to driver assistance systems of the vehicle viavehicle bus 7.User interface device 2 receives data from these driver assistance systems viavehicle bus 7, and processes these data so that they are displayed graphically to the driver or the vehicle occupants viadisplay 1. To this end,user interface device 2 produces graphic data for objects displayable ondisplay 1, which, inter alia, graphically represent the information of the driver assistance systems.User interface device 2 is furthermore connected viavehicle bus 7 to various information, communication and entertainment devices of the vehicle. The varied information from these devices of the vehicle is processed inuser interface device 2 and converted into graphic data for a graphical representation. For animations of the graphical representation on the display, the user interface device includes acomputing unit 3, which accessessystem clock 8 in producing intermediate images. - In the following, the display of information on
display 1 in the vehicle is explained in detail: - The graphic data, which
user interface device 2 produces fordisplay 1, relate to a portion of a list. Multiple entries from this list are represented as objects arranged on avirtual ring section 5 shown in perspective. The perspective representation showsvirtual ring section 5 as though the observer were looking at an angle from above ontoring section 5. Anobject 6 in the foreground is shown as the largest in the perspective view. Further back,additional objects 6 are shown, partially hidden byobjects 6 situated in front of them. One or more devices of the vehicle are assigned toobjects 6 or to the associated list entries.Objects 6 shown inFIG. 2 either give only an indication of the associated vehicle device or additionally already provide informational content of this device. - In the view shown in
FIG. 2 ,object 6 represented in the foreground refers to the control of the air conditioning of the interior of the vehicle. The object adjacent to the right and behind thisobject 6 is assigned to the music playback device of the vehicle, and the object adjacent to the left ofobject 6 associated with the air-conditioning control is assigned to the navigation system of the vehicle. An object situated further back relates to the settings of the display device itself. - Below the
virtual ring section 5 shown in perspective are buttons for controlling the graphic data produced byuser interface device 2 fordisplay 1. Using afirst button 9, the user is able to return in any display state to the initial state, in which objects 6 are displayed as situated on thevirtual ring section 5 shown in perspective. Usingbuttons virtual ring section 5 may be rotated in carousel-like fashion by a position clockwise or counterclockwise, as will be explained later in detail.Buttons object 6, and viabutton 14 the user is able to access e.g. directly the settings menu for the graphics. - Starting from the view of
display 1 shown inFIG. 2 , any of theobjects 6 shown onvirtual ring section 5 may be selected and activated using input device 4. If the input device includes a touch-sensitive film, the user can tap e.g. using his finger on theobject 6 to be activated. If input device 4 detects e.g. the position of the finger or the hand of the user in front ofdisplay 1, it suffices for the user to bring his finger near the position ofdisplay 1, at which theobject 6 to be activated is represented. If anobject 6 has been activated with the aid of input device 4, then in a smooth transition,virtual ring section 5 with theobjects 6 situated on it is faded out and only activatedobject 6 is still represented. Whenvirtual ring section 5 andobjects 6 situated upon it are faded out until solely the activatedobject 6 is shown, activatedobject 6 displaces theother objects 6 ofring section 5 and is enlarged continuously until only activatedobject 6 is still displayed. The graphic data for this smooth transition are calculated by theuser interface device 2. When producing the intermediate images,user interface device 2 usessystem time 8 in order to ensure that the animation is reproduced smoothly without jerky movements. -
FIG. 3 for example shows a view ofdisplay 1, which is shown when activating theobject 6 associated with the air conditioning. In this case, the representation ofobject 6 for the air conditioning shown inFIG. 2 is magnified.Buttons 9 through 14 are preserved in the process. If the user actuatesbutton 9 from this view shown inFIG. 3 , then the representation ofobject 6 for the air conditioning is continuously scaled down, and theother objects 6 ofvirtual ring section 5 appear again, until the view shown inFIG. 2 is displayed. - If, on the other hand,
button object 6 for the air conditioning is similarly scaled down and theother objects 6 ofvirtual ring section 5 are displayed again as shown inFIG. 2 . Subsequently, however,virtual ring section 5 is rotated clockwise or counterclockwise in carousel-like fashion to the next position, whereupon thenext object 6 now positioned in the foreground is activated and in a smooth transition is shown enlarged and as displacing theother objects 6, until only thisnext object 6 is displayed. -
FIG. 4 shows object 6 associated with the settings of the graphics system itself. When activating thisobject 6, various settings may be adjusted for the graphic display using input device 4. - The list entries, to which
graphic objects 6 belong, are part of a hierarchical structure, as described above, since the activation of a list entry or agraphic object 6 results in the display of more detailed information. Furthermore, the activation of agraphic object 6 or a list entry may result e.g. in the playback of an audio file. In this case too, the playback of the audio file corresponds to a lower level of the hierarchy. When displaying such a hierarchically organized data set, the transition between two static representations ondisplay 1 is of particular importance. Particularly if the display device is installed in a motor vehicle, it is very important that the driver of the motor vehicle in particular is able to detect the information represented ondisplay 1 quickly, simply and intuitively, even if he directs his eyes ontodisplay 1 only for brief intervals. For this purpose it is particularly important that the viewer is at all times able to orient himself well within the hierarchical structure. It has been shown that abrupt changes of displayed images are disadvantageous for this purpose. It has been shown in particular that smooth transitions between two static representations, in which intermediate images are displayed, which emphasize the transition between the two static representations, facilitate orientation within the structure as well as the grasp of the informational content. - Furthermore, it is particularly important how the intermediate images visualize the transition between two static representations. In this regard, a linear transition from one static representation to the next static representation is not suitable for use in a motor vehicle. Rather, intermediate images are advantageous, which illustrate a movement of
objects 6 that is at least partially slowed down. The parameters for the choice of the slowdown and the duration of the transition must also be adapted for use in the vehicle because conventional animations known from the field of computers are usually not usable. - The following describes in detail how the user is able to rotate the
objects 6 situated onvirtual ring section 5 in carousel-like fashion on thisvirtual ring section 5 with the aid of input device 4. - The starting point of the rotation is the arrangement of
objects 6 onvirtual ring section 5 shown inFIG. 2 .Objects 6 are thus represented on a ring-shaped one-dimensional line. By a first operator action via input device 4, the user now specifies the total rotational angle αRotation, by which objects 6 are to be rotated on the ring as a whole, of which only aring section 5 is shown. For this purpose, the user enters a line ondisplay 1 using input device 4 for example. If input device 4 includes a touch-sensitive film, the user may run his finger over the film in order to specify this line. This line for entering the first operator action runs in particular substantially on the line on which objects 6 are represented, or parallel to this line. If input device 4 is able to detect a gesture of the user or a movement of his hand or his finger in front ofdisplay 1, the line is entered in this manner. If the line is not oriented in the direction of the width ofdisplay 1, but rather runs aslant, preferably only the length of the horizontal component of the entered line oriented in parallel to the width ofdisplay 1 is used, as long as the deviation from the horizontal does not exceed 45°. The total rotational angle αRotation is calculated from the ratio of the length of this horizontal component of the line to the total width ofdisplay 1. In fact, in the perspective view ondisplay 1, objects 6 are shifted only in the plane ofdisplay 1. Rotational angle α is thus transformed accordingly byuser interface device 2 when producing the graphic data. - If input device 4 is a separate operating element having a rotary encoder, then the total rotational angle may also be input via the rotational angle of the rotary encoder.
- Since only a portion of the list is shown on
display 1, the shown portion changes as the virtual ring is rotated. In the rotation, objects 6 disappear on the one side ofring section 5, whilenew objects 6 are displayed on the other side ofring section 5, which belong to list entries that were originally not displayed. In this manner, the first operator action may be used to run through the entries of the list consecutively, the entries of the list being shown on the basis ofgraphic objects 6. - After the total rotational angle αRotation has been entered by the user with the aid of input device 4, a
computing unit 3 ofuser interface device 2 produces an animation, which includes the graphic data for successive intermediate images, which represent successive rotational angles α forobjects 6. The transition from the initial state having an initial angle of the objects to a final state, in which the objects have been rotated by total angle αRotation, occurs by a decelerated movement, which is calculated by the following steps: -
- In this instance, t is the system time, which
computing unit 3 retrieves fromsystem clock 8. At the start of the rotation, the system time is zero. In the final state of the rotation, system time t corresponds to the duration d of the total rotation about total angle=angle×rotation. Steps (i) and (ii) are repeated until system time t is greater than the duration d of the total rotation. -
FIG. 5 shows the time characteristic of rotational angle α for the intermediate images if anobject 6 is rotated by a total rotational angle αRotation of 30° in 2.5 seconds. It can be seen that the rotational speed is initially relatively fast and is then slowed down. - In the present exemplary embodiment, the total rotational angle αRotation is additionally adapted to a rotational angle, which corresponds to a carousel-like rotation of
objects 6 onvirtual ring section 5, at which in the final state anobject 6 appears in the foreground, as shown inFIG. 2 . The rotational angle input by the user via input device 4 is thus rounded up or down to these discrete rotational angles. - If an object is rotated into the foreground by the user via input device 4, this
object 6 is automatically activated after the expiration of a time interval. Following the expiration of the time interval, thisobject 6 is thus magnified, while at the same time thevirtual ring section 5 with theother objects 6 vanishes until only the activatedobject 6 is still displayed. - The data of the speedometer of the motor vehicle may be transmitted to
user interface device 2 and thus computingunit 3 viavehicle bus 7. These data are used by computingunit 3 to ascertain the duration of the rotation. The higher the speed of the motor vehicle, the longer is the duration of the total rotation. - Furthermore, a second operator action may be performed in the method described herein or the device described herein. In this operator action, the user performs an operator movement that differs from the operator movement of the first operator action. In the present case, a vertical operator movement upward or downward is performed. For example, the user may move the finger upward or downward on the touch-sensitive surface. The second operator movement in particular starts with a touch of the touch-sensitive surface at a position at which an
object 6, i.e. an entry of the portion of the list, is displayed. Thus, for the second operator action—in contrast to conventional systems—it is not necessary to display e.g. a separate index in order to skip list entries. It suffices to perform an operator movement that has a direction of movement that is clearly distinguished from the direction of movement of the first operator movement. Furthermore, the user may perform a corresponding gesture in front ofdisplay area 1. Finally, it is possible for the user to actuate corresponding mechanical operating elements for controlling the movement of anobject 6 in a corresponding direction. - The second operator action has the consequence that
user interface device 2 changes the graphic data in such a way that multiple list entries are skipped in the display of the portion of the list so as to reach a list entry, which, or the associatedgraphic object 6 of which, is not contained in the displayed portion of the list entries that were displayed while performing the second operator action. Following the execution of the second operator action, agraphic object 6 is thus displayed in the foreground, which was previously not displayed. The total ring with the list entries was transported directly into a rotated state, in which another portion of the list is shown, without it having been necessary to run through the individual list entries consecutively. - The second operator action is explained in detail in the following with reference to
FIGS. 6 through 8 : - The starting point is the display of a portion of a list on the display area of
display 1, as it is represented inFIG. 6 . This informational display substantially corresponds to the informational display represented inFIG. 2 of the list entries on the basis ofgraphic objects 6. In the state shown inFIG. 6 , three list entries G22, G23, and G24 are shown by way of example. - The entries of this list are divided into classes, the individual classes in turn forming a list. In the present case, this may be an alphabetic sorting of concepts associated with the list entries for example. In this case, list entries G22, G23 and G24 are the entries no. 22 through 24 beginning with the letter G. The list entries having other initial letters are sorted accordingly.
- By a first operator action, as explained above, the user is now able to bring about a change in the displayed list entries by an operator movement in the direction of double-arrow H. If he moves the finger on the touch-sensitive surface to the left for example, the displayed portion of the list will move to the left as well such that list entries on the left side disappear, while new list entries emerge on the right side. Accordingly, when performing an operator movement to the right, list entries on the right side disappear, while new list entries emerge on the left side.
- In the second operator action, the user may perform a movement upward in the direction of arrow A or downward in the direction of arrow B. An operator action upward in the direction of arrow A has the consequence that the displayed portion of the list entries is changed immediately such that multiple list entries are skipped and the first list entry of the next initial letter is reached, i.e. as shown in
FIG. 7 , list entry F1. Furthermore, the displayed portion of the list may change such that the last list entry or the last entries beginning with the letter G is or are not included in the displayed portion. - If the second operator action by contrast includes an operator movement downward in the direction of arrow B, the first list entry beginning with the next lower initial letter is reached directly, i.e. as shown in
FIG. 8 , list entry H1. A portion of the list is thus displayed, which begins with the list entry H1 and thereupon shows additional list entries beginning with the letter H. -
User interface device 2 may furthermore be coupled to anacceleration sensor 16, which, alternatively or additionally, may be used as an input device.Acceleration sensor 16 is able to detect an acceleration of the device, which in this case is arranged in particular as a mobile device, as well as the direction of acceleration. A movement of the device, includingdisplay area 1, in space corresponds in this case to the previously explained movement on the touch-sensitive surface ofdisplay area 1. In this case, a movement to the left makes it possible to scroll through the list in one direction for example, while a movement to the right makes it possible to scroll through the list in the other direction. An upward movement results in multiple list entries being skipped forward in the list, while a movement downward results in multiple list entries being skipped backward, as was explained above. - 1 display
- 2 user interface device
- 3 computing unit
- 4 input device
- 5 virtual ring section
- 6 graphic objects
- 7 vehicle bus
- 8 system clock
- 9-14 buttons of an
object 6 - 15 memory
- 16 acceleration sensor
Claims (12)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/716,109 US20180018074A1 (en) | 2008-10-21 | 2017-09-26 | Method and device for displaying information arranged in lists |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE102008052485A DE102008052485A1 (en) | 2008-10-21 | 2008-10-21 | Method and device for displaying list-ordered information |
DE102008052485.9 | 2008-10-21 | ||
US13/125,175 US9804764B2 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information arranged in lists |
PCT/EP2009/060249 WO2010046147A1 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information sorted into lists |
US15/716,109 US20180018074A1 (en) | 2008-10-21 | 2017-09-26 | Method and device for displaying information arranged in lists |
Related Parent Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/125,175 Continuation US9804764B2 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information arranged in lists |
PCT/EP2009/060249 Continuation WO2010046147A1 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information sorted into lists |
Publications (1)
Publication Number | Publication Date |
---|---|
US20180018074A1 true US20180018074A1 (en) | 2018-01-18 |
Family
ID=41510831
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/125,175 Active 2030-10-24 US9804764B2 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information arranged in lists |
US15/716,109 Abandoned US20180018074A1 (en) | 2008-10-21 | 2017-09-26 | Method and device for displaying information arranged in lists |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/125,175 Active 2030-10-24 US9804764B2 (en) | 2008-10-21 | 2009-08-06 | Method and device for displaying information arranged in lists |
Country Status (8)
Country | Link |
---|---|
US (2) | US9804764B2 (en) |
EP (1) | EP2350799B1 (en) |
KR (3) | KR20160084504A (en) |
CN (2) | CN102203710A (en) |
DE (1) | DE102008052485A1 (en) |
MX (1) | MX2011004124A (en) |
RU (1) | RU2477511C2 (en) |
WO (1) | WO2010046147A1 (en) |
Families Citing this family (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB9722766D0 (en) | 1997-10-28 | 1997-12-24 | British Telecomm | Portable computers |
US7469381B2 (en) | 2007-01-07 | 2008-12-23 | Apple Inc. | List scrolling and document translation, scaling, and rotation on a touch-screen display |
US7487444B2 (en) | 2002-03-19 | 2009-02-03 | Aol Llc | Reformatting columns of content for display |
US7844915B2 (en) | 2007-01-07 | 2010-11-30 | Apple Inc. | Application programming interfaces for scrolling operations |
US9767438B2 (en) * | 2009-07-14 | 2017-09-19 | Bitvore Corp. | Presentation and sorting of email |
US8972467B2 (en) | 2010-08-31 | 2015-03-03 | Sovanta Ag | Method for selecting a data set from a plurality of data sets by means of an input device |
US8767019B2 (en) | 2010-08-31 | 2014-07-01 | Sovanta Ag | Computer-implemented method for specifying a processing operation |
DE102011011802A1 (en) | 2011-02-19 | 2012-08-23 | Volkswagen Ag | Method and device for providing a user interface, in particular in a vehicle |
DE102011112448A1 (en) * | 2011-09-03 | 2013-03-07 | Volkswagen Aktiengesellschaft | Method and device for providing a user interface, in particular in a vehicle |
DE102011116141A1 (en) | 2011-10-15 | 2013-04-18 | Volkswagen Aktiengesellschaft | A method of displaying information in a vehicle and display device for a vehicle |
US9645733B2 (en) | 2011-12-06 | 2017-05-09 | Google Inc. | Mechanism for switching between document viewing windows |
USD703687S1 (en) | 2011-12-28 | 2014-04-29 | Target Brands, Inc. | Display screen with graphical user interface |
USD711400S1 (en) | 2011-12-28 | 2014-08-19 | Target Brands, Inc. | Display screen with graphical user interface |
USD715818S1 (en) | 2011-12-28 | 2014-10-21 | Target Brands, Inc. | Display screen with graphical user interface |
USD705790S1 (en) | 2011-12-28 | 2014-05-27 | Target Brands, Inc. | Display screen with graphical user interface |
USD703685S1 (en) | 2011-12-28 | 2014-04-29 | Target Brands, Inc. | Display screen with graphical user interface |
USD705792S1 (en) * | 2011-12-28 | 2014-05-27 | Target Brands, Inc. | Display screen with graphical user interface |
USD706793S1 (en) | 2011-12-28 | 2014-06-10 | Target Brands, Inc. | Display screen with graphical user interface |
USD703686S1 (en) | 2011-12-28 | 2014-04-29 | Target Brands, Inc. | Display screen with graphical user interface |
USD706794S1 (en) | 2011-12-28 | 2014-06-10 | Target Brands, Inc. | Display screen with graphical user interface |
USD711399S1 (en) | 2011-12-28 | 2014-08-19 | Target Brands, Inc. | Display screen with graphical user interface |
JP5799861B2 (en) * | 2012-03-09 | 2015-10-28 | コニカミノルタ株式会社 | Function setting device and control program for function setting device |
DE102012008778A1 (en) | 2012-04-28 | 2013-10-31 | Volkswagen Aktiengesellschaft | Method and device for visualizing a hierarchically sortable data stock, in particular in a vehicle |
US9696879B2 (en) * | 2012-09-07 | 2017-07-04 | Google Inc. | Tab scrubbing using navigation gestures |
FR2995836B1 (en) * | 2012-09-27 | 2015-05-22 | Valeo Systemes Thermiques | CONTROL MODULE |
DE102013000880A1 (en) | 2013-01-10 | 2014-07-10 | Volkswagen Aktiengesellschaft | Method and apparatus for providing a user interface in a vehicle |
JP5975928B2 (en) * | 2013-04-05 | 2016-08-23 | カルソニックカンセイ株式会社 | Vehicle display device |
DE102013007576A1 (en) | 2013-05-02 | 2014-11-06 | Volkswagen Aktiengesellschaft | Method and apparatus for selecting an object of a list |
DE102013013225B4 (en) * | 2013-08-08 | 2019-08-29 | Audi Ag | Motor vehicle with switchable operating device |
US10766366B2 (en) | 2013-08-20 | 2020-09-08 | Volkswagen Ag | Operating method for an operating and display device in a vehicle and operating and display device in a vehicle |
DE102014208502A1 (en) * | 2014-05-07 | 2015-11-12 | Volkswagen Aktiengesellschaft | User interface and method for switching between screen views of a user interface |
US10380669B2 (en) * | 2014-10-30 | 2019-08-13 | Walmart Apollo, Llc | Product browsing system and method |
US9984403B2 (en) | 2014-10-30 | 2018-05-29 | Wal-Mart Stores, Inc. | Electronic shopping cart processing system and method |
DE102014018116A1 (en) * | 2014-12-09 | 2016-06-09 | Daimler Ag | Information entertainment system for a motor vehicle and method of operating such |
USD776131S1 (en) * | 2015-03-30 | 2017-01-10 | Intuit Inc. | Display device with transfer animation for a user interface |
USD764495S1 (en) * | 2015-03-30 | 2016-08-23 | Intuit Inc. | Display device with initiating transfer animation for a user interface |
CN109040413A (en) * | 2017-06-12 | 2018-12-18 | 阿里巴巴集团控股有限公司 | Display methods, the device and system of data |
CN109426475A (en) * | 2017-08-31 | 2019-03-05 | 中兴通讯股份有限公司 | Screen display method and device |
CN108198272B (en) * | 2017-12-29 | 2020-11-20 | 深圳市元征科技股份有限公司 | Data processing method and equipment thereof |
US11416917B2 (en) * | 2020-08-04 | 2022-08-16 | Contextlogic, Inc. | Content carousel |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060026521A1 (en) * | 2004-07-30 | 2006-02-02 | Apple Computer, Inc. | Gestures for touch sensitive input devices |
US20130254308A1 (en) * | 2010-04-29 | 2013-09-26 | British Broadcasting Corporation | Content provision system |
Family Cites Families (28)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6043809A (en) * | 1997-09-23 | 2000-03-28 | Compaq Computer Corporation | Computer keyboard scroll bar control |
JPH11187443A (en) * | 1997-12-25 | 1999-07-09 | Sony Corp | Portable radio information terminal equipment, screen operation method, record medium, and microcomputer |
US7469381B2 (en) * | 2007-01-07 | 2008-12-23 | Apple Inc. | List scrolling and document translation, scaling, and rotation on a touch-screen display |
DE19944324A1 (en) | 1999-09-15 | 2001-03-22 | Audi Ag | Multi-function control device |
DE10058244C2 (en) | 2000-11-19 | 2003-02-06 | Hertz Inst Heinrich | Measuring method for determining the position of an object in front of a screen and device for carrying out the method |
JP4701564B2 (en) | 2001-08-31 | 2011-06-15 | ソニー株式会社 | Menu display device and menu display method |
US7837562B2 (en) * | 2002-06-20 | 2010-11-23 | Igt | Display panel for a gaming apparatus |
DE10305341A1 (en) | 2003-02-10 | 2004-08-19 | Ident Technology Ag | Switching signal provision procedure e.g. for operating motor vehicle functional units, requires setting up signal for coupling into user within framework of actuating a switching device |
JP2008135075A (en) * | 2002-08-28 | 2008-06-12 | Sony Corp | Electronic apparatus, spatial displacement computing device, and spatial displacement computation method |
JP4482269B2 (en) * | 2002-08-28 | 2010-06-16 | ソニー株式会社 | Electronic device apparatus, signal compensation apparatus, and signal compensation method |
SE0202664L (en) | 2002-09-09 | 2003-11-04 | Zenterio Ab | Graphical user interface for navigation and selection from various selectable options presented on a monitor |
BRPI0410134A (en) | 2003-05-08 | 2006-05-16 | Thomson Licensing | method and apparatus for alphabetic text navigation |
RU2336553C2 (en) * | 2003-08-21 | 2008-10-20 | Майкрософт Корпорейшн | System and method for support of applications that are minimised with expanded set of functions |
DE10341580A1 (en) * | 2003-09-09 | 2005-03-31 | Siemens Ag | Input device for a data processing system |
US8028250B2 (en) * | 2004-08-31 | 2011-09-27 | Microsoft Corporation | User interface having a carousel view for representing structured data |
US7728823B2 (en) | 2004-09-24 | 2010-06-01 | Apple Inc. | System and method for processing raw data of track pad device |
DE102004048956A1 (en) | 2004-10-07 | 2006-04-27 | Ident Technology Ag | Signal transferring method using human body, integrates condenser device that is used as signal interface in oscillating circuit, where circuit is operated as parallel or serial oscillating circuit in receive or transmit mode, respectively |
US7810043B2 (en) * | 2005-07-27 | 2010-10-05 | Microsoft Corporation | Media user interface left/right navigation |
US20070028268A1 (en) | 2005-07-27 | 2007-02-01 | Microsoft Corporation | Media user interface start menu |
US8769408B2 (en) | 2005-10-07 | 2014-07-01 | Apple Inc. | Intelligent media navigation |
US20070132789A1 (en) * | 2005-12-08 | 2007-06-14 | Bas Ording | List scrolling in response to moving contact over list of index symbols |
US20070157126A1 (en) | 2006-01-04 | 2007-07-05 | Tschirhart Michael D | Three-dimensional display and control image |
US7269002B1 (en) * | 2006-03-29 | 2007-09-11 | Matthew Scott Turner | Vehicle recess integrated docking station |
KR101479769B1 (en) * | 2006-08-01 | 2015-01-08 | 엘지전자 주식회사 | Touch screen apparatus and file search method thereof |
US8091045B2 (en) * | 2007-01-07 | 2012-01-03 | Apple Inc. | System and method for managing lists |
KR20080068781A (en) * | 2007-01-20 | 2008-07-24 | 엘지전자 주식회사 | Electronic device with touch screen and method of displaying information using same |
US20100001960A1 (en) * | 2008-07-02 | 2010-01-07 | Sling Media, Inc. | Systems and methods for gestural interaction with user interface objects |
US20100175026A1 (en) * | 2009-01-05 | 2010-07-08 | Bortner Christopher F | System and method for graphical content and media management, sorting, and retrieval |
-
2008
- 2008-10-21 DE DE102008052485A patent/DE102008052485A1/en not_active Withdrawn
-
2009
- 2009-08-06 KR KR1020167018051A patent/KR20160084504A/en active Application Filing
- 2009-08-06 US US13/125,175 patent/US9804764B2/en active Active
- 2009-08-06 KR KR1020177019858A patent/KR20170086141A/en not_active Application Discontinuation
- 2009-08-06 CN CN2009801418553A patent/CN102203710A/en active Pending
- 2009-08-06 CN CN201510480013.8A patent/CN105034809B/en active Active
- 2009-08-06 MX MX2011004124A patent/MX2011004124A/en active IP Right Grant
- 2009-08-06 KR KR1020117009401A patent/KR101646293B1/en active IP Right Grant
- 2009-08-06 EP EP09781591.4A patent/EP2350799B1/en active Active
- 2009-08-06 WO PCT/EP2009/060249 patent/WO2010046147A1/en active Application Filing
- 2009-08-06 RU RU2011120464/08A patent/RU2477511C2/en not_active IP Right Cessation
-
2017
- 2017-09-26 US US15/716,109 patent/US20180018074A1/en not_active Abandoned
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20060026521A1 (en) * | 2004-07-30 | 2006-02-02 | Apple Computer, Inc. | Gestures for touch sensitive input devices |
US20130254308A1 (en) * | 2010-04-29 | 2013-09-26 | British Broadcasting Corporation | Content provision system |
Also Published As
Publication number | Publication date |
---|---|
US9804764B2 (en) | 2017-10-31 |
WO2010046147A1 (en) | 2010-04-29 |
CN102203710A (en) | 2011-09-28 |
DE102008052485A1 (en) | 2010-04-22 |
RU2477511C2 (en) | 2013-03-10 |
CN105034809B (en) | 2018-04-06 |
KR20170086141A (en) | 2017-07-25 |
US20110265036A1 (en) | 2011-10-27 |
KR20160084504A (en) | 2016-07-13 |
KR101646293B1 (en) | 2016-08-05 |
CN105034809A (en) | 2015-11-11 |
MX2011004124A (en) | 2011-05-19 |
KR20110084197A (en) | 2011-07-21 |
EP2350799A1 (en) | 2011-08-03 |
EP2350799B1 (en) | 2017-11-01 |
RU2011120464A (en) | 2012-12-20 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20180018074A1 (en) | Method and device for displaying information arranged in lists | |
US8477108B2 (en) | Method for displaying information in a vehicle and display device for a vehicle | |
US10449857B2 (en) | Information reproduction system for a vehicle and method for providing information for the user of a vehicle | |
US9758150B2 (en) | Method and device for displaying information | |
US9013470B2 (en) | Method and device for displaying information, particularly in a vehicle | |
US9878618B2 (en) | Information playback system and method for information playback | |
US8677271B2 (en) | Method for displaying information in a motor vehicle and display device for a motor vehicle | |
US9802484B2 (en) | Method and display device for transitioning display information | |
KR101609024B1 (en) | Method and device for providing a user interface | |
US9019311B2 (en) | Method for displaying a two-sided two-dimensional object on a display in a motor vehicle and display device for a motor vehicle | |
US20130057400A1 (en) | Method and device for displaying information in a vehicle | |
US20180307405A1 (en) | Contextual vehicle user interface | |
US10139988B2 (en) | Method and device for displaying information arranged in lists | |
CN105677163A (en) | Concentrated control system for vehicle | |
US9530182B2 (en) | Method for displaying information in particular in a vehicle and display unit for a vehicle | |
US11014449B2 (en) | Method and device for displaying information, in particular in a vehicle | |
US11099715B2 (en) | Method and device for providing a user interface in a vehicle | |
US20180232115A1 (en) | In-vehicle input device and in-vehicle input device control method | |
WO2020196561A1 (en) | Operation device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: VOLKSWAGEN AG, GERMANY Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOEHNE, SVEN;REEL/FRAME:043705/0666 Effective date: 20110627 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STCV | Information on status: appeal procedure |
Free format text: NOTICE OF APPEAL FILED |
|
STCV | Information on status: appeal procedure |
Free format text: EXAMINER'S ANSWER TO APPEAL BRIEF MAILED |
|
STCV | Information on status: appeal procedure |
Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS |
|
STCV | Information on status: appeal procedure |
Free format text: BOARD OF APPEALS DECISION RENDERED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |