US20180018074A1 - Method and device for displaying information arranged in lists - Google Patents

Method and device for displaying information arranged in lists Download PDF

Info

Publication number
US20180018074A1
US20180018074A1 US15/716,109 US201715716109A US2018018074A1 US 20180018074 A1 US20180018074 A1 US 20180018074A1 US 201715716109 A US201715716109 A US 201715716109A US 2018018074 A1 US2018018074 A1 US 2018018074A1
Authority
US
United States
Prior art keywords
list
operator action
displayed
entries
movement
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/716,109
Inventor
Sven Hoehne
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Volkswagen AG
Original Assignee
Volkswagen AG
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Volkswagen AG filed Critical Volkswagen AG
Priority to US15/716,109 priority Critical patent/US20180018074A1/en
Assigned to VOLKSWAGEN AG reassignment VOLKSWAGEN AG ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HOEHNE, SVEN
Publication of US20180018074A1 publication Critical patent/US20180018074A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K35/00Arrangement of adaptations of instruments
    • B60K35/10
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60KARRANGEMENT OR MOUNTING OF PROPULSION UNITS OR OF TRANSMISSIONS IN VEHICLES; ARRANGEMENT OR MOUNTING OF PLURAL DIVERSE PRIME-MOVERS IN VEHICLES; AUXILIARY DRIVES FOR VEHICLES; INSTRUMENTATION OR DASHBOARDS FOR VEHICLES; ARRANGEMENTS IN CONNECTION WITH COOLING, AIR INTAKE, GAS EXHAUST OR FUEL SUPPLY OF PROPULSION UNITS IN VEHICLES
    • B60K37/00Dashboards
    • B60K37/04Arrangement of fittings on dashboard
    • B60K37/06Arrangement of fittings on dashboard of controls, e.g. controls knobs
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/0482Interaction with lists of selectable items, e.g. menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • B60K2350/1024
    • B60K2350/1028
    • B60K2360/143
    • B60K2360/1438

Definitions

  • the present invention relates to a method for displaying information, in which a portion of a list is displayed on a display area, which includes a plurality of list entries.
  • the displayed portion of the list is changed by a first operator action of a user, the first operator action running consecutively through the entries of the list.
  • the present invention furthermore relates to a device for displaying information including a display device having a display area, a memory for storing a list, which includes a plurality of list entries, of which a portion is displayable on the display area, and an operating device, which is operable by a user, in order to change the displayed portion of the list by a first operator action, the first operator action running consecutively through the entries of the list.
  • Mobile devices in consumer electronics such as e.g. devices for playing back audio and/or video files have memories, in which very large amounts of data may be stored.
  • Mobile music players for example, are able to store several thousands of musical pieces.
  • the mobile devices there arises the problem of how the user is able to access these large amounts of data in a simple manner.
  • the mobile devices often have a display and an operating device.
  • the operating device may include a touch-sensitive surface of the display or may be an operating element that is separate from the display.
  • the data sets stored in the device may be stored at least in part as a list, of which a portion may be displayed on the display. With the aid of the operating device, the user is able to scroll through the list in order to reach the desired list entry.
  • the display of information in a vehicle also entails the problem that very large data sets must be visualized in the vehicle.
  • Modern vehicles include for example a plurality of driver assistance systems, the information of which must be displayed in the vehicle.
  • vehicles often include a navigation system.
  • Such a navigation system is able to display digital geographic road maps including a route and possibly various additional information.
  • modern vehicles often include communication and multimedia applications, including a mobile telephone interface and devices for playing back music and voice. For these applications it must also be possible to display information in the vehicle.
  • the display of information in a vehicle and the operation of the manifold devices of the vehicle entail very specific requirements. In the vehicle, it is, among others, the driver who takes in the information and operates the respective device.
  • the information should thus be represented in the vehicle in such a way that while taking in the information the driver is not distracted from driving.
  • the represented information should therefore be ascertainable intuitively and quickly by the driver such that he has to take his eyes off the driving action only very briefly when taking in the information.
  • the operation of the vehicle devices should be as simple and intuitive as possible such that the driver is able to operate the devices even while driving. If the operation is supported or guided by a display, then the display should occur in such a way that the driver has to view the display only very briefly in order to carry out the operation.
  • the representation of information in a vehicle often occurs in that a portion of the total information is displayed on the display area in the vehicle and that the user is able to change the portion.
  • the total information may be a list having a plurality of list entries for example. A portion of the entries of the entire list is displayed.
  • the user is able to scroll so as to display other list entries.
  • the user may perform for example an operator movement of a certain extent.
  • the user may furthermore actuate an operating element for a certain operating duration.
  • the quantity or the measure of the change of the displayed information i.e. in this case the number of items, by which the displayed list entries change, depends in this case on the extent of the operator movement or on the operating duration.
  • Example embodiments of the present invention provide a method and a device of the type mentioned at the outset, by which the displayed portion of the list may be changed in a manner that is simple and intuitive for the user such that the operator action required for this purpose may be performed as quickly as possible. Furthermore, as little attention as possible is to be exacted from the user.
  • a list should be understood as a data set that is divisible into a plurality of data units.
  • the data units then correspond to the list entries.
  • the data units may be brought into a certain sequence such that they then represent a list.
  • the user is able to run through the entries of the list consecutively.
  • a portion of the list i.e. multiple list entries, may be represented on the display area on the basis of a graphic object.
  • the list may include a plurality of audio or video files for example. Furthermore, it may pertain to entries of a telephone or address book, entries of a menu of a hierarchical menu structure for operating a device or other data units, which may be integrated in a list and represented abstractly on the basis of a graphic object.
  • the list may be scanned much more quickly in order to reach a desired list entry since it is not necessary to run through the entire list entry by entry before reaching the desired entry, as the second operator action rather makes it possible to skip multiple entries in order to reach the desired entry more quickly.
  • the first operator action may include steering a movement of a list entry in a first direction and the second operator action may include steering a movement of a list entry in a second direction, which differs from the first direction.
  • the second direction runs in particular substantially perpendicular to the first direction.
  • a list entry may be moved by the operator action in the horizontal direction in order to scroll through the individual list entries.
  • a vertical movement of the list entry has the result that multiple list entries are skipped so as to reach a list entry that was not yet displayed when the respective list entry was moved in the vertical direction.
  • any other directional combinations are possible for the two operator actions as long as the two directions for the operator actions differ.
  • the entries of the list may be divided into classes.
  • the second operator action takes one to a list entry of another class.
  • the classes in particular also form a list, the individual classes including sorted list entries.
  • the second operator action preferably takes one to the first list entry of the next or previous class.
  • An example of such a class division is an alphabetical sorting of list entries.
  • a class in this case includes all list entries beginning with a particular letter.
  • the first operator action is in particular used to scroll through the list.
  • the second operator action may be used to abbreviate the scrolling by skipping multiple list entries.
  • At least one operator action may include a movement on touch-sensitive surface of an input device. Many other operator actions are possible, however, which substantially depend on the type of input device. If the input device allows for so-called gesture detection, then the operator action may include a gesture performed by the hand of a user. If the input device is a mechanical operating element, then the operator action is to actuate this operating element, e.g. to turn it or to press on it for a certain time.
  • the touch-sensitive surface of the input device may be provided on the display area.
  • the entries of the portion of the list may be displayed in a one-dimensional line on the display area.
  • the first operator action in this case may include a movement on the touch-sensitive surface, which runs substantially in the direction of the one-dimensional line.
  • the second operator action may include a movement on the touch-sensitive surface, which runs substantially in a direction perpendicular to the one-dimensional line.
  • the movements may run exactly in the direction of the one-dimensional line or perpendicular to it. They may also run in an angular range of e.g. ⁇ 5 degrees or ⁇ 10 degrees around these directions.
  • the second operator action may furthermore be a movement, which starts with a touch of the touch-sensitive surface at a position, at which an entry of the portion of the list is displayed.
  • At least one operator action may include a movement of the display area in space, and the acceleration of the display area in space may be detected.
  • the user moves the device, which includes the display area, e.g. in a particular direction.
  • This movement includes an acceleration of the display area, which is detected.
  • the direction in which a list is to be run through is derivable from the direction of movement.
  • the movement in a different direction may be interpreted as a second operator action, in which multiple list entries are skipped.
  • the displayed list entries may be represented as arranged on a virtual ring or ring section represented in perspective.
  • the first operator action rotates the displayed list entries on the virtual ring or ring section in carousel-like fashion.
  • Rotating should be understood as rotating the virtual ring or ring section. Since the three-dimensional virtual ring on the display is represented in a two-dimensional perspective view, the objects are actually shifted with respect to their position on the display.
  • the user is able to actuate the operating device by a second operator action, by which multiple list entries are skipped in order to arrive at a list entry that is not contained in the displayed portion of list entries that were displayed while performing the second operator action.
  • the device may be able to implement the method steps described above partially or entirely.
  • the operating device of the display device may include a user interface device and an input device.
  • the user interface device As a function of an input via the input device, the user interface device is able to produce graphic data, which control the display device such that a particular display is represented on the display area.
  • the user interface device may include a computing unit, which produces graphic data for intermediate images for the animation. It is also possible, however, for these intermediate images to be already stored in advance in the memory.
  • the first or second operator action which is implemented by the operating device of the device into a change of the display on the display area, includes in particular the operator actions described above with reference to the method.
  • the input device may include a touch-sensitive surface. This surface may be part of a so-called touch pad. Furthermore, the surface may be provided on the display area, i.e. the display device is equipped with a touch screen.
  • the input device may be a device for detecting and evaluating a gesture of a part of the body of a user, which is performed in front of the display area.
  • the input device may include in particular a receiving device, to which a signal from the body part of the user may be transmitted capacitively when the body part is in proximity of the receiving device. The position of the body part may be detected via this capacitive coupling. From the change of this position over time, a gesture of the user may be inferred.
  • the input device may include an infrared light source or a receiver for reflected infrared light for detecting the gesture of the body part of the user. In this case as well, the position of the body part and its change over time are detected and interpreted as a gesture.
  • an approach sensor may be provided, which detects the operating intention of a user on the basis of the approach of a part of the body of the user.
  • the device may include an acceleration sensor, which is able to detect the acceleration of the display area in space.
  • the operating device is able to interpret the data detection of the acceleration sensor so as to assign it to the first or to the second operator action.
  • FIG. 1 schematically shows display device according to an example embodiment of the present invention and the coupling of this display device to the electronic system of a motor vehicle.
  • FIG. 2 shows a display of information produced by the method according to an example embodiment of the present invention in a motor vehicle, in which multiple objects are represented.
  • FIG. 3 shows a display of information of an activated object produced by the method according to an example embodiment of the present invention.
  • FIG. 4 shows a display of information of another object produced by the method according to an example embodiment of the present invention.
  • FIG. 5 shows the time characteristic of the rotational angle of an object when rotating on the virtual ring.
  • FIGS. 6 through 8 show the change of the displayed list entries in a second operator action.
  • An exemplary embodiment described below relates to a display device and a method for displaying information in a vehicle, in particular a motor vehicle.
  • Example embodiments of the present invention may be utilized in the same manner in a mobile or stationary consumer electronics device, in particular a device for playing back audio and video files.
  • the display device includes a display 1 for the graphical representation of information on a display area.
  • Display 1 may be a matrix display, e.g. an LCD (liquid crystal display), especially a color display using TFT (thin-film transistor) technology.
  • the display may be a so-called twisted nematic-liquid crystal display (TN-LCD), a super twisted nematic (STN) display, a double-layer STN, an FLC (ferroelectric liquid crystal) display or an SSFLC (surface stabilized ferroelectric liquid crystal) display.
  • Display 1 has an associated back-lighting (not shown), which may be provided by one or more light-emitting diodes.
  • Display 1 is freely programmable, i.e. any desired graphic data may be produced, which are represented on display 1 .
  • display 1 is mounted in an area of the vehicle that is clearly visible at least for the driver. If the operation of the devices of the vehicle is coupled directly to the position of display 1 such that the user must bring e.g. his hand or his finger at least near display 1 in order to make inputs, display 1 is positioned in such a way that the driver of the vehicle may reach it readily with his hand or his finger. Display 1 may be accommodated in the center console of the vehicle for example.
  • Display 1 is connected to a user interface device 2 , which is able to produce graphic data for graphic objects 6 representable on display 1 . Furthermore, user interface device 2 is connected to an input device 4 , via which the user is able to control devices of the vehicle, the information of which is displayed on display 1 . User interface device 2 and input device 4 form an operating device, which may be actuated by a user and which implements the input of a user in graphic data for representation on display 1 .
  • User interface device 2 is furthermore connected to a memory 15 .
  • data for a list as well as the graphic objects 6 associated with the individual list entries may be stored in this memory.
  • Input device 4 may be for example a device for detecting and evaluating a gesture of a part of a user's body.
  • the hand of the user may perform the gesture in front of display 1 for example.
  • the three-dimensional position of the hand is detected in a specific detection area in front of display 1 , without it being necessary to touch display 1 .
  • the admissible detection area depends on the placement of display 1 in the motor vehicle. The area should be selected such that the presence of the hand of a user in this detection area may be associated unequivocally with an operative control of input device 4 .
  • the boundary of the detection area may be e.g. 40 cm to 10 cm in front of display 1 .
  • Input device 4 detects the position and the movement of the hand of the user in the detection area. In the process, various gestures performed by the hand are recognized and interpreted as inputs.
  • Input device 4 may include e.g. infrared light sources and infrared light receivers, which detect the infrared light reflected by the hand. Details of such an input device are described in German Patent Document No. 100 58 244 which is incorporated herein by reference. Further input devices, which may be used in conjunction with the display device, are described in the following publications: German Patent Document No. 103 05 341 and German Patent Document No. 10 2004 048 956.
  • the position of the hand and its change over time may also be detected by an optical system.
  • a light-emitting diode emits e.g. square-wave, amplitude-modulated light. This light is reflected by the object to be detected, i.e. the hand, and after the reflection reaches a photodiode.
  • Another light-emitting diode likewise emits square-wave, amplitude-modulated light to the photodiode, which light is phase-shifted by 180° however.
  • the two light signals are superposed and cancel each other out if they have exactly the same amplitude.
  • the light emission of the second diode is regulated via a control loop such that the total received signal again adds up to zero. If the position of the object changes, this also causes a change in the light component which arrives at the photodiode from the first light-emitting diode via the reflection on the object. This brings about a correction of the intensity of the second light-emitting diode through the control loop.
  • the control signal is therefore a measure for the reflection of the light, which is emitted by the first diode, on the object. In this manner, it is possible to derive from the control signal a signal that is characteristic for the position of the object.
  • the input device may be a touch-sensitive film, which is provided on display 1 .
  • the film makes it possible to detect the position at which the user touches display 1 situated behind the film.
  • the film may be arranged e.g. as a resistive touch film, a capacitive touch film or piezoelectric film.
  • the film may be arranged so as to measure a heat flow emanating e.g. from the finger of a user.
  • Various inputs may be obtained from the development of the touch of the film over time. In the simplest case, for example, a touch of the film at a specific position may be allocated to a graphic object shown on display 1 . Furthermore, sliding movements of the finger over the film may be interpreted. In particular, in this manner the user is able to define a line on display 1 , in that he touches the film at one point, slides on the film toward another point and removes the finger from the film at the other point.
  • a separate operating element may be used as input device.
  • the separate operating element is a mechanical operating element.
  • a rotary switch may be provided, by which objects shown on the display are controllable and are selectable by pressing the rotary switch.
  • an angle of rotation may also be input directly by the rotary switch, as will be explained later.
  • separate press switches may be disposed around the rotary switch, the arrangement of display fields on the display, which are associated with the press switches, corresponding at least schematically to the arrangement of the press switches.
  • the informational system may include a multifunction operating device as described in European Patent Document No. 1 212 208 for example.
  • User interface device 2 is further coupled to a system clock 8 and if applicable to a vehicle bus 7 .
  • User interface device 2 may be connected e.g. to driver assistance systems of the vehicle via vehicle bus 7 .
  • User interface device 2 receives data from these driver assistance systems via vehicle bus 7 , and processes these data so that they are displayed graphically to the driver or the vehicle occupants via display 1 .
  • user interface device 2 produces graphic data for objects displayable on display 1 , which, inter alia, graphically represent the information of the driver assistance systems.
  • User interface device 2 is furthermore connected via vehicle bus 7 to various information, communication and entertainment devices of the vehicle. The varied information from these devices of the vehicle is processed in user interface device 2 and converted into graphic data for a graphical representation.
  • the user interface device includes a computing unit 3 , which accesses system clock 8 in producing intermediate images.
  • the graphic data, which user interface device 2 produces for display 1 relate to a portion of a list. Multiple entries from this list are represented as objects arranged on a virtual ring section 5 shown in perspective.
  • the perspective representation shows virtual ring section 5 as though the observer were looking at an angle from above onto ring section 5 .
  • An object 6 in the foreground is shown as the largest in the perspective view. Further back, additional objects 6 are shown, partially hidden by objects 6 situated in front of them.
  • One or more devices of the vehicle are assigned to objects 6 or to the associated list entries. Objects 6 shown in FIG. 2 either give only an indication of the associated vehicle device or additionally already provide informational content of this device.
  • object 6 represented in the foreground refers to the control of the air conditioning of the interior of the vehicle.
  • the object adjacent to the right and behind this object 6 is assigned to the music playback device of the vehicle, and the object adjacent to the left of object 6 associated with the air-conditioning control is assigned to the navigation system of the vehicle.
  • An object situated further back relates to the settings of the display device itself.
  • buttons for controlling the graphic data produced by user interface device 2 for display 1 are buttons for controlling the graphic data produced by user interface device 2 for display 1 .
  • a first button 9 the user is able to return in any display state to the initial state, in which objects 6 are displayed as situated on the virtual ring section 5 shown in perspective.
  • buttons 10 and 11 objects 6 on virtual ring section 5 may be rotated in carousel-like fashion by a position clockwise or counterclockwise, as will be explained later in detail.
  • Buttons 12 and 13 may be used to actuate functions within an object 6
  • the user is able to access e.g. directly the settings menu for the graphics.
  • any of the objects 6 shown on virtual ring section 5 may be selected and activated using input device 4 .
  • the input device includes a touch-sensitive film, the user can tap e.g. using his finger on the object 6 to be activated.
  • input device 4 detects e.g. the position of the finger or the hand of the user in front of display 1 , it suffices for the user to bring his finger near the position of display 1 , at which the object 6 to be activated is represented. If an object 6 has been activated with the aid of input device 4 , then in a smooth transition, virtual ring section 5 with the objects 6 situated on it is faded out and only activated object 6 is still represented.
  • activated object 6 displaces the other objects 6 of ring section 5 and is enlarged continuously until only activated object 6 is still displayed.
  • the graphic data for this smooth transition are calculated by the user interface device 2 .
  • user interface device 2 uses system time 8 in order to ensure that the animation is reproduced smoothly without jerky movements.
  • FIG. 3 for example shows a view of display 1 , which is shown when activating the object 6 associated with the air conditioning.
  • the representation of object 6 for the air conditioning shown in FIG. 2 is magnified. Buttons 9 through 14 are preserved in the process. If the user actuates button 9 from this view shown in FIG. 3 , then the representation of object 6 for the air conditioning is continuously scaled down, and the other objects 6 of virtual ring section 5 appear again, until the view shown in FIG. 2 is displayed.
  • object 6 for the air conditioning is similarly scaled down and the other objects 6 of virtual ring section 5 are displayed again as shown in FIG. 2 . Subsequently, however, virtual ring section 5 is rotated clockwise or counterclockwise in carousel-like fashion to the next position, whereupon the next object 6 now positioned in the foreground is activated and in a smooth transition is shown enlarged and as displacing the other objects 6 , until only this next object 6 is displayed.
  • FIG. 4 shows object 6 associated with the settings of the graphics system itself. When activating this object 6 , various settings may be adjusted for the graphic display using input device 4 .
  • the list entries, to which graphic objects 6 belong, are part of a hierarchical structure, as described above, since the activation of a list entry or a graphic object 6 results in the display of more detailed information. Furthermore, the activation of a graphic object 6 or a list entry may result e.g. in the playback of an audio file. In this case too, the playback of the audio file corresponds to a lower level of the hierarchy. When displaying such a hierarchically organized data set, the transition between two static representations on display 1 is of particular importance.
  • the display device is installed in a motor vehicle, it is very important that the driver of the motor vehicle in particular is able to detect the information represented on display 1 quickly, simply and intuitively, even if he directs his eyes onto display 1 only for brief intervals. For this purpose it is particularly important that the viewer is at all times able to orient himself well within the hierarchical structure. It has been shown that abrupt changes of displayed images are disadvantageous for this purpose. It has been shown in particular that smooth transitions between two static representations, in which intermediate images are displayed, which emphasize the transition between the two static representations, facilitate orientation within the structure as well as the grasp of the informational content.
  • intermediate images visualize the transition between two static representations.
  • a linear transition from one static representation to the next static representation is not suitable for use in a motor vehicle.
  • intermediate images are advantageous, which illustrate a movement of objects 6 that is at least partially slowed down.
  • the parameters for the choice of the slowdown and the duration of the transition must also be adapted for use in the vehicle because conventional animations known from the field of computers are usually not usable.
  • the starting point of the rotation is the arrangement of objects 6 on virtual ring section 5 shown in FIG. 2 .
  • Objects 6 are thus represented on a ring-shaped one-dimensional line.
  • the user By a first operator action via input device 4 , the user now specifies the total rotational angle ⁇ Rotation , by which objects 6 are to be rotated on the ring as a whole, of which only a ring section 5 is shown.
  • the user enters a line on display 1 using input device 4 for example. If input device 4 includes a touch-sensitive film, the user may run his finger over the film in order to specify this line.
  • This line for entering the first operator action runs in particular substantially on the line on which objects 6 are represented, or parallel to this line.
  • the line is entered in this manner. If the line is not oriented in the direction of the width of display 1 , but rather runs aslant, preferably only the length of the horizontal component of the entered line oriented in parallel to the width of display 1 is used, as long as the deviation from the horizontal does not exceed 45°.
  • the total rotational angle ⁇ Rotation is calculated from the ratio of the length of this horizontal component of the line to the total width of display 1 . In fact, in the perspective view on display 1 , objects 6 are shifted only in the plane of display 1 . Rotational angle ⁇ is thus transformed accordingly by user interface device 2 when producing the graphic data.
  • the total rotational angle may also be input via the rotational angle of the rotary encoder.
  • the shown portion changes as the virtual ring is rotated.
  • objects 6 disappear on the one side of ring section 5 , while new objects 6 are displayed on the other side of ring section 5 , which belong to list entries that were originally not displayed.
  • the first operator action may be used to run through the entries of the list consecutively, the entries of the list being shown on the basis of graphic objects 6 .
  • a computing unit 3 of user interface device 2 produces an animation, which includes the graphic data for successive intermediate images, which represent successive rotational angles ⁇ for objects 6 .
  • the transition from the initial state having an initial angle of the objects to a final state, in which the objects have been rotated by total angle ⁇ Rotation occurs by a decelerated movement, which is calculated by the following steps:
  • t is the system time, which computing unit 3 retrieves from system clock 8 .
  • the system time is zero.
  • FIG. 5 shows the time characteristic of rotational angle ⁇ for the intermediate images if an object 6 is rotated by a total rotational angle ⁇ Rotation of 30° in 2.5 seconds. It can be seen that the rotational speed is initially relatively fast and is then slowed down.
  • the total rotational angle ⁇ Rotation is additionally adapted to a rotational angle, which corresponds to a carousel-like rotation of objects 6 on virtual ring section 5 , at which in the final state an object 6 appears in the foreground, as shown in FIG. 2 .
  • the rotational angle input by the user via input device 4 is thus rounded up or down to these discrete rotational angles.
  • this object 6 is automatically activated after the expiration of a time interval. Following the expiration of the time interval, this object 6 is thus magnified, while at the same time the virtual ring section 5 with the other objects 6 vanishes until only the activated object 6 is still displayed.
  • the data of the speedometer of the motor vehicle may be transmitted to user interface device 2 and thus computing unit 3 via vehicle bus 7 . These data are used by computing unit 3 to ascertain the duration of the rotation. The higher the speed of the motor vehicle, the longer is the duration of the total rotation.
  • a second operator action may be performed in the method described herein or the device described herein.
  • the user performs an operator movement that differs from the operator movement of the first operator action.
  • a vertical operator movement upward or downward is performed.
  • the user may move the finger upward or downward on the touch-sensitive surface.
  • the second operator movement in particular starts with a touch of the touch-sensitive surface at a position at which an object 6 , i.e. an entry of the portion of the list, is displayed.
  • an object 6 i.e. an entry of the portion of the list
  • the second operator action has the consequence that user interface device 2 changes the graphic data in such a way that multiple list entries are skipped in the display of the portion of the list so as to reach a list entry, which, or the associated graphic object 6 of which, is not contained in the displayed portion of the list entries that were displayed while performing the second operator action.
  • a graphic object 6 is thus displayed in the foreground, which was previously not displayed.
  • the total ring with the list entries was transported directly into a rotated state, in which another portion of the list is shown, without it having been necessary to run through the individual list entries consecutively.
  • the starting point is the display of a portion of a list on the display area of display 1 , as it is represented in FIG. 6 .
  • This informational display substantially corresponds to the informational display represented in FIG. 2 of the list entries on the basis of graphic objects 6 .
  • three list entries G 22 , G 23 , and G 24 are shown by way of example.
  • the entries of this list are divided into classes, the individual classes in turn forming a list.
  • this may be an alphabetic sorting of concepts associated with the list entries for example.
  • list entries G 22 , G 23 and G 24 are the entries no. 22 through 24 beginning with the letter G.
  • the list entries having other initial letters are sorted accordingly.
  • the user may perform a movement upward in the direction of arrow A or downward in the direction of arrow B.
  • An operator action upward in the direction of arrow A has the consequence that the displayed portion of the list entries is changed immediately such that multiple list entries are skipped and the first list entry of the next initial letter is reached, i.e. as shown in FIG. 7 , list entry F 1 .
  • the displayed portion of the list may change such that the last list entry or the last entries beginning with the letter G is or are not included in the displayed portion.
  • the second operator action by contrast includes an operator movement downward in the direction of arrow B
  • the first list entry beginning with the next lower initial letter is reached directly, i.e. as shown in FIG. 8 , list entry H 1 .
  • a portion of the list is thus displayed, which begins with the list entry H 1 and thereupon shows additional list entries beginning with the letter H.
  • User interface device 2 may furthermore be coupled to an acceleration sensor 16 , which, alternatively or additionally, may be used as an input device.
  • Acceleration sensor 16 is able to detect an acceleration of the device, which in this case is arranged in particular as a mobile device, as well as the direction of acceleration.
  • a movement of the device, including display area 1 , in space corresponds in this case to the previously explained movement on the touch-sensitive surface of display area 1 .
  • a movement to the left makes it possible to scroll through the list in one direction for example, while a movement to the right makes it possible to scroll through the list in the other direction.
  • An upward movement results in multiple list entries being skipped forward in the list, while a movement downward results in multiple list entries being skipped backward, as was explained above.

Abstract

In a method for displaying information, a portion of a list is displayed on a display area, which includes a plurality of list entries, and the displayed portion of the list is changed by a first operator action of a user, the entries of the list being run through consecutively by the first operator action. By a second operator action multiple list entries are skipped so as to reach a list entry that is not contained in the displayed portion of the list entries that were displayed while performing the second operator action. In a device for displaying information, the above method steps are implementable.

Description

    CROSS REFERENCE TO RELATED APPLICATION
  • The present application is a continuation application of U.S. patent application Ser. No. 13/125,175, filed on Jul. 15, 2011, as a national stage application of PCT/EP2009/060249, filed on Aug. 6, 2009, which claims priority to German Patent Application No. 10 2008 052 485.9, filed on Oct. 21, 2008, each of which is incorporated herein in its entirety by reference thereto.
  • FIELD OF THE INVENTION
  • The present invention relates to a method for displaying information, in which a portion of a list is displayed on a display area, which includes a plurality of list entries. The displayed portion of the list is changed by a first operator action of a user, the first operator action running consecutively through the entries of the list. The present invention furthermore relates to a device for displaying information including a display device having a display area, a memory for storing a list, which includes a plurality of list entries, of which a portion is displayable on the display area, and an operating device, which is operable by a user, in order to change the displayed portion of the list by a first operator action, the first operator action running consecutively through the entries of the list.
  • BACKGROUND INFORMATION
  • Mobile devices in consumer electronics such as e.g. devices for playing back audio and/or video files have memories, in which very large amounts of data may be stored. Mobile music players, for example, are able to store several thousands of musical pieces. In such mobile devices there arises the problem of how the user is able to access these large amounts of data in a simple manner. For navigation, the mobile devices often have a display and an operating device. The operating device may include a touch-sensitive surface of the display or may be an operating element that is separate from the display. The data sets stored in the device may be stored at least in part as a list, of which a portion may be displayed on the display. With the aid of the operating device, the user is able to scroll through the list in order to reach the desired list entry. If the list contains very many entries, however, there arises the problem that it takes very long to reach an entry that is arranged in the list very far away from the displayed list entries. For this reason it has been proposed to group the list entries hierarchically. This entails the disadvantage, however, that the user must navigate through a deeply branched hierarchical menu structure in order to arrive at a desired list entry.
  • Furthermore, the display of information in a vehicle also entails the problem that very large data sets must be visualized in the vehicle. Modern vehicles include for example a plurality of driver assistance systems, the information of which must be displayed in the vehicle. Furthermore, vehicles often include a navigation system. Such a navigation system is able to display digital geographic road maps including a route and possibly various additional information. Finally, modern vehicles often include communication and multimedia applications, including a mobile telephone interface and devices for playing back music and voice. For these applications it must also be possible to display information in the vehicle. The display of information in a vehicle and the operation of the manifold devices of the vehicle entail very specific requirements. In the vehicle, it is, among others, the driver who takes in the information and operates the respective device. The information should thus be represented in the vehicle in such a way that while taking in the information the driver is not distracted from driving. The represented information should therefore be ascertainable intuitively and quickly by the driver such that he has to take his eyes off the driving action only very briefly when taking in the information. Likewise, the operation of the vehicle devices should be as simple and intuitive as possible such that the driver is able to operate the devices even while driving. If the operation is supported or guided by a display, then the display should occur in such a way that the driver has to view the display only very briefly in order to carry out the operation.
  • The representation of information in a vehicle often occurs in that a portion of the total information is displayed on the display area in the vehicle and that the user is able to change the portion. The total information may be a list having a plurality of list entries for example. A portion of the entries of the entire list is displayed. By an operator action, the user is able to scroll so as to display other list entries. In the operator action, the user may perform for example an operator movement of a certain extent. The user may furthermore actuate an operating element for a certain operating duration. The quantity or the measure of the change of the displayed information, i.e. in this case the number of items, by which the displayed list entries change, depends in this case on the extent of the operator movement or on the operating duration.
  • If a very long list having many entries is to be accessed in the vehicle, there exists the problem that the operator action for scrolling to very distant list entries requires much time and attention on the part of the user. In the application in a vehicle, however, this is disadvantageous, as explained above.
  • SUMMARY
  • Example embodiments of the present invention provide a method and a device of the type mentioned at the outset, by which the displayed portion of the list may be changed in a manner that is simple and intuitive for the user such that the operator action required for this purpose may be performed as quickly as possible. Furthermore, as little attention as possible is to be exacted from the user.
  • In the method according to example embodiments of the present invention, by a second operator action, multiple list entries are skipped so as to reach a list entry that is not contained in the displayed portion of the list entries that were displayed while performing the second operator action.
  • A list should be understood as a data set that is divisible into a plurality of data units. The data units then correspond to the list entries. The data units may be brought into a certain sequence such that they then represent a list. By the first operator action, the user is able to run through the entries of the list consecutively. For this purpose, a portion of the list, i.e. multiple list entries, may be represented on the display area on the basis of a graphic object.
  • The list may include a plurality of audio or video files for example. Furthermore, it may pertain to entries of a telephone or address book, entries of a menu of a hierarchical menu structure for operating a device or other data units, which may be integrated in a list and represented abstractly on the basis of a graphic object.
  • Using the second operator action, which the method described herein makes possible, the list may be scanned much more quickly in order to reach a desired list entry since it is not necessary to run through the entire list entry by entry before reaching the desired entry, as the second operator action rather makes it possible to skip multiple entries in order to reach the desired entry more quickly.
  • The first operator action may include steering a movement of a list entry in a first direction and the second operator action may include steering a movement of a list entry in a second direction, which differs from the first direction. The second direction runs in particular substantially perpendicular to the first direction. For example, a list entry may be moved by the operator action in the horizontal direction in order to scroll through the individual list entries. In this case, a vertical movement of the list entry has the result that multiple list entries are skipped so as to reach a list entry that was not yet displayed when the respective list entry was moved in the vertical direction. Of course, any other directional combinations are possible for the two operator actions as long as the two directions for the operator actions differ.
  • The entries of the list may be divided into classes. In this case, the second operator action takes one to a list entry of another class. The classes in particular also form a list, the individual classes including sorted list entries. In this case, the second operator action preferably takes one to the first list entry of the next or previous class. An example of such a class division is an alphabetical sorting of list entries. A class in this case includes all list entries beginning with a particular letter.
  • The first operator action is in particular used to scroll through the list. The second operator action may be used to abbreviate the scrolling by skipping multiple list entries.
  • At least one operator action may include a movement on touch-sensitive surface of an input device. Many other operator actions are possible, however, which substantially depend on the type of input device. If the input device allows for so-called gesture detection, then the operator action may include a gesture performed by the hand of a user. If the input device is a mechanical operating element, then the operator action is to actuate this operating element, e.g. to turn it or to press on it for a certain time.
  • The touch-sensitive surface of the input device may be provided on the display area. The entries of the portion of the list may be displayed in a one-dimensional line on the display area. The first operator action in this case may include a movement on the touch-sensitive surface, which runs substantially in the direction of the one-dimensional line.
  • The second operator action may include a movement on the touch-sensitive surface, which runs substantially in a direction perpendicular to the one-dimensional line. For this purpose, it is not necessary for the movements to run exactly in the direction of the one-dimensional line or perpendicular to it. They may also run in an angular range of e.g. ±5 degrees or ±10 degrees around these directions.
  • If the touch-sensitive surface of the input device is provided on the display area, then the second operator action may furthermore be a movement, which starts with a touch of the touch-sensitive surface at a position, at which an entry of the portion of the list is displayed.
  • At least one operator action may include a movement of the display area in space, and the acceleration of the display area in space may be detected. In this case, the user moves the device, which includes the display area, e.g. in a particular direction. This movement includes an acceleration of the display area, which is detected. The direction in which a list is to be run through is derivable from the direction of movement. The movement in a different direction may be interpreted as a second operator action, in which multiple list entries are skipped.
  • The displayed list entries may be represented as arranged on a virtual ring or ring section represented in perspective. In this case, the first operator action rotates the displayed list entries on the virtual ring or ring section in carousel-like fashion.
  • Rotating should be understood as rotating the virtual ring or ring section. Since the three-dimensional virtual ring on the display is represented in a two-dimensional perspective view, the objects are actually shifted with respect to their position on the display.
  • In a device according to example embodiments of the present invention for displaying information, the user is able to actuate the operating device by a second operator action, by which multiple list entries are skipped in order to arrive at a list entry that is not contained in the displayed portion of list entries that were displayed while performing the second operator action.
  • The device may be able to implement the method steps described above partially or entirely.
  • The operating device of the display device may include a user interface device and an input device. As a function of an input via the input device, the user interface device is able to produce graphic data, which control the display device such that a particular display is represented on the display area. For changing the graphic data, for example for producing an animation, the user interface device may include a computing unit, which produces graphic data for intermediate images for the animation. It is also possible, however, for these intermediate images to be already stored in advance in the memory.
  • The first or second operator action, which is implemented by the operating device of the device into a change of the display on the display area, includes in particular the operator actions described above with reference to the method. For this purpose, the input device may include a touch-sensitive surface. This surface may be part of a so-called touch pad. Furthermore, the surface may be provided on the display area, i.e. the display device is equipped with a touch screen. Moreover, the input device may be a device for detecting and evaluating a gesture of a part of the body of a user, which is performed in front of the display area. For this purpose, the input device may include in particular a receiving device, to which a signal from the body part of the user may be transmitted capacitively when the body part is in proximity of the receiving device. The position of the body part may be detected via this capacitive coupling. From the change of this position over time, a gesture of the user may be inferred.
  • Furthermore, the input device may include an infrared light source or a receiver for reflected infrared light for detecting the gesture of the body part of the user. In this case as well, the position of the body part and its change over time are detected and interpreted as a gesture.
  • Furthermore, an approach sensor may be provided, which detects the operating intention of a user on the basis of the approach of a part of the body of the user.
  • Furthermore, the device may include an acceleration sensor, which is able to detect the acceleration of the display area in space. The operating device is able to interpret the data detection of the acceleration sensor so as to assign it to the first or to the second operator action.
  • Example embodiments of the present invention are explained in more detail below with reference to the Figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 schematically shows display device according to an example embodiment of the present invention and the coupling of this display device to the electronic system of a motor vehicle.
  • FIG. 2 shows a display of information produced by the method according to an example embodiment of the present invention in a motor vehicle, in which multiple objects are represented.
  • FIG. 3 shows a display of information of an activated object produced by the method according to an example embodiment of the present invention.
  • FIG. 4 shows a display of information of another object produced by the method according to an example embodiment of the present invention.
  • FIG. 5 shows the time characteristic of the rotational angle of an object when rotating on the virtual ring.
  • FIGS. 6 through 8 show the change of the displayed list entries in a second operator action.
  • DETAILED DESCRIPTION
  • An exemplary embodiment described below relates to a display device and a method for displaying information in a vehicle, in particular a motor vehicle. Example embodiments of the present invention, however, may be utilized in the same manner in a mobile or stationary consumer electronics device, in particular a device for playing back audio and video files.
  • The display device includes a display 1 for the graphical representation of information on a display area. Display 1 may be a matrix display, e.g. an LCD (liquid crystal display), especially a color display using TFT (thin-film transistor) technology. Furthermore, the display may be a so-called twisted nematic-liquid crystal display (TN-LCD), a super twisted nematic (STN) display, a double-layer STN, an FLC (ferroelectric liquid crystal) display or an SSFLC (surface stabilized ferroelectric liquid crystal) display. Display 1 has an associated back-lighting (not shown), which may be provided by one or more light-emitting diodes. Display 1 is freely programmable, i.e. any desired graphic data may be produced, which are represented on display 1.
  • In particular, display 1 is mounted in an area of the vehicle that is clearly visible at least for the driver. If the operation of the devices of the vehicle is coupled directly to the position of display 1 such that the user must bring e.g. his hand or his finger at least near display 1 in order to make inputs, display 1 is positioned in such a way that the driver of the vehicle may reach it readily with his hand or his finger. Display 1 may be accommodated in the center console of the vehicle for example.
  • Display 1 is connected to a user interface device 2, which is able to produce graphic data for graphic objects 6 representable on display 1. Furthermore, user interface device 2 is connected to an input device 4, via which the user is able to control devices of the vehicle, the information of which is displayed on display 1. User interface device 2 and input device 4 form an operating device, which may be actuated by a user and which implements the input of a user in graphic data for representation on display 1.
  • User interface device 2 is furthermore connected to a memory 15. In particular data for a list as well as the graphic objects 6 associated with the individual list entries may be stored in this memory.
  • Input device 4 may be for example a device for detecting and evaluating a gesture of a part of a user's body. The hand of the user may perform the gesture in front of display 1 for example. For this purpose, the three-dimensional position of the hand is detected in a specific detection area in front of display 1, without it being necessary to touch display 1. The admissible detection area depends on the placement of display 1 in the motor vehicle. The area should be selected such that the presence of the hand of a user in this detection area may be associated unequivocally with an operative control of input device 4. The boundary of the detection area may be e.g. 40 cm to 10 cm in front of display 1. If the hand of the user is brought up closer than this threshold value to display 1, this is detected by input device 4 or by a separate approach sensor, and the approach is interpreted as an operating intention. For instance, this may result in a change in the manner in which the objects represented on display 1 are rendered. Input device 4 detects the position and the movement of the hand of the user in the detection area. In the process, various gestures performed by the hand are recognized and interpreted as inputs.
  • Input device 4 may include e.g. infrared light sources and infrared light receivers, which detect the infrared light reflected by the hand. Details of such an input device are described in German Patent Document No. 100 58 244 which is incorporated herein by reference. Further input devices, which may be used in conjunction with the display device, are described in the following publications: German Patent Document No. 103 05 341 and German Patent Document No. 10 2004 048 956.
  • Furthermore, the position of the hand and its change over time may also be detected by an optical system. In this system, a light-emitting diode emits e.g. square-wave, amplitude-modulated light. This light is reflected by the object to be detected, i.e. the hand, and after the reflection reaches a photodiode. Another light-emitting diode likewise emits square-wave, amplitude-modulated light to the photodiode, which light is phase-shifted by 180° however. At the photodiode, the two light signals are superposed and cancel each other out if they have exactly the same amplitude. If the signals do not cancel each other out at the photodiode, the light emission of the second diode is regulated via a control loop such that the total received signal again adds up to zero. If the position of the object changes, this also causes a change in the light component which arrives at the photodiode from the first light-emitting diode via the reflection on the object. This brings about a correction of the intensity of the second light-emitting diode through the control loop. The control signal is therefore a measure for the reflection of the light, which is emitted by the first diode, on the object. In this manner, it is possible to derive from the control signal a signal that is characteristic for the position of the object.
  • In addition, the input device may be a touch-sensitive film, which is provided on display 1. The film makes it possible to detect the position at which the user touches display 1 situated behind the film. The film may be arranged e.g. as a resistive touch film, a capacitive touch film or piezoelectric film. Furthermore, the film may be arranged so as to measure a heat flow emanating e.g. from the finger of a user. Various inputs may be obtained from the development of the touch of the film over time. In the simplest case, for example, a touch of the film at a specific position may be allocated to a graphic object shown on display 1. Furthermore, sliding movements of the finger over the film may be interpreted. In particular, in this manner the user is able to define a line on display 1, in that he touches the film at one point, slides on the film toward another point and removes the finger from the film at the other point.
  • Finally, a separate operating element may be used as input device. In particular, the separate operating element is a mechanical operating element. For instance, a rotary switch may be provided, by which objects shown on the display are controllable and are selectable by pressing the rotary switch. Furthermore, an angle of rotation may also be input directly by the rotary switch, as will be explained later. In addition, separate press switches may be disposed around the rotary switch, the arrangement of display fields on the display, which are associated with the press switches, corresponding at least schematically to the arrangement of the press switches. The informational system may include a multifunction operating device as described in European Patent Document No. 1 212 208 for example.
  • User interface device 2 is further coupled to a system clock 8 and if applicable to a vehicle bus 7. User interface device 2 may be connected e.g. to driver assistance systems of the vehicle via vehicle bus 7. User interface device 2 receives data from these driver assistance systems via vehicle bus 7, and processes these data so that they are displayed graphically to the driver or the vehicle occupants via display 1. To this end, user interface device 2 produces graphic data for objects displayable on display 1, which, inter alia, graphically represent the information of the driver assistance systems. User interface device 2 is furthermore connected via vehicle bus 7 to various information, communication and entertainment devices of the vehicle. The varied information from these devices of the vehicle is processed in user interface device 2 and converted into graphic data for a graphical representation. For animations of the graphical representation on the display, the user interface device includes a computing unit 3, which accesses system clock 8 in producing intermediate images.
  • In the following, the display of information on display 1 in the vehicle is explained in detail:
  • The graphic data, which user interface device 2 produces for display 1, relate to a portion of a list. Multiple entries from this list are represented as objects arranged on a virtual ring section 5 shown in perspective. The perspective representation shows virtual ring section 5 as though the observer were looking at an angle from above onto ring section 5. An object 6 in the foreground is shown as the largest in the perspective view. Further back, additional objects 6 are shown, partially hidden by objects 6 situated in front of them. One or more devices of the vehicle are assigned to objects 6 or to the associated list entries. Objects 6 shown in FIG. 2 either give only an indication of the associated vehicle device or additionally already provide informational content of this device.
  • In the view shown in FIG. 2, object 6 represented in the foreground refers to the control of the air conditioning of the interior of the vehicle. The object adjacent to the right and behind this object 6 is assigned to the music playback device of the vehicle, and the object adjacent to the left of object 6 associated with the air-conditioning control is assigned to the navigation system of the vehicle. An object situated further back relates to the settings of the display device itself.
  • Below the virtual ring section 5 shown in perspective are buttons for controlling the graphic data produced by user interface device 2 for display 1. Using a first button 9, the user is able to return in any display state to the initial state, in which objects 6 are displayed as situated on the virtual ring section 5 shown in perspective. Using buttons 10 and 11, objects 6 on virtual ring section 5 may be rotated in carousel-like fashion by a position clockwise or counterclockwise, as will be explained later in detail. Buttons 12 and 13 may be used to actuate functions within an object 6, and via button 14 the user is able to access e.g. directly the settings menu for the graphics.
  • Starting from the view of display 1 shown in FIG. 2, any of the objects 6 shown on virtual ring section 5 may be selected and activated using input device 4. If the input device includes a touch-sensitive film, the user can tap e.g. using his finger on the object 6 to be activated. If input device 4 detects e.g. the position of the finger or the hand of the user in front of display 1, it suffices for the user to bring his finger near the position of display 1, at which the object 6 to be activated is represented. If an object 6 has been activated with the aid of input device 4, then in a smooth transition, virtual ring section 5 with the objects 6 situated on it is faded out and only activated object 6 is still represented. When virtual ring section 5 and objects 6 situated upon it are faded out until solely the activated object 6 is shown, activated object 6 displaces the other objects 6 of ring section 5 and is enlarged continuously until only activated object 6 is still displayed. The graphic data for this smooth transition are calculated by the user interface device 2. When producing the intermediate images, user interface device 2 uses system time 8 in order to ensure that the animation is reproduced smoothly without jerky movements.
  • FIG. 3 for example shows a view of display 1, which is shown when activating the object 6 associated with the air conditioning. In this case, the representation of object 6 for the air conditioning shown in FIG. 2 is magnified. Buttons 9 through 14 are preserved in the process. If the user actuates button 9 from this view shown in FIG. 3, then the representation of object 6 for the air conditioning is continuously scaled down, and the other objects 6 of virtual ring section 5 appear again, until the view shown in FIG. 2 is displayed.
  • If, on the other hand, button 10 or 11 is actuated, object 6 for the air conditioning is similarly scaled down and the other objects 6 of virtual ring section 5 are displayed again as shown in FIG. 2. Subsequently, however, virtual ring section 5 is rotated clockwise or counterclockwise in carousel-like fashion to the next position, whereupon the next object 6 now positioned in the foreground is activated and in a smooth transition is shown enlarged and as displacing the other objects 6, until only this next object 6 is displayed.
  • FIG. 4 shows object 6 associated with the settings of the graphics system itself. When activating this object 6, various settings may be adjusted for the graphic display using input device 4.
  • The list entries, to which graphic objects 6 belong, are part of a hierarchical structure, as described above, since the activation of a list entry or a graphic object 6 results in the display of more detailed information. Furthermore, the activation of a graphic object 6 or a list entry may result e.g. in the playback of an audio file. In this case too, the playback of the audio file corresponds to a lower level of the hierarchy. When displaying such a hierarchically organized data set, the transition between two static representations on display 1 is of particular importance. Particularly if the display device is installed in a motor vehicle, it is very important that the driver of the motor vehicle in particular is able to detect the information represented on display 1 quickly, simply and intuitively, even if he directs his eyes onto display 1 only for brief intervals. For this purpose it is particularly important that the viewer is at all times able to orient himself well within the hierarchical structure. It has been shown that abrupt changes of displayed images are disadvantageous for this purpose. It has been shown in particular that smooth transitions between two static representations, in which intermediate images are displayed, which emphasize the transition between the two static representations, facilitate orientation within the structure as well as the grasp of the informational content.
  • Furthermore, it is particularly important how the intermediate images visualize the transition between two static representations. In this regard, a linear transition from one static representation to the next static representation is not suitable for use in a motor vehicle. Rather, intermediate images are advantageous, which illustrate a movement of objects 6 that is at least partially slowed down. The parameters for the choice of the slowdown and the duration of the transition must also be adapted for use in the vehicle because conventional animations known from the field of computers are usually not usable.
  • The following describes in detail how the user is able to rotate the objects 6 situated on virtual ring section 5 in carousel-like fashion on this virtual ring section 5 with the aid of input device 4.
  • The starting point of the rotation is the arrangement of objects 6 on virtual ring section 5 shown in FIG. 2. Objects 6 are thus represented on a ring-shaped one-dimensional line. By a first operator action via input device 4, the user now specifies the total rotational angle αRotation, by which objects 6 are to be rotated on the ring as a whole, of which only a ring section 5 is shown. For this purpose, the user enters a line on display 1 using input device 4 for example. If input device 4 includes a touch-sensitive film, the user may run his finger over the film in order to specify this line. This line for entering the first operator action runs in particular substantially on the line on which objects 6 are represented, or parallel to this line. If input device 4 is able to detect a gesture of the user or a movement of his hand or his finger in front of display 1, the line is entered in this manner. If the line is not oriented in the direction of the width of display 1, but rather runs aslant, preferably only the length of the horizontal component of the entered line oriented in parallel to the width of display 1 is used, as long as the deviation from the horizontal does not exceed 45°. The total rotational angle αRotation is calculated from the ratio of the length of this horizontal component of the line to the total width of display 1. In fact, in the perspective view on display 1, objects 6 are shifted only in the plane of display 1. Rotational angle α is thus transformed accordingly by user interface device 2 when producing the graphic data.
  • If input device 4 is a separate operating element having a rotary encoder, then the total rotational angle may also be input via the rotational angle of the rotary encoder.
  • Since only a portion of the list is shown on display 1, the shown portion changes as the virtual ring is rotated. In the rotation, objects 6 disappear on the one side of ring section 5, while new objects 6 are displayed on the other side of ring section 5, which belong to list entries that were originally not displayed. In this manner, the first operator action may be used to run through the entries of the list consecutively, the entries of the list being shown on the basis of graphic objects 6.
  • After the total rotational angle αRotation has been entered by the user with the aid of input device 4, a computing unit 3 of user interface device 2 produces an animation, which includes the graphic data for successive intermediate images, which represent successive rotational angles α for objects 6. The transition from the initial state having an initial angle of the objects to a final state, in which the objects have been rotated by total angle αRotation, occurs by a decelerated movement, which is calculated by the following steps:
  • n = t d - 1 ; ( i ) α = b + α Rotation · ( 1 - n 4 ) . ( ii )
  • In this instance, t is the system time, which computing unit 3 retrieves from system clock 8. At the start of the rotation, the system time is zero. In the final state of the rotation, system time t corresponds to the duration d of the total rotation about total angle=angle×rotation. Steps (i) and (ii) are repeated until system time t is greater than the duration d of the total rotation.
  • FIG. 5 shows the time characteristic of rotational angle α for the intermediate images if an object 6 is rotated by a total rotational angle αRotation of 30° in 2.5 seconds. It can be seen that the rotational speed is initially relatively fast and is then slowed down.
  • In the present exemplary embodiment, the total rotational angle αRotation is additionally adapted to a rotational angle, which corresponds to a carousel-like rotation of objects 6 on virtual ring section 5, at which in the final state an object 6 appears in the foreground, as shown in FIG. 2. The rotational angle input by the user via input device 4 is thus rounded up or down to these discrete rotational angles.
  • If an object is rotated into the foreground by the user via input device 4, this object 6 is automatically activated after the expiration of a time interval. Following the expiration of the time interval, this object 6 is thus magnified, while at the same time the virtual ring section 5 with the other objects 6 vanishes until only the activated object 6 is still displayed.
  • The data of the speedometer of the motor vehicle may be transmitted to user interface device 2 and thus computing unit 3 via vehicle bus 7. These data are used by computing unit 3 to ascertain the duration of the rotation. The higher the speed of the motor vehicle, the longer is the duration of the total rotation.
  • Furthermore, a second operator action may be performed in the method described herein or the device described herein. In this operator action, the user performs an operator movement that differs from the operator movement of the first operator action. In the present case, a vertical operator movement upward or downward is performed. For example, the user may move the finger upward or downward on the touch-sensitive surface. The second operator movement in particular starts with a touch of the touch-sensitive surface at a position at which an object 6, i.e. an entry of the portion of the list, is displayed. Thus, for the second operator action—in contrast to conventional systems—it is not necessary to display e.g. a separate index in order to skip list entries. It suffices to perform an operator movement that has a direction of movement that is clearly distinguished from the direction of movement of the first operator movement. Furthermore, the user may perform a corresponding gesture in front of display area 1. Finally, it is possible for the user to actuate corresponding mechanical operating elements for controlling the movement of an object 6 in a corresponding direction.
  • The second operator action has the consequence that user interface device 2 changes the graphic data in such a way that multiple list entries are skipped in the display of the portion of the list so as to reach a list entry, which, or the associated graphic object 6 of which, is not contained in the displayed portion of the list entries that were displayed while performing the second operator action. Following the execution of the second operator action, a graphic object 6 is thus displayed in the foreground, which was previously not displayed. The total ring with the list entries was transported directly into a rotated state, in which another portion of the list is shown, without it having been necessary to run through the individual list entries consecutively.
  • The second operator action is explained in detail in the following with reference to FIGS. 6 through 8:
  • The starting point is the display of a portion of a list on the display area of display 1, as it is represented in FIG. 6. This informational display substantially corresponds to the informational display represented in FIG. 2 of the list entries on the basis of graphic objects 6. In the state shown in FIG. 6, three list entries G22, G23, and G24 are shown by way of example.
  • The entries of this list are divided into classes, the individual classes in turn forming a list. In the present case, this may be an alphabetic sorting of concepts associated with the list entries for example. In this case, list entries G22, G23 and G24 are the entries no. 22 through 24 beginning with the letter G. The list entries having other initial letters are sorted accordingly.
  • By a first operator action, as explained above, the user is now able to bring about a change in the displayed list entries by an operator movement in the direction of double-arrow H. If he moves the finger on the touch-sensitive surface to the left for example, the displayed portion of the list will move to the left as well such that list entries on the left side disappear, while new list entries emerge on the right side. Accordingly, when performing an operator movement to the right, list entries on the right side disappear, while new list entries emerge on the left side.
  • In the second operator action, the user may perform a movement upward in the direction of arrow A or downward in the direction of arrow B. An operator action upward in the direction of arrow A has the consequence that the displayed portion of the list entries is changed immediately such that multiple list entries are skipped and the first list entry of the next initial letter is reached, i.e. as shown in FIG. 7, list entry F1. Furthermore, the displayed portion of the list may change such that the last list entry or the last entries beginning with the letter G is or are not included in the displayed portion.
  • If the second operator action by contrast includes an operator movement downward in the direction of arrow B, the first list entry beginning with the next lower initial letter is reached directly, i.e. as shown in FIG. 8, list entry H1. A portion of the list is thus displayed, which begins with the list entry H1 and thereupon shows additional list entries beginning with the letter H.
  • User interface device 2 may furthermore be coupled to an acceleration sensor 16, which, alternatively or additionally, may be used as an input device. Acceleration sensor 16 is able to detect an acceleration of the device, which in this case is arranged in particular as a mobile device, as well as the direction of acceleration. A movement of the device, including display area 1, in space corresponds in this case to the previously explained movement on the touch-sensitive surface of display area 1. In this case, a movement to the left makes it possible to scroll through the list in one direction for example, while a movement to the right makes it possible to scroll through the list in the other direction. An upward movement results in multiple list entries being skipped forward in the list, while a movement downward results in multiple list entries being skipped backward, as was explained above.
  • LIST OF REFERENCE NUMERALS
  • 1 display
  • 2 user interface device
  • 3 computing unit
  • 4 input device
  • 5 virtual ring section
  • 6 graphic objects
  • 7 vehicle bus
  • 8 system clock
  • 9-14 buttons of an object 6
  • 15 memory
  • 16 acceleration sensor

Claims (12)

What is claimed is:
1. A method for displaying information, comprising:
displaying, on a display area, a portion of a list including a plurality of list entries:
changing the displayed portion of the list by a first operator action of a user including selecting a list entry and steering a movement of the selected list entry in a first direction;
consecutively running through the entries of the list by the first operator action; and
skipping multiple list entries by a second operator action including steering a movement of a list entry in a second direction different from the first direction, to reach a list entry that is not included in the displayed portion of the list entries displayed while performing the second operator action;
wherein at least one operator action includes a movement on a touch-sensitive surface of an input device arranged on the display area; and
wherein the second operator action includes a movement that starts with a touch of the touch-sensitive surface at a position at which a list entry of the portion of the list is displayed.
2. The method according to claim 1, wherein the entries of each class are sorted alphabetically when displayed on the display area.
3. The method according to claim 1, further comprising scrolling through the list by the first operator action.
4. The method according to claim 1, wherein the list entries of the portion of the list are displayed in a one-dimensional line on the display area, the first operator action includes a movement on the touch-sensitive surface substantially in a direction of the one-dimensional line, and the second operator action includes a movement on the touch-sensitive surface substantially in a direction perpendicular to the one-dimensional line.
5. The method according to claim 1, wherein at least one operator action includes a movement of the display area in space, the method further comprising detecting an acceleration of the display area in space.
6. The method according to claim 1, wherein the displayed list entries are represented as situated on at least one of (a) a virtual ring and (b) a virtual ring section shown in perspective, and the displayed list entries on the at least one of (a) the virtual ring and (b) the virtual ring section are rotated in carousel-like fashion by the first operator action.
7. A device for displaying information, comprising:
a display device having a display area;
a memory adapted to store a list that includes a plurality of list entries, a portion of the list displayable on the display area; and
an operating device operable by a user to change the displayed portion of the list by a first operator action including selecting a list entry and steering a movement of the selected list entry in a first direction, the entries of the list being run through consecutively in response to the first operator action;
the operating device operable by the user by a second operator action including steering a movement of a list entry in a second direction different from the first direction, in response to the second operator action multiple list entries are skipped to reach a first list entry that is not included in the displayed portion of the list entries displayed while performing the second operator action; and
wherein at least one operator action includes a movement on a touch-sensitive surface of an input device arranged on the display area; and
wherein the second operator action includes a movement that starts with a touch of the touch-sensitive surface at a position at which a list entry of the portion of the list is displayed.
8. The device according to claim 7, wherein the operating device includes an input device having a touch-sensitive surface.
9. The device according to claim 7, wherein the device includes an acceleration sensor adapted to detect an acceleration of the display area in space.
10. The device according to claim 7, wherein the list entries of the portion of the list are displayed in a one-dimensional line on the display area, the first operator action includes a movement on the touch-sensitive surface substantially in a direction of the one-dimensional line, and the second operator action includes a movement on the touch-sensitive surface substantially in a direction perpendicular to the one-dimensional line.
11. The device according to claim 7, wherein at least one operator action includes a movement of the display area in space, the method further comprising detecting an acceleration of the display area in space.
12. The device according to claim 7, wherein the displayed list entries are represented as situated on at least one of (a) a virtual ring and (b) a virtual ring section shown in perspective, and the displayed list entries on the at least one of (a) the virtual ring and (b) the virtual ring section are rotated in carousel-like fashion by the first operator action.
US15/716,109 2008-10-21 2017-09-26 Method and device for displaying information arranged in lists Abandoned US20180018074A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/716,109 US20180018074A1 (en) 2008-10-21 2017-09-26 Method and device for displaying information arranged in lists

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
DE102008052485A DE102008052485A1 (en) 2008-10-21 2008-10-21 Method and device for displaying list-ordered information
DE102008052485.9 2008-10-21
US13/125,175 US9804764B2 (en) 2008-10-21 2009-08-06 Method and device for displaying information arranged in lists
PCT/EP2009/060249 WO2010046147A1 (en) 2008-10-21 2009-08-06 Method and device for displaying information sorted into lists
US15/716,109 US20180018074A1 (en) 2008-10-21 2017-09-26 Method and device for displaying information arranged in lists

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US13/125,175 Continuation US9804764B2 (en) 2008-10-21 2009-08-06 Method and device for displaying information arranged in lists
PCT/EP2009/060249 Continuation WO2010046147A1 (en) 2008-10-21 2009-08-06 Method and device for displaying information sorted into lists

Publications (1)

Publication Number Publication Date
US20180018074A1 true US20180018074A1 (en) 2018-01-18

Family

ID=41510831

Family Applications (2)

Application Number Title Priority Date Filing Date
US13/125,175 Active 2030-10-24 US9804764B2 (en) 2008-10-21 2009-08-06 Method and device for displaying information arranged in lists
US15/716,109 Abandoned US20180018074A1 (en) 2008-10-21 2017-09-26 Method and device for displaying information arranged in lists

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US13/125,175 Active 2030-10-24 US9804764B2 (en) 2008-10-21 2009-08-06 Method and device for displaying information arranged in lists

Country Status (8)

Country Link
US (2) US9804764B2 (en)
EP (1) EP2350799B1 (en)
KR (3) KR20160084504A (en)
CN (2) CN102203710A (en)
DE (1) DE102008052485A1 (en)
MX (1) MX2011004124A (en)
RU (1) RU2477511C2 (en)
WO (1) WO2010046147A1 (en)

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB9722766D0 (en) 1997-10-28 1997-12-24 British Telecomm Portable computers
US7469381B2 (en) 2007-01-07 2008-12-23 Apple Inc. List scrolling and document translation, scaling, and rotation on a touch-screen display
US7487444B2 (en) 2002-03-19 2009-02-03 Aol Llc Reformatting columns of content for display
US7844915B2 (en) 2007-01-07 2010-11-30 Apple Inc. Application programming interfaces for scrolling operations
US9767438B2 (en) * 2009-07-14 2017-09-19 Bitvore Corp. Presentation and sorting of email
US8972467B2 (en) 2010-08-31 2015-03-03 Sovanta Ag Method for selecting a data set from a plurality of data sets by means of an input device
US8767019B2 (en) 2010-08-31 2014-07-01 Sovanta Ag Computer-implemented method for specifying a processing operation
DE102011011802A1 (en) 2011-02-19 2012-08-23 Volkswagen Ag Method and device for providing a user interface, in particular in a vehicle
DE102011112448A1 (en) * 2011-09-03 2013-03-07 Volkswagen Aktiengesellschaft Method and device for providing a user interface, in particular in a vehicle
DE102011116141A1 (en) 2011-10-15 2013-04-18 Volkswagen Aktiengesellschaft A method of displaying information in a vehicle and display device for a vehicle
US9645733B2 (en) 2011-12-06 2017-05-09 Google Inc. Mechanism for switching between document viewing windows
USD703687S1 (en) 2011-12-28 2014-04-29 Target Brands, Inc. Display screen with graphical user interface
USD711400S1 (en) 2011-12-28 2014-08-19 Target Brands, Inc. Display screen with graphical user interface
USD715818S1 (en) 2011-12-28 2014-10-21 Target Brands, Inc. Display screen with graphical user interface
USD705790S1 (en) 2011-12-28 2014-05-27 Target Brands, Inc. Display screen with graphical user interface
USD703685S1 (en) 2011-12-28 2014-04-29 Target Brands, Inc. Display screen with graphical user interface
USD705792S1 (en) * 2011-12-28 2014-05-27 Target Brands, Inc. Display screen with graphical user interface
USD706793S1 (en) 2011-12-28 2014-06-10 Target Brands, Inc. Display screen with graphical user interface
USD703686S1 (en) 2011-12-28 2014-04-29 Target Brands, Inc. Display screen with graphical user interface
USD706794S1 (en) 2011-12-28 2014-06-10 Target Brands, Inc. Display screen with graphical user interface
USD711399S1 (en) 2011-12-28 2014-08-19 Target Brands, Inc. Display screen with graphical user interface
JP5799861B2 (en) * 2012-03-09 2015-10-28 コニカミノルタ株式会社 Function setting device and control program for function setting device
DE102012008778A1 (en) 2012-04-28 2013-10-31 Volkswagen Aktiengesellschaft Method and device for visualizing a hierarchically sortable data stock, in particular in a vehicle
US9696879B2 (en) * 2012-09-07 2017-07-04 Google Inc. Tab scrubbing using navigation gestures
FR2995836B1 (en) * 2012-09-27 2015-05-22 Valeo Systemes Thermiques CONTROL MODULE
DE102013000880A1 (en) 2013-01-10 2014-07-10 Volkswagen Aktiengesellschaft Method and apparatus for providing a user interface in a vehicle
JP5975928B2 (en) * 2013-04-05 2016-08-23 カルソニックカンセイ株式会社 Vehicle display device
DE102013007576A1 (en) 2013-05-02 2014-11-06 Volkswagen Aktiengesellschaft Method and apparatus for selecting an object of a list
DE102013013225B4 (en) * 2013-08-08 2019-08-29 Audi Ag Motor vehicle with switchable operating device
US10766366B2 (en) 2013-08-20 2020-09-08 Volkswagen Ag Operating method for an operating and display device in a vehicle and operating and display device in a vehicle
DE102014208502A1 (en) * 2014-05-07 2015-11-12 Volkswagen Aktiengesellschaft User interface and method for switching between screen views of a user interface
US10380669B2 (en) * 2014-10-30 2019-08-13 Walmart Apollo, Llc Product browsing system and method
US9984403B2 (en) 2014-10-30 2018-05-29 Wal-Mart Stores, Inc. Electronic shopping cart processing system and method
DE102014018116A1 (en) * 2014-12-09 2016-06-09 Daimler Ag Information entertainment system for a motor vehicle and method of operating such
USD776131S1 (en) * 2015-03-30 2017-01-10 Intuit Inc. Display device with transfer animation for a user interface
USD764495S1 (en) * 2015-03-30 2016-08-23 Intuit Inc. Display device with initiating transfer animation for a user interface
CN109040413A (en) * 2017-06-12 2018-12-18 阿里巴巴集团控股有限公司 Display methods, the device and system of data
CN109426475A (en) * 2017-08-31 2019-03-05 中兴通讯股份有限公司 Screen display method and device
CN108198272B (en) * 2017-12-29 2020-11-20 深圳市元征科技股份有限公司 Data processing method and equipment thereof
US11416917B2 (en) * 2020-08-04 2022-08-16 Contextlogic, Inc. Content carousel

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20130254308A1 (en) * 2010-04-29 2013-09-26 British Broadcasting Corporation Content provision system

Family Cites Families (28)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6043809A (en) * 1997-09-23 2000-03-28 Compaq Computer Corporation Computer keyboard scroll bar control
JPH11187443A (en) * 1997-12-25 1999-07-09 Sony Corp Portable radio information terminal equipment, screen operation method, record medium, and microcomputer
US7469381B2 (en) * 2007-01-07 2008-12-23 Apple Inc. List scrolling and document translation, scaling, and rotation on a touch-screen display
DE19944324A1 (en) 1999-09-15 2001-03-22 Audi Ag Multi-function control device
DE10058244C2 (en) 2000-11-19 2003-02-06 Hertz Inst Heinrich Measuring method for determining the position of an object in front of a screen and device for carrying out the method
JP4701564B2 (en) 2001-08-31 2011-06-15 ソニー株式会社 Menu display device and menu display method
US7837562B2 (en) * 2002-06-20 2010-11-23 Igt Display panel for a gaming apparatus
DE10305341A1 (en) 2003-02-10 2004-08-19 Ident Technology Ag Switching signal provision procedure e.g. for operating motor vehicle functional units, requires setting up signal for coupling into user within framework of actuating a switching device
JP2008135075A (en) * 2002-08-28 2008-06-12 Sony Corp Electronic apparatus, spatial displacement computing device, and spatial displacement computation method
JP4482269B2 (en) * 2002-08-28 2010-06-16 ソニー株式会社 Electronic device apparatus, signal compensation apparatus, and signal compensation method
SE0202664L (en) 2002-09-09 2003-11-04 Zenterio Ab Graphical user interface for navigation and selection from various selectable options presented on a monitor
BRPI0410134A (en) 2003-05-08 2006-05-16 Thomson Licensing method and apparatus for alphabetic text navigation
RU2336553C2 (en) * 2003-08-21 2008-10-20 Майкрософт Корпорейшн System and method for support of applications that are minimised with expanded set of functions
DE10341580A1 (en) * 2003-09-09 2005-03-31 Siemens Ag Input device for a data processing system
US8028250B2 (en) * 2004-08-31 2011-09-27 Microsoft Corporation User interface having a carousel view for representing structured data
US7728823B2 (en) 2004-09-24 2010-06-01 Apple Inc. System and method for processing raw data of track pad device
DE102004048956A1 (en) 2004-10-07 2006-04-27 Ident Technology Ag Signal transferring method using human body, integrates condenser device that is used as signal interface in oscillating circuit, where circuit is operated as parallel or serial oscillating circuit in receive or transmit mode, respectively
US7810043B2 (en) * 2005-07-27 2010-10-05 Microsoft Corporation Media user interface left/right navigation
US20070028268A1 (en) 2005-07-27 2007-02-01 Microsoft Corporation Media user interface start menu
US8769408B2 (en) 2005-10-07 2014-07-01 Apple Inc. Intelligent media navigation
US20070132789A1 (en) * 2005-12-08 2007-06-14 Bas Ording List scrolling in response to moving contact over list of index symbols
US20070157126A1 (en) 2006-01-04 2007-07-05 Tschirhart Michael D Three-dimensional display and control image
US7269002B1 (en) * 2006-03-29 2007-09-11 Matthew Scott Turner Vehicle recess integrated docking station
KR101479769B1 (en) * 2006-08-01 2015-01-08 엘지전자 주식회사 Touch screen apparatus and file search method thereof
US8091045B2 (en) * 2007-01-07 2012-01-03 Apple Inc. System and method for managing lists
KR20080068781A (en) * 2007-01-20 2008-07-24 엘지전자 주식회사 Electronic device with touch screen and method of displaying information using same
US20100001960A1 (en) * 2008-07-02 2010-01-07 Sling Media, Inc. Systems and methods for gestural interaction with user interface objects
US20100175026A1 (en) * 2009-01-05 2010-07-08 Bortner Christopher F System and method for graphical content and media management, sorting, and retrieval

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060026521A1 (en) * 2004-07-30 2006-02-02 Apple Computer, Inc. Gestures for touch sensitive input devices
US20130254308A1 (en) * 2010-04-29 2013-09-26 British Broadcasting Corporation Content provision system

Also Published As

Publication number Publication date
US9804764B2 (en) 2017-10-31
WO2010046147A1 (en) 2010-04-29
CN102203710A (en) 2011-09-28
DE102008052485A1 (en) 2010-04-22
RU2477511C2 (en) 2013-03-10
CN105034809B (en) 2018-04-06
KR20170086141A (en) 2017-07-25
US20110265036A1 (en) 2011-10-27
KR20160084504A (en) 2016-07-13
KR101646293B1 (en) 2016-08-05
CN105034809A (en) 2015-11-11
MX2011004124A (en) 2011-05-19
KR20110084197A (en) 2011-07-21
EP2350799A1 (en) 2011-08-03
EP2350799B1 (en) 2017-11-01
RU2011120464A (en) 2012-12-20

Similar Documents

Publication Publication Date Title
US20180018074A1 (en) Method and device for displaying information arranged in lists
US8477108B2 (en) Method for displaying information in a vehicle and display device for a vehicle
US10449857B2 (en) Information reproduction system for a vehicle and method for providing information for the user of a vehicle
US9758150B2 (en) Method and device for displaying information
US9013470B2 (en) Method and device for displaying information, particularly in a vehicle
US9878618B2 (en) Information playback system and method for information playback
US8677271B2 (en) Method for displaying information in a motor vehicle and display device for a motor vehicle
US9802484B2 (en) Method and display device for transitioning display information
KR101609024B1 (en) Method and device for providing a user interface
US9019311B2 (en) Method for displaying a two-sided two-dimensional object on a display in a motor vehicle and display device for a motor vehicle
US20130057400A1 (en) Method and device for displaying information in a vehicle
US20180307405A1 (en) Contextual vehicle user interface
US10139988B2 (en) Method and device for displaying information arranged in lists
CN105677163A (en) Concentrated control system for vehicle
US9530182B2 (en) Method for displaying information in particular in a vehicle and display unit for a vehicle
US11014449B2 (en) Method and device for displaying information, in particular in a vehicle
US11099715B2 (en) Method and device for providing a user interface in a vehicle
US20180232115A1 (en) In-vehicle input device and in-vehicle input device control method
WO2020196561A1 (en) Operation device

Legal Events

Date Code Title Description
AS Assignment

Owner name: VOLKSWAGEN AG, GERMANY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HOEHNE, SVEN;REEL/FRAME:043705/0666

Effective date: 20110627

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STCV Information on status: appeal procedure

Free format text: NOTICE OF APPEAL FILED

STCV Information on status: appeal procedure

Free format text: EXAMINER'S ANSWER TO APPEAL BRIEF MAILED

STCV Information on status: appeal procedure

Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS

STCV Information on status: appeal procedure

Free format text: BOARD OF APPEALS DECISION RENDERED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION