WO2014054208A1 - 操作装置 - Google Patents
操作装置 Download PDFInfo
- Publication number
- WO2014054208A1 WO2014054208A1 PCT/JP2013/004701 JP2013004701W WO2014054208A1 WO 2014054208 A1 WO2014054208 A1 WO 2014054208A1 JP 2013004701 W JP2013004701 W JP 2013004701W WO 2014054208 A1 WO2014054208 A1 WO 2014054208A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- operating
- space
- movement
- finger
- image
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3664—Details of the user input interface, e.g. buttons, knobs or sliders, including those provided on a touch screen; remote controllers; input using gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0354—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of 2D relative movements between the device, or an operating part thereof, and a plane or surface, e.g. 2D mice, trackballs, pens or pucks
- G06F3/03547—Touch pads, in which fingers can move on a surface
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/044—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
Definitions
- This disclosure relates to an operation device that operates an image unit displayed on a display screen by input to the operation unit.
- Patent Document 1 discloses a technique for moving an image portion such as a navigation pointer and a radio main screen displayed on a display screen in association with an operation performed on a remote touchpad portion.
- the user interface device disclosed in Patent Document 1 relates a remote touch pad unit that detects an operation of moving an operator's finger and the like, and a finger operation detected by the remote touch pad unit to movement of a map, a pointer, and the like. And a control unit.
- control unit obtains the distance from the remote touch pad unit to the finger.
- the control unit displays the operation with the finger detected by the remote touch pad unit on the display screen. Associate it with the operation that moves the top pointer.
- the control unit performs an operation with the finger detected by the remote touch pad unit. Is associated with the switching operation for switching from the radio main screen to the operation standby screen.
- Patent Document 1 is not provided with any configuration that assists the input by the operator's finger when performing a switching operation for switching from the radio main screen to the operation standby screen. Therefore, it has been difficult for the operator to grasp the range in which the switching operation can be performed. If the range for performing the switching operation is difficult to understand in this way, the operator may not be able to easily perform the input for performing the switching operation.
- An object of the present disclosure is to provide an operation device that facilitates input by an operator in an operation space separated from an operation unit.
- the operating device operates the image displayed on the display screen by inputting the operation body to the operation surface.
- the operating device includes a detecting unit that detects movement of the operating body, an acquiring unit that acquires an operating body distance from the operating surface to the operating body, and a first operating space in which the operating body distance is less than a predetermined threshold distance. Distinguishing the first movement of the operating tool detected in step 2 from the second movement of the operating tool detected in the second operating space in which the operating tool distance exceeds the threshold distance.
- the operator when performing an input in the second operation space in which the movement of the operating tool is associated with the second image, the operator is provided with an input assist formed at a position separated from the operation surface by a threshold distance or more. You can receive input assistance by touching the part. With such assistance, the operator can easily grasp the position of the second operation space that is defined apart from the operation surface. Therefore, the operator can easily perform input in the second operation space.
- the operating device is provided as follows.
- the operating device operates the image displayed on the display screen by inputting the operation body to the operation surface.
- the operating device includes: means for detecting movement of the operating body; means for acquiring an operating body distance from the operating surface to the operating body; and a first operation in which the operating body distance is less than a predetermined threshold distance.
- the first movement and the second movement are associated with a second image portion different from the first image portion and the first image portion displayed on the display screen, respectively, and the first image portion or the first movement Means for changing a display mode of at least one of the two image portions; and means for assisting the second movement of the operating body in the second space.
- FIG. 6 is a diagram for explaining the configuration of the remote control device according to the first embodiment, and is a cross-sectional view taken along line VI-VI in FIG.
- the remote control device In the remote control device according to the first embodiment, it is a diagram for explaining that the image portion to be operated is switched by the operating body distance, It is a figure for demonstrating a series of icon selection operation by an operator, In the remote operation device according to the first embodiment, it is a flowchart showing an operation mode selection process performed by the operation control unit, In the remote operation device according to the first embodiment, it is a diagram for explaining the relationship between the sensitivity value detected by the touch sensor and the operation state determined by the operation control unit, It is a figure which shows each sensitivity threshold value memorize
- FIG. 13 is a diagram for explaining a configuration of a remote control device according to a second embodiment, and is a cross-sectional view taken along line XIII-XIII in FIG. It is a figure for demonstrating a series of icon selection operation by an operator, It is sectional drawing for demonstrating the structure of the remote control apparatus by 3rd embodiment, It is a figure for demonstrating a series of icon selection operation by an operator,
- a flowchart showing an operation mode selection process performed by the operation control unit It is sectional drawing for demonstrating the structure of the remote control apparatus by 4th embodiment, It is sectional drawing for demonstrating the structure of the remote control apparatus by 5th embodiment, It is sectional drawing for demonstrating the structure of the remote control apparatus by 6th embodiment, It is a figure which shows the modification of FIG.
- a remote control device 100 is mounted on a vehicle and constitutes a display system 10 together with a navigation device 50 and the like as shown in FIG.
- the remote operation device 100 is installed at a position adjacent to the palm rest 39 at the center console of the vehicle, and exposes the operation surface 70 in a range that can be easily reached by the operator.
- An operation with an index finger (hereinafter simply referred to as “finger”) F or the like of the operator's hand is input to the operation surface 70.
- the navigation device 50 is installed in the instrument panel of the vehicle in such a manner that the display screen 52 is exposed so that the operator can see it, and the display screen 52 is viewed from the driver's seat.
- Various display images 60 are displayed on the display screen 52.
- the display image 60 shown in FIG. 3 is one of a plurality of display images displayed on the display screen 52, and shows an air conditioning menu image for operating the air conditioning equipment mounted on the vehicle.
- the display image 60 includes a plurality of icons 63 associated with a predetermined function, a focus 62 for selecting the icons 63, a background portion 64 serving as a background of the icons 63 and the focus 62, and the like.
- the position where the focus 62 is displayed on the display screen 52 corresponds to the position where the finger F touches on the operation surface 70 shown in FIG.
- the above display image 60 is generated when the navigation device 50 (see FIG. 1) superimposes a plurality of drawing layers. Specifically, the display image 60 is generated by superimposing a background layer L2 on which the background portion 64 is drawn, an object layer L3 on which the icon 63 is drawn, a focus layer L1 on which the focus 62 is drawn, and the like. Yes.
- the sizes of the layers L1 to L3 are defined so as to match the size of the display screen 52.
- the remote control device 100 is connected to a Controller Area Network (CAN) bus 90, an external battery 95, and the like.
- the CAN bus 90 is a transmission path used for data transmission between in-vehicle devices in an in-vehicle communication network formed by connecting a plurality of in-vehicle devices mounted on a vehicle.
- the remote control device 100 can communicate with the navigation device 50 located remotely via the CAN bus 90.
- the remote operation device 100 includes power interfaces 21 and 22, a communication control unit 23, a communication interface 24, a touch sensor 31, an operation control unit 33, and the like.
- Each power interface 21, 22 stabilizes the power supplied from the battery 95 and supplies it to the operation control unit 33.
- One power interface 21 is always supplied with power from the battery 95.
- the other power interface 22 is supplied with electric power from the battery 95 when the switch 93 is energized when the accessory (ACC) power source of the vehicle is turned on.
- the communication control unit 23 and the communication interface 24 are configured to output information processed by the operation control unit 33 to the CAN bus 90 and to acquire information output to the CAN bus 90 from other in-vehicle devices.
- the communication control unit 23 and the communication interface 24 are connected to each other by a transmission signal line TX and a reception signal line RX.
- the touch sensor 31 is of a capacitive type and is formed in a rectangular flat plate shape.
- the touch sensor 31 is also referred to as a touch panel 31.
- the touch sensor 31 is formed by arranging electrodes extending along the x-axis direction and electrodes extending along the y-axis direction in FIG. 5 in a grid pattern.
- the touch sensor 31 is connected to the operation control unit 33.
- the touch sensor 31 detects an operation with the finger F (see FIG. 2) on the operation surface 70 and outputs it to the operation control unit 33.
- the operation control unit 33 is also referred to as a control circuit 33, and includes a processor that performs various types of arithmetic processing, a RAM that functions as a work area for arithmetic processing, a flash memory that stores programs used for arithmetic processing, and the like. Yes.
- the operation control unit 33 is connected to the power interfaces 21 and 22, the communication control unit 23, the touch sensor 31, and the like.
- the operation control unit 33 illustrated in FIG. 1 acquires a sensitivity value (see FIG. 10) of the sensor 31 by executing a predetermined program and measuring the potential of each electrode of the touch sensor 31.
- the operation control unit 33 includes an x-coordinate and a y-coordinate indicating the relative position of the finger F with respect to the operation surface 70, and z corresponding to a distance from the operation surface 70 to the finger F (hereinafter referred to as “operation object distance d”).
- operation object distance d a distance from the operation surface 70 to the finger F
- the coordinates are detected by a calculation process based on the sensitivity value.
- the operation control unit 33 outputs the x coordinate and the y coordinate indicating the relative position of the finger F to the CAN bus 90 through the communication control unit 23 and the communication interface 24.
- the navigation device 50 includes a display control unit 53 and a liquid crystal display 51.
- the display control unit 53 includes a processor that performs various types of arithmetic processing, a RAM that functions as a work area for arithmetic processing, a graphic processor that performs image drawing processing, a graphic RAM that functions as a work area for drawing processing, and the like.
- the display control unit 53 has a flash memory for storing data used for arithmetic processing and drawing processing, a communication interface connected to the CAN bus 90, and a video output interface for outputting drawn image data to the liquid crystal display 51. is doing.
- the display control unit 53 draws a display image 60 to be displayed on the display screen 52 based on information acquired from the CAN bus 90. Then, the display control unit 53 sequentially outputs the image data of the drawn display image 60 to the liquid crystal display 51 through the video output interface.
- the liquid crystal display 51 is a dot matrix type display that realizes color display by controlling a plurality of pixels arranged on the display screen 52.
- the liquid crystal display 51 displays video by continuously forming image data sequentially acquired from the display control unit 53 on the display screen 52.
- the operation surface 70 is formed in a recess 80 provided in the remote operation device 100.
- the recess 80 is recessed in a rectangular shape from the surrounding surface 85 surrounding the recess 80.
- the recess 80 is formed with a first bottom surface 81 and a second bottom surface 83 provided with steps, and a plurality of side wall surfaces 82, 84, 86.
- the first bottom surface 81 and the second bottom surface 83 are formed in a planar shape along the xy plane.
- the first bottom surface 81 is formed at a position deeper from the surrounding surface 85 than the second bottom surface 83.
- the second bottom surface 83 is formed at an intermediate position between the first bottom surface 81 and the surrounding surface 85 in the z-axis direction. Further, the second bottom surface 83 is provided in the y-axis direction and in front of the vehicle traveling direction.
- the intermediate side wall surface 82 is formed between the first bottom surface 81 and the second bottom surface 83.
- the intermediate side wall surface 82 is inclined toward the outer peripheral side of the first bottom surface 81 as it is separated from the first bottom surface 81 along the z-axis direction.
- the front side wall surface 84 is formed between the second bottom surface 83 and the surrounding surface 85.
- the front side wall surface 84 is inclined toward the outer peripheral side of the second bottom surface 83 as it is separated from the second bottom surface 83 along the z-axis direction.
- the rear side wall surface 86 is formed between the first bottom surface 81 and the surrounding surface 85.
- the rear side wall surface 86 is inclined to the outer peripheral side of the first bottom surface 81 as it is separated from the first bottom surface 81 along the z-axis direction.
- the operation surface 70 formed in the above recess 80 is constituted by a first operation surface 72, a second operation surface 74, and the like.
- the first operation surface 72 is formed in a rectangular shape on the first bottom surface 81.
- the second operation surface 74 is formed in a rectangular shape on the second bottom surface 83 and is located away from the first operation surface 72 by a distance equal to or greater than a first threshold distance Dth1 (see FIG. 7A) described later.
- the area of the second operation surface 74 is narrower than that of the first operation surface 72.
- the touch sensor 31 is opposed to the entire area of the first operation surface 72 and the second operation surface 74 in the z-axis direction. As a result, the touch sensor 31 detects not only a moving operation for moving the finger F in a region facing the first operation surface 72 but also a moving operation for moving the finger F in a region facing the second operation surface 74. It is possible. Therefore, the second operation surface 74 covers a part of the capacitive touch sensor (31), and also functions as a guide unit that guides an input operation to the capacitive touch sensor (31). .
- the operation mode is switched depending on the relative position of the finger F that inputs the movement operation and the operation object distance d.
- toe F is changed.
- the deep operation mode is set when the finger F is located in the first operation space Sp1.
- the first operation space Sp1 is a space in which the operation body distance d is less than the first threshold distance Dth1, and is defined as a region facing the first operation surface 72.
- the first threshold distance Dth1 is slightly shorter than the size of the step between the first operation surface 72 and the second operation surface 74, and is set to about 0.5 to 1 cm, for example.
- the moving operation for moving the finger F along the xy plane is defined as “deep operation”, and the position of the focus 62 displayed on the display screen 52 can be moved. .
- the shallow operation mode As shown in FIG. 7B, the moving operation with the finger F moves the plurality of submenu images 164 on the display screen 52 in the horizontal direction (hereinafter referred to as “scroll”). Associated with the scrolling control.
- the deep operation mode is set when the finger F is located in the second operation space Sp2.
- the second operation space Sp2 is a space in which the operation body distance d is equal to or greater than the first threshold distance Dth1 and less than the second threshold distance Dth2, and is defined as a region facing the second operation surface 74.
- the second threshold distance Dth2 is slightly shorter than the depth from the surrounding surface 85 to the first operation surface 72, and is set to about 2 to 3 cm, for example.
- the display on the display screen 52 is switched to a state in which a plurality of submenu images 164 including the air conditioning menu image can be selected.
- the movement operation for moving the finger F along the xy plane in the second operation space Sp2 is defined as “shallow operation”, and a plurality of submenu images 164 displayed on the display screen 52 are displayed. It can be scrolled.
- the second operation surface 74 is positioned closer to the second operation space Sp2 than the first operation surface 72 by the setting of the first threshold distance Dth1 and the second threshold distance Dth2, and the second operation space. It becomes a form located in Sp2. Therefore, the operator can perform the shallow portion operation while bringing the finger F into contact with the second operation surface 74. In this way, the second operation surface 74 can assist input by the finger F in the second operation space Sp2.
- Non-proximity mode In the non-proximity mode, the moving operation with the finger F is not associated with any image unit 61 on the display screen 52.
- the non-proximity mode is set when the finger F is not positioned in either the first operation space Sp1 (see FIG. 7A) or the second operation space Sp2.
- the space excluding the first operation space Sp1 and the second operation space Sp2 is set as a non-proximity space.
- FIG. 8A shows a state in which the operator has started the operation of bringing the finger F closer to the operation surfaces 72 and 74.
- the operator who is going to start the icon selection operation moves the finger F, which is located far from the first operation surface 72 and the second operation surface 74, toward the operation surfaces 72 and 72.
- the remote control device 100 since the remote control device 100 is in the standby state, the movement related to the operation of the finger F is not performed with respect to the focus 62 and the like of the display screen 52.
- FIG. 8B shows a state in which the finger F is moved from the non-proximity space to the second operation space Sp2. From this state, when the operator inputs an operation of hitting the second operation surface 74 (hereinafter referred to as “tap operation”), the association between the shallow portion operation with the finger F and the scroll control is started. As the operation mode is changed to the shallow operation mode, the display on the display screen 52 is switched to a state in which the plurality of submenu images 164 can be scrolled.
- the operator inputs an operation of tracing the second operation surface 74 with the finger F to the second operation surface 74.
- the operator displays a submenu image 164 (for example, an air conditioning menu image) including an arbitrary icon 63 on the display screen 52 by a shallow operation in the second operation space Sp2 in which the finger F is moved along the x-axis direction. It can be moved to the center.
- a submenu image 164 for example, an air conditioning menu image
- FIG. 8D shows a state in which the finger F is moved from the second operation space Sp2 to the first operation space Sp1 on the first operation surface 72.
- the operation mode of the remote operation device 100 is switched from the shallow operation mode to the deep operation mode.
- the movement operation is associated with the focus control, and the sub menu image 164 (air conditioning menu image, see FIG. 8C) displayed in the center portion of the display screen 52 is displayed on the entire display screen 52. .
- the operator can superimpose the focus 62 on the arbitrary icon 63 by inputting a deep operation of tracing the first operation surface 72 with the finger F.
- the operator can select the arbitrary icon 63 by inputting a tap operation to the first operation surface 72 in a state where the focus 62 is superimposed on the arbitrary icon 63.
- the operator who has completed the selection operation of the icon 63 moves the finger F to the non-proximity space as shown in FIG. 8A. Thereby, the operation mode is switched to the non-proximity mode. Then, after the operation mode is switched to the non-proximity mode, when a predetermined threshold time Tth elapses, the remote operation device 100 enters a state of waiting for an icon selection operation by the operator from the next time.
- the operation mode selection process performed by the operation control unit 33 in order to realize the above icon selection operation will be described in detail with reference to FIGS.
- the operation mode selection process shown in FIG. 9 is started by the operation control unit 33 (see FIG. 1) when the ACC power supply of the vehicle is turned on.
- each section is expressed as S101, for example.
- each section can be divided into a plurality of subsections, while a plurality of sections can be combined into one section.
- each section configured in this manner can be referred to as a device, module, or means.
- each of the above sections or a combination thereof includes not only (i) a section of software combined with a hardware unit (eg, a computer), but also (ii) hardware (eg, an integrated circuit, As a section of (wiring logic circuit), it can be realized with or without the function of related devices.
- the hardware section can be configured inside the microcomputer.
- S ⁇ b> 101 the presence / absence of a tap operation on one of the first operation surface 72 and the second operation surface 74 is determined based on the change in output acquired from the touch sensor 31. If it is determined in S101 that there is no tap operation, the standby state of the remote operation device 100 is maintained by repeating the determination in S101. On the other hand, if it is determined in S101 that there is a tap operation, the process proceeds to S102.
- an acquisition process for acquiring a sensitivity value detected by each electrode of the touch sensor 31 is performed, and the process proceeds to S103.
- the x, y, and z coordinates (hereinafter, “input position coordinates”) indicating the position of the finger F in the three-dimensional direction with respect to the operation surface 32 are calculated from the sensitivity value acquired in S102. , The process proceeds to S104.
- the sensitivity value is a value that increases as the capacitance stored between the operation surface 32 and the finger F increases. Therefore, the coordinates in the x-axis direction and the y-axis direction at which the sensitivity value is maximized indicate the relative position of the finger F on the first operation surface 72 and the second operation surface 74.
- the sensitivity value decreases as the operating tool distance d (see FIG. 7A) decreases, and increases as the operating tool distance d increases. Therefore, the maximum sensitivity value corresponds to the operating tool distance d, and thus the coordinate in the z-axis direction (see FIG. 6C).
- the sensitivity threshold Hth1 corresponding to the first threshold distance Dth1 and the sensitivity threshold Hth2 corresponding to the second threshold distance Dth2 are stored in the operation control unit 33 (see FIG. 1) in advance. Has been.
- the operation control unit 33 performs processing for comparing the maximum sensitivity value acquired in S103 with the sensitivity threshold values Hth1 and Hth2.
- S104 it is determined whether or not the finger F is in the first operation space Sp1. That is, it is determined whether or not the relative position of the finger F is on the first operation surface 72 and the sensitivity value is greater than or equal to the first sensitivity threshold value Hth1. If an affirmative determination is made in S104, the process proceeds to S105. In S105, the operation mode is set to the deep operation mode, and the process returns to S102.
- S106 when a negative determination is made in S104, it is determined whether or not the relative position of the finger F is on the first operation surface 72 and the sensitivity value is less than the first sensitivity threshold Hth1. If a positive determination is made in S106, the process proceeds to S107. In S107, the operation mode is set to the non-proximity mode, and the process proceeds to S108. In S107, when the operation mode is switched from one of the deep operation mode and the shallow operation mode to the non-proximity mode, counting of the elapsed time t after shifting to the non-proximity mode is started.
- S108 it is determined whether or not the elapsed time t at which the counting was started in S107 of this time or S107 before the previous time is equal to or greater than a predetermined threshold time Tth. If a positive determination is made in S108, the process returns to S101. Thereby, the remote control device 100 shifts to a standby state. On the other hand, if a negative determination is made in S108, the process returns to S102.
- S109 when a negative determination is made in S106, it is determined whether or not the finger F is in the second operation space Sp2. That is, it is determined whether or not the relative position of the finger F is on the second operation surface 74 and the sensitivity value is less than the first sensitivity threshold value Hth1 and greater than or equal to the second sensitivity threshold value Hth2. If a negative determination is made in S109, it is estimated that the finger F is in the non-proximity space, and the process proceeds to S107 described above. On the other hand, if a positive determination is made in S109, the process proceeds to S110. In S110, the operation mode is set to the shallow operation mode, and the process returns to S102.
- the operation control unit 33 that has changed the operation mode in S105, S107, and S110 described above outputs a command signal notifying the change of the operation mode to the CAN bus 90 through the communication control unit 23 and the communication interface 24.
- the display control unit 53 that has acquired the command signal activates the drawing layer corresponding to each operation mode based on the signal.
- the operation control unit 33 sets the operation mode to the deep operation mode
- the display control unit 53 selects the focus layer L1 as an active drawing layer.
- the operation control unit 33 can change or operate the display of the focus 62 by associating the deep operation by the finger F with the focus 62 (focus control).
- the display mode including display content, display position, display color, display shape, display type, display specification, and display size
- the display mode including display content, display position, display color, display shape, display type, display specification, and display size
- the operation control unit 33 sets the operation mode to the shallow operation mode
- the display control unit 53 sets a submenu layer (not shown) on which a plurality of submenu images 164 are drawn as an active drawing layer. select.
- the operation control unit 33 associates the shallow portion operation with the finger F with the plurality of submenu images 164 and can change or operate the display of the submenu images 164 (scroll control).
- the display mode including display content, display position, display color, display shape, display type, display specification, and display size
- the display mode including display content, display position, display color, display shape, display type, display specification, and display size
- the operation control unit 33 sets the operation mode to the non-proximity mode
- the display control unit 53 sets the active drawing layer to “none”. As a result, the movement operation of the finger F is not associated with any image unit 61.
- the operator when inputting the shallow portion operation in the second operation space Sp2, the operator touches the second operation surface 74 with the finger F to assist the input. Can be received. Such assistance makes it easier for the operator to grasp the position of the second operation space Sp ⁇ b> 2 that is defined apart from the first operation surface 72. Therefore, the operator can easily perform input in the second operation space Sp2.
- the operator by moving the finger F back and forth between the first operation surface 72 and the second operation surface 74, the operator can select the image unit 61 to be operated as the focus 62 and the submenu. It can be switched in the image 164.
- the first operation space Sp1 and the second operation space Sp2 are divided forward and backward along the zx plane, so that the operator can more easily grasp the operation spaces Sp1 and Sp2. Therefore, the operator can perform a series of icon selections by blind operation without paying attention to the hand.
- the operator can use the surrounding surface 85 and the palm rest 39.
- Input to each of the operation surfaces 72 and 74 can be performed after stabilizing the hand. Therefore, the input in each operation space Sp1, Sp2 by an operator becomes still easier.
- the capacitive touch sensor 31 used in the first embodiment is substantially orthogonal to the first operation surface 72 as well as the movement of the finger F in the x and y axis directions along the first operation surface 72.
- the movement of the finger F in the z-axis direction can also be detected. Therefore, the touch sensor 31 described above is suitable as a configuration for detecting both the relative position of the finger F on the xy plane and the operating body distance d.
- the icon 63 in the selected sub menu image 164 can be selected.
- the display image 60 has a hierarchical structure that switches to a state in which a more detailed function can be selected by performing one selection.
- the physical depth in the recess 80 of each of the operation surfaces 72 and 74 operated by the finger F is set so as to correspond to the hierarchical depth of the image unit 61 associated with the operation by the finger F. The As a result, the operator can easily grasp the change of the display image 60 accompanying the input of the finger F sensuously.
- the touch sensor 31 and the operation control unit 33 are also referred to as “detection means” and “acquisition means” in cooperation.
- the operation control unit 33 is also referred to as “association means”.
- the first operation surface 72 is also referred to as an “operation surface”.
- the second operation surface 74 is also referred to as an “input auxiliary part” or a “guide part”.
- the first bottom surface 81 is also referred to as a “bottom surface”.
- the operator's finger F is also referred to as an “operation body”.
- the remote operation device 100 is also referred to as an “operation device”.
- the focus 62 is also referred to as “first image portion”, and the submenu image 164 is also referred to as “second image portion”. The “first image portion” and the “second image portion” may be changed as appropriate.
- the second embodiment of the present disclosure shown in FIGS. 12 to 14 is a modification of the first embodiment.
- the shape of the recess 280 is different from the shape of the recess 80 (see FIG. 6) of the first embodiment.
- the recess 280 of the second embodiment is recessed from the surrounding surface 85 into a two-stage cylindrical hole shape.
- the recess 280 is formed with a first bottom surface 281 and a second bottom surface 283 that are provided with steps, and two inner peripheral wall surfaces 282 and 284.
- the first bottom surface 281 and the second bottom surface 283 are formed in a planar shape along the xy plane.
- the first bottom surface 281 is formed in a substantially circular shape at a position deeper from the surrounding surface 85 than the second bottom surface 283.
- the second bottom surface 283 is formed on the outer peripheral side of the first bottom surface 281 and is formed in an annular shape that is coaxial with the first bottom surface 281.
- the second bottom surface 283 is located between the first bottom surface 281 and the surrounding surface 85 in the z-axis direction.
- the two inner peripheral wall surfaces 282 and 284 are both formed in a cylindrical surface shape. Of these inner peripheral wall surfaces 282 and 284, the deeper inner peripheral wall surface 282 that is closer to the first bottom surface 281 than the surrounding surface 85 connects the outer edge of the first bottom surface 281 and the inner edge of the second bottom surface 283. . On the other hand, the shallow inner circumferential wall surface 284 that is closer to the surrounding surface 85 than the first bottom surface 281 connects the outer edge of the second bottom surface 283 and the surrounding surface 85.
- Each of the inner peripheral wall surfaces 282 and 284 may be formed along the z-axis direction, or may be formed in a tapered shape whose diameter decreases from the surrounding surface 85 toward the first bottom surface 281.
- a first operation surface 272 and a second operation surface 274 are formed in the recess 280.
- the first operation surface 272 is formed in a circular shape on the first bottom surface 281.
- the second operation surface 274 is formed in an annular shape over the entire area of the second bottom surface 283.
- the second operation surface 274 extends in a belt shape in the circumferential direction along the outer edge portion of the first bottom surface 281.
- the touch sensor 231 is formed in a disk shape that faces the entire area of the first operation surface 272 and the second operation surface 274 in the z-axis direction. As a result, the touch sensor 231 detects not only the moving operation for moving the finger F in the region facing the first operation surface 272 but also the moving operation for moving the finger F in the region facing the second operation surface 274. It is possible.
- the first operation space Sp1 in the remote operation device 200 is defined as a range in which the operation pair distance d is less than the first threshold distance Dth1 in the region facing the first operation surface 272. Yes.
- the first threshold distance Dth1 is set to be slightly shorter than the size of the step between the first operation surface 272 and the second operation surface 274. Therefore, the first operation space Sp1 is defined as a region surrounded by the deep side inner peripheral wall surface 282.
- the second operation space Sp2 is defined as a range in which the operation pair distance d is not less than the first threshold distance Dth1 and less than the second threshold distance Dth2 in the region facing the second operation surface 274.
- the second threshold distance Dth2 is set to be slightly shorter than the depth from the surrounding surface 85 to the first operation surface 272. Therefore, the second operation space Sp ⁇ b> 2 is defined on the outer peripheral side of the deep portion inner peripheral wall surface 282 and in a region surrounded by the shallow portion inner peripheral wall surface 284.
- a display image 260 shown in FIG. 14B is an audio menu image for operating an audio device mounted on the vehicle.
- the audio menu image includes a track icon 263 associated with audio data, a pointer 262 for selecting the track icon 263, and a focus 62 that emphasizes the track icon 263.
- the operator in FIG. 14A inputs an operation of tracing the second operation surface 274 with the finger F to the second operation surface 274 in the second operation space Sp2.
- Such shallow operations by the operator are associated with the track icons 263 on the display screen 52, and these track icons 263 are integrally scrolled up and down (scroll control). Therefore, by inputting the shallow portion operation of tracing the second operation surface 274 with the finger F, the operator can move the arbitrary track icon 263 to the center of the display screen 52.
- FIG. 14B the operator moves the finger F from the second operation space Sp2 to the first operation space Sp1.
- a pointer 262 is displayed on the display screen 52.
- the operator in FIG. 14B inputs an operation of tracing the first operation surface 272 with the finger F to the first operation surface 272.
- Such deep operation by the operator is associated with the pointer 262 of the display screen 52, and enables a change in display mode for moving the pointer 262 (pointer control). Therefore, by inputting a deep operation of tracing the first operation surface 272 with the finger F, the operator can superimpose the pointer 262 on an arbitrary track icon 263.
- the track icon 263 on which the pointer 262 is superimposed is surrounded by the focus 62.
- the operator can select the arbitrary track icon 263 by inputting a tap operation to the first operation surface 272 in a state where the pointer 262 is superimposed on the arbitrary track icon 263.
- the operator can perform the shallow portion operation while touching the second operation surface 274. Since the position of the second operation space Sp2 can be grasped by the support of the second operation surface 274, input in the operation space Sp2 becomes easy.
- the second operation surface 274 extends in a strip shape while being curved along the outer edge of the first operation surface 272. Therefore, the operator can draw a substantially arc-shaped trajectory with the finger F while fixing the hand with either the palm rest 39 (see FIG. 2) or the surrounding surface 85 (see FIG. 12). In this way, by matching the shape of the second operation surface 274 with the shape of a trajectory in which the finger F can be easily moved, input by the operator's finger F becomes even easier.
- the first operation surface 272 is also referred to as “operation surface”.
- the second operation surface 274 is also referred to as an “input auxiliary part” or a “guide part”.
- the first bottom surface 281 is also referred to as a “bottom surface”.
- the pointer 262 is also referred to as “first image portion”.
- the track icon 263 is also referred to as “second image portion”. The “first image portion” and the “second image portion” may be changed as appropriate.
- the remote operation device 200 is also referred to as an “operation device”.
- the third embodiment of the present disclosure shown in FIGS. 15 to 17 is a modification of the second embodiment.
- the touch sensor 331 is formed in a disk shape that is opposed to the entire area of the first bottom surface 281 in the z-axis direction.
- the second bottom surface 283 forms a shallow operation guide surface 375 instead of the second operation surface 274 (see FIG. 13).
- the shallow operation guide surface 375 can assist the input by the finger F in the second operation space Sp2 described later by supporting the operator's hand.
- the shallow operation guide surface 375 is located in the vicinity of the virtual boundary surface BP between the first operation space Sp1 and the second operation space Sp2, and is formed along the boundary surface BP. With such a configuration, the shallow portion operation guide surface 375 can indicate the position of the boundary surface BP to the operator who performs the blind operation.
- the operation surface of the third embodiment that is substantially the same as the first operation surface 272 (see FIG. 13) of the second embodiment and is formed on the first bottom surface 281 is an operation surface 370. Yes.
- the touch sensor 331 detects a moving operation for moving the finger F in a region facing the operation surface 370.
- the first operation space Sp1 is defined as a range in which the operation pair distance d is less than the first threshold distance Dth1 in the region facing the operation surface 370.
- the second operation space Sp2 is defined as a range in which the operation pair distance d is not less than the first threshold distance Dth1 and less than the second threshold distance Dth2 in the region facing the operation surface 370. That is, the second operation space Sp ⁇ b> 2 is not formed in the region facing the second bottom surface 283.
- a contact operation mode and an aerial operation mode are defined as a plurality of operation modes together with the non-proximity mode.
- the contact operation mode is an operation mode when the finger F is positioned in the first operation space Sp1, and corresponds to the deep operation mode in the first embodiment.
- the aerial operation mode is an operation mode when the finger F is positioned in the second operation space Sp2, and corresponds to the shallow operation mode in the first embodiment.
- a display image 360 shown in FIG. 16B is a navigation image showing a route to the destination set by the operator.
- the navigation image includes a plurality of icons 63 associated with a predetermined function, a pointer 262 for selecting the icon 63, a map 364 indicating the form of roads around the vehicle, and the like.
- the display image 360 includes a focus 62 that emphasizes the icon 63 on which the pointer 262 is superimposed.
- the operator in FIG. 16A inputs the movement operation of the finger F in the second operation space Sp2 in a state where, for example, the thumb and the little finger other than the index finger F are placed on the shallow operation guide surface 375.
- the movement operation for moving the finger F along the xy plane in the second operation space Sp2 is defined as “aerial operation” and is associated with the map 364 displayed on the display screen 52, and
- the map 364 can be moved vertically and horizontally (map control). Therefore, the operator can move the map 364 by inputting an aerial operation so that the arbitrary icon 63 is displayed at an easily selectable position.
- the operator moves the finger F from the second operation space Sp2 to the operation surface 370 in the first operation space Sp1.
- a pointer 262 is displayed on the display screen 52.
- the operator in FIG. 16B inputs an operation of tracing the operation surface 370 with the finger F to the operation surface 370.
- the moving operation for moving the finger F along the xy plane in the first operation space Sp1 is defined as a “contact operation” and is associated with the pointer operation. Therefore, by inputting the contact operation of tracing the operation surface 370 with the finger F, the operator can superimpose the pointer 262 on the arbitrary icon 63. Under such a state, by inputting a tap operation to the operation surface 370, the operator can select an arbitrary icon 63 that is superimposed on the pointer 262 and surrounded by the focus 62.
- the presence / absence of a tap operation on the operation surface 370 is determined based on a change in output acquired from the touch sensor 331. If it is determined in S301 that there is no tap operation, the standby state of the remote operation device 300 is maintained by repeating the determination in S301. On the other hand, if it is determined in S301 that there has been a tap operation, the processes of S302 and S303 that are substantially the same as S102 and S103 (see FIG. 9) of the first embodiment are performed.
- S304 it is determined whether or not the finger F is in the first operation space Sp1. That is, it is determined whether or not the sensitivity value is greater than or equal to the first sensitivity threshold value Hth1. If an affirmative determination is made in S304, the process proceeds to S305. In S305, the operation mode is set to the contact operation mode, and the process returns to S302.
- S306 when a negative determination is made in S304, it is determined whether or not the finger F is in the second operation space Sp2. That is, it is determined whether or not the sensitivity value is less than the first sensitivity threshold value Hth1 and greater than or equal to the second sensitivity threshold value Hth2. If a positive determination is made in S306, the process proceeds to S307. In S307, the operation mode is set to the aerial operation mode, and the process returns to S302.
- S309 it is determined whether the elapsed time t at which the counting was started in S308 of this time or S308 before the previous time is equal to or greater than a predetermined threshold time Tth. If a positive determination is made in S309, the process returns to S301. Thereby, the remote control device 300 shifts to a standby state. On the other hand, if a negative determination is made in S309, the process returns to S302.
- the shallow operation guide surface 375 formed on the second bottom surface 283 is provided.
- the position of the boundary surface BP can be indicated. Therefore, the operator can easily grasp the position of the boundary surface BP during the blind operation. Therefore, the remote operation device 300 can acquire high operability for the aerial operation.
- the operator stabilizes the hand by touching the shallow operation guide surface 375 with a thumb, a little finger, etc., and performs an input with the finger F in the second operation space Sp2 adjacent to the shallow operation guide surface 375. It can be carried out. Even when the cost is reduced by suppressing the detection range of the touch sensor 331, the input in the second operation space Sp2 is further facilitated by exerting the auxiliary function by the shallow operation guide surface 375. It becomes possible.
- the shallow portion operation guide surface 375 is also referred to as an “input auxiliary portion”.
- the map 364 is also referred to as a “second image portion”.
- the “second image portion” may be changed as appropriate.
- the remote operation device 300 is also referred to as an “operation device”.
- the fourth embodiment of the present disclosure shown in FIG. 18 is a modification in which the control of the third embodiment is combined with the configuration of the recess 80 or the like in the first embodiment.
- the touch sensor 431 is smaller than the touch sensor 31 (see FIG. 6) of the first embodiment, and is rectangular with respect to the entire area of the first bottom surface 81 in the z-axis direction. It is formed into a shape.
- the touch sensor 431 detects a moving operation for moving the finger F in a region facing the operation surface 470 formed on the first bottom surface 81.
- the first threshold distance Dth1 is set to, for example, about 0.3 cm so as to be equal to or less than half of the depth from the second bottom surface 83 to the first bottom surface 81. Therefore, the first operation space Sp ⁇ b> 1 is defined as a range in the vicinity of the first bottom surface 81 in the region facing the operation surface 470. On the other hand, the second operation space Sp ⁇ b> 2 is expanded in the z-axis direction to include a range closer to the first bottom surface 81 than the second bottom surface 83.
- the shallow operation guide surface 475 formed on the second bottom surface 83 can indicate the position of the second operation space Sp2 to the operator who performs the blind operation. Therefore, the operator can reliably input the moving operation in the second operation space Sp2 by moving the finger F in accordance with the position of the shallow operation guide surface 475 in the z-axis direction. Therefore, the remote operation device 400 can acquire high operability for the aerial operation.
- the second operation space Sp2 is enlarged, input to the second operation space Sp2 by the blind operation is further facilitated.
- the operability of the remote operation device 400 can be further improved by the synergistic input assisting action of the set value of the first threshold distance Dth1 and the provision of the shallow operation guide surface 475.
- the shallow portion operation guide surface 475 is also referred to as an “input auxiliary portion”.
- the remote operation device 400 is also referred to as an “operation device”.
- the fifth embodiment of the present disclosure shown in FIG. 19 is another modification of the first embodiment.
- the recess 80 (see FIG. 5) is omitted from the remote control device 500 according to the fifth embodiment.
- the first operation surface 572 is formed on a flat surface having no step with the surrounding surface 85.
- the second operation surface 574 is formed by a movable member 576 in a region facing the first operation surface 572.
- the movable member 576 is a member that supports the second operation surface 574, and has a L-shaped cross section in the zy plane.
- the movable member 576 includes an input surface portion 578, a support column portion 579, and a hinge portion 577.
- the input surface portion 578 is formed in a rectangular plate shape along the xy plane and is spaced apart from the first operation surface 572. Of the both surfaces of the input surface portion 578, the upper surface on the side away from the first operation surface 572 forms the second operation surface 574.
- the column portion 579 extends toward the surrounding surface 85 from the front end separated from the palm rest 39 among the both ends of the input surface portion 578 in the y-axis direction.
- the support column 579 is formed in a rectangular plate shape along the zx plane.
- the hinge portion 577 is formed at the front end of the column portion 579 in the extending direction. The hinge 577 allows the movable member 576 to rotate around the x axis. Therefore, the movable member 576 can be tilted forward in the y-axis direction so as to be separated from the first operation surface 572 together with the second operation surface 574 (see the two-dot chain line in FIG. 19).
- the first threshold distance Dth1 is defined according to the height from the first operation surface 572 to the input surface portion 578. Therefore, the first operation space Sp1 is defined in a range lower than the input surface portion 578 in the z-axis direction in the region facing the first operation surface 572.
- the second operation space Sp2 is defined as a range in which the operation tool distance d is less than the second threshold distance Dth2 in the region facing the second operation surface 574.
- the remote operation device 500 can obtain high operability with respect to the input to the second operation surface 574.
- the second operation surface 574 is provided in a region facing the first operation surface 572. Therefore, an increase in the size of the remote operation device 500 due to the addition of the second operation surface 574 can be avoided. In addition, since the second operation surface 574 is separated from the first operation surface 572, a situation in which the area of the first operation surface 572 is narrowed can be avoided. Therefore, in the remote operation device 500, not only the operability of input in the second operation space Sp2, but also the operability of input in the first operation space Sp1 is ensured.
- the second operation surface 574 is retracted from the area facing the first operation surface 572 when the switching of the image unit 61 (see FIG. 7) to be operated is unnecessary. obtain. Therefore, when the second operation surface 574 is unnecessary, a situation in which the input to the first operation surface 572 is hindered can be avoided.
- the first operation surface 572 is also referred to as “operation surface”.
- the second operation surface 574 is also referred to as an “input auxiliary part” or a “guide part”.
- the movable member 576 is also referred to as a “movable support portion”.
- the remote operation device 500 is also referred to as an “operation device”.
- FIG. 20 The sixth embodiment of the present disclosure shown in FIG. 20 is a modification of the fifth embodiment.
- a recess 680 is formed in the remote control device 600 according to the sixth embodiment.
- a first operation surface 672 is formed on the bottom surface 681 of the recess 680.
- the second operation surface 674 is formed on a lid-like lid member 676 that covers a part of the opening of the recess 680.
- the second operation surface 674 is formed on the upper surface on the side away from the first operation surface 672 out of both surfaces of the lid member 676.
- the lid member 676 is a member that supports the second operation surface 674 and is formed in a plate shape along the xy plane.
- the lid member 676 is provided so as to be movable in the y-axis direction, and can be accommodated in the peripheral portion of the recess 680 by moving forward away from the palm rest 39 (see the two-dot chain line in FIG. 20). Accordingly, the lid member 676 can be separated from the first operation surface 672 together with the second operation surface 674, and the second operation surface 674 can be retracted from the region facing the first operation surface 672.
- the lid member 676 is provided with a claw portion 679 for facilitating movement.
- the claw portion 679 protrudes from the rear end of the lid member 676 in the y-axis direction in a direction away from the bottom surface 681 along the z-axis direction. The operator can pull out the accommodated lid member 676 rearward in proximity to the palm rest 39 by placing the finger F on the claw portion 679.
- the first threshold distance Dth1 is defined according to the height from the first operation surface 672 to the lid member 676. Therefore, the first operation space Sp ⁇ b> 1 is defined in a range lower than the lid member 676 in the z-axis direction in the region facing the first operation surface 672.
- the second operation space Sp2 is defined as a range in which the operation tool distance d is less than the second threshold distance Dth2 in the region facing the second operation surface 674.
- the remote operation device 600 can obtain high operability with respect to the input to the second operation surface 674.
- the second operation surface 674 can be retracted from the region facing the first operation surface 672. Therefore, when the second operation surface 674 is unnecessary, a situation in which the input to the first operation surface 672 is hindered can be avoided.
- the first operation surface 672 is also referred to as an “operation surface”.
- the second operation surface 674 is also referred to as an “input auxiliary part” or a “guide part”.
- the lid member 676 is also referred to as a “movable part”.
- the remote operation device 600 is also referred to as an “operation device”.
- the second operation surface 574 is provided on a fixing member 776 erected from the surrounding surface 85.
- the fixing member 776 has an input surface portion 578 and a column portion 779.
- the column portion 779 is fixed to the surrounding surface 85.
- the second operation surface 574 may not be retractable from the region facing the first operation surface 572.
- a push button for arbitrarily selecting the icon 63 is provided in the vicinity of the recess 80 or the like. The operator can select the function of the icon 63 by pressing the push button with the pointer 262 and the focus 62 superimposed on the arbitrary icon 63.
- a pressure-sensitive touch sensor that detects an operation with the finger F by detecting the pressure applied by the finger F is used as the “detecting means”.
- an infrared sensor that measures the operating tool distance d by infrared rays is used as “acquiring means”.
- a camera that captures the vicinity of each operation surface and an image analysis circuit that acquires the operating object distance d by analyzing an image captured by the camera are used as “acquisition means”. Yes.
- a display using a plasma display panel, a display using an organic EL, and the like form a “display screen”.
- the navigation device 50 includes the display screen 52 as a configuration.
- hysteresis is provided for the first threshold distance Dth1 and the second threshold distance Dth2.
- the first threshold distance Dth1 is extended based on the finger F being positioned in the first operation space Sp1.
- the second threshold distance Dth2 is extended based on the finger F being located in the second operation space Sp2.
- the plurality of “first image portion” and “second image portion” exemplified may be changed as appropriate.
- a plurality of functions also referred to as sections, means, or devices
- functions such as “association section” and “acquisition section” may be provided by a hardware circuit that performs a predetermined function without depending on a program.
- the present disclosure is applied to a remote control device used in a display system mounted on a vehicle.
- the present disclosure can also be applied to a so-called touch panel type operation device configured integrally with a display screen.
- the operation device to which the present disclosure is applied can be adopted not only for vehicles but also for all display systems used for various transportation equipment and various information terminals.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Automation & Control Theory (AREA)
- User Interface Of Digital Computer (AREA)
- Position Input By Displaying (AREA)
- Navigation (AREA)
Abstract
Description
本開示の第一実施形態による遠隔操作装置100は、車両に搭載され、図1に示されるように、ナビゲーション装置50等と共に表示システム10を構成している。遠隔操作装置100は、図2に示されるように、車両のセンターコンソールにてパームレスト39と隣接する位置に設置され、操作者の手の届き易い範囲に操作面70を露出させている。この操作面70には、操作者の手の人差し指(以下、単に「指」という)F等による操作が入力される。ナビゲーション装置50は、操作者から目視されるよう表示画面52を露出させ且つ表示画面52を運転席から見られる姿勢にて、車両のインスツルメントパネル内に設置されている。この表示画面52には、種々の表示画像60が表示される。
深部操作モードでは、図7Aに示すように、指Fによる移動操作は、表示画面52のフォーカス62を移動させるフォーカス制御に関連付けられる。こうした深部操作モードとされるのは、第一操作空間Sp1内に指Fが位置していた場合である。この第一操作空間Sp1は、操作体距離dが第一閾値距離Dth1未満となる空間であって、第一操作面72と対向する領域に規定されている。第一閾値距離Dth1は、第一操作面72及び第二操作面74間の段差の大きさよりも僅かに短く、例えば0.5~1cm程度に設定されている。
浅部操作モードでは、図7Bに示すように、指Fによる移動操作は、表示画面52の複数のサブメニュー画像164を水平方向に移動(以下、「スクロール」という)させるスクロール制御に関連付けられる。こうした深部操作モードとされるのは、第二操作空間Sp2内に指Fが位置していた場合である。この第二操作空間Sp2は、操作体距離dが第一閾値距離Dth1以上且つ第二閾値距離Dth2未満となる空間であって、第二操作面74と対向している領域に規定されている。第二閾値距離Dth2は、囲繞面85から第一操作面72までの深さよりも僅かに短くされており、例えば2~3cm程度に設定されている。
非近接モードでは、指Fによる移動操作は、表示画面52のいずれの画像部61とも関連付けられない。こうした非近接モードとされるのは、第一操作空間Sp1(図7A参照)及び第二操作空間Sp2のいずれにも指Fが位置していない場合である。このように、第一操作空間Sp1及び第二操作空間Sp2を除く空間を、非近接空間とする。
図12~14に示される本開示の第二実施形態は、第一実施形態の変形例である。第二実施形態による遠隔操作装置200では、凹所280の形状が、第一実施形態の凹所80(図6参照)の形状と異なっている。図12,13に示すように、第二実施形態の凹所280は、囲繞面85から二段筒孔状に凹んでいる。凹所280には、互いに段差を設けられた第一底面281及び第二底面283と、二つの内周壁面282,284とが形成されている。
図15~17に示される本開示の第三実施形態は、第二実施形態の変形例である。図15に示すように、第三実施形態による遠隔操作装置300において、タッチセンサ331は、第一底面281の全域とz軸方向において相対する円盤状に形成されている。こうしたタッチセンサ331の縮小により、第二底面283は、第二操作面274(図13参照)に替えて、浅部操作ガイド面375を形成している。浅部操作ガイド面375は、操作者の手を支持することで、後述する第二操作空間Sp2内での指Fによる入力を補助することができる。浅部操作ガイド面375は、第一操作空間Sp1と第二操作空間Sp2との間における仮想の境界面BPの近傍に位置し、当該境界面BPに沿って形成されている。こうした構成により、浅部操作ガイド面375は、ブラインド操作を行う操作者に、境界面BPの位置を示すことができる。一方で、第二実施形態の第一操作面272(図13参照)と実質的に同一であって、第一底面281に形成される第三実施形態の操作面は、操作面370とされている。この操作面370と相対する領域にて指Fを移動させる移動操作を、タッチセンサ331は検出する。
図18に示される本開示の第四実施形態は、第一実施形態における凹所80等の構成に、第三実施形態の制御を組み合わせた変形例である。第四実施形態による遠隔操作装置400において、タッチセンサ431は、第一実施形態のタッチセンサ31(図6参照)よりも縮小されており、第一底面81の全域とz軸方向において相対する矩形状に形成されている。タッチセンサ431は、第一底面81に形成された操作面470と対向する領域にて指Fを移動させる移動操作を検出する。
図19に示される本開示の第五実施形態は、第一実施形態の別の変形例である。第五実施形態による遠隔操作装置500からは、凹所80(図5参照)が省略されている。これにより、第一操作面572は、囲繞面85と段差のない平面上に形成されている。
図20に示される本開示の第六実施形態は、第五実施形態の変形例である。第六実施形態による遠隔操作装置600には、凹所680が形成されている。そして、この凹所680の底面681には、第一操作面672が形成されている。一方、第二操作面674は、凹所680の開口の一部を覆う蓋状の蓋部材676に形成されている。第二操作面674は、蓋部材676の両面のうち、第一操作面672から離間した側となる上面に形成されている。
以上、本開示による複数の実施形態について説明したが、本開示は、上記実施形態に限定して解釈されるものではなく、本開示の要旨を逸脱しない範囲内において種々の実施形態及び組み合わせに適用することができる。
Claims (12)
- 操作体(F)による入力が、操作面(72,272,370,470,572,672)に対してなされることによって、表示画面(52)に表示された画像(61)を操作する操作装置であって、
前記操作体の移動を検出する検出手段(31,33)と、
前記操作面から前記操作体までの操作体距離(d)を取得する取得手段(31,33)と、
前記操作体距離が予め規定された閾値距離(Dth1)未満となる第一操作空間(Sp1)内において検出された前記操作体の第一移動と、前記操作体距離が前記閾値距離を超える第二操作空間(Sp2)内で検出された前記操作体の第二移動とを区別し、
当該操作体の前記第一移動と前記第二移動を、前記表示画面上に表示された第一画像部(62、262)、と前記第一画像部とは異なる第二画像部(164,263,364)に、それぞれ、関連付けて、前記第一画像部もしくは前記第二画像部の少なくとも一方の表示態様を変化させる関連付け手段(33,S105,S110,S305,S307)と、
前記操作面よりも前記閾値距離以上離れた位置に形成された入力補助部(74,274,375,475,574,674)と、を備えることを特徴とする操作装置(100、200、300、400、500、600)。 - 前記入力補助部は、前記第二操作空間内に位置し、
前記検出手段は、前記入力補助部に沿った前記操作体の前記第二移動をさらに検出することを特徴とする請求項1に記載の操作装置。 - 前記入力補助部は、前記第一操作空間と前記第二操作空間との間における仮想の境界面(BP)に沿って形成され、
前記関連付け手段は、前記操作面と対向し且つ前記入力補助部と隣接する前記第二操作空間内での前記操作体の前記第二移動を、前記第二画像部に関連付けることを特徴とする請求項1に記載の操作装置。 - 前記境界面は、前記入力補助部よりも前記操作面に近接していることを特徴とする請求項3に記載の操作装置。
- 前記入力補助部は、前記操作面と対向する領域内に設けられ、且つ当該操作面から離間していることを特徴とする請求項1~4のいずれか一項に記載の操作装置。
- 前記入力補助部は、前記操作面の外縁に沿う帯状に形成されることを特徴とする請求項1~5のいずいれか一項に記載の操作装置。
- 前記入力補助部を支持し、当該入力補助部と共に前記操作面から離間する方向に移動する可動支持部(576,676)、をさらに備えることを特徴とする請求項1~6のいずれか一項に記載の操作装置。
- 前記操作面は、前記操作装置に設けられた凹所(80,280,680)の底面(81,281,681)に形成されることを特徴とする請求項1~7のいずれか一項に記載の操作装置。
- 前記検出手段は、前記操作体の移動と共に前記操作体距離を検出することにより、前記取得手段を兼ねることを特徴とする請求項1~8のいずれか一項に記載の操作装置。
- 前記入力補助部が、前記操作面から前記閾値距離以上離れた直上方領域と、該直上方領域と隣接する隣接領域の少なくとも何れか一方に位置していることを特徴とする請求項1に記載の操作装置。
- 平面状の静電容量式タッチパネル(31,231)と、
前記静電容量式タッチパネルの一部領域を覆い、前記静電容量式タッチパネルへの入力操作をガイドするガイド部(74,274,574,674)と
を備えたことを特徴とする操作装置(100、200、500、600)。 - 操作体(F)による入力が、操作面(72,272,370,470,572,672)に対してなされることによって、表示画面(52)に表示された画像(61)を操作する操作装置であって、
前記操作体の移動を検出する手段(31,33)と、
前記操作面から前記操作体までの操作体距離(d)を取得する手段(31,33)と、
前記操作体距離が予め規定された閾値距離(Dth1)未満となる第一操作空間(Sp1)内において検出された前記操作体の移動と、前記操作体距離が前記閾値距離を超える第二操作空間(Sp2)内で検出された前記操作体の移動とを区別して、当該操作体の前記第一移動と前記第二移動を、前記表示画面上に表示された第一画像部(62、262)と前記第一画像部とは異なる第二画像部(164,263,364)のそれぞれに関連付けて、前記第一画像部もしくは前記第二画像部の少なくとも一方の表示態様を変化させる手段(33,S105,S110,S305,S307)と、
前記操作体の前記第二空間内の前記第二移動を補助する手段(74,274,375,475,574,674)と、
を備えることを特徴とする操作装置(100、200、300、400、500、600)。
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/427,732 US20150242102A1 (en) | 2012-10-02 | 2013-08-02 | Manipulating apparatus |
CN201380051769.XA CN104704459B (zh) | 2012-10-02 | 2013-08-02 | 操作装置 |
DE112013004843.4T DE112013004843T5 (de) | 2012-10-02 | 2013-08-02 | Manipulationsvorrichtung |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012-220661 | 2012-10-02 | ||
JP2012220661A JP5751233B2 (ja) | 2012-10-02 | 2012-10-02 | 操作デバイス |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014054208A1 true WO2014054208A1 (ja) | 2014-04-10 |
Family
ID=50434556
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/004701 WO2014054208A1 (ja) | 2012-10-02 | 2013-08-02 | 操作装置 |
Country Status (5)
Country | Link |
---|---|
US (1) | US20150242102A1 (ja) |
JP (1) | JP5751233B2 (ja) |
CN (1) | CN104704459B (ja) |
DE (1) | DE112013004843T5 (ja) |
WO (1) | WO2014054208A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106605190A (zh) * | 2014-08-27 | 2017-04-26 | 株式会社电装 | 车辆用触摸面板以及车辆用输入接口 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5790578B2 (ja) * | 2012-04-10 | 2015-10-07 | 株式会社デンソー | 表示システム、表示装置、及び操作装置 |
JP6179253B2 (ja) | 2013-08-02 | 2017-08-16 | 株式会社デンソー | 入力装置 |
EP3343286B1 (en) * | 2016-12-27 | 2021-08-18 | Canon Kabushiki Kaisha | Display control apparatus and control method therefor |
JP2019211979A (ja) * | 2018-06-04 | 2019-12-12 | 本田技研工業株式会社 | 表示装置、表示制御方法およびプログラム |
CN111666029B (zh) * | 2020-05-28 | 2024-09-20 | 阿波罗智联(北京)科技有限公司 | 车机地图的操作方法、装置、设备和可读存储介质 |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003312373A (ja) * | 2002-04-24 | 2003-11-06 | Toyota Motor Corp | 入力装置 |
JP2007164695A (ja) * | 2005-12-16 | 2007-06-28 | Toyota Motor Corp | 車載遠隔操作装置 |
JP2007287118A (ja) * | 2006-03-23 | 2007-11-01 | Wacom Co Ltd | 情報処理装置、操作入力方法及びセンス部品 |
JP3143462U (ja) * | 2007-05-15 | 2008-07-24 | 宏達國際電子股▲ふん▼有限公司 | 切り替え可能なユーザ・インターフェースを有する電子デバイスおよび便利なタッチ操作機能を有する電子デバイス |
JP2008541222A (ja) * | 2005-06-02 | 2008-11-20 | サムスン エレクトロニクス カンパニー リミテッド | ユーザ命令を3次元的に入力可能な電子装置 |
JP2010191892A (ja) * | 2009-02-20 | 2010-09-02 | Sony Corp | 情報処理装置、表示制御方法、及びプログラム |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7903094B2 (en) * | 2006-06-23 | 2011-03-08 | Wacom Co., Ltd | Information processing apparatus, operation input method, and sensing device |
US20100079413A1 (en) * | 2008-09-29 | 2010-04-01 | Denso Corporation | Control device |
CA2674663A1 (en) * | 2008-10-08 | 2010-04-08 | Research In Motion Limited | A method and handheld electronic device having dual mode touchscreen-based navigation |
JP4793422B2 (ja) * | 2008-10-10 | 2011-10-12 | ソニー株式会社 | 情報処理装置、情報処理方法、情報処理システムおよび情報処理用プログラム |
WO2010089980A1 (ja) * | 2009-02-06 | 2010-08-12 | パナソニック株式会社 | 画像表示装置 |
KR101560718B1 (ko) * | 2009-05-29 | 2015-10-15 | 엘지전자 주식회사 | 이동 단말기 및 이동 단말기에서의 정보 표시 방법 |
KR101092722B1 (ko) * | 2009-12-02 | 2011-12-09 | 현대자동차주식회사 | 차량의 멀티미디어 시스템 조작용 사용자 인터페이스 장치 |
JP2011150413A (ja) * | 2010-01-19 | 2011-08-04 | Sony Corp | 情報処理装置、操作入力方法及び操作入力プログラム |
US20110292268A1 (en) * | 2010-05-26 | 2011-12-01 | T-Mobile Usa, Inc. | Multi-region touchpad device |
JP2012058921A (ja) * | 2010-09-07 | 2012-03-22 | Sony Corp | 情報処理装置、情報処理方法およびプログラム |
JP5649169B2 (ja) * | 2010-11-22 | 2015-01-07 | インターナショナル・ビジネス・マシーンズ・コーポレーションInternational Business Machines Corporation | タッチパネルにおけるドラッグ操作でオブジェクトを移動させる方法、装置及びコンピュータプログラム |
JP2012118675A (ja) * | 2010-11-30 | 2012-06-21 | Pentel Corp | 静電容量型タッチパネル用入力補助具 |
KR101781908B1 (ko) * | 2011-03-24 | 2017-09-26 | 엘지전자 주식회사 | 이동 단말기 및 그것의 제어방법 |
-
2012
- 2012-10-02 JP JP2012220661A patent/JP5751233B2/ja active Active
-
2013
- 2013-08-02 WO PCT/JP2013/004701 patent/WO2014054208A1/ja active Application Filing
- 2013-08-02 CN CN201380051769.XA patent/CN104704459B/zh not_active Expired - Fee Related
- 2013-08-02 US US14/427,732 patent/US20150242102A1/en not_active Abandoned
- 2013-08-02 DE DE112013004843.4T patent/DE112013004843T5/de not_active Ceased
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003312373A (ja) * | 2002-04-24 | 2003-11-06 | Toyota Motor Corp | 入力装置 |
JP2008541222A (ja) * | 2005-06-02 | 2008-11-20 | サムスン エレクトロニクス カンパニー リミテッド | ユーザ命令を3次元的に入力可能な電子装置 |
JP2007164695A (ja) * | 2005-12-16 | 2007-06-28 | Toyota Motor Corp | 車載遠隔操作装置 |
JP2007287118A (ja) * | 2006-03-23 | 2007-11-01 | Wacom Co Ltd | 情報処理装置、操作入力方法及びセンス部品 |
JP3143462U (ja) * | 2007-05-15 | 2008-07-24 | 宏達國際電子股▲ふん▼有限公司 | 切り替え可能なユーザ・インターフェースを有する電子デバイスおよび便利なタッチ操作機能を有する電子デバイス |
JP2010191892A (ja) * | 2009-02-20 | 2010-09-02 | Sony Corp | 情報処理装置、表示制御方法、及びプログラム |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106605190A (zh) * | 2014-08-27 | 2017-04-26 | 株式会社电装 | 车辆用触摸面板以及车辆用输入接口 |
CN106605190B (zh) * | 2014-08-27 | 2019-06-28 | 株式会社电装 | 车辆用触摸面板以及车辆用输入接口 |
Also Published As
Publication number | Publication date |
---|---|
US20150242102A1 (en) | 2015-08-27 |
CN104704459B (zh) | 2017-10-13 |
CN104704459A (zh) | 2015-06-10 |
JP2014074955A (ja) | 2014-04-24 |
DE112013004843T5 (de) | 2015-06-18 |
JP5751233B2 (ja) | 2015-07-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9489500B2 (en) | Manipulation apparatus | |
US9778764B2 (en) | Input device | |
US9411424B2 (en) | Input device for operating graphical user interface | |
WO2014054208A1 (ja) | 操作装置 | |
JP5656046B2 (ja) | 車両用情報表示装置 | |
JP5858059B2 (ja) | 入力装置 | |
JP2011116304A (ja) | ユーザインタフェース装置 | |
WO2013153750A1 (ja) | 表示システム、表示装置、及び操作装置 | |
JP5754483B2 (ja) | 操作デバイス | |
JP5954145B2 (ja) | 入力装置 | |
JP5772804B2 (ja) | 操作デバイス、及び操作デバイスの操作教示方法 | |
WO2016031152A1 (ja) | 車両用入力インターフェイス | |
JP2008009596A (ja) | 入力装置 | |
WO2014162698A1 (ja) | 入力装置 | |
WO2016031148A1 (ja) | 車両用タッチパッドおよび車両用入力インターフェイス | |
JP5966125B2 (ja) | 車両用情報表示装置 | |
JP5921698B2 (ja) | 表示装置 | |
JP2018010583A (ja) | 操作支援装置及びコンピュータプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13843232 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14427732 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 112013004843 Country of ref document: DE Ref document number: 1120130048434 Country of ref document: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 13843232 Country of ref document: EP Kind code of ref document: A1 |