US20150169160A1 - Input apparatus, control apparatus, control system, electronic apparatus, and control method - Google Patents
Input apparatus, control apparatus, control system, electronic apparatus, and control method Download PDFInfo
- Publication number
- US20150169160A1 US20150169160A1 US14/572,884 US201414572884A US2015169160A1 US 20150169160 A1 US20150169160 A1 US 20150169160A1 US 201414572884 A US201414572884 A US 201414572884A US 2015169160 A1 US2015169160 A1 US 2015169160A1
- Authority
- US
- United States
- Prior art keywords
- scroll
- screen
- input
- movement
- user
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
Definitions
- the present invention relates to an input apparatus for operating a GUI (Graphical User Interface), a control apparatus for controlling the GUI in accordance with information transmitted from the input apparatus, a control system including those apparatuses, an electronic apparatus, and a control method.
- GUI Graphic User Interface
- Pointing devices are used as controllers for GUIs widely used in PCs (Personal Computers). Not just as HIs (Human Interfaces) of PCs of the related art, the GUIs are now starting to be used as interfaces for AV equipment and game machines used in living rooms etc. with, for example, televisions as image media.
- Various pointing devices that a user is capable of operating 3-dimensionally are proposed as controllers for the GUIs of this type (see, for example, Japanese Patent Application Laid-open No. 2001-56743 (paragraphs (0030) and (0031), FIG. 3; hereinafter, referred to as Patent Document 1) and Japanese Examined Patent Publication No. Hei 6-7371 (P. 3, 11.18-20 on left-hand column; hereinafter, referred to as Patent Document 2)).
- Patent Document 1 discloses an input apparatus including angular velocity gyroscopes of two axes, that is, two angular velocity sensors.
- the angular velocity sensors detect angular velocities about two orthogonal axes, and a signal as positional information of a cursor or the like displayed by a display means is generated in accordance with the angular velocities.
- the signal is transmitted to a control apparatus, and the control apparatus controls display so that the cursor moves on a screen in response to the signal.
- Patent Document 2 discloses an input apparatus (space mouse) including three acceleration sensors (of three axes) and three angular velocity sensors (of three axes) (gyro).
- an input apparatus it is also possible to structure an input apparatus, a control apparatus, and the like so that an image displayed on a screen is scrolled when the input apparatus is operated 3-dimensionally.
- the image displayed on the screen is scrolled in accordance with a 3-dimensional operation of the input apparatus.
- an input apparatus including a detection means, a change means, and a transmission means.
- the detection means detects a movement amount of a user operation in an arbitrary direction.
- the change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- the transmission means transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- a scroll direction of the image can be biased in directions such as a horizontal-axis direction and a vertical-axis direction on the screen.
- an image can be prevented from being scrolled in a direction unintended by a user on the screen, with the result that an operational feeling for the user in scrolling an image can be improved.
- the input apparatus may further include a judgment means.
- the judgment means judges a direction of the user operation based on the detected detection value.
- the change means may change the ratio of the first movement amount to the second movement amount in accordance with the judged direction of the user operation.
- the scroll direction of the image can be biased appropriately in accordance with a direction of the user operation.
- the change means may change the ratio of the first movement amount to the second movement amount so that a scroll direction of the image is biased in at least the first direction on the screen and the second direction on the screen.
- scroll direction of the image can be biased in the first direction and the second direction on the screen in the embodiment of the present invention, an operational feeling in scrolling an image can be additionally improved.
- the change means may change the ratio so that, when the judged direction of the user operation is within a first angle range from the first operation direction, the scroll direction is biased in the first direction, and change the ratio so that, when the judged direction of the user operation is within a second angle range from the second operation direction, the scroll direction is biased in the second direction.
- the scroll direction can be biased in the first direction on the screen.
- the scroll direction can be biased in the second direction on the screen.
- the input apparatus may further include an angle range control means.
- the angle range control means variably controls the first angle range and the second angle range.
- the angle range control means may variably control the first angle range and the second angle range in accordance with the direction of the user operation.
- the first angle range and the second angle range can be changed appropriately in accordance with a direction of the user operation.
- the angle range control means may control the first angle range and the second angle range so that the first angle range is widened when the direction of the user operation is within a first modified angle range from the first operation direction and the second angle range is widened when the direction of the user operation is within a second modified angle range from the second operation direction.
- the second angle range may be wider than the first angle range.
- the change means may change the ratio of the first movement amount to the second movement amount so that the scroll direction of the image is restricted to at least the first direction on the screen and the second direction on the screen.
- the change means may change the ratio of the first movement amount to the second movement amount so that the scroll direction of the image is restricted to directions that respectively form predetermined angles with respect to the first direction on the screen and the second direction on the screen.
- the detection means may be a sensor that detects the user operation in space.
- a control apparatus controlling display of scroll of an image displayed on a screen in accordance with information transmitted from an input apparatus including a detection means for detecting a movement amount of a user operation in an arbitrary direction and a transmission means for transmitting the information on a related value related to a detection value detected by the detection means, the control apparatus including a reception means, a change means, and a display control means.
- the reception means receives the information.
- the change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- the display control means controls the display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- the “related value related to a detection value” may be a detection value itself or an operational value calculated based on the detection value.
- a scroll direction of the image can be biased in directions including the first direction and the second direction on the screen.
- an image can be prevented from being scrolled in a direction unintended by the user on the screen, with the result that an operational feeling for the user in scrolling an image can be improved.
- a control system including an input apparatus and a control apparatus.
- the input apparatus includes a detection means, a change means, and a transmission means.
- the detection means detects a movement amount of a user operation in an arbitrary direction.
- the change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- the transmission means transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- the control apparatus includes a reception means and a display control means.
- the reception means receives the scroll information.
- the display control means controls display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- a control system including an input apparatus and a control apparatus.
- the input apparatus includes a detection means and a transmission means.
- the detection means detects a movement amount of a user operation in an arbitrary direction.
- the transmission means transmits information on a related value related to a detection value detected by the detection means.
- the control apparatus includes a reception means, a change means, and a display control means.
- the reception means receives the information.
- the change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- the display control means controls display on the screen so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- an electronic apparatus including a display section, a detection means, a change means, and a display control means.
- the display section displays a screen.
- the detection means detects a movement amount of a user operation in an arbitrary direction.
- the change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- the display control means controls display on the screen so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- a control method including detecting a movement amount of a user operation in an arbitrary direction.
- a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction is changed, the first movement amount and the second movement amount corresponding to a detection value detected.
- Display on the screen is controlled so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- an input apparatus including a detection section, a change section, and a transmission section.
- the detection section detects a movement amount of a user operation in an arbitrary direction.
- the change section changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection section.
- the transmission section transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- a control apparatus controlling display of scroll of an image displayed on a screen in accordance with information transmitted from an input apparatus including a detection means for detecting a movement amount of a user operation in an arbitrary direction and a transmission means for transmitting the information on a related value related to a detection value detected by the detection means, the control apparatus including a reception section, a change section, and a display control section.
- the reception section receives the information.
- the change section changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- the display control section controls the display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- elements described as “ . . . means” may be realized by hardware, or may be realized by both software and hardware.
- the hardware includes at least a storage device for storing a software program.
- the hardware is constituted by selectively using at least one of a sensor, a CPU (Central Processing Unit), an MPU (Micro Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), a DSP (Digital Signal Processor), an FPGA (Field Programmable Gate Array), an ASIC (Application Specific Integrated Circuit), a NIC (Network Interface Card), a WNIC (Wireless NIC), a modem, an optical disc, a magnetic disk, and a flash memory.
- a CPU Central Processing Unit
- MPU Micro Processing Unit
- RAM Random Access Memory
- ROM Read Only Memory
- DSP Digital Signal Processor
- FPGA Field Programmable Gate Array
- ASIC Application Specific Integrated Circuit
- NIC Network Interface Card
- WNIC Wireless NIC
- modem an optical disc, a magnetic disk, and a flash memory.
- an input apparatus a control apparatus, a control system, an electronic apparatus, and a control method that are capable of improving an operational feeling in scrolling an image displayed on a screen can be provided.
- FIG. 1 is a diagram showing a control system according to an embodiment of the present invention
- FIG. 2 is a perspective diagram showing an input apparatus
- FIG. 3 is a diagram schematically showing an internal structure of the input apparatus
- FIG. 4 is a block diagram showing an electrical structure of the input apparatus
- FIG. 5 is a diagram showing an example of a screen displayed on a display apparatus
- FIG. 6 is a diagram showing a state where a user is holding the input apparatus
- FIGS. 7A and 7B are explanatory diagrams showing typical examples of ways of moving the input apparatus and ways a pointer moves on a screen accordingly;
- FIG. 8 is a perspective diagram showing a sensor unit
- FIG. 9 is a diagram for explaining an operation of the control system that is carried out when the pointer moves on the screen in accordance with a 3-dimensional operation made by the user (pointer mode);
- FIG. 10 is a flowchart showing an operation of the input apparatus according to the embodiment of the present invention.
- FIGS. 11A and 11B are diagrams for explaining relationships between weighting factors ⁇ and ⁇ and scroll tilt directions
- FIG. 12 is a diagram showing an operation of the input apparatus according to another embodiment of the present invention.
- FIGS. 13A and 13B are diagrams showing relationships between operation directions of the input apparatus and scroll directions in a case where the processing shown in FIG. 12 is executed;
- FIG. 14 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention.
- FIG. 15 is a diagram for explaining a first angle range and a second angle range
- FIGS. 16A and 16B are diagrams showing relationships between the operation directions of the input apparatus and scroll directions in a case where the processing shown in FIG. 14 is executed;
- FIG. 17 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention.
- FIGS. 18A and 183 are diagrams showing temporal changes of ranges of the first angle range and the second angle range in a case where the processing shown in FIG. 17 is executed;
- FIG. 19 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention.
- FIG. 20 is a diagram for explaining a first modified angle range and a second modified angle range
- FIG. 21 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention.
- FIG. 22 is a diagram for explaining a third angle range
- FIG. 23 are diagrams showing relationships between the operation directions of the input apparatus and scroll directions in a case where the processing shown in FIG. 21 is executed;
- FIGS. 24A and 24B are diagrams each showing a relationship between an operation direction of the input apparatus and a direction in which an image is scrolled;
- FIG. 25 is a flowchart showing an operation of the input apparatus of the control system according to another embodiment of the present invention.
- FIG. 26 is a diagram showing an image and a small-size screen displayed on the screen
- FIG. 27 is a diagram showing an image and scrollbars displayed on the screen.
- FIG. 28 is a diagram showing an image and a reference point displayed on the screen.
- FIG. 1 is a diagram showing a control system according to a first embodiment of the present invention.
- a control system 100 includes a display apparatus 5 , a control apparatus 40 , and an input apparatus 1 .
- FIG. 2 is a perspective diagram showing the input apparatus 1 .
- the input apparatus 1 is of a size that a user is capable of holding.
- the input apparatus 1 includes a casing 10 .
- the input apparatus 1 includes an operation section 23 (see FIG. 6 ) including a button 11 and a button 12 adjacent to the button 11 that are provided at a center of an upper portion of the casing 10 , and a button 13 provided at a side portion of the casing 10 .
- buttons 11 , 12 , and 13 are each a press-type button.
- the operation section 23 is not limited to the press-type button, and a bar-type operation section that is operated with one end as a fulcrum, or a slide-type operation section may also be used.
- Each of the buttons 11 , 12 , and 13 includes a built-in switch (not shown) which detects an operation of the user with respect to the operation section and outputs an operation signal.
- a switch As the switch that outputs an operation signal, an optical sensor or a capacitance sensor may be used.
- the button 11 has a function corresponding to a left button of a planar-operation-type mouse used for a PC, and the button 12 adjacent to the button 11 has a function corresponding to a right button of a mouse, for example.
- an operation of selecting an icon 4 may be carried out by clicking the button 11
- an operation of opening a file may be carried out by double-clicking the button 11 .
- the button 13 has a function as a switch button for switching a pointer mode to a scroll mode and vice versa.
- the “pointer mode” is a mode in which a pointer 2 displayed on a screen 3 (see FIG. 5 ) is moved in accordance with a movement of the casing 10 .
- the “scroll mode” is a mode in which an image 6 displayed on the screen 3 is scrolled in accordance with the movement of the casing 10 .
- FIG. 3 is a diagram schematically showing an internal structure of the input apparatus 1 .
- FIG. 4 is a block diagram showing an electrical structure of the input apparatus 1 .
- the input apparatus 1 includes a sensor unit 17 , a control unit 30 , and batteries 14 .
- FIG. 8 is a perspective diagram showing the sensor unit 17 .
- the sensor unit 17 includes an acceleration sensor unit 16 for detecting accelerations in different angles such as along two orthogonal axes (X′ axis and Y′ axis).
- the acceleration sensor unit 16 includes two sensors, that is, a first acceleration sensor 161 and a second acceleration sensor 162 .
- the sensor unit 17 further includes an angular velocity sensor unit 15 for detecting angular accelerations about the two orthogonal axes.
- the angular velocity sensor unit 15 includes two sensors, that is, a first angular velocity sensor 151 and a second angular velocity sensor 152 .
- the acceleration sensor unit 16 and the angular velocity sensor unit 15 are packaged and mounted on a circuit board 25 .
- first angular velocity sensor 151 and the second angular velocity sensor 152 a vibration gyro sensor for detecting Coriolis force in proportion to an angular velocity is used.
- first acceleration sensor 161 and the second acceleration sensor 162 any sensor such as a piezoresistive sensor, a piezoelectric sensor, or a capacitance sensor may be used.
- Each of the angular velocity sensors 151 and 152 is not limited to the vibration gyro sensor, and a rotary top gyro sensor, a ring laser gyro sensor, a gas rate gyro sensor, a geomagnetic gyro sensor, and the like may also be used.
- a longitudinal direction of the casing 10 is referred to as Z′ direction
- a thickness direction of the casing 10 is referred to as X′ direction
- a width direction of the casing 10 is referred to as Y′ direction for convenience.
- the sensor unit 17 is incorporated into the casing 10 such that a surface of the circuit board 25 on which the acceleration sensor unit 16 and the angular velocity sensor unit 15 are mounted becomes substantially parallel to an X′-Y′ plane.
- the sensor units 16 and 15 each detect physical amounts with respect to the two axes, that is, the X′ axis and the Y′ axis.
- a coordinate system that moves along with the input apparatus 1 that is, a coordinate system fixed to the input apparatus 1 is expressed using the X′ axis, Y′ axis, and Z′ axis
- a coordinate system stationary on earth that is, an inertial coordinate system is expressed using the X axis, Y axis, and Z axis.
- a rotational direction about the X′ axis is sometimes referred to as pitch direction
- a rotational direction about the Y′ axis is sometimes referred to as yaw direction
- a rotational direction about the Z′ axis (roll axis) is sometimes referred to as roll direction.
- the control unit 30 includes a main substrate 18 , an MPU 19 (Micro Processing Unit) (or CPU) mounted on the main substrate 18 , a crystal oscillator 20 , a transceiver 21 , and an antenna 22 printed on the main substrate 18 .
- MPU 19 Micro Processing Unit
- CPU Central Processing Unit
- the MPU 19 includes a built-in volatile or nonvolatile memory requisite therefor.
- the MPU 19 is input with a detection signal from the sensor unit 17 , an operation signal from the operation section, and the like, and executes various types of operational processing in order to generate predetermined control signals in response to those input signals.
- the memory may be provided separate from the MPU 19 .
- the sensor unit 17 outputs analog signals.
- the MPU 19 includes an A/D (Analog/Digital) converter.
- the sensor unit 17 may be a unit that includes the A/D converter.
- the transceiver 21 (transmission means) transmits the control signals generated in the MPU 19 as RF radio signals to the control apparatus 40 via the antenna 22 .
- the transceiver 21 is also capable of receiving various signals transmitted from the control apparatus 40 .
- the crystal oscillator 20 generates clocks and supplies them to the MPU 19 .
- the batteries 14 dry cell batteries, rechargeable batteries, and the like are used.
- the control apparatus 40 includes an MPU 35 (or CPU), a RAM 36 , a ROM 37 , a video RAM 41 , a display control section 42 , an antenna 39 , and a transceiver 38 .
- the transceiver 38 receives the control signal transmitted from the input apparatus 1 via the antenna 39 (reception means).
- the transceiver 38 is also capable of transmitting various predetermined signals to the input apparatus 1 .
- the MPU 35 analyzes the control signal and executes various types of operational processing.
- the display control section 42 mainly generates screen data to be displayed on the screen 3 of the display apparatus 5 under control of the MPU 35 .
- the video RAM 41 serves as a work area of the display control section 42 and temporarily stores the generated screen data.
- the control apparatus 40 may be an apparatus dedicated to the input apparatus 1 , or may be a PC or the like.
- the control apparatus 40 is not limited to the apparatus dedicated to the input apparatus 1 , and may be a computer integrally formed with the display apparatus 5 , audiovisual equipment, a projector, a game device, a car navigation system, or the like.
- Examples of the display apparatus 5 include a liquid crystal display and an EL (Electro-Luminescence) display.
- the display apparatus 5 may alternatively be an apparatus integrally formed with a display and capable of receiving television broadcasts and the like, or an apparatus in which such a display and the control apparatus 40 are integrated.
- FIG. 5 is a diagram showing an example of the screen 3 displayed on the display apparatus 5 .
- GUIs such as icons 4 and the pointer 2 are displayed on the screen 3 .
- the icons are images on the screen 3 representing functions of programs, execution commands, file contents, and the like on the computer.
- an image 6 such as a web image including a plurality of letters 7 is displayed, for example.
- FIG. 6 is a diagram showing a state where a user is holding the input apparatus 1 .
- the input apparatus 1 may include, as the operation section 23 , in addition to the buttons 11 , 12 , and 13 , various operation buttons 29 such as those provided to a remote controller for operating a television or the like and a power switch 28 , for example.
- Command signals generated when the user moves the input apparatus 1 in the air or operates the operation section 23 while holding the input apparatus 1 as shown in the figure are output to the control apparatus 40 , and the control apparatus 40 controls the GUI.
- FIGS. 7A and 7B are explanatory diagrams therefor.
- the user holds the input apparatus 1 so as to aim the buttons 11 and 12 side of the input apparatus 1 at the display apparatus 5 side.
- the user holds the input apparatus 1 so that a thumb is located on an upper side and a pinky is located on a lower side as in handshakes.
- the circuit board 25 (see FIG. 8 ) of the sensor unit 17 is close to being in parallel with the screen 3 of the display apparatus 5 , and the two axes as detection axes of the sensor unit 17 respectively correspond to the horizontal axis (X axis) and the vertical axis (Y axis) on the screen 3 .
- the position of the input apparatus 1 as shown in FIGS. 7A and 7B is referred to as reference position.
- the control apparatus 40 controls display of the pointer 2 so as to move the pointer 2 in the vertical direction on the screen 3 .
- the control apparatus 40 controls display of the pointer 2 so as to move the pointer 2 in the horizontal direction on the screen 3 .
- FIG. 9 is a flowchart showing the operation of the control system 100 in this case.
- biaxial angular velocity signals are output from the angular velocity sensor unit.
- the MPU 19 acquires angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) from the angular velocity signals (Step 101 ).
- biaxial acceleration signals are output from the acceleration sensor unit 16 .
- the MPU 19 acquires acceleration values (a x , a y ) from the biaxial acceleration signals (Step 102 ).
- the MPU 19 typically carries out the process of acquiring angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) (Step 101 ) and the process of acquiring acceleration values (a x , a y ) (Step 102 ) in sync.
- the process of acquiring angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) and the process of acquiring acceleration values (a x , a y ) do not always need to be carried out in sync (at the same time).
- the acceleration values (a x , a y ) may be obtained after the angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) are obtained, or the angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) may be obtained after the acceleration values (a x , a y ) are obtained.
- the MPU 19 calculates velocity values (first velocity value V x and second velocity value V y ) by a predetermined operation (Step 103 ).
- the first velocity value V x is a velocity value in a direction along the X′ axis
- the second velocity value V y is a velocity value in a direction along the Y′ axis.
- the MPU 19 obtains radius gyrations (R ⁇ , R ⁇ ) of the movement of the input apparatus 1 by dividing the acceleration values (a x , a y ) by angular acceleration values ( ⁇ ⁇ , ⁇ ⁇ ), and calculates velocity values by multiplying the radius gyrations (R ⁇ , R ⁇ ) by the angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ).
- the radius gyrations (R ⁇ , R ⁇ ) may be obtained by dividing acceleration change rates ( ⁇ a x , ⁇ a y ) by angular acceleration change rates ( ⁇ ( ⁇ ⁇ ), ⁇ ( ⁇ ⁇ )).
- An effect of gravity accelerations can be removed when the radius gyrations (R ⁇ , R ⁇ ) are calculated by dividing the acceleration change rates ( ⁇ a x , ⁇ a y ) by the angular acceleration change rates ( ⁇ ( ⁇ ⁇ ), ⁇ ( ⁇ ⁇ )).
- the MPU 19 calculates the velocity values by, for example, integrating the acceleration values (a x , a y ) while using the angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) as an adjunct for the integration operation.
- the velocity values (V x , V y ) do not always need to be calculated by the calculation method above.
- the velocity values (V x , V y ) may be calculated by simply integrating the acceleration values (a x , a y ).
- the detected angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) may be used as they are as the velocity values (V x , V y ).
- the MPU 19 transmits information on the calculated velocity values (V x , V y ) to the control apparatus 40 via the transceiver 21 and the antenna 22 (Step 104 ).
- the MPU 35 of the control apparatus 40 receives the information on the velocity values (V x , V y ) via the antenna 39 and the transceiver 38 (Step 105 ).
- the input apparatus 1 transmits the velocity values (V x , V y ) every predetermined number of clocks, that is, every time a predetermined time passes, so the control apparatus 40 receives the velocity values every predetermined number of clocks.
- the MPU 35 of the control apparatus 40 Upon receiving the velocity values, the MPU 35 of the control apparatus 40 generates new coordinate values (X(t), Y(t)) by adding the velocity values to coordinate values using Equations (1) and (2) below (Step 106 ). The MPU 35 controls display on the screen so that the pointer 2 moves to a position corresponding to the generated coordinate values (Step 107 ).
- the calculation of the velocity values (V x , V y ) may be executed by the control apparatus 40 .
- the input apparatus 1 transmits information on the angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) and the acceleration values (a x , a y ) to the control apparatus 40 via the transceiver 21 and the antenna 22 .
- the control apparatus 40 calculates the velocity values (V x , V y ). The method of calculating the velocity values is as described above.
- FIG. 10 is a flowchart showing an operation of the input apparatus.
- the MPU 19 acquires angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) and acceleration values (a x , a y ) from the angular velocity sensor unit 15 and the acceleration sensor unit 16 (Step 201 ). Based on the acquired angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) and acceleration values (a x , a y ), the MPU 19 calculates velocity values (V x , V y ) (Step 202 ).
- the MPU 19 Upon calculating the velocity values (V x , V y ), the MPU 19 judges whether an operation signal from a switch (not shown) provided to the button 13 is input (Step 203 ). When the user has not pressed the button 13 and an operation signal from the switch is not yet input (NO in Step 203 ), the MPU 19 transmits the calculated velocity values (V x , V y ) to the control apparatus 40 as information on a movement amount of the pointer 2 (Step 204 ). Upon transmitting information on the velocity values (V x , V y ), the MPU 19 returns to Step 201 .
- the MPU 35 of the control apparatus 40 Upon receiving the information on the velocity values (V x , V y ), the MPU 35 of the control apparatus 40 generates new coordinate values and controls display on the screen 3 so that the pointer 2 moves to a position corresponding to the generated coordinate values (pointer mode).
- an operation signal is output from the switch to be input to the MPU 19 (YES in Step 203 ).
- the MPU 19 multiplies the first velocity value V x and the second velocity value V y by a first weighting factor ⁇ and a second weighting factor ⁇ , respectively, as expressed in Equations (3) and (4) below to thus calculate a first modified velocity value V x ′ and a second modified velocity value V y ′ (Step 205 ).
- V x ′ ⁇ V x (3)
- V y ′ ⁇ V y (4)
- the weighting factors ( ⁇ , ⁇ ) are typically different values and stored in a memory (not shown), for example.
- the MPU 19 changes a ratio of the first velocity value V x to the second velocity value V y (ratio change means).
- the weighting factors ( ⁇ , ⁇ ) can take various values.
- a scroll direction can be biased in a vertical-axis (Y-axis) direction or a horizontal-axis (X-axis) direction on the screen 3 . Details on relationships between the weighting factors ( ⁇ , ⁇ ) and scroll tilt directions will be described later.
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 206 ). Upon transmitting the information on the modified velocity values (V x ′, V y ′), the MPU 19 returns to Step 201 .
- the MPU 35 of the control apparatus 40 receives the transmitted information on the modified velocity values (V x ′, V y ′).
- the MPU 35 controls display so that the letters 7 inside the image 6 are scrolled at a velocity corresponding to the received modified velocity values (V x ′, V y ′) (scroll mode).
- the image 6 as a scroll target include a web image, a map, and an EPG (Electronic Program Guide).
- a signal transmitted from the input apparatus 1 to the control apparatus 40 contains, in addition to the information on the modified velocity values (V x ′, V y ′), a signal for causing the control apparatus 40 to control display of scroll. Accordingly, since the control apparatus 40 can distinctively recognize the pointer mode and the scroll mode, display of scroll on the screen can be controlled when the modified velocity values (V x ′, V y ′) are transmitted. It should be noted that as another method used for the control apparatus 40 to distinctively recognize the pointer mode and the scroll mode, there is a method of transmitting a mode switch signal that indicates that a mode has been switched.
- control apparatus 40 can distinctively recognize the pointer mode and the scroll mode also by transmission of a signal indicating that the button 13 has been pressed (e.g., press code). Any method may be adopted for the method used for the control apparatus 40 to distinctively recognize the pointer mode and the scroll mode.
- FIGS. 11A and 11B are diagrams for explaining the relationships between the weighting factors ( ⁇ , ⁇ ) and scroll tilt directions.
- the weighting factors ( ⁇ , ⁇ ) are set to, for example, (1 ⁇ 3, 1), (1 ⁇ 2, 1), (1 ⁇ 2, 2), (1 ⁇ 2, 3), (1, 2), (1, 3), or (1, 4).
- the weighting factors ( ⁇ , ⁇ ) are not limited to those values and may of course take other values.
- the scroll direction can be biased in the vertical-axis direction on the screen. Accordingly, an operational feeling in scroll operations can be improved in a case where the image 6 is long in the vertical-axis direction on the screen 3 as a whole, for example. Since the image 6 such as a web image is, in many cases, long in the vertical-axis direction on the screen 3 as a whole in particular, an operational feeling in scrolling the image 6 such as a web image can be improved.
- the scroll direction of the image 6 is biased in the horizontal-axis (X-axis) direction on the screen 3 with respect to the operation direction of the input apparatus 1 .
- the weighting factors ( ⁇ , ⁇ ) are set to, for example, (4, 1), (3, 1), (2, 1), (3, 1 ⁇ 2), (2, 1 ⁇ 2), (1, 1 ⁇ 2), or (1, 1 ⁇ 3).
- the weighting factors ( ⁇ , ⁇ ) are not limited to those values and may of course take other values.
- the scroll direction can be biased in the horizontal-axis direction on the screen. Accordingly, an operational feeling in scroll operations can be improved in a case where the image 6 is long in the horizontal-axis direction on the screen 3 as a whole, for example.
- weighting factors such that either the first weighting factor ⁇ or the second weighting factor ⁇ is set to 0 like (1, 0) and (0, 1).
- the MPU 19 multiplies the first and second velocity values (V x , V y ) by 1 and 0, respectively, to thus calculate the first and second modified velocity values (V x ′, V y ′) in Step 205 . Then, the MPU 19 transmits information on the calculated modified velocity values (V x ′, V y ′) to the control apparatus 40 (Step 206 ). In this case, the image 6 is scrolled only in the vertical-axis direction and not in the horizontal-axis direction on the screen 3 . In other words, the scroll direction is restricted to the vertical-axis direction on the screen 3 .
- the image 6 is scrolled only in the horizontal-axis direction and not in the vertical-axis direction on the screen 3 .
- the scroll direction is restricted to the horizontal-axis direction on the screen 3 .
- the expression “scroll direction is biased” means that, as shown in FIGS. 11A and 11B , the scroll direction is biased in a predetermined axial direction on the screen 3 (e.g., vertical-axis direction).
- the expression “scroll direction is restricted” means that the scroll direction is biased at maximum to a predetermined axial direction on the screen and scroll cannot be performed in any other directions.
- weighting factors ( ⁇ , ⁇ ) are set such that either the first weighting factor ⁇ or the second weighting factor ⁇ is set to 1 like (0, 1), (1 ⁇ 2, 1), (1, 2), (2, 1), (1, 1 ⁇ 2), and (1, 0).
- the first velocity value V x is multiplied by 1 ⁇ 2 and reduced, and a first modified velocity value V x ′ is thus obtained (Step 205 ).
- the second velocity value V y is multiplied by 1 to thus obtain a second modified velocity value V y ′.
- the value obtained by multiplying the second velocity value V y by 1 (second modified velocity value V y ′) is the second velocity value V y itself, so the second modified velocity value V y ′ does not need to be calculated.
- the MPU 19 only needs to transmit the first modified velocity value V x ′ and the second velocity value V y to the control apparatus 40 as scroll information in Step 206 .
- the processing shown in FIG. 10 may be mainly executed by the control apparatus 40 .
- the control apparatus 40 receives information on velocity values (V x , V y ) transmitted from the input apparatus 1 .
- the MPU 35 of the control apparatus 40 calculates modified velocity values (V x ′, V y ′) by multiplying the received velocity values (V x , V y ) by the weighting factors ( ⁇ , ⁇ ). Then, the MPU 35 controls display on the screen so that the image displayed on the screen is scrolled at a velocity corresponding to the modified velocity values (V x ′, V y ′). It should be noted that processing according to embodiments and modified examples of the present invention to be described hereinbelow can all be applied as processing of the control apparatus 40 .
- the first embodiment above has described a case where the scroll direction is biased in (restricted to) a uniaxial direction of one of the horizontal-axis direction and the vertical-axis direction on the screen 3 .
- the second embodiment is different from the first embodiment in that the scroll direction is biased in (restricted to) biaxial directions of the horizontal-axis direction and the vertical-axis direction on the screen 3 . Therefore, that point will mainly be described.
- FIG. 12 is a flowchart showing an operation of the input apparatus 1 according to the second embodiment.
- Steps 301 to 304 processes that are the same as those of Steps 201 to 204 of FIG. 10 are executed.
- the button 13 is not pressed (NO in Step 303 )
- information on velocity values is transmitted from the input apparatus 1 (Step 304 )
- the pointer 2 displayed on the screen 3 is moved at a velocity corresponding to the velocity values.
- the MPU 19 Upon being input with the operation signal, the MPU 19 judges whether an absolute value of the first velocity value
- the MPU 19 sets the first weighting factor ⁇ to be larger than the second weighting factor ⁇ (Step 306 ).
- the MPU 19 set the first weighting factor ⁇ to be smaller than the second weighting factor ⁇ (Step 307 ). Values determined in advance are used as the weighting factors ( ⁇ , ⁇ ) set in Steps 306 and 307 .
- the weighting factors ( ⁇ , ⁇ ) set in Step 306 are, for example, (1, 1 ⁇ 2)
- the weighting factors ( ⁇ , ⁇ ) set in Step 307 are, for example, (1 ⁇ 2, 1).
- the weighting factors ( ⁇ , ⁇ ) set in Steps 306 and 307 there are, for example, [(1, 0) and (0, 1)], [(1, 1 ⁇ 3) and (1 ⁇ 3, 1)], [(1, 2) and (2, 1)], and [(1, 3) and (3, 1)].
- the combination is not limited to those combinations, and other values may be used instead.
- the MPU 19 Upon setting the weighting factors ( ⁇ , ⁇ ), the MPU 19 multiplies the first and second velocity values (V x , V y ) by the first and second weighting factors ( ⁇ , ⁇ ), respectively, to thus calculate first and second modified velocity values (V x ′, V y ′) (Step 308 ).
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 309 ).
- the MPU 35 of the control apparatus 40 controls display so that the letters 7 in the image 6 are scrolled at a velocity corresponding to the received modified velocity values (V x ′, V y ′).
- FIGS. 13A and 13B are diagrams showing relationships between operation directions of the input apparatus 1 and scroll directions in a case where the processing shown in FIG. 12 is executed.
- FIG. 13A shows relationships between operation directions of the input apparatus 1 and scroll directions in a case where a combination of weighting factors set in Steps 306 and 307 is, for example, [(1, 1 ⁇ 2) and (1 ⁇ 2, 1)] or [(2, 1) and (1, 2)].
- FIG. 13B shows relationships between operation directions of the input apparatus 1 and scroll directions in a case where 0 (or value that is substantially 0) is used as in [(1, 0) and (0, 1)] and [(2, 0) and (0, 2)], for example.
- a scroll direction of an image on the screen 3 is biased in the horizontal-axis (X-axis) direction on the screen.
- the scroll direction of the image on the screen 3 is biased in the vertical-axis (Y-axis) direction on the screen.
- a scroll direction of the image 6 is restricted to the horizontal-axis (X-axis) direction on the screen.
- the scroll direction of the image 6 is restricted to the vertical-axis (Y-axis) direction on the screen.
- the third embodiment mainly describes points different from those of the second embodiment above.
- FIG. 14 is a flowchart showing an operation of the input apparatus 1 according to the third embodiment.
- Steps 401 to 404 processes that are the same as those of Steps 301 to 304 of FIG. 12 are executed.
- the pointer 2 moves on the screen 3 in accordance with the 3-dimensional operation.
- an operation signal is output from the switch provided to the button 13 and input to the MPU 19 (YES in Step 403 ).
- the MPU 19 calculates a tilt angle of a combined vector of the first velocity value and the second velocity value using Equation (5) below (Step 405 ). By calculating the combined vector tilt angle, the MPU 19 judges an operation direction (movement direction) of the input apparatus 1 .
- the MPU 19 Upon calculating the combined vector tilt angle ⁇ , the MPU 19 judges whether the combined vector tilt angle ⁇ is an angle within a first angle range (Step 406 ).
- FIG. 15 is a diagram for explaining the first angle range and the second angle range.
- the first angle range indicates a range within a predetermined angle from 0 degree (or 180 degrees; same holds true for descriptions below) (e.g., 0 ⁇ 30 degrees).
- the second angle range indicates a range within a predetermined angle from 90 degrees (or 270 degrees; same holds true for descriptions below) (e.g., 90 ⁇ 60 degrees).
- the input apparatus 1 stores the first angle range and the second angle range as shown in FIG. 15 in a memory.
- the horizontal-axis direction within the angle ranges shown in FIG. 15 corresponds to a movement direction (operation direction) of the input apparatus 1 in the horizontal-axis direction
- the vertical-axis direction corresponds to the movement direction (operation direction) of the input apparatus 1 in the vertical-axis direction.
- the first angle range and the second angle range can be set variously, but in the description on FIG. 14 , the first angle range is assumed to be an angle range of 0 ⁇ 30 degrees and the second angle range is assumed to be an angle range of 90 ⁇ 60 degrees for convenience.
- the MPU 19 may judge whether the combined vector tilt angle ⁇ is an angle within the second angle range in Step 406 .
- the MPU 19 sets the first weighting factor ⁇ to be larger than the second weighting factor ⁇ (Step 407 ).
- the MPU 19 sets the first weighting factor ⁇ to be smaller than the second weighting factor ⁇ (Step 408 ).
- the MPU 19 Upon setting the weighting factors ( ⁇ , ⁇ ), the MPU 19 multiplies the first and second velocity values (V x , V y ) by the first and second weighting factors ( ⁇ , ⁇ ), respectively, to thus calculate first and second modified velocity values (V x ′, V y ′) (Step 409 ).
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 410 ).
- the MPU 35 of the control apparatus 40 controls display so that the letters 7 in the image 6 are scrolled at a velocity corresponding to the received modified velocity values (V x ′, V y ′).
- FIGS. 16A and 16B are diagrams showing relationships between operation directions of the input apparatus 1 and scroll directions in a case where the processing shown in FIG. 14 is executed.
- FIG. 16A is a diagram showing relationships between operation directions of the input apparatus 1 and scroll directions in a case where a combination of weighting factors set in Steps 407 and 408 is, for example, [(1, 1 ⁇ 2) and (1 ⁇ 2, 1)] or [(2, 1) and (1, 2)].
- FIG. 16B is a diagram showing relationships between operation directions of the input apparatus 1 and scroll directions in a case where 0 (or value that is substantially 0) is used for the weighting factors ( ⁇ , ⁇ ) as in [(1, 0) and (0, 1)] and [(2, 0) and (0, 2)], for example.
- a scroll direction of the image on the screen 3 is biased in the horizontal-axis (X-axis) direction on the screen.
- the scroll direction of the image on the screen 3 is biased in the vertical-axis (Y-axis) direction on the screen.
- the image 6 can be scrolled in the vertical-axis direction on the screen 3 with high sensitivity.
- an operational feeling in scroll operations can be additionally improved in a case where the image 6 is long in the vertical-axis direction on the screen 3 as a whole.
- first angle range and the second angle range can be set variously as described above.
- Examples of the combination of the first angle range and the second angle range include combinations of (0 ⁇ 35 degrees, 90 ⁇ 55 degrees) and (0 ⁇ 40 degrees, 90 ⁇ 50 degrees).
- the first angle range may be set to be larger than the second angle range.
- Examples of the combination of the first angle range and the second angle range in this case include combinations of (0 ⁇ 60 degrees, 90 ⁇ 30 degrees), (0 ⁇ 55 degrees, 90 ⁇ 35 degrees), and (0 ⁇ 50 degrees, 90 ⁇ 40 degrees).
- the fourth embodiment is different from the third embodiment above in that the first angle range and the second angle range are controlled variably. Therefore, that point will mainly be described.
- FIG. 17 is a flowchart showing an operation of the input apparatus 1 according to the fourth embodiment.
- the MPU 19 upon calculating velocity values based on acquired acceleration values and angular velocity values (Steps 501 and 502 ), the MPU 19 stores the calculated velocity values in the memory (Step 503 ). Next, the MPU 19 judges whether an operation signal from the switch of the button 13 is input (Step 504 ). When judged that an operation signal is not yet input (NO in Step 504 ), the MPU 19 transmits information on the velocity values as information on a movement amount of the pointer 2 (Step 505 ).
- the MPU 19 reads out velocity values of past n histories that are stored in the memory. Then, the MPU 19 calculates a combined vector of the read-out velocity values (Step 506 ). Typically, the MPU 19 obtains a sum ⁇ V x and sum ⁇ V y of first velocity values V x and second velocity values V y of past n histories that are stored in the memory and calculates a combined vector.
- the MPU 19 Upon calculating the combined vector, the MPU 19 calculates a combined vector tilt angle ⁇ ′ by Equation (6) below (Step 507 ).
- the MPU 19 Upon calculating the combined vector tilt angle ⁇ ′, the MPU 19 judges whether the combined vector tilt angle ⁇ ′ is an angle within the first angle range (Step 508 ). When judged that the combined vector tilt angle ⁇ ′ is an angle within the first angle range (YES in Step 508 ), the MPU 19 widens the first angle range (Step 509 ) (angle range control means). In this case, the second angle range is narrowed. Upon widening the first angle range, the MPU 19 sets the first weighting factor ⁇ to be larger than the second weighting factor ⁇ (Step 510 ).
- the MPU 19 narrows the first angle range (Step 511 ). In this case, the second angle range is widened. Upon narrowing the first angle range, the MPU 19 sets the first weighting factor ⁇ to be smaller than the second weighting factor ⁇ (Step 512 ).
- the MPU 19 Upon setting the weighting factors ( ⁇ , ⁇ ), the MPU 19 multiplies the first and second velocity values (V x , V y ) by the first and second weighting factors ( ⁇ , ⁇ ), respectively, to thus calculate first and second modified velocity values (V x ′, V y ′) (Step 513 ).
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 514 ).
- FIGS. 18A and 18B are diagrams showing temporal changes of ranges of the first angle range and the second angle range in a case where the processing shown in FIG. 17 is executed.
- FIG. 18A is a diagram showing temporal changes of the first angle range and the second angle range in a case where the user operates the input apparatus 1 in the horizontal-axis (X′-axis) direction.
- FIG. 18B is a diagram showing temporal changes of the first angle range and the second angle range in a case where the user operates the input apparatus 1 in the vertical-axis (Y′-axis) direction.
- the first angle range is gradually widened.
- the second angle range is in a widened state when an arm is swung. Therefore, even when the user swings an arm and operates the input apparatus 1 in an oblique direction, scroll in the vertical-axis direction is prioritized on the screen.
- the first angle range and the second angle range are controlled variably in the input apparatus 1 according to the fourth embodiment, an operational feeing for the user in operating the image 6 displayed on the screen 3 can be additionally improved.
- FIG. 19 is a flowchart showing an operation of the input apparatus 1 according to the fifth embodiment.
- Steps 601 to 605 processes that are the same as those of Steps 501 to 505 of FIG. 17 are executed, and by the user operating the input apparatus 1 in a state where the button 13 is not pressed, the pointer 2 moves on the screen 3 .
- the MPU 19 reads out velocity values (V x , V y ) of past n histories that are stored in the memory and calculates a combined vector of the read-out velocity values (V x , V y ) (Step 606 ).
- the MPU 19 obtains sums of first velocity values and second velocity values of past n histories that are stored in the memory and calculates a combined vector.
- the MPU 19 Upon calculating the combined vector of the velocity values, the MPU 19 calculates a combined vector tilt angle ⁇ ′ by Equation (6) above (Step 607 ). Next, the MPU 19 judges whether the combined vector tilt angle ⁇ ′ is an angle within a first modified angle range (Step 608 ).
- FIG. 20 is a diagram for explaining the first modified angle range and second modified angle range.
- the first modified angle range is an angle range for changing the first angle range and the second angle range and indicates an angle range of, for example, ⁇ 45 degrees from 0 degree (or 180 degrees; same holds true for descriptions below).
- the second modified angle range is an angle range for changing the first angle range and the second angle range and indicates an angle range of, for example, ⁇ 45 degrees from 90 degrees (or 270 degrees; same holds true for descriptions below).
- the horizontal-axis direction within the modified angle ranges shown in FIG. 20 corresponds to a movement direction (operation direction) of the input apparatus 1 in the horizontal-axis direction
- the vertical-axis direction corresponds to the movement direction (operation direction) of the input apparatus 1 in the vertical-axis direction.
- the first modified angle range and the second modified angle range are fixed and do not fluctuate by the combined vector tilt angle ⁇ ′.
- the first modified angle range and the second modified angle range are not limited to the range of 0 ⁇ 45 (or 90 ⁇ 45) degrees.
- the first modified angle range and the second modified angle range can be changed as appropriate.
- the MPU 19 widens the first angle range (Step 609 ). In this case, the second angle range is narrowed.
- the MPU 19 narrows the first angle range (Step 610 ). In this case, the second angle range is widened.
- the MPU 19 judges whether the combined vector tilt angle ⁇ ′ is an angle within the first angle range (Step 611 ). When judged that the combined vector tilt angle ⁇ ′ is an angle within the first angle range (YES in Step 611 ), the MPU 19 sets the first weighting factor ⁇ to be larger than the second weighting factor ⁇ (Step 612 ).
- the MPU 19 sets the first weighting factor ⁇ to be smaller than the second weighting factor ⁇ (Step 613 ).
- the MPU 19 Upon setting the weighting factors ( ⁇ , ⁇ ), the MPU 19 multiplies the first and second velocity values (V x , V y ) by the first and second weighting factors ( ⁇ , ⁇ ), respectively, to thus calculate first and second modified velocity values (V x ′, V y ′) (Step 614 ).
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 615 ).
- the first angle range and the second angle range are controlled variably based on the first modified angle range and the second modified angle range as fixed values.
- the first angle range and the second angle range can be widened/narrowed as appropriate.
- the above embodiments have described a case where the scroll direction is biased in (restricted to) a uniaxial direction or biaxial directions on the screen.
- the sixth embodiment is different from the above embodiments in that the scroll direction is restricted to directions along four axes on the screen 3 . Therefore, that point will mainly be described.
- FIG. 21 is a flowchart showing an operation of the input apparatus 1 according to this embodiment.
- Steps 701 to 704 information on velocity values is transmitted as information on a movement amount of the pointer 2 when the button 13 is not pressed.
- the MPU 19 calculates a tilt angle ⁇ of a combined vector of velocity values (V x , V y ) using Equation (5) above (Step 705 ).
- the MPU 19 Upon calculating the combined vector tilt angle ⁇ , the MPU 19 judges whether the combined vector tilt angle ⁇ is within a third angle range (Step 706 ).
- FIG. 22 is a diagram for explaining the third angle range.
- an angle range is divided into the first angle range, the second angle range, and the third angle range.
- the first angle range is, for example, a range within 0 ⁇ 22.5 degrees or 180 ⁇ 22.5 degrees.
- the second angle range is, for example, a range within 90 ⁇ 22.5 degrees or 270 ⁇ 22.5 degrees.
- the third angle range is, for example, a range within 45 ⁇ 22.5 degrees, 135 ⁇ 22.5 degrees, 225 ⁇ 22.5 degrees, or 315 ⁇ 22.5 degrees. It should be noted that ranges of the first angle range, the second angle range, and the third angle range can be changed as appropriate. Angles to be a reference of the third angle range (broken lines of FIG. 22 ) can also be changed as appropriate.
- the horizontal-axis direction in the angle ranges shown in FIG. 22 corresponds to a movement direction (operation direction) of the input apparatus 1 in the horizontal-axis direction
- the vertical-axis direction corresponds to a movement direction (operation direction) of the input apparatus 1 in the vertical-axis direction.
- the MPU 19 references a table and sets the weighting factors ( ⁇ , ⁇ ) (Step 710 ).
- the weighting factors ( ⁇ , ⁇ ) read out from the table are not constant and are values determined in relation to velocity values (V x , V y ).
- the weighting factors ( ⁇ , ⁇ ) are stored in the table as values for restricting the scroll direction to directions at angles of ⁇ 45 degrees from the vertical-axis direction on the screen. It should be noted that the weighting factors ( ⁇ , ⁇ ) set in Step 710 may be calculated by a program.
- Step 706 When judged in Step 706 that the combined vector tilt angle ⁇ is not an angle within the third angle range (NO in Step 706 ), the MPU 19 judges whether the combined vector tilt angle ⁇ is an angle within the first angle range (Step 707 ). When the combined vector tilt angle ⁇ is an angle within the first angle range (YES in Step 707 ), the MPU 19 sets the first weighting factor ⁇ to 1 and the second weighting factor 3 to 0 (Step 708 ).
- the MPU 19 sets the first weighting factor ⁇ to 0 and the second weighting factor ⁇ to 1 (Step 709 ).
- the MPU 19 Upon setting the weighting factors ( ⁇ , ⁇ ), the MPU 19 multiplies the first and second velocity values (V x , V y ) by the first and second weighting factors ( ⁇ , ⁇ ), respectively, to thus calculate first and second modified velocity values (V x ′, V y ′) (Step 711 ).
- the MPU 19 Upon calculating the modified velocity values (V x ′, V y ′), the MPU 19 transmits information on the modified velocity values (V x ′, V y ′) to the control apparatus 40 as scroll information (Step 712 ).
- FIG. 23 is a diagram showing relationships between operation directions of the input apparatus 1 and scroll directions in a case where the processing shown in FIG. 21 is executed.
- a scroll direction of the image 6 is restricted to the horizontal-axis (X-axis) direction on the screen.
- the scroll direction of the image 6 is restricted to the vertical-axis (Y-axis) direction on the screen.
- the scroll direction of the image 6 is restricted to a direction at an angle of +45 degrees from the horizontal axis on the screen.
- the scroll direction of the image 6 is restricted to a direction at an angle of ⁇ 45 degrees from the horizontal axis on the screen.
- the scroll direction is restricted to directions along four axes of the horizontal-axis direction, the vertical-axis direction, the direction at an angle of +45 degrees from the horizontal axis, and the direction at an angle of ⁇ 45 degrees from the horizontal axis on the screen.
- an operational feeling in scroll operations in a case where the image 6 such as a map that is long in the vertical-axis direction and the horizontal-axis direction on the screen 3 as a whole is operated can be improved.
- the scroll direction can be restricted to various directions.
- the combination of directions to which scroll is restricted include a combination of the horizontal-axis direction, the vertical-axis direction, and directions at angles of ⁇ 30 degrees from the horizontal-axis direction and a combination of the horizontal-axis direction, the vertical-axis direction, and directions at angles of ⁇ 60 degrees from the horizontal-axis direction. It is of course possible to use other combinations.
- the number of restriction axes on the screen 3 is also not limited to four (four axes).
- the number of restriction axes may be three (three axes) or five (five axes) or more.
- the sixth embodiment has described a case where the scroll direction on the screen 3 is restricted. However, it is also possible to bias the scroll direction on the screen 3 .
- first angle range, the second angle range, and the third angle range may be controlled variably.
- control system 100 according to a seventh embodiment of the present invention will be described.
- FIGS. 24A and 24B are diagrams each showing a relationship between the operation direction of the input apparatus 1 and a direction in which the image 6 is scrolled.
- FIG. 24A is a diagram showing a case where the image 6 is scrolled in a direction in which the input apparatus 1 is operated
- FIG. 24B is a diagram showing a case where the image 6 is scrolled in an opposite direction from the direction in which the input apparatus 1 is operated.
- the inventors of the present invention have conducted a user test, which revealed that there are both users who feel that scroll of an image in a direction in which the input apparatus 1 is operated provides a better operational feeling and users who feel that scroll of an image in an opposite direction from the direction in which the input apparatus 1 is operated provides a better operational feeling.
- the input apparatus 1 executes processing for improving an operational feeling regarding a direction of scrolling the image 6 .
- FIG. 25 is a flowchart showing an operation of the input apparatus 1 of the control system 100 according to this embodiment.
- the input apparatus 1 calculates velocity values (V x , V y ) based on acquired angular velocity values ( ⁇ ⁇ , ⁇ ⁇ ) and acceleration values (a x , a y ) (Steps 801 and 802 ).
- the MPU 19 judges whether an operation signal from the switch provided to the button 13 is input (Step 803 ).
- the MPU 19 transmits information on the velocity values (V x , V y ). In this case, the pointer 2 moves on the screen 3 in accordance with a movement of the input apparatus 1 .
- the input apparatus 1 transmits information on the velocity values (V x , V y ) and a small-size screen display signal (Step 805 ).
- the MPU 35 of the control apparatus 40 Upon receiving the small-size screen display signal from the input apparatus 1 , the MPU 35 of the control apparatus 40 controls display on the screen 3 so that a small-size screen 8 is displayed on the screen 3 . Moreover, upon receiving the information on the velocity values (V x , V y ), the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the image 6 is scrolled at a velocity corresponding to the velocity values (V x , V y ). It should be noted that since a small-size screen display signal is transmitted from the input apparatus 1 during the scroll mode, the MPU 35 can distinctively recognize the velocity values (V x , V y ) transmitted in Step 804 and the velocity values (V x , V y ) transmitted in Step 805 .
- FIG. 26 is a diagram showing the image 6 and small-size screen 8 displayed on the screen. As shown in FIG. 26 , the small-size screen 8 is displayed at a lower right-hand corner of the image 6 , for example. It should be noted that a position at which the small-size screen 8 is displayed may be any position as long as it does not lower visibility of the image 6 .
- the small-size screen 8 is sectioned into a first area 8 a (area in slashes in FIG. 26 ) corresponding to the entire image 6 and a second area 8 b corresponding to a part of the image 6 currently being displayed on the screen.
- the MPU 35 of the control apparatus 40 controls display so that the image 6 is scrolled downwardly at a velocity corresponding to the velocity values (V x , V y ). In other words, the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the image 6 is scrolled in an opposite direction from a vector direction of the velocity values (V x , V y ). In addition, the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the second area 8 b moves upwardly in an area in which the small-size screen 8 is displayed. In other words, the MPU 35 of the control apparatus 40 controls display so that the image 6 moves in an opposite direction from a direction in which the image 6 is scrolled.
- the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the image 6 is scrolled in an opposite direction from the direction in which the input apparatus 1 is operated and the second area 8 b moves in a direction in which the input apparatus 1 is operated.
- the user can scroll an image displayed on a screen by merely operating the second area 8 b in the small-size screen 8 . Accordingly, since it becomes possible to perform scroll operations intuitionally, an operational feeling in scroll operations can be improved. Moreover, since the small-size screen 8 is displayed while the button 13 is pressed (during scroll mode), it does not lower visibility during the pointer mode.
- the input apparatus 1 may transmit modified velocity values (V x ′, V y ′) instead of velocity values (V x , V y ) in Step 805 .
- the processing described in the above embodiments can all be applied to this embodiment.
- the scroll direction of the image 6 is biased in (restricted to) the horizontal-axis direction or the vertical-axis direction on the screen, an operational feeling in scroll operations can be additionally improved. The same holds true for modified examples to be described later.
- the input apparatus 1 of the control system 100 transmits information on velocity values (V x , V y ) and a scrollbar display signal in Step 805 shown in FIG. 25 .
- the control apparatus 40 Upon receiving the scrollbar display signal, the control apparatus 40 displays a scrollbar 9 on the screen 3 .
- FIG. 27 is a diagram showing the image 6 and scrollbar 9 displayed on the screen 3 .
- the scrollbar 9 is displayed at a lower end and rightward end on the screen 3 . It should be noted that positions at which the scrollbar 9 is displayed may be any position as long as it does not lower visibility of the image 6 .
- the scrollbar 9 includes an ordinate-axis scrollbar 9 a and an abscissa-axis scrollbar 9 b.
- the MPU 35 of the control apparatus 40 controls display so that the image 6 is scrolled downwardly at a velocity corresponding to velocity values (V x , V y ) transmitted in Step 805 .
- the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the image 6 is scrolled in an opposite direction from a vector direction of the velocity values (V x , V y ).
- the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the ordinate-axis scrollbar 9 a moves upwardly.
- the MPU 35 of the control apparatus 40 controls display so that the ordinate-axis scrollbar 9 a moves in an opposite direction from the direction in which the image 6 is scrolled.
- the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the image 6 is scrolled in an opposite direction from a direction in which the input apparatus 1 is operated and the ordinate-axis scrollbar 9 a and the abscissa-axis scrollbar 9 b are moved in directions in which the input apparatus 1 is operated.
- the user can scroll the image 6 displayed on the screen by merely operating the scrollbar 9 , with the result that an operational feeling in scroll operations can be improved. Moreover, since the scrollbar 9 is displayed while the button 13 is pressed (during scroll mode), it does not lower visibility during the pointer mode.
- the input apparatus 1 of the control system 100 transmits information on velocity values (V x , V y ) and a reference point display signal in Step 805 shown in FIG. 25 .
- the control apparatus 40 Upon receiving the reference point display signal, the control apparatus 40 displays a reference point 43 on the image 6 when the pointer 2 displayed on the screen 3 is positioned on the image 6 , for example.
- FIG. 28 is a diagram showing the image 6 and reference point 43 displayed on the screen 3 .
- the reference point 43 is displayed as, for example, a circular point. It should be noted that a shape of the reference point 43 is not particularly limited.
- the reference point 43 is displayed at a position at which the pointer 2 is positioned at a time the button 13 is pressed.
- the MPU 35 of the control apparatus 40 Upon displaying the reference point 43 on the screen 3 , the MPU 35 of the control apparatus 40 generates coordinate values of the pointer 2 based on information on velocity values (V x , V y ) transmitted from the input apparatus 1 in Step 805 . Then, the MPU 35 of the control apparatus 40 controls display so that the pointer 2 moves on the screen. In other words, in the control system 100 according to the second modified example, the pointer 2 also moves during the scroll mode.
- the MPU 35 of the control apparatus 40 adds the velocity values (V x , V y ) transmitted from the input apparatus 1 in Step 805 to thus generate integration values.
- the MPU 35 of the control apparatus 40 controls display on the screen so that the image 6 is scrolled at a velocity corresponding to the integration values.
- the MPU 35 of the control apparatus 40 controls display on the screen 3 so that the pointer 2 moves in the same direction as a vector direction of the velocity values (V x , V y ) and the image 6 is scrolled in the same direction as the vector direction of the velocity values (V x , V y ).
- the user can scroll the image 6 with the pointer 2 as a guide.
- an operational feeling can be improved.
- the present invention is applicable to input apparatuses such as a planar-operation-type mouse, a touchpad, a joystick, and a pen tablet.
- the present invention may be applied to a slide-resistance-type input apparatus that detects a movement of an operation section inside an opening formed on a casing by a slide resistance.
- the present invention may be applied to an optical input apparatus that calculates a movement amount and operation direction of a finger of a user by irradiating light onto a semicircular operation section provided at an upper portion of a casing and detecting reflected light.
- the present invention may be applied to an electronic apparatus including any of the input apparatuses described above (e.g., laptop PC including touchpad).
- the present invention may be applied to a handheld apparatus that includes a display section, for example.
- a handheld apparatus that includes a display section, for example.
- an image displayed on the display section is scrolled when the user moves a main body of the handheld apparatus.
- the user moves the pointer by moving the main body of the handheld apparatus.
- the handheld apparatus include a PDA (Personal Digital Assistance), a cellular phone, a portable music player, and a digital camera.
- the input apparatus 1 has transmitted input information to the control apparatus 40 wirelessly.
- the input information may be transmitted by wire.
- the pointer 2 that moves on the screen in accordance with the movement of the input apparatus 1 has been represented as an image of an arrow.
- the image of the pointer 2 is not limited to the arrow and may be a simple circle, square, or the like, or a character image or any other images.
- the input apparatus 1 may include, for example, acceleration sensors of three orthogonal axes and angular velocity sensors of three orthogonal axes, and even with only one of the above, the processing shown in the above embodiments can be realized.
- the input apparatus 1 includes a uniaxial acceleration sensor or a uniaxial angular velocity sensor is also conceivable.
- a screen on which a plurality of GUIs as pointing targets of the pointer 2 displayed on the screen 3 are arranged uniaxially is conceivable.
- the input apparatus 1 may include a geomagnetic sensor, an image sensor, and the like instead of the acceleration sensors and the angular velocity sensors.
- each of the angular velocity sensor unit 15 and the acceleration sensor unit 16 of the sensor unit 17 do not necessarily need to be mutually orthogonal like the X′ axis and the Y′ axis described above.
- accelerations respectively projected in the mutually-orthogonal axial directions can be obtained by a calculation that uses a trigonometric function.
- angular velocities about the mutually-orthogonal axes can be obtained by a calculation that uses the trigonometric function.
- the X′ and Y′ detection axes of the angular velocity sensor unit 15 and the X′ and Y′ detection axes of the acceleration sensor unit 16 of the sensor unit 17 described in the above embodiments match.
- those detection axes do not necessarily need to match.
- the angular velocity sensor unit 15 and the acceleration sensor unit 16 may be mounted while being deviated a predetermined rotation angle within a main surface of the substrate so that the detection axes of the angular velocity sensor unit 15 and the acceleration sensor unit 16 do not match.
- accelerations and angular velocities with respect to the respective axes can be obtained by a calculation that uses the trigonometric function.
- the input apparatus 1 is operated 3-dimensionally has been described.
- the present invention is not limited thereto, and the input apparatus may be operated while a part of the casing 10 is in contact with a table, for example.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Position Input By Displaying (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
An input apparatus includes a detection section, a change section, and a transmission section. The detection section detects a movement amount of a user operation in an arbitrary direction. The change section changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection section. The transmission section transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
Description
- This application is a continuation of and claims the benefit under 35 U.S.C. §120 of U.S. patent application Ser. No. 12/645,732, titled “INPUT APPARATUS, CONTROL APPARATUS, CONTROL SYSTEM, ELECTRONIC APPARATUS, AND CONTROL METHOD,” filed on Dec. 23, 2009, which claims the benefit under 35 U.S.C. §119 of Japanese Patent Application 2008-331617, filed on Dec. 25, 2008, each of which is hereby incorporated by reference in its entirety.
- 1. Field of the Invention
- The present invention relates to an input apparatus for operating a GUI (Graphical User Interface), a control apparatus for controlling the GUI in accordance with information transmitted from the input apparatus, a control system including those apparatuses, an electronic apparatus, and a control method.
- 2. Description of the Related Art
- Pointing devices, particularly a mouse and a touchpad, are used as controllers for GUIs widely used in PCs (Personal Computers). Not just as HIs (Human Interfaces) of PCs of the related art, the GUIs are now starting to be used as interfaces for AV equipment and game machines used in living rooms etc. with, for example, televisions as image media. Various pointing devices that a user is capable of operating 3-dimensionally are proposed as controllers for the GUIs of this type (see, for example, Japanese Patent Application Laid-open No. 2001-56743 (paragraphs (0030) and (0031), FIG. 3; hereinafter, referred to as Patent Document 1) and Japanese Examined Patent Publication No. Hei 6-7371 (P. 3, 11.18-20 on left-hand column; hereinafter, referred to as Patent Document 2)).
-
Patent Document 1 discloses an input apparatus including angular velocity gyroscopes of two axes, that is, two angular velocity sensors. When a user holds the input apparatus in hand and swings it vertically and laterally, for example, the angular velocity sensors detect angular velocities about two orthogonal axes, and a signal as positional information of a cursor or the like displayed by a display means is generated in accordance with the angular velocities. The signal is transmitted to a control apparatus, and the control apparatus controls display so that the cursor moves on a screen in response to the signal. -
Patent Document 2 discloses an input apparatus (space mouse) including three acceleration sensors (of three axes) and three angular velocity sensors (of three axes) (gyro). - With the input apparatuses disclosed in
Patent Documents - Incidentally, it is also possible to structure an input apparatus, a control apparatus, and the like so that an image displayed on a screen is scrolled when the input apparatus is operated 3-dimensionally. In this case, the image displayed on the screen is scrolled in accordance with a 3-dimensional operation of the input apparatus.
- In this case, however, since the input apparatus is operated in space without any guide, there is a problem that a scroll direction of the image on the screen is not settled if a movement of the input apparatus is converted into scroll as it is, thus leading to a poor operational feeling. For example, even when a user is meaning to move the input apparatus vertically in space, the input apparatus also moves horizontally against a will of the user. As a result, the input apparatus also detects a movement in the horizontal direction in addition to the movement in the vertical direction. If the movement of the input apparatus is converted into scroll as it is in this case, the image on the screen is scrolled in a direction unintended by the user, thus resulting in a problem of a poor operational feeling.
- In view of the circumstances as described above, there is a need for an input apparatus, a control apparatus, a control system, an electronic apparatus, and a control method that are capable of improving an operational feeling in scrolling an image displayed on a screen.
- According to an embodiment of the present invention, there is provided an input apparatus including a detection means, a change means, and a transmission means.
- The detection means detects a movement amount of a user operation in an arbitrary direction.
- The change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- The transmission means transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- In the embodiment of the present invention, since the ratio of the first movement amount to the second movement amount is changed, a scroll direction of the image can be biased in directions such as a horizontal-axis direction and a vertical-axis direction on the screen. As a result, an image can be prevented from being scrolled in a direction unintended by a user on the screen, with the result that an operational feeling for the user in scrolling an image can be improved.
- The input apparatus may further include a judgment means.
- The judgment means judges a direction of the user operation based on the detected detection value.
- In this case, the change means may change the ratio of the first movement amount to the second movement amount in accordance with the judged direction of the user operation.
- With this structure, the scroll direction of the image can be biased appropriately in accordance with a direction of the user operation.
- In the input apparatus, the change means may change the ratio of the first movement amount to the second movement amount so that a scroll direction of the image is biased in at least the first direction on the screen and the second direction on the screen.
- Since the scroll direction of the image can be biased in the first direction and the second direction on the screen in the embodiment of the present invention, an operational feeling in scrolling an image can be additionally improved.
- In the input apparatus, the change means may change the ratio so that, when the judged direction of the user operation is within a first angle range from the first operation direction, the scroll direction is biased in the first direction, and change the ratio so that, when the judged direction of the user operation is within a second angle range from the second operation direction, the scroll direction is biased in the second direction.
- Assuming that, for example, the first angle range is ±45 degrees from the first operation direction and the second angle range is ±45 degrees from the second operation direction, if a direction of the user operation is within ±45 degrees from the first operation direction, the scroll direction can be biased in the first direction on the screen. On the other hand, if the direction of the user operation is within ±45 degrees from the second operation direction, the scroll direction can be biased in the second direction on the screen.
- The input apparatus may further include an angle range control means.
- The angle range control means variably controls the first angle range and the second angle range.
- In the input apparatus, the angle range control means may variably control the first angle range and the second angle range in accordance with the direction of the user operation.
- With this structure, the first angle range and the second angle range can be changed appropriately in accordance with a direction of the user operation.
- In the input apparatus, the angle range control means may control the first angle range and the second angle range so that the first angle range is widened when the direction of the user operation is within a first modified angle range from the first operation direction and the second angle range is widened when the direction of the user operation is within a second modified angle range from the second operation direction.
- With this structure, when an input operation is made in a direction biased in the first operation direction corresponding to the first direction on the screen (direction within first modified angle range), an image is easily scrolled in the first direction on the screen, whereas it becomes difficult to scroll the image in the second direction on the screen. On the other hand, when an input operation is made in a direction biased in the second operation direction corresponding to the second direction on the screen (direction within second modified angle range), an image is easily scrolled in the second direction on the screen, whereas it becomes difficult to scroll the image in the first direction on the screen. As described above, in the embodiment of the present invention, since the first angle range and the second angle range can be changed appropriately in accordance with a direction of the user operation, an operational feeling for the user in scrolling an image can be additionally improved.
- In the input apparatus, the second angle range may be wider than the first angle range.
- With this structure, when an input operation is made in an oblique direction with respect to the first operation direction and the second operation direction (e.g., direction at angle of 45 degrees from second operation direction), scroll in the second direction is prioritized over the first direction. As a result, an operational feeling in scrolling an image that is long in the second direction on the screen as described above, for example, can be improved.
- In the input apparatus, the change means may change the ratio of the first movement amount to the second movement amount so that the scroll direction of the image is restricted to at least the first direction on the screen and the second direction on the screen.
- In the input apparatus, the change means may change the ratio of the first movement amount to the second movement amount so that the scroll direction of the image is restricted to directions that respectively form predetermined angles with respect to the first direction on the screen and the second direction on the screen.
- In the input apparatus, the detection means may be a sensor that detects the user operation in space.
- According to an embodiment of the present invention, there is provided a control apparatus controlling display of scroll of an image displayed on a screen in accordance with information transmitted from an input apparatus including a detection means for detecting a movement amount of a user operation in an arbitrary direction and a transmission means for transmitting the information on a related value related to a detection value detected by the detection means, the control apparatus including a reception means, a change means, and a display control means.
- The reception means receives the information.
- The change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- The display control means controls the display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- The “related value related to a detection value” may be a detection value itself or an operational value calculated based on the detection value.
- In the embodiment of the present invention, since the ratio of the first movement amount to the second movement amount is changed, a scroll direction of the image can be biased in directions including the first direction and the second direction on the screen. As a result, an image can be prevented from being scrolled in a direction unintended by the user on the screen, with the result that an operational feeling for the user in scrolling an image can be improved.
- According to an embodiment of the present invention, there is provided a control system including an input apparatus and a control apparatus.
- The input apparatus includes a detection means, a change means, and a transmission means.
- The detection means detects a movement amount of a user operation in an arbitrary direction.
- The change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- The transmission means transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- The control apparatus includes a reception means and a display control means.
- The reception means receives the scroll information.
- The display control means controls display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- According to another embodiment of the present invention, there is provided a control system including an input apparatus and a control apparatus.
- The input apparatus includes a detection means and a transmission means.
- The detection means detects a movement amount of a user operation in an arbitrary direction.
- The transmission means transmits information on a related value related to a detection value detected by the detection means.
- The control apparatus includes a reception means, a change means, and a display control means.
- The reception means receives the information.
- The change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- The display control means controls display on the screen so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- According to an embodiment of the present invention, there is provided an electronic apparatus including a display section, a detection means, a change means, and a display control means.
- The display section displays a screen.
- The detection means detects a movement amount of a user operation in an arbitrary direction.
- The change means changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection means.
- The display control means controls display on the screen so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- According to an embodiment of the present invention, there is provided a control method including detecting a movement amount of a user operation in an arbitrary direction.
- A ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction is changed, the first movement amount and the second movement amount corresponding to a detection value detected.
- Display on the screen is controlled so that an image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- According to an embodiment of the present invention, there is provided an input apparatus including a detection section, a change section, and a transmission section.
- The detection section detects a movement amount of a user operation in an arbitrary direction.
- The change section changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on a screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to a detection value detected by the detection section.
- The transmission section transmits the first movement amount and the second movement amount whose ratio has been changed as scroll information of an image displayed on the screen.
- According to an embodiment of the present invention, there is provided a control apparatus controlling display of scroll of an image displayed on a screen in accordance with information transmitted from an input apparatus including a detection means for detecting a movement amount of a user operation in an arbitrary direction and a transmission means for transmitting the information on a related value related to a detection value detected by the detection means, the control apparatus including a reception section, a change section, and a display control section.
- The reception section receives the information.
- The change section changes a ratio of a first movement amount as a movement amount in a first operation direction corresponding to a first direction on the screen to a second movement amount as a movement amount in a second operation direction corresponding to a second direction on the screen different from the first direction, the first movement amount and the second movement amount corresponding to the detected detection value.
- The display control section controls the display on the screen so that the image displayed on the screen is scrolled in accordance with the first movement amount and the second movement amount whose ratio has been changed.
- In the descriptions above, elements described as “ . . . means” may be realized by hardware, or may be realized by both software and hardware. In the case of realization by both the software and hardware, the hardware includes at least a storage device for storing a software program.
- Typically, the hardware is constituted by selectively using at least one of a sensor, a CPU (Central Processing Unit), an MPU (Micro Processing Unit), a RAM (Random Access Memory), a ROM (Read Only Memory), a DSP (Digital Signal Processor), an FPGA (Field Programmable Gate Array), an ASIC (Application Specific Integrated Circuit), a NIC (Network Interface Card), a WNIC (Wireless NIC), a modem, an optical disc, a magnetic disk, and a flash memory.
- As described above, according to the embodiments of the present invention, an input apparatus, a control apparatus, a control system, an electronic apparatus, and a control method that are capable of improving an operational feeling in scrolling an image displayed on a screen can be provided.
- These and other objects, features and advantages of the present invention will become more apparent in light of the following detailed description of best mode embodiments thereof, as illustrated in the accompanying drawings.
-
FIG. 1 is a diagram showing a control system according to an embodiment of the present invention; -
FIG. 2 is a perspective diagram showing an input apparatus; -
FIG. 3 is a diagram schematically showing an internal structure of the input apparatus; -
FIG. 4 is a block diagram showing an electrical structure of the input apparatus; -
FIG. 5 is a diagram showing an example of a screen displayed on a display apparatus; -
FIG. 6 is a diagram showing a state where a user is holding the input apparatus; -
FIGS. 7A and 7B are explanatory diagrams showing typical examples of ways of moving the input apparatus and ways a pointer moves on a screen accordingly; -
FIG. 8 is a perspective diagram showing a sensor unit; -
FIG. 9 is a diagram for explaining an operation of the control system that is carried out when the pointer moves on the screen in accordance with a 3-dimensional operation made by the user (pointer mode); -
FIG. 10 is a flowchart showing an operation of the input apparatus according to the embodiment of the present invention; -
FIGS. 11A and 11B are diagrams for explaining relationships between weighting factors α and β and scroll tilt directions; -
FIG. 12 is a diagram showing an operation of the input apparatus according to another embodiment of the present invention; -
FIGS. 13A and 13B are diagrams showing relationships between operation directions of the input apparatus and scroll directions in a case where the processing shown inFIG. 12 is executed; -
FIG. 14 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention; -
FIG. 15 is a diagram for explaining a first angle range and a second angle range; -
FIGS. 16A and 16B are diagrams showing relationships between the operation directions of the input apparatus and scroll directions in a case where the processing shown inFIG. 14 is executed; -
FIG. 17 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention; -
FIGS. 18A and 183 are diagrams showing temporal changes of ranges of the first angle range and the second angle range in a case where the processing shown inFIG. 17 is executed; -
FIG. 19 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention; -
FIG. 20 is a diagram for explaining a first modified angle range and a second modified angle range; -
FIG. 21 is a flowchart showing an operation of the input apparatus according to another embodiment of the present invention; -
FIG. 22 is a diagram for explaining a third angle range; -
FIG. 23 are diagrams showing relationships between the operation directions of the input apparatus and scroll directions in a case where the processing shown inFIG. 21 is executed; -
FIGS. 24A and 24B are diagrams each showing a relationship between an operation direction of the input apparatus and a direction in which an image is scrolled; -
FIG. 25 is a flowchart showing an operation of the input apparatus of the control system according to another embodiment of the present invention; -
FIG. 26 is a diagram showing an image and a small-size screen displayed on the screen; -
FIG. 27 is a diagram showing an image and scrollbars displayed on the screen; and -
FIG. 28 is a diagram showing an image and a reference point displayed on the screen. - Hereinafter, embodiments of the present invention will be described with reference to the drawings.
-
FIG. 1 is a diagram showing a control system according to a first embodiment of the present invention. Acontrol system 100 includes adisplay apparatus 5, acontrol apparatus 40, and aninput apparatus 1. -
FIG. 2 is a perspective diagram showing theinput apparatus 1. Theinput apparatus 1 is of a size that a user is capable of holding. Theinput apparatus 1 includes acasing 10. Further, theinput apparatus 1 includes an operation section 23 (seeFIG. 6 ) including abutton 11 and abutton 12 adjacent to thebutton 11 that are provided at a center of an upper portion of thecasing 10, and abutton 13 provided at a side portion of thecasing 10. - Typically, the
buttons operation section 23 is not limited to the press-type button, and a bar-type operation section that is operated with one end as a fulcrum, or a slide-type operation section may also be used. Each of thebuttons - The
button 11 has a function corresponding to a left button of a planar-operation-type mouse used for a PC, and thebutton 12 adjacent to thebutton 11 has a function corresponding to a right button of a mouse, for example. For example, an operation of selecting an icon 4 (seeFIG. 5 ) may be carried out by clicking thebutton 11, and an operation of opening a file may be carried out by double-clicking thebutton 11. - The
button 13 has a function as a switch button for switching a pointer mode to a scroll mode and vice versa. The “pointer mode” is a mode in which apointer 2 displayed on a screen 3 (seeFIG. 5 ) is moved in accordance with a movement of thecasing 10. The “scroll mode” is a mode in which animage 6 displayed on thescreen 3 is scrolled in accordance with the movement of thecasing 10. -
FIG. 3 is a diagram schematically showing an internal structure of theinput apparatus 1.FIG. 4 is a block diagram showing an electrical structure of theinput apparatus 1. - The
input apparatus 1 includes asensor unit 17, acontrol unit 30, andbatteries 14. -
FIG. 8 is a perspective diagram showing thesensor unit 17. - The
sensor unit 17 includes anacceleration sensor unit 16 for detecting accelerations in different angles such as along two orthogonal axes (X′ axis and Y′ axis). Specifically, theacceleration sensor unit 16 includes two sensors, that is, afirst acceleration sensor 161 and asecond acceleration sensor 162. - The
sensor unit 17 further includes an angularvelocity sensor unit 15 for detecting angular accelerations about the two orthogonal axes. Specifically, the angularvelocity sensor unit 15 includes two sensors, that is, a firstangular velocity sensor 151 and a secondangular velocity sensor 152. Theacceleration sensor unit 16 and the angularvelocity sensor unit 15 are packaged and mounted on acircuit board 25. - As each of the first
angular velocity sensor 151 and the secondangular velocity sensor 152, a vibration gyro sensor for detecting Coriolis force in proportion to an angular velocity is used. As each of thefirst acceleration sensor 161 and thesecond acceleration sensor 162, any sensor such as a piezoresistive sensor, a piezoelectric sensor, or a capacitance sensor may be used. Each of theangular velocity sensors - In descriptions on
FIGS. 2 and 3 , a longitudinal direction of thecasing 10 is referred to as Z′ direction, a thickness direction of thecasing 10 is referred to as X′ direction, and a width direction of thecasing 10 is referred to as Y′ direction for convenience. In this case, thesensor unit 17 is incorporated into thecasing 10 such that a surface of thecircuit board 25 on which theacceleration sensor unit 16 and the angularvelocity sensor unit 15 are mounted becomes substantially parallel to an X′-Y′ plane. As described above, thesensor units - In the specification, a coordinate system that moves along with the
input apparatus 1, that is, a coordinate system fixed to theinput apparatus 1 is expressed using the X′ axis, Y′ axis, and Z′ axis, whereas a coordinate system stationary on earth, that is, an inertial coordinate system is expressed using the X axis, Y axis, and Z axis. Moreover, in descriptions below, with regard to a movement of theinput apparatus 1, a rotational direction about the X′ axis is sometimes referred to as pitch direction, a rotational direction about the Y′ axis is sometimes referred to as yaw direction, and a rotational direction about the Z′ axis (roll axis) is sometimes referred to as roll direction. - The
control unit 30 includes amain substrate 18, an MPU 19 (Micro Processing Unit) (or CPU) mounted on themain substrate 18, acrystal oscillator 20, atransceiver 21, and anantenna 22 printed on themain substrate 18. - The
MPU 19 includes a built-in volatile or nonvolatile memory requisite therefor. TheMPU 19 is input with a detection signal from thesensor unit 17, an operation signal from the operation section, and the like, and executes various types of operational processing in order to generate predetermined control signals in response to those input signals. The memory may be provided separate from theMPU 19. - Typically, the
sensor unit 17 outputs analog signals. In this case, theMPU 19 includes an A/D (Analog/Digital) converter. However, thesensor unit 17 may be a unit that includes the A/D converter. - The transceiver 21 (transmission means) transmits the control signals generated in the
MPU 19 as RF radio signals to thecontrol apparatus 40 via theantenna 22. Thetransceiver 21 is also capable of receiving various signals transmitted from thecontrol apparatus 40. - The
crystal oscillator 20 generates clocks and supplies them to theMPU 19. As thebatteries 14, dry cell batteries, rechargeable batteries, and the like are used. - The
control apparatus 40 includes an MPU 35 (or CPU), aRAM 36, aROM 37, avideo RAM 41, adisplay control section 42, anantenna 39, and atransceiver 38. - The
transceiver 38 receives the control signal transmitted from theinput apparatus 1 via the antenna 39 (reception means). Thetransceiver 38 is also capable of transmitting various predetermined signals to theinput apparatus 1. TheMPU 35 analyzes the control signal and executes various types of operational processing. Thedisplay control section 42 mainly generates screen data to be displayed on thescreen 3 of thedisplay apparatus 5 under control of theMPU 35. Thevideo RAM 41 serves as a work area of thedisplay control section 42 and temporarily stores the generated screen data. - The
control apparatus 40 may be an apparatus dedicated to theinput apparatus 1, or may be a PC or the like. Thecontrol apparatus 40 is not limited to the apparatus dedicated to theinput apparatus 1, and may be a computer integrally formed with thedisplay apparatus 5, audiovisual equipment, a projector, a game device, a car navigation system, or the like. - Examples of the
display apparatus 5 include a liquid crystal display and an EL (Electro-Luminescence) display. Thedisplay apparatus 5 may alternatively be an apparatus integrally formed with a display and capable of receiving television broadcasts and the like, or an apparatus in which such a display and thecontrol apparatus 40 are integrated. -
FIG. 5 is a diagram showing an example of thescreen 3 displayed on thedisplay apparatus 5. GUIs such asicons 4 and thepointer 2 are displayed on thescreen 3. The icons are images on thescreen 3 representing functions of programs, execution commands, file contents, and the like on the computer. Moreover, on thescreen 3, animage 6 such as a web image including a plurality ofletters 7 is displayed, for example. -
FIG. 6 is a diagram showing a state where a user is holding theinput apparatus 1. As shown inFIG. 6 , theinput apparatus 1 may include, as theoperation section 23, in addition to thebuttons various operation buttons 29 such as those provided to a remote controller for operating a television or the like and apower switch 28, for example. Command signals generated when the user moves theinput apparatus 1 in the air or operates theoperation section 23 while holding theinput apparatus 1 as shown in the figure are output to thecontrol apparatus 40, and thecontrol apparatus 40 controls the GUI. - Next, a description will be given on typical examples of ways of moving the
input apparatus 1 and ways thepointer 2 moves on thescreen 3 accordingly.FIGS. 7A and 7B are explanatory diagrams therefor. - As shown in
FIGS. 7A and 7B , the user holds theinput apparatus 1 so as to aim thebuttons input apparatus 1 at thedisplay apparatus 5 side. The user holds theinput apparatus 1 so that a thumb is located on an upper side and a pinky is located on a lower side as in handshakes. In this state, the circuit board 25 (seeFIG. 8 ) of thesensor unit 17 is close to being in parallel with thescreen 3 of thedisplay apparatus 5, and the two axes as detection axes of thesensor unit 17 respectively correspond to the horizontal axis (X axis) and the vertical axis (Y axis) on thescreen 3. Hereinafter, the position of theinput apparatus 1 as shown inFIGS. 7A and 7B is referred to as reference position. - As shown in
FIG. 7A , when the user moves a wrist or an arm in the vertical direction, that is, the pitch direction from the reference position, thesecond acceleration sensor 162 detects an acceleration ay in the Y′-axis direction and the secondangular velocity sensor 152 detects an angular velocity ωθ about the X′ axis. Based on those physical amounts, thecontrol apparatus 40 controls display of thepointer 2 so as to move thepointer 2 in the vertical direction on thescreen 3. - Meanwhile, as shown in
FIG. 7B , when the user moves the wrist or the arm in the lateral direction, that is, the yaw direction from the reference position, thefirst acceleration sensor 161 detects an acceleration ax in the X′-axis direction and the firstangular velocity sensor 151 detects an angular velocity ωψ about the Y′ axis. Based on the thus-detected physical amounts, thecontrol apparatus 40 controls display of thepointer 2 so as to move thepointer 2 in the horizontal direction on thescreen 3. - (Description on Operation)
- Next, an operation of the
control system 100 structured as described above will be described. - First, an operation of the
control system 100 in a case where thepointer 2 moves on thescreen 3 in accordance with a 3-dimensional operation made by the user (pointer mode) will be described briefly.FIG. 9 is a flowchart showing the operation of thecontrol system 100 in this case. - As shown in
FIG. 9 , when the user presses thepower supply switch 28 and the power of theinput apparatus 1 is thus turned on, for example, biaxial angular velocity signals are output from the angular velocity sensor unit. TheMPU 19 acquires angular velocity values (ωψ, ωθ) from the angular velocity signals (Step 101). - Further, upon turning on the power of the
input apparatus 1, biaxial acceleration signals are output from theacceleration sensor unit 16. TheMPU 19 acquires acceleration values (ax, ay) from the biaxial acceleration signals (Step 102). - The
MPU 19 typically carries out the process of acquiring angular velocity values (ωψ, ωθ) (Step 101) and the process of acquiring acceleration values (ax, ay) (Step 102) in sync. However, the process of acquiring angular velocity values (ωψ, ωθ) and the process of acquiring acceleration values (ax, ay) do not always need to be carried out in sync (at the same time). For example, the acceleration values (ax, ay) may be obtained after the angular velocity values (ωψ, ωθ) are obtained, or the angular velocity values (ωψ, ωθ) may be obtained after the acceleration values (ax, ay) are obtained. - Based on the acceleration values (ax, ay) and the angular velocity values (ωψ, ωθ), the
MPU 19 calculates velocity values (first velocity value Vx and second velocity value Vy) by a predetermined operation (Step 103). The first velocity value Vx is a velocity value in a direction along the X′ axis, and the second velocity value Vy is a velocity value in a direction along the Y′ axis. - As a method of calculating velocity values, there is a method in which the
MPU 19 obtains radius gyrations (Rψ, Rθ) of the movement of theinput apparatus 1 by dividing the acceleration values (ax, ay) by angular acceleration values (Δωψ, Δωθ), and calculates velocity values by multiplying the radius gyrations (Rψ, Rθ) by the angular velocity values (ωψ, ωθ). Alternatively, the radius gyrations (Rψ, Rθ) may be obtained by dividing acceleration change rates (Δax, Δay) by angular acceleration change rates (Δ(Δωψ), Δ(Δωθ)). An effect of gravity accelerations can be removed when the radius gyrations (Rψ, Rθ) are calculated by dividing the acceleration change rates (Δax, Δay) by the angular acceleration change rates (Δ(Δωψ), Δ(Δωθ)). - As another example of the method of calculating the velocity values (Vx, Vy), there is a method in which the
MPU 19 calculates the velocity values by, for example, integrating the acceleration values (ax, ay) while using the angular velocity values (ωψ, ωθ) as an adjunct for the integration operation. - By calculating the velocity values by the calculation method described above, an operational feeling of the
input apparatus 1 that matches an intuition of the user can be obtained, and moreover, the movement of thepointer 2 on thescreen 3 also accurately matches the movement of theinput apparatus 1. However, the velocity values (Vx, Vy) do not always need to be calculated by the calculation method above. For example, it is also possible for the velocity values (Vx, Vy) to be calculated by simply integrating the acceleration values (ax, ay). Alternatively, the detected angular velocity values (ωψ, ωθ) may be used as they are as the velocity values (Vx, Vy). - The
MPU 19 transmits information on the calculated velocity values (Vx, Vy) to thecontrol apparatus 40 via thetransceiver 21 and the antenna 22 (Step 104). - The
MPU 35 of thecontrol apparatus 40 receives the information on the velocity values (Vx, Vy) via theantenna 39 and the transceiver 38 (Step 105). In this case, theinput apparatus 1 transmits the velocity values (Vx, Vy) every predetermined number of clocks, that is, every time a predetermined time passes, so thecontrol apparatus 40 receives the velocity values every predetermined number of clocks. - Upon receiving the velocity values, the
MPU 35 of thecontrol apparatus 40 generates new coordinate values (X(t), Y(t)) by adding the velocity values to coordinate values using Equations (1) and (2) below (Step 106). TheMPU 35 controls display on the screen so that thepointer 2 moves to a position corresponding to the generated coordinate values (Step 107). -
X(t)=X(t−1)+V x (1) -
Y(t)=Y(t−1)+V y (2) - It should be noted that the calculation of the velocity values (Vx, Vy) may be executed by the
control apparatus 40. In this case, theinput apparatus 1 transmits information on the angular velocity values (ωθ, ωθ) and the acceleration values (ax, ay) to thecontrol apparatus 40 via thetransceiver 21 and theantenna 22. Based on the information on the angular velocity values (ωθ, ωθ) and the acceleration values (ax, ay) received via theantenna 39 and thetransceiver 38, thecontrol apparatus 40 calculates the velocity values (Vx, Vy). The method of calculating the velocity values is as described above. - Next, an operation of the input apparatus during the pointer mode and the scroll mode will be described.
-
FIG. 10 is a flowchart showing an operation of the input apparatus. - As shown in
FIG. 10 , theMPU 19 acquires angular velocity values (ωψ, ωθ) and acceleration values (ax, ay) from the angularvelocity sensor unit 15 and the acceleration sensor unit 16 (Step 201). Based on the acquired angular velocity values (ωψ, ωθ) and acceleration values (ax, ay), theMPU 19 calculates velocity values (Vx, Vy) (Step 202). - Upon calculating the velocity values (Vx, Vy), the
MPU 19 judges whether an operation signal from a switch (not shown) provided to thebutton 13 is input (Step 203). When the user has not pressed thebutton 13 and an operation signal from the switch is not yet input (NO in Step 203), theMPU 19 transmits the calculated velocity values (Vx, Vy) to thecontrol apparatus 40 as information on a movement amount of the pointer 2 (Step 204). Upon transmitting information on the velocity values (Vx, Vy), theMPU 19 returns to Step 201. - Upon receiving the information on the velocity values (Vx, Vy), the
MPU 35 of thecontrol apparatus 40 generates new coordinate values and controls display on thescreen 3 so that thepointer 2 moves to a position corresponding to the generated coordinate values (pointer mode). - When the user presses the
button 13, an operation signal is output from the switch to be input to the MPU 19 (YES in Step 203). Upon being input with the operation signal, theMPU 19 multiplies the first velocity value Vx and the second velocity value Vy by a first weighting factor α and a second weighting factor β, respectively, as expressed in Equations (3) and (4) below to thus calculate a first modified velocity value Vx′ and a second modified velocity value Vy′ (Step 205). -
V x ′=αV x (3) -
V y ′=βV y (4) - Here, the weighting factors (α, β) are typically different values and stored in a memory (not shown), for example. By multiplying the different weighting factors (α, β) to the velocity values (Vx, Vy), the
MPU 19 changes a ratio of the first velocity value Vx to the second velocity value Vy (ratio change means). The weighting factors (α, β) can take various values. By setting the weighting factors (α, β) as appropriate, a scroll direction can be biased in a vertical-axis (Y-axis) direction or a horizontal-axis (X-axis) direction on thescreen 3. Details on relationships between the weighting factors (α, β) and scroll tilt directions will be described later. - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 206). Upon transmitting the information on the modified velocity values (Vx′, Vy′), theMPU 19 returns to Step 201. - The
MPU 35 of thecontrol apparatus 40 receives the transmitted information on the modified velocity values (Vx′, Vy′). When theimage 6 displayed on thescreen 3 is in an active state or thepointer 2 is positioned inside theimage 6 on thescreen 3, for example, theMPU 35 controls display so that theletters 7 inside theimage 6 are scrolled at a velocity corresponding to the received modified velocity values (Vx′, Vy′) (scroll mode). It should be noted that examples of theimage 6 as a scroll target include a web image, a map, and an EPG (Electronic Program Guide). - By the processing shown in
FIG. 10 , by the user operating the input apparatus 3-dimensionally while pressing thebutton 13, theimage 6 displayed on thescreen 3 is scrolled in a direction biased in the vertical-axis direction or the horizontal-axis direction. - When the information on the modified velocity values (Vx′, Vy′) is transmitted in Step 206, a signal transmitted from the
input apparatus 1 to thecontrol apparatus 40 contains, in addition to the information on the modified velocity values (Vx′, Vy′), a signal for causing thecontrol apparatus 40 to control display of scroll. Accordingly, since thecontrol apparatus 40 can distinctively recognize the pointer mode and the scroll mode, display of scroll on the screen can be controlled when the modified velocity values (Vx′, Vy′) are transmitted. It should be noted that as another method used for thecontrol apparatus 40 to distinctively recognize the pointer mode and the scroll mode, there is a method of transmitting a mode switch signal that indicates that a mode has been switched. Alternatively, thecontrol apparatus 40 can distinctively recognize the pointer mode and the scroll mode also by transmission of a signal indicating that thebutton 13 has been pressed (e.g., press code). Any method may be adopted for the method used for thecontrol apparatus 40 to distinctively recognize the pointer mode and the scroll mode. - (Relationships Between Weighting Factors (α, β) and Scroll Tilt Directions)
- Next, relationships between the weighting factors (α, β) and scroll tilt directions will be described.
-
FIGS. 11A and 11B are diagrams for explaining the relationships between the weighting factors (α, β) and scroll tilt directions. - As shown in
FIG. 11A , when the first weighting factor α is set to be smaller than the second weighting factor β, a scroll direction of theimage 6 is biased in the vertical-axis (Y-axis) direction on thescreen 3 with respect to an operation direction (movement direction) of theinput apparatus 1. In this case, the weighting factors (α, β) are set to, for example, (⅓, 1), (½, 1), (½, 2), (½, 3), (1, 2), (1, 3), or (1, 4). The weighting factors (α, β) are not limited to those values and may of course take other values. - By thus setting the first weighting factor α to be smaller than the second weighting factor β, the scroll direction can be biased in the vertical-axis direction on the screen. Accordingly, an operational feeling in scroll operations can be improved in a case where the
image 6 is long in the vertical-axis direction on thescreen 3 as a whole, for example. Since theimage 6 such as a web image is, in many cases, long in the vertical-axis direction on thescreen 3 as a whole in particular, an operational feeling in scrolling theimage 6 such as a web image can be improved. - As shown in
FIG. 11B , when the first weighting factor α is set to be larger than the second weighting factor β, the scroll direction of theimage 6 is biased in the horizontal-axis (X-axis) direction on thescreen 3 with respect to the operation direction of theinput apparatus 1. In this case, the weighting factors (α, β) are set to, for example, (4, 1), (3, 1), (2, 1), (3, ½), (2, ½), (1, ½), or (1, ⅓). The weighting factors (α, β) are not limited to those values and may of course take other values. - By thus setting the first weighting factor α to be larger than the second weighting factor β, the scroll direction can be biased in the horizontal-axis direction on the screen. Accordingly, an operational feeling in scroll operations can be improved in a case where the
image 6 is long in the horizontal-axis direction on thescreen 3 as a whole, for example. - Here, it is also possible to set the weighting factors such that either the first weighting factor α or the second weighting factor β is set to 0 like (1, 0) and (0, 1).
- For example, when the weighting factors (α, β) are (1, 0), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by 1 and 0, respectively, to thus calculate the first and second modified velocity values (Vx′, Vy′) in Step 205. Then, theMPU 19 transmits information on the calculated modified velocity values (Vx′, Vy′) to the control apparatus 40 (Step 206). In this case, theimage 6 is scrolled only in the vertical-axis direction and not in the horizontal-axis direction on thescreen 3. In other words, the scroll direction is restricted to the vertical-axis direction on thescreen 3. - Similarly, when the weighting factors (α, β) are (0, 1), for example, the
image 6 is scrolled only in the horizontal-axis direction and not in the vertical-axis direction on thescreen 3. In other words, the scroll direction is restricted to the horizontal-axis direction on thescreen 3. - It should be noted that in the specification, the expression “scroll direction is biased” means that, as shown in
FIGS. 11A and 11B , the scroll direction is biased in a predetermined axial direction on the screen 3 (e.g., vertical-axis direction). On the other hand, the expression “scroll direction is restricted” means that the scroll direction is biased at maximum to a predetermined axial direction on the screen and scroll cannot be performed in any other directions. - Next, a description will be given on a case where the weighting factors (α, β) are set such that either the first weighting factor α or the second weighting factor β is set to 1 like (0, 1), (½, 1), (1, 2), (2, 1), (1, ½), and (1, 0).
- When the weighting factors (α, β) are, for example, (½, 1), the first velocity value Vx is multiplied by ½ and reduced, and a first modified velocity value Vx′ is thus obtained (Step 205). Moreover, the second velocity value Vy is multiplied by 1 to thus obtain a second modified velocity value Vy′. The value obtained by multiplying the second velocity value Vy by 1 (second modified velocity value Vy′) is the second velocity value Vy itself, so the second modified velocity value Vy′ does not need to be calculated. In this case, the
MPU 19 only needs to transmit the first modified velocity value Vx′ and the second velocity value Vy to thecontrol apparatus 40 as scroll information in Step 206. - In other words, when either one of the weighting factors (α, β) is 1, one of the modified velocity values (Vx′, Vy′) corresponding to one of the velocity values (Vx, Vy) to which 1 is multiplied does not need to be calculated. Accordingly, a calculation amount can be reduced, with the result that power consumption of the
input apparatus 1 can be reduced. - The processing shown in
FIG. 10 may be mainly executed by thecontrol apparatus 40. - In this case, the
control apparatus 40 receives information on velocity values (Vx, Vy) transmitted from theinput apparatus 1. Upon receiving the information on the velocity values (Vx, Vy), theMPU 35 of thecontrol apparatus 40 calculates modified velocity values (Vx′, Vy′) by multiplying the received velocity values (Vx, Vy) by the weighting factors (α, β). Then, theMPU 35 controls display on the screen so that the image displayed on the screen is scrolled at a velocity corresponding to the modified velocity values (Vx′, Vy′). It should be noted that processing according to embodiments and modified examples of the present invention to be described hereinbelow can all be applied as processing of thecontrol apparatus 40. - Next, a second embodiment of the present invention will be described. The first embodiment above has described a case where the scroll direction is biased in (restricted to) a uniaxial direction of one of the horizontal-axis direction and the vertical-axis direction on the
screen 3. The second embodiment is different from the first embodiment in that the scroll direction is biased in (restricted to) biaxial directions of the horizontal-axis direction and the vertical-axis direction on thescreen 3. Therefore, that point will mainly be described. -
FIG. 12 is a flowchart showing an operation of theinput apparatus 1 according to the second embodiment. - As shown in
FIG. 12 , in Steps 301 to 304, processes that are the same as those of Steps 201 to 204 ofFIG. 10 are executed. In other words, when judged that thebutton 13 is not pressed (NO in Step 303), information on velocity values is transmitted from the input apparatus 1 (Step 304), and thepointer 2 displayed on thescreen 3 is moved at a velocity corresponding to the velocity values. - When the user presses the
button 13, an operation signal is output from the switch provided to thebutton 13 and input to the MPU 19 (YES in Step 303). - Upon being input with the operation signal, the
MPU 19 judges whether an absolute value of the first velocity value |Vx| is larger than an absolute value of the second velocity value |Vy|. By comparing the absolute value of the first velocity value |Vx| and the absolute value of the second velocity value |Vy| in Step 305, theMPU 19 judges an operation direction (movement direction) of the input apparatus 1 (judgment means). Specifically, when the absolute value of the first velocity value |Vx| is larger than the absolute value of the second velocity value |Vy|, theMPU 19 judges that theinput apparatus 1 is being operated in a direction biased in the X′-axis direction. Similarly, when the absolute value of the second velocity value |Vy| is larger than the absolute value of the first velocity value |Vx|, theMPU 19 judges that theinput apparatus 1 is being operated in a direction biased in the Y′-axis direction. - When judged that the absolute value of the first velocity value |Vx| is larger than the absolute value of the second velocity value |Vy| (YES in Step 305), the
MPU 19 sets the first weighting factor α to be larger than the second weighting factor β (Step 306). On the other hand, when judged that the absolute value of the first velocity value |Vx| is smaller than the absolute value of the second velocity value |Vy| (NO in Step 305), theMPU 19 set the first weighting factor α to be smaller than the second weighting factor β (Step 307). Values determined in advance are used as the weighting factors (α, β) set in Steps 306 and 307. For example, the weighting factors (α, β) set in Step 306 are, for example, (1, ½), and the weighting factors (α, β) set in Step 307 are, for example, (½, 1). As other combinations of the weighting factors (α, β) set in Steps 306 and 307, there are, for example, [(1, 0) and (0, 1)], [(1, ⅓) and (⅓, 1)], [(1, 2) and (2, 1)], and [(1, 3) and (3, 1)]. However, the combination is not limited to those combinations, and other values may be used instead. - Upon setting the weighting factors (α, β), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by the first and second weighting factors (α, β), respectively, to thus calculate first and second modified velocity values (Vx′, Vy′) (Step 308). - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 309). - Upon receiving the transmitted information on the modified velocity values (Vx′, Vy′), the
MPU 35 of thecontrol apparatus 40 controls display so that theletters 7 in theimage 6 are scrolled at a velocity corresponding to the received modified velocity values (Vx′, Vy′). -
FIGS. 13A and 13B are diagrams showing relationships between operation directions of theinput apparatus 1 and scroll directions in a case where the processing shown inFIG. 12 is executed.FIG. 13A shows relationships between operation directions of theinput apparatus 1 and scroll directions in a case where a combination of weighting factors set in Steps 306 and 307 is, for example, [(1, ½) and (½, 1)] or [(2, 1) and (1, 2)].FIG. 13B shows relationships between operation directions of theinput apparatus 1 and scroll directions in a case where 0 (or value that is substantially 0) is used as in [(1, 0) and (0, 1)] and [(2, 0) and (0, 2)], for example. - As shown in
FIG. 13A , when the user operates theinput apparatus 1 in a direction within an angle range of ±45 degrees from a direction along the X′-axis direction, a scroll direction of an image on thescreen 3 is biased in the horizontal-axis (X-axis) direction on the screen. On the other hand, when the user operates theinput apparatus 1 in a direction within an angle range of ±45 degrees from a direction along the Y′-axis direction, the scroll direction of the image on thescreen 3 is biased in the vertical-axis (Y-axis) direction on the screen. - As shown in
FIG. 13B , if 0 is used for the weighting factors (α, β), when the user operates theinput apparatus 1 in a direction within an angle range of ±45 degrees from a direction along the X′-axis direction, a scroll direction of theimage 6 is restricted to the horizontal-axis (X-axis) direction on the screen. On the other hand, when the user operates theinput apparatus 1 in a direction within an angle range of ±45 degrees from a direction along the Y′-axis direction, the scroll direction of theimage 6 is restricted to the vertical-axis (Y-axis) direction on the screen. - As described above, since the scroll direction can be biased (restricted) appropriately in accordance with the operation direction of the
input apparatus 1 in theinput apparatus 1 according to the second embodiment, an operational feeling in scroll operations can be additionally improved. - Next, an input apparatus according to a third embodiment of the present invention will be described.
- The third embodiment mainly describes points different from those of the second embodiment above.
-
FIG. 14 is a flowchart showing an operation of theinput apparatus 1 according to the third embodiment. - As shown in
FIG. 14 , in Steps 401 to 404, processes that are the same as those of Steps 301 to 304 ofFIG. 12 are executed. In this case, by the user operating theinput apparatus 1 3-dimensionally in a state where thebutton 13 is not pressed, thepointer 2 moves on thescreen 3 in accordance with the 3-dimensional operation. - When the
button 13 is pressed, an operation signal is output from the switch provided to thebutton 13 and input to the MPU 19 (YES in Step 403). Upon being input with the operation signal, theMPU 19 calculates a tilt angle of a combined vector of the first velocity value and the second velocity value using Equation (5) below (Step 405). By calculating the combined vector tilt angle, theMPU 19 judges an operation direction (movement direction) of theinput apparatus 1. -
arctan(V y /V x)=ξ (5) - Upon calculating the combined vector tilt angle ξ, the
MPU 19 judges whether the combined vector tilt angle ξ is an angle within a first angle range (Step 406). - Now, the first angle range and a second angle range will be described.
-
FIG. 15 is a diagram for explaining the first angle range and the second angle range. - As shown in
FIG. 15 , the first angle range indicates a range within a predetermined angle from 0 degree (or 180 degrees; same holds true for descriptions below) (e.g., 0±30 degrees). The second angle range indicates a range within a predetermined angle from 90 degrees (or 270 degrees; same holds true for descriptions below) (e.g., 90±60 degrees). Theinput apparatus 1 stores the first angle range and the second angle range as shown inFIG. 15 in a memory. The horizontal-axis direction within the angle ranges shown inFIG. 15 corresponds to a movement direction (operation direction) of theinput apparatus 1 in the horizontal-axis direction, and the vertical-axis direction corresponds to the movement direction (operation direction) of theinput apparatus 1 in the vertical-axis direction. - The first angle range and the second angle range can be set variously, but in the description on
FIG. 14 , the first angle range is assumed to be an angle range of 0±30 degrees and the second angle range is assumed to be an angle range of 90±60 degrees for convenience. - It should be noted that the
MPU 19 may judge whether the combined vector tilt angle ξ is an angle within the second angle range in Step 406. - When judged that the combined vector tilt angle ξ is an angle within the first angle range (YES in Step 406), the
MPU 19 sets the first weighting factor α to be larger than the second weighting factor β (Step 407). On the other hand, when judged that the combined vector tilt angle ξ is not an angle within the first angle range (NO in Step 406), theMPU 19 sets the first weighting factor α to be smaller than the second weighting factor β (Step 408). - Upon setting the weighting factors (α, β), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by the first and second weighting factors (α, β), respectively, to thus calculate first and second modified velocity values (Vx′, Vy′) (Step 409). - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 410). - Upon receiving the transmitted information on the modified velocity values (Vx′, Vy′), the
MPU 35 of thecontrol apparatus 40 controls display so that theletters 7 in theimage 6 are scrolled at a velocity corresponding to the received modified velocity values (Vx′, Vy′). -
FIGS. 16A and 16B are diagrams showing relationships between operation directions of theinput apparatus 1 and scroll directions in a case where the processing shown inFIG. 14 is executed.FIG. 16A is a diagram showing relationships between operation directions of theinput apparatus 1 and scroll directions in a case where a combination of weighting factors set in Steps 407 and 408 is, for example, [(1, ½) and (½, 1)] or [(2, 1) and (1, 2)].FIG. 16B is a diagram showing relationships between operation directions of theinput apparatus 1 and scroll directions in a case where 0 (or value that is substantially 0) is used for the weighting factors (α, β) as in [(1, 0) and (0, 1)] and [(2, 0) and (0, 2)], for example. - As shown in
FIG. 16A , when the user operates theinput apparatus 1 in a direction within an angle range of ±30 degrees from the direction along the X′-axis direction, a scroll direction of the image on thescreen 3 is biased in the horizontal-axis (X-axis) direction on the screen. On the other hand, when the user operates theinput apparatus 1 in a direction within an angle range of ±60 degrees from the direction along the Y′-axis direction, the scroll direction of the image on thescreen 3 is biased in the vertical-axis (Y-axis) direction on the screen. - As shown in
FIG. 16B , if 0 (or value that is substantially 0) is used for the weighting factors (α, β), when the user operates theinput apparatus 1 in a direction within an angle range of ±30 degrees from the direction along the X′-axis direction, a scroll direction of theimage 6 is restricted to the horizontal-axis (X-axis) direction on the screen. On the other hand, when the user operates theinput apparatus 1 in a direction within an angle range of ±60 degrees from the direction along the Y′-axis direction, the scroll direction of theimage 6 is restricted to the vertical-axis (Y-axis) direction on thescreen 3. - As described above, since the second angle range is set to be larger than the first angle range in the
input apparatus 1 according to the third embodiment, theimage 6 can be scrolled in the vertical-axis direction on thescreen 3 with high sensitivity. As a result, an operational feeling in scroll operations can be additionally improved in a case where theimage 6 is long in the vertical-axis direction on thescreen 3 as a whole. - Here, the first angle range and the second angle range can be set variously as described above. Examples of the combination of the first angle range and the second angle range include combinations of (0±35 degrees, 90±55 degrees) and (0±40 degrees, 90±50 degrees).
- Alternatively, the first angle range may be set to be larger than the second angle range. Examples of the combination of the first angle range and the second angle range in this case include combinations of (0±60 degrees, 90±30 degrees), (0±55 degrees, 90±35 degrees), and (0±50 degrees, 90±40 degrees). When the first angle range is set to be larger than the second angle range, the
image 6 can be scrolled in the horizontal-axis direction on thescreen 3 with high sensitivity. As a result, an operational feeling in scroll operations can be additionally improved in a case where theimage 6 is long in the horizontal-axis direction on thescreen 3 as a whole. - Next, an input apparatus according to a fourth embodiment of the present invention will be described.
- The fourth embodiment is different from the third embodiment above in that the first angle range and the second angle range are controlled variably. Therefore, that point will mainly be described.
-
FIG. 17 is a flowchart showing an operation of theinput apparatus 1 according to the fourth embodiment. - As shown in
FIG. 17 , upon calculating velocity values based on acquired acceleration values and angular velocity values (Steps 501 and 502), theMPU 19 stores the calculated velocity values in the memory (Step 503). Next, theMPU 19 judges whether an operation signal from the switch of thebutton 13 is input (Step 504). When judged that an operation signal is not yet input (NO in Step 504), theMPU 19 transmits information on the velocity values as information on a movement amount of the pointer 2 (Step 505). - On the other hand, when the user presses the
button 13 and an operation signal from the switch of thebutton 13 is input (YES in Step 504), theMPU 19 reads out velocity values of past n histories that are stored in the memory. Then, theMPU 19 calculates a combined vector of the read-out velocity values (Step 506). Typically, theMPU 19 obtains a sum ΣVx and sum ΣVy of first velocity values Vx and second velocity values Vy of past n histories that are stored in the memory and calculates a combined vector. - Upon calculating the combined vector, the
MPU 19 calculates a combined vector tilt angle ξ′ by Equation (6) below (Step 507). -
arctan=ΣV y /ΣV x)=ξ′ (6) - Upon calculating the combined vector tilt angle ξ′, the
MPU 19 judges whether the combined vector tilt angle ξ′ is an angle within the first angle range (Step 508). When judged that the combined vector tilt angle ξ′ is an angle within the first angle range (YES in Step 508), theMPU 19 widens the first angle range (Step 509) (angle range control means). In this case, the second angle range is narrowed. Upon widening the first angle range, theMPU 19 sets the first weighting factor α to be larger than the second weighting factor β (Step 510). - On the other hand, when judged that the combined vector tilt angle ξ′ is not an angle within the first angle range (NO in Step 508), that is, when judged that the combined vector tilt angle ξ′ is an angle within the second angle range, the
MPU 19 narrows the first angle range (Step 511). In this case, the second angle range is widened. Upon narrowing the first angle range, theMPU 19 sets the first weighting factor α to be smaller than the second weighting factor β (Step 512). - Upon setting the weighting factors (α, β), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by the first and second weighting factors (α, β), respectively, to thus calculate first and second modified velocity values (Vx′, Vy′) (Step 513). - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 514). -
FIGS. 18A and 18B are diagrams showing temporal changes of ranges of the first angle range and the second angle range in a case where the processing shown inFIG. 17 is executed.FIG. 18A is a diagram showing temporal changes of the first angle range and the second angle range in a case where the user operates theinput apparatus 1 in the horizontal-axis (X′-axis) direction.FIG. 18B is a diagram showing temporal changes of the first angle range and the second angle range in a case where the user operates theinput apparatus 1 in the vertical-axis (Y′-axis) direction. - As shown in
FIG. 18A , when the user operates theinput apparatus 1 in the horizontal-axis direction, the first angle range is gradually widened. As a result, when the user operates theinput apparatus 1 in the horizontal-axis direction, it becomes easier with time to perform a scroll operation in the horizontal-axis direction with respect to the operation direction of theinput apparatus 1. - As shown in
FIG. 183 , when the user operates theinput apparatus 1 in the vertical-axis direction, it becomes easier with time to perform a scroll operation in the vertical-axis direction with respect to the operation direction of theinput apparatus 1. - For example, when the user holds the
input apparatus 1 and moves it in the vertical-axis direction from the reference position, the user might swing his/her arm in an oblique direction from the vertical-axis direction. However, in theinput apparatus 1 according to the fourth embodiment, the second angle range is in a widened state when an arm is swung. Therefore, even when the user swings an arm and operates theinput apparatus 1 in an oblique direction, scroll in the vertical-axis direction is prioritized on the screen. Thus, since the first angle range and the second angle range are controlled variably in theinput apparatus 1 according to the fourth embodiment, an operational feeing for the user in operating theimage 6 displayed on thescreen 3 can be additionally improved. - In the description on
FIG. 17 , a case where a combined vector is calculated by obtaining sums of first velocity values Vx and second velocity values Vy of past n histories in Step 506 has been described. However, it is also possible for theMPU 19 to calculate mean values of first velocity values Vx and second velocity values Vy of past n histories in Step 506. Alternatively, a moving average of the first and second velocity values may be obtained. Alternatively, a value passed through an LPF (Lowpass Filter) (hereinafter, referred to as LPF-passed value) may be used as the velocity value in Step 506. When an IIR (Infinite Impulse Response) filter or an FIR (Finite Impulse Response) filter is used as the LPF, the LPF-passed value only needs to be stored in the memory in Step 503. - Next, a fifth embodiment of the present invention will be described. In a description on the fifth embodiment, points different from those of the fourth embodiment will be mainly described.
-
FIG. 19 is a flowchart showing an operation of theinput apparatus 1 according to the fifth embodiment. - As shown in
FIG. 19 , in Steps 601 to 605, processes that are the same as those of Steps 501 to 505 ofFIG. 17 are executed, and by the user operating theinput apparatus 1 in a state where thebutton 13 is not pressed, thepointer 2 moves on thescreen 3. - When the user presses the
button 13 and an operation signal from the switch is input (YES in Step 604), theMPU 19 reads out velocity values (Vx, Vy) of past n histories that are stored in the memory and calculates a combined vector of the read-out velocity values (Vx, Vy) (Step 606). Typically, theMPU 19 obtains sums of first velocity values and second velocity values of past n histories that are stored in the memory and calculates a combined vector. - Upon calculating the combined vector of the velocity values, the
MPU 19 calculates a combined vector tilt angle ξ′ by Equation (6) above (Step 607). Next, theMPU 19 judges whether the combined vector tilt angle ξ′ is an angle within a first modified angle range (Step 608). -
FIG. 20 is a diagram for explaining the first modified angle range and second modified angle range. The first modified angle range is an angle range for changing the first angle range and the second angle range and indicates an angle range of, for example, ±45 degrees from 0 degree (or 180 degrees; same holds true for descriptions below). The second modified angle range is an angle range for changing the first angle range and the second angle range and indicates an angle range of, for example, ±45 degrees from 90 degrees (or 270 degrees; same holds true for descriptions below). The horizontal-axis direction within the modified angle ranges shown inFIG. 20 corresponds to a movement direction (operation direction) of theinput apparatus 1 in the horizontal-axis direction, and the vertical-axis direction corresponds to the movement direction (operation direction) of theinput apparatus 1 in the vertical-axis direction. - The first modified angle range and the second modified angle range are fixed and do not fluctuate by the combined vector tilt angle ξ′.
- The first modified angle range and the second modified angle range are not limited to the range of 0±45 (or 90±45) degrees. The first modified angle range and the second modified angle range can be changed as appropriate.
- It should be noted that it is also possible to judge whether the combined vector tilt angle ξ′ is an angle within the second modified angle range in Step 608.
- When judged that the combined vector tilt angle ξ′ is an angle within the first modified angle range (YES in Step 608), the
MPU 19 widens the first angle range (Step 609). In this case, the second angle range is narrowed. On the other hand, when judged that the combined vector tilt angle ξ′ is not an angle within the first modified angle range (NO in Step 608), that is, when judged that the combined vector tilt angle ξ′ is an angle within the second modified angle range, theMPU 19 narrows the first angle range (Step 610). In this case, the second angle range is widened. - Next, the
MPU 19 judges whether the combined vector tilt angle ξ′ is an angle within the first angle range (Step 611). When judged that the combined vector tilt angle ξ′ is an angle within the first angle range (YES in Step 611), theMPU 19 sets the first weighting factor α to be larger than the second weighting factor β (Step 612). - On the other hand, when judged that the combined vector tilt angle ξ′ is not an angle within the first angle range (NO in Step 611), that is, when judged that the combined vector tilt angle ξ′ is an angle within the second angle range, the
MPU 19 sets the first weighting factor α to be smaller than the second weighting factor β (Step 613). - Upon setting the weighting factors (α, β), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by the first and second weighting factors (α, β), respectively, to thus calculate first and second modified velocity values (Vx′, Vy′) (Step 614). - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 615). - In the fifth embodiment, the first angle range and the second angle range are controlled variably based on the first modified angle range and the second modified angle range as fixed values. As a result, the first angle range and the second angle range can be widened/narrowed as appropriate.
- Next, a sixth embodiment of the present invention will be described.
- The above embodiments have described a case where the scroll direction is biased in (restricted to) a uniaxial direction or biaxial directions on the screen. On the other hand, the sixth embodiment is different from the above embodiments in that the scroll direction is restricted to directions along four axes on the
screen 3. Therefore, that point will mainly be described. -
FIG. 21 is a flowchart showing an operation of theinput apparatus 1 according to this embodiment. - As shown in
FIG. 21 , in Steps 701 to 704, information on velocity values is transmitted as information on a movement amount of thepointer 2 when thebutton 13 is not pressed. - When the user presses the
button 13 and an operation signal from the switch is input (YES in Step 703), theMPU 19 calculates a tilt angle ξ of a combined vector of velocity values (Vx, Vy) using Equation (5) above (Step 705). - Upon calculating the combined vector tilt angle ξ, the
MPU 19 judges whether the combined vector tilt angle ξ is within a third angle range (Step 706). -
FIG. 22 is a diagram for explaining the third angle range. - As shown in
FIG. 22 , in theinput apparatus 1 of this embodiment, an angle range is divided into the first angle range, the second angle range, and the third angle range. The first angle range is, for example, a range within 0±22.5 degrees or 180±22.5 degrees. The second angle range is, for example, a range within 90±22.5 degrees or 270±22.5 degrees. The third angle range is, for example, a range within 45±22.5 degrees, 135±22.5 degrees, 225±22.5 degrees, or 315±22.5 degrees. It should be noted that ranges of the first angle range, the second angle range, and the third angle range can be changed as appropriate. Angles to be a reference of the third angle range (broken lines ofFIG. 22 ) can also be changed as appropriate. The horizontal-axis direction in the angle ranges shown inFIG. 22 corresponds to a movement direction (operation direction) of theinput apparatus 1 in the horizontal-axis direction, and the vertical-axis direction corresponds to a movement direction (operation direction) of theinput apparatus 1 in the vertical-axis direction. - When judged that the combined vector tilt angle ξ is within the third angle range (YES in Step 706), the
MPU 19 references a table and sets the weighting factors (α, β) (Step 710). In this case, the weighting factors (α, β) read out from the table are not constant and are values determined in relation to velocity values (Vx, Vy). The weighting factors (α, β) are stored in the table as values for restricting the scroll direction to directions at angles of ±45 degrees from the vertical-axis direction on the screen. It should be noted that the weighting factors (α, β) set in Step 710 may be calculated by a program. - When judged in Step 706 that the combined vector tilt angle ξ is not an angle within the third angle range (NO in Step 706), the
MPU 19 judges whether the combined vector tilt angle ξ is an angle within the first angle range (Step 707). When the combined vector tilt angle ξ is an angle within the first angle range (YES in Step 707), theMPU 19 sets the first weighting factor α to 1 and thesecond weighting factor 3 to 0 (Step 708). - On the other hand, when judged that the combined vector tilt angle ξ is not an angle within the first angle range (NO in Step 707), that is, when judged that the combined vector tilt angle ξ is an angle within the second angle range, the
MPU 19 sets the first weighting factor α to 0 and the second weighting factor β to 1 (Step 709). - Upon setting the weighting factors (α, β), the
MPU 19 multiplies the first and second velocity values (Vx, Vy) by the first and second weighting factors (α, β), respectively, to thus calculate first and second modified velocity values (Vx′, Vy′) (Step 711). - Upon calculating the modified velocity values (Vx′, Vy′), the
MPU 19 transmits information on the modified velocity values (Vx′, Vy′) to thecontrol apparatus 40 as scroll information (Step 712). -
FIG. 23 is a diagram showing relationships between operation directions of theinput apparatus 1 and scroll directions in a case where the processing shown inFIG. 21 is executed. - As shown in
FIG. 23 , when the user operates theinput apparatus 1 in a direction within an angle range of ±22.5 degrees from a direction along the X′-axis direction, a scroll direction of theimage 6 is restricted to the horizontal-axis (X-axis) direction on the screen. When the user operates theinput apparatus 1 in a direction within an angle range of ±22.5 degrees from a direction along the Y′-axis direction, the scroll direction of theimage 6 is restricted to the vertical-axis (Y-axis) direction on the screen. When the user operates theinput apparatus 1 in a direction within an angle of ±22.5 degrees from a direction at an angle of +45 degrees from the X′-axis direction, the scroll direction of theimage 6 is restricted to a direction at an angle of +45 degrees from the horizontal axis on the screen. When the user operates theinput apparatus 1 in a direction within an angle of ±22.5 degrees from a direction at an angle of −45 degrees from the X′-axis direction, the scroll direction of theimage 6 is restricted to a direction at an angle of −45 degrees from the horizontal axis on the screen. - As described above, in the sixth embodiment, the scroll direction is restricted to directions along four axes of the horizontal-axis direction, the vertical-axis direction, the direction at an angle of +45 degrees from the horizontal axis, and the direction at an angle of −45 degrees from the horizontal axis on the screen. As a result, an operational feeling in scroll operations in a case where the
image 6 such as a map that is long in the vertical-axis direction and the horizontal-axis direction on thescreen 3 as a whole is operated can be improved. - The sixth embodiment has been described assuming that the directions to which the scroll is restricted are the horizontal-axis direction, the vertical-axis direction, and the directions at angles of ±45 degrees from the horizontal-axis direction on the screen. However, the directions to which the scroll is restricted are not limited thereto. By setting the weighting factors (α, β) stored in the table as appropriate in Step 710, the scroll direction can be restricted to various directions. Examples of the combination of directions to which scroll is restricted include a combination of the horizontal-axis direction, the vertical-axis direction, and directions at angles of ±30 degrees from the horizontal-axis direction and a combination of the horizontal-axis direction, the vertical-axis direction, and directions at angles of ±60 degrees from the horizontal-axis direction. It is of course possible to use other combinations.
- The number of restriction axes on the
screen 3 is also not limited to four (four axes). The number of restriction axes may be three (three axes) or five (five axes) or more. - The sixth embodiment has described a case where the scroll direction on the
screen 3 is restricted. However, it is also possible to bias the scroll direction on thescreen 3. - Moreover, the first angle range, the second angle range, and the third angle range may be controlled variably.
- Next, the
control system 100 according to a seventh embodiment of the present invention will be described. - In the seventh and subsequent embodiments, processing related to an operation direction of the
input apparatus 1 and a direction in which an image is scrolled will be described. - In the 3-dimensional
operation input apparatus 1, whether to scroll theimage 6 in a direction in which theinput apparatus 1 is operated or scroll theimage 6 in an opposite direction from the direction in which theinput apparatus 1 is operated sometimes becomes a problem. -
FIGS. 24A and 24B are diagrams each showing a relationship between the operation direction of theinput apparatus 1 and a direction in which theimage 6 is scrolled.FIG. 24A is a diagram showing a case where theimage 6 is scrolled in a direction in which theinput apparatus 1 is operated, andFIG. 24B is a diagram showing a case where theimage 6 is scrolled in an opposite direction from the direction in which theinput apparatus 1 is operated. - The inventors of the present invention have conducted a user test, which revealed that there are both users who feel that scroll of an image in a direction in which the
input apparatus 1 is operated provides a better operational feeling and users who feel that scroll of an image in an opposite direction from the direction in which theinput apparatus 1 is operated provides a better operational feeling. - In this regard, the
input apparatus 1 according to the seventh embodiment executes processing for improving an operational feeling regarding a direction of scrolling theimage 6. -
FIG. 25 is a flowchart showing an operation of theinput apparatus 1 of thecontrol system 100 according to this embodiment. - As shown in
FIG. 25 , theinput apparatus 1 calculates velocity values (Vx, Vy) based on acquired angular velocity values (ωψ, ωθ) and acceleration values (ax, ay) (Steps 801 and 802). Upon calculating the velocity values (Vx, Vy), theMPU 19 judges whether an operation signal from the switch provided to thebutton 13 is input (Step 803). - When judged that the operation signal is not input (NO in Step 803), the
MPU 19 transmits information on the velocity values (Vx, Vy). In this case, thepointer 2 moves on thescreen 3 in accordance with a movement of theinput apparatus 1. - When the user presses the
button 13, theinput apparatus 1 transmits information on the velocity values (Vx, Vy) and a small-size screen display signal (Step 805). - Upon receiving the small-size screen display signal from the
input apparatus 1, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that a small-size screen 8 is displayed on thescreen 3. Moreover, upon receiving the information on the velocity values (Vx, Vy), theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that theimage 6 is scrolled at a velocity corresponding to the velocity values (Vx, Vy). It should be noted that since a small-size screen display signal is transmitted from theinput apparatus 1 during the scroll mode, theMPU 35 can distinctively recognize the velocity values (Vx, Vy) transmitted in Step 804 and the velocity values (Vx, Vy) transmitted in Step 805. -
FIG. 26 is a diagram showing theimage 6 and small-size screen 8 displayed on the screen. As shown inFIG. 26 , the small-size screen 8 is displayed at a lower right-hand corner of theimage 6, for example. It should be noted that a position at which the small-size screen 8 is displayed may be any position as long as it does not lower visibility of theimage 6. - The small-
size screen 8 is sectioned into afirst area 8 a (area in slashes inFIG. 26 ) corresponding to theentire image 6 and asecond area 8 b corresponding to a part of theimage 6 currently being displayed on the screen. - When the user holds the
input apparatus 1 and swings it upwardly from the reference position, theMPU 35 of thecontrol apparatus 40 controls display so that theimage 6 is scrolled downwardly at a velocity corresponding to the velocity values (Vx, Vy). In other words, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that theimage 6 is scrolled in an opposite direction from a vector direction of the velocity values (Vx, Vy). In addition, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that thesecond area 8 b moves upwardly in an area in which the small-size screen 8 is displayed. In other words, theMPU 35 of thecontrol apparatus 40 controls display so that theimage 6 moves in an opposite direction from a direction in which theimage 6 is scrolled. - In other words, the
MPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that theimage 6 is scrolled in an opposite direction from the direction in which theinput apparatus 1 is operated and thesecond area 8 b moves in a direction in which theinput apparatus 1 is operated. - By the processing as described above, the user can scroll an image displayed on a screen by merely operating the
second area 8 b in the small-size screen 8. Accordingly, since it becomes possible to perform scroll operations intuitionally, an operational feeling in scroll operations can be improved. Moreover, since the small-size screen 8 is displayed while thebutton 13 is pressed (during scroll mode), it does not lower visibility during the pointer mode. - The
input apparatus 1 may transmit modified velocity values (Vx′, Vy′) instead of velocity values (Vx, Vy) in Step 805. The processing described in the above embodiments can all be applied to this embodiment. As a result, since the scroll direction of theimage 6 is biased in (restricted to) the horizontal-axis direction or the vertical-axis direction on the screen, an operational feeling in scroll operations can be additionally improved. The same holds true for modified examples to be described later. - Next, a first modified example of the
control system 100 according to the seventh embodiment will be described. - The
input apparatus 1 of thecontrol system 100 according to the first modified example transmits information on velocity values (Vx, Vy) and a scrollbar display signal in Step 805 shown inFIG. 25 . - Upon receiving the scrollbar display signal, the
control apparatus 40 displays a scrollbar 9 on thescreen 3. -
FIG. 27 is a diagram showing theimage 6 and scrollbar 9 displayed on thescreen 3. As shown inFIG. 27 , the scrollbar 9 is displayed at a lower end and rightward end on thescreen 3. It should be noted that positions at which the scrollbar 9 is displayed may be any position as long as it does not lower visibility of theimage 6. - The scrollbar 9 includes an ordinate-
axis scrollbar 9 a and an abscissa-axis scrollbar 9 b. - When the user holds the
input apparatus 1 and swings it upwardly from the reference position, theMPU 35 of thecontrol apparatus 40 controls display so that theimage 6 is scrolled downwardly at a velocity corresponding to velocity values (Vx, Vy) transmitted in Step 805. In other words, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that theimage 6 is scrolled in an opposite direction from a vector direction of the velocity values (Vx, Vy). Moreover, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that the ordinate-axis scrollbar 9 a moves upwardly. Specifically, theMPU 35 of thecontrol apparatus 40 controls display so that the ordinate-axis scrollbar 9 a moves in an opposite direction from the direction in which theimage 6 is scrolled. - When the user moves the
input apparatus 1 in a right-hand direction on thescreen 3 from the reference position, the image is scrolled in a left-hand direction, and the abscissa-axis scrollbar 9 b is moved in the right-hand direction on thescreen 3. - In other words, the
MPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that theimage 6 is scrolled in an opposite direction from a direction in which theinput apparatus 1 is operated and the ordinate-axis scrollbar 9 a and the abscissa-axis scrollbar 9 b are moved in directions in which theinput apparatus 1 is operated. - By the processing as described above, the user can scroll the
image 6 displayed on the screen by merely operating the scrollbar 9, with the result that an operational feeling in scroll operations can be improved. Moreover, since the scrollbar 9 is displayed while thebutton 13 is pressed (during scroll mode), it does not lower visibility during the pointer mode. - Next, a second modified example of the
control system 100 according to the seventh embodiment of the present invention will be described. - The
input apparatus 1 of thecontrol system 100 according to the second modified example transmits information on velocity values (Vx, Vy) and a reference point display signal in Step 805 shown inFIG. 25 . - Upon receiving the reference point display signal, the
control apparatus 40 displays areference point 43 on theimage 6 when thepointer 2 displayed on thescreen 3 is positioned on theimage 6, for example. -
FIG. 28 is a diagram showing theimage 6 andreference point 43 displayed on thescreen 3. Thereference point 43 is displayed as, for example, a circular point. It should be noted that a shape of thereference point 43 is not particularly limited. Thereference point 43 is displayed at a position at which thepointer 2 is positioned at a time thebutton 13 is pressed. - Upon displaying the
reference point 43 on thescreen 3, theMPU 35 of thecontrol apparatus 40 generates coordinate values of thepointer 2 based on information on velocity values (Vx, Vy) transmitted from theinput apparatus 1 in Step 805. Then, theMPU 35 of thecontrol apparatus 40 controls display so that thepointer 2 moves on the screen. In other words, in thecontrol system 100 according to the second modified example, thepointer 2 also moves during the scroll mode. - Moreover, the
MPU 35 of thecontrol apparatus 40 adds the velocity values (Vx, Vy) transmitted from theinput apparatus 1 in Step 805 to thus generate integration values. TheMPU 35 of thecontrol apparatus 40 controls display on the screen so that theimage 6 is scrolled at a velocity corresponding to the integration values. - When the user holds the
input apparatus 1 and swings it upwardly from the reference position, thepointer 2 is moved upwardly on thescreen 3 and theimage 6 is scrolled upwardly. In other words, theMPU 35 of thecontrol apparatus 40 controls display on thescreen 3 so that thepointer 2 moves in the same direction as a vector direction of the velocity values (Vx, Vy) and theimage 6 is scrolled in the same direction as the vector direction of the velocity values (Vx, Vy). - By the processing as described above, the user can scroll the
image 6 with thepointer 2 as a guide. As a result, since intuitional operations can be made, an operational feeling can be improved. - The embodiment of the present invention is not limited to the above embodiments and various modifications can be made.
- For example, it is possible to execute processing that inhibits, when the
button 13 is started to be pressed, an image displayed on thescreen 3 from being scrolled during a predetermined time period (first time period) since the start of the press. Accordingly, it is possible to prevent the image from being scrolled in a direction unintended by the user due to the input apparatus being moved when the user presses thebutton 13. - The present invention is applicable to input apparatuses such as a planar-operation-type mouse, a touchpad, a joystick, and a pen tablet. Alternatively, the present invention may be applied to a slide-resistance-type input apparatus that detects a movement of an operation section inside an opening formed on a casing by a slide resistance. Alternatively, the present invention may be applied to an optical input apparatus that calculates a movement amount and operation direction of a finger of a user by irradiating light onto a semicircular operation section provided at an upper portion of a casing and detecting reflected light. Alternatively, the present invention may be applied to an electronic apparatus including any of the input apparatuses described above (e.g., laptop PC including touchpad).
- The present invention may be applied to a handheld apparatus that includes a display section, for example. In this case, an image displayed on the display section is scrolled when the user moves a main body of the handheld apparatus. Alternatively, the user moves the pointer by moving the main body of the handheld apparatus. Examples of the handheld apparatus include a PDA (Personal Digital Assistance), a cellular phone, a portable music player, and a digital camera.
- The
input apparatus 1 according to the above embodiments has transmitted input information to thecontrol apparatus 40 wirelessly. However, the input information may be transmitted by wire. - In the above embodiments, the
pointer 2 that moves on the screen in accordance with the movement of theinput apparatus 1 has been represented as an image of an arrow. However, the image of thepointer 2 is not limited to the arrow and may be a simple circle, square, or the like, or a character image or any other images. - The above embodiments have described about the biaxial acceleration sensor unit and the biaxial angular velocity sensor unit. However, the present invention is not limited thereto, and the
input apparatus 1 may include, for example, acceleration sensors of three orthogonal axes and angular velocity sensors of three orthogonal axes, and even with only one of the above, the processing shown in the above embodiments can be realized. Alternatively, an embodiment in which theinput apparatus 1 includes a uniaxial acceleration sensor or a uniaxial angular velocity sensor is also conceivable. When provided with the uniaxial acceleration sensor or uniaxial angular velocity sensor, typically a screen on which a plurality of GUIs as pointing targets of thepointer 2 displayed on thescreen 3 are arranged uniaxially is conceivable. - Alternatively, the
input apparatus 1 may include a geomagnetic sensor, an image sensor, and the like instead of the acceleration sensors and the angular velocity sensors. - The detection axes of each of the angular
velocity sensor unit 15 and theacceleration sensor unit 16 of thesensor unit 17 do not necessarily need to be mutually orthogonal like the X′ axis and the Y′ axis described above. In this case, accelerations respectively projected in the mutually-orthogonal axial directions can be obtained by a calculation that uses a trigonometric function. Similarly, angular velocities about the mutually-orthogonal axes can be obtained by a calculation that uses the trigonometric function. - Descriptions have been given on the case where the X′ and Y′ detection axes of the angular
velocity sensor unit 15 and the X′ and Y′ detection axes of theacceleration sensor unit 16 of thesensor unit 17 described in the above embodiments match. However, those detection axes do not necessarily need to match. For example, in a case where the angularvelocity sensor unit 15 and theacceleration sensor unit 16 are mounted on a substrate, the angularvelocity sensor unit 15 and theacceleration sensor unit 16 may be mounted while being deviated a predetermined rotation angle within a main surface of the substrate so that the detection axes of the angularvelocity sensor unit 15 and theacceleration sensor unit 16 do not match. In this case, accelerations and angular velocities with respect to the respective axes can be obtained by a calculation that uses the trigonometric function. - In the above embodiments, the case where the
input apparatus 1 is operated 3-dimensionally has been described. However, the present invention is not limited thereto, and the input apparatus may be operated while a part of thecasing 10 is in contact with a table, for example. - It should be understood by those skilled in the art that various modifications, combinations, sub-combinations and alterations may occur depending on design requirements and other factors insofar as they are within the scope of the appended claims or the equivalents thereof.
Claims (22)
1. An information processing apparatus comprising:
an input unit configured to receive a mode selection input from a user;
a switching unit configured to switch between a pointing mode and a scroll mode based on the mode selection input; and
a scroll control unit configured to scroll in a scroll direction in the scroll mode.
2. The information processing apparatus of claim 1 , wherein the scroll control unit is configured to restrict the scroll direction to a horizontal direction based on movement of the user in three-dimensional space.
3. The information processing apparatus of claim 1 , wherein the scroll control unit is configured to restrict the scroll direction to a vertical direction based on movement of the user in three-dimensional space.
4. The information processing apparatus of claim 1 , further comprising a direction weighting unit configured to determine a first direction weighting factor that restricts the scroll direction in a first scroll direction based on movement of the user in three-dimensional space.
5. The information processing apparatus of claim 4 , wherein the direction weighting unit is configured to determine a second direction weighting factor that restricts the scroll direction in a second scroll direction, different from the first scroll direction, based on movement of the user in three-dimensional space.
6. The information processing apparatus of claim 1 , wherein the input unit is further configured to receive a movement input based on movement of the user in three-dimensional space.
7. The information processing apparatus of claim 6 , wherein the scroll control unit is further configured to restrict the scroll direction based on the movement input received by the input unit.
8. An information processing method comprising:
receiving, by an input unit, a mode selection input from a user;
switching, by a switching unit, between a pointing mode and a scroll mode based on the mode selection input; and
scrolling, by a scroll control unit, in a scroll direction in the scroll mode.
9. The information processing method of claim 8 , wherein scrolling includes restricting the scroll direction to a horizontal direction based on movement of the user in three-dimensional space.
10. The information processing method of claim 8 , wherein scrolling includes restricting the scroll direction to a vertical direction based on movement of the user in three-dimensional space.
11. The information processing method of claim 8 , further comprising determining a first direction weighting factor that restricts the scroll direction in a first scroll direction based on movement of the user in three-dimensional space.
12. The information processing method of claim 11 , further comprising determining a second direction weighting factor that restricts the scroll direction in a second scroll direction, different from the first scroll direction, based on movement of the user in three-dimensional space.
13. The information processing method of claim 8 , further comprising receiving, by the input unit, a movement input based on movement of the user in three-dimensional space.
14. The information processing method of claim 13 , wherein scrolling includes restricting the scroll direction based on the movement input received by the input unit.
15. An information processing apparatus comprising:
an input unit configured to receive from a user a mode selection input indicative of a pointing mode or a scroll mode and to receive a movement input based on movement of the user in three-dimensional space;
a switching unit configured to operate in the pointing mode or the scroll mode based on the mode selection input; and
a scroll control unit configured to scroll in a scroll direction in the scroll mode based on the movement input received by the input unit.
16. The information processing apparatus of claim 15 , wherein the scroll control unit is configured to restrict the scroll direction to a horizontal direction based on the movement input received by the input unit.
17. The information processing apparatus of claim 15 , wherein the scroll control unit is configured to restrict the scroll direction to a vertical direction based on the movement input received by the input unit.
18. The information processing apparatus of claim 15 , further comprising a direction weighting unit configured to determine a first direction weighting factor that restricts the scroll direction in a first scroll direction based on the movement input received by the input unit.
19. The information processing apparatus of claim 18 , wherein the direction weighting unit is configured to determine a second direction weighting factor that restricts the scroll direction in a second scroll direction, different from the first scroll direction, based on the movement input received by the input unit.
20. The information processing apparatus of claim 15 , wherein the scroll control unit is further configured to restrict the scroll direction based on the movement input received by the input unit.
21. An information processing apparatus comprising processing circuitry configured to:
receive from a user a mode selection input indicative of a pointing mode or a scroll mode and to receive a movement input based on movement of the user in three-dimensional space;
operate in the pointing mode or the scroll mode based on the mode selection input; and
scroll in a scroll direction in the scroll mode based on the movement input.
22. A computer-readable storage device encoded with computer-executable instructions that, when executed by a processing apparatus, perform a method comprising:
receiving from a user a mode selection input indicative of a pointing mode or a scroll mode and receiving a movement input based on movement of the user in three-dimensional space;
operating in the pointing mode or the scroll mode based on the mode selection input; and
scrolling in a scroll direction in the scroll mode based on the movement input.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/572,884 US20150169160A1 (en) | 2008-12-25 | 2014-12-17 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008331617A JP2010152761A (en) | 2008-12-25 | 2008-12-25 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
JP2008-331617 | 2008-12-25 | ||
US12/645,732 US9152246B2 (en) | 2008-12-25 | 2009-12-23 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
US14/572,884 US20150169160A1 (en) | 2008-12-25 | 2014-12-17 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/645,732 Continuation US9152246B2 (en) | 2008-12-25 | 2009-12-23 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20150169160A1 true US20150169160A1 (en) | 2015-06-18 |
Family
ID=41716380
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/645,732 Expired - Fee Related US9152246B2 (en) | 2008-12-25 | 2009-12-23 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
US14/572,884 Abandoned US20150169160A1 (en) | 2008-12-25 | 2014-12-17 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/645,732 Expired - Fee Related US9152246B2 (en) | 2008-12-25 | 2009-12-23 | Input apparatus, control apparatus, control system, electronic apparatus, and control method |
Country Status (6)
Country | Link |
---|---|
US (2) | US9152246B2 (en) |
EP (1) | EP2202615A3 (en) |
JP (1) | JP2010152761A (en) |
KR (1) | KR20100075770A (en) |
CN (1) | CN101763182B (en) |
TW (1) | TW201028913A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9569012B2 (en) | 2008-12-24 | 2017-02-14 | Sony Corporation | Input apparatus, control apparatus, and control method for input apparatus |
US10474250B2 (en) | 2008-10-31 | 2019-11-12 | Sony Corporation | Input device and method and program |
US10682577B2 (en) * | 2016-07-14 | 2020-06-16 | Yoshinori Matsumoto | Mobile terminal with call function or texting function, or a function as a game terminal, and game program and game method executed by mobile terminal |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8564544B2 (en) | 2006-09-06 | 2013-10-22 | Apple Inc. | Touch screen device, method, and graphical user interface for customizing display of content category icons |
US8493344B2 (en) | 2009-06-07 | 2013-07-23 | Apple Inc. | Devices, methods, and graphical user interfaces for accessibility using a touch-sensitive surface |
US20120327009A1 (en) * | 2009-06-07 | 2012-12-27 | Apple Inc. | Devices, methods, and graphical user interfaces for accessibility using a touch-sensitive surface |
KR20120124874A (en) * | 2011-05-06 | 2012-11-14 | 삼성전자주식회사 | Method and apparatus for motion sensing with independant grip direction |
US9569057B2 (en) * | 2012-01-05 | 2017-02-14 | Sony Corporation | Information processing apparatus and method for outputting a guiding operation to a user |
US8881269B2 (en) | 2012-03-31 | 2014-11-04 | Apple Inc. | Device, method, and graphical user interface for integrating recognition of handwriting gestures with a screen reader |
TWI463371B (en) * | 2012-06-20 | 2014-12-01 | Pixart Imaging Inc | Gesture detection apparatus and method for determining continuous gesture depending on velocity |
TWI515632B (en) | 2012-06-26 | 2016-01-01 | 緯創資通股份有限公司 | Touch-and-play input device and operating method thereof |
TWI467467B (en) * | 2012-10-29 | 2015-01-01 | Pixart Imaging Inc | Method and apparatus for controlling object movement on screen |
US10191643B2 (en) | 2012-11-29 | 2019-01-29 | Facebook, Inc. | Using clamping to modify scrolling |
US9965162B2 (en) * | 2012-11-29 | 2018-05-08 | Facebook, Inc. | Scrolling across boundaries in a structured document |
CN105208420A (en) * | 2015-08-24 | 2015-12-30 | 小米科技有限责任公司 | Television control method and device and electronic equipment |
KR102180169B1 (en) * | 2017-05-19 | 2020-11-18 | (주)아이티버스 | Smart controller |
WO2019087942A1 (en) * | 2017-10-31 | 2019-05-09 | 富士フイルム株式会社 | Operation device, and operation method and operation program therefor |
JP6682079B2 (en) * | 2017-11-01 | 2020-04-15 | 株式会社ソニー・インタラクティブエンタテインメント | Information processing apparatus and signal conversion method |
CN108109186B (en) * | 2017-11-30 | 2021-06-11 | 维沃移动通信有限公司 | Video file processing method and device and mobile terminal |
US11329722B2 (en) | 2020-03-27 | 2022-05-10 | Relative Dynamics Incorporated | Optical terminals |
WO2024128337A1 (en) * | 2022-12-13 | 2024-06-20 | 삼성전자 주식회사 | Method and device for providing ultra-wideband communication-based service |
Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020051018A1 (en) * | 2000-10-26 | 2002-05-02 | Nan-Ting Yeh | Apparatus and method for browser interface operation |
US20040021694A1 (en) * | 2002-08-01 | 2004-02-05 | Apple Computer, Inc. | Mode activated scrolling |
US20040141011A1 (en) * | 2000-07-07 | 2004-07-22 | Smethers Paul A. | Graphical user interface features of a browser in a hand-held wireless communication device |
US20040189720A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US20040218104A1 (en) * | 2003-05-01 | 2004-11-04 | Smith Gregory C. | Multimedia user interface |
US20060026535A1 (en) * | 2004-07-30 | 2006-02-02 | Apple Computer Inc. | Mode-based graphical user interfaces for touch sensitive input devices |
US20060250358A1 (en) * | 2005-05-04 | 2006-11-09 | Hillcrest Laboratories, Inc. | Methods and systems for scrolling and pointing in user interfaces |
US20070139443A1 (en) * | 2005-12-12 | 2007-06-21 | Sonny Computer Entertainment Inc. | Voice and video control of interactive electronically simulated environment |
US20080155473A1 (en) * | 2006-12-21 | 2008-06-26 | Canon Kabushiki Kaisha | Scrolling interface |
US20080222558A1 (en) * | 2007-03-08 | 2008-09-11 | Samsung Electronics Co., Ltd. | Apparatus and method of providing items based on scrolling |
US20090064031A1 (en) * | 2007-09-04 | 2009-03-05 | Apple Inc. | Scrolling techniques for user interfaces |
US20090158203A1 (en) * | 2007-12-14 | 2009-06-18 | Apple Inc. | Scrolling displayed objects using a 3D remote controller in a media system |
US20090231275A1 (en) * | 2005-01-30 | 2009-09-17 | Simtrix Limited | Computer mouse peripheral |
US20100017732A1 (en) * | 2008-04-24 | 2010-01-21 | Nintendo Co., Ltd. | Computer-readable storage medium having object display order changing program stored therein and apparatus |
US20100131904A1 (en) * | 2008-11-21 | 2010-05-27 | Microsoft Corporation | Tiltable user interface |
US20110047459A1 (en) * | 2007-10-08 | 2011-02-24 | Willem Morkel Van Der Westhuizen | User interface |
US20130061175A1 (en) * | 2006-09-06 | 2013-03-07 | Michael Matas | Portable Electronic Device for Photo Management |
US20130152012A1 (en) * | 2004-08-30 | 2013-06-13 | Microsoft Corporation | Scrolling web pages using direct interaction |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
IL92220A (en) | 1989-11-06 | 1993-02-21 | Ibm Israel | Three-dimensional computer input device |
JPH0436794A (en) * | 1990-06-01 | 1992-02-06 | Nippon Telegr & Teleph Corp <Ntt> | Cursor control device |
JPH05210461A (en) * | 1992-01-31 | 1993-08-20 | Sony Corp | Display device and remote manipulator for controlling display device |
JP2664614B2 (en) | 1992-02-20 | 1997-10-15 | ジ・ベ・エス ソシエテ アノニム | Cervical spine correction, fixation, clamping and retraction devices |
US5453758A (en) * | 1992-07-31 | 1995-09-26 | Sony Corporation | Input apparatus |
JP3857504B2 (en) * | 1993-05-21 | 2006-12-13 | ソニー株式会社 | Input device, control system |
US5454043A (en) * | 1993-07-30 | 1995-09-26 | Mitsubishi Electric Research Laboratories, Inc. | Dynamic and static hand gesture recognition through low-level image analysis |
US5633657A (en) * | 1994-10-11 | 1997-05-27 | Falcon; Fernando D. | Mouse driver arrangement for providing advanced scrolling capabilities to a conventional mouse |
US5745719A (en) * | 1995-01-19 | 1998-04-28 | Falcon; Fernando D. | Commands functions invoked from movement of a control input device |
KR100301016B1 (en) * | 1998-10-27 | 2001-09-06 | 윤종용 | Method for selecting on-screen menu and apparatus thereof |
JP3980966B2 (en) * | 2002-08-21 | 2007-09-26 | シャープ株式会社 | Presentation display device |
KR100486739B1 (en) * | 2003-06-27 | 2005-05-03 | 삼성전자주식회사 | Wearable phone and method using the same |
US20050030279A1 (en) * | 2003-08-08 | 2005-02-10 | Liang Fu | Multi-functional pointing and control device |
KR100937572B1 (en) * | 2004-04-30 | 2010-01-19 | 힐크레스트 래보래토리스, 인크. | Free space pointing device and method |
US7761814B2 (en) * | 2004-09-13 | 2010-07-20 | Microsoft Corporation | Flick gesture |
JP2006236163A (en) * | 2005-02-28 | 2006-09-07 | Kyocera Mita Corp | Scrolling controller and scrolling control program |
JP2008181248A (en) * | 2007-01-23 | 2008-08-07 | Kyocera Mita Corp | Display controller and display control program |
JP5434591B2 (en) * | 2007-06-20 | 2014-03-05 | ソニー株式会社 | Control device, input device, control system, handheld information processing device, and control method |
US8203578B2 (en) * | 2007-10-30 | 2012-06-19 | Alpine Electronics, Inc. | Map scroll method and apparatus for conducting smooth map scroll operation for navigation system |
JP4458301B2 (en) * | 2007-12-05 | 2010-04-28 | コニカミノルタビジネステクノロジーズ株式会社 | Image forming apparatus |
JP4292228B1 (en) * | 2007-12-27 | 2009-07-08 | 株式会社東芝 | Information processing device |
KR101400230B1 (en) * | 2008-03-11 | 2014-05-28 | 삼성전자주식회사 | Three dimensional pointing input apparatus and method thereof |
EP2347321B1 (en) * | 2008-09-25 | 2013-09-18 | Movea S.A | Command by gesture interface |
US8649554B2 (en) * | 2009-05-01 | 2014-02-11 | Microsoft Corporation | Method to control perspective for a camera-controlled computer |
US8638989B2 (en) * | 2012-01-17 | 2014-01-28 | Leap Motion, Inc. | Systems and methods for capturing motion in three-dimensional space |
-
2008
- 2008-12-25 JP JP2008331617A patent/JP2010152761A/en not_active Ceased
-
2009
- 2009-11-23 TW TW98139716A patent/TW201028913A/en unknown
- 2009-12-22 EP EP20090252861 patent/EP2202615A3/en not_active Withdrawn
- 2009-12-23 US US12/645,732 patent/US9152246B2/en not_active Expired - Fee Related
- 2009-12-23 CN CN200910259711.XA patent/CN101763182B/en not_active Expired - Fee Related
- 2009-12-24 KR KR20090131088A patent/KR20100075770A/en not_active Application Discontinuation
-
2014
- 2014-12-17 US US14/572,884 patent/US20150169160A1/en not_active Abandoned
Patent Citations (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040141011A1 (en) * | 2000-07-07 | 2004-07-22 | Smethers Paul A. | Graphical user interface features of a browser in a hand-held wireless communication device |
US20020051018A1 (en) * | 2000-10-26 | 2002-05-02 | Nan-Ting Yeh | Apparatus and method for browser interface operation |
US20040021694A1 (en) * | 2002-08-01 | 2004-02-05 | Apple Computer, Inc. | Mode activated scrolling |
US20040189720A1 (en) * | 2003-03-25 | 2004-09-30 | Wilson Andrew D. | Architecture for controlling a computer using hand gestures |
US20040218104A1 (en) * | 2003-05-01 | 2004-11-04 | Smith Gregory C. | Multimedia user interface |
US20060026535A1 (en) * | 2004-07-30 | 2006-02-02 | Apple Computer Inc. | Mode-based graphical user interfaces for touch sensitive input devices |
US20130152012A1 (en) * | 2004-08-30 | 2013-06-13 | Microsoft Corporation | Scrolling web pages using direct interaction |
US20090231275A1 (en) * | 2005-01-30 | 2009-09-17 | Simtrix Limited | Computer mouse peripheral |
US20060250358A1 (en) * | 2005-05-04 | 2006-11-09 | Hillcrest Laboratories, Inc. | Methods and systems for scrolling and pointing in user interfaces |
US20070139443A1 (en) * | 2005-12-12 | 2007-06-21 | Sonny Computer Entertainment Inc. | Voice and video control of interactive electronically simulated environment |
US20130061175A1 (en) * | 2006-09-06 | 2013-03-07 | Michael Matas | Portable Electronic Device for Photo Management |
US20080155473A1 (en) * | 2006-12-21 | 2008-06-26 | Canon Kabushiki Kaisha | Scrolling interface |
US20080222558A1 (en) * | 2007-03-08 | 2008-09-11 | Samsung Electronics Co., Ltd. | Apparatus and method of providing items based on scrolling |
US20090064031A1 (en) * | 2007-09-04 | 2009-03-05 | Apple Inc. | Scrolling techniques for user interfaces |
US20110047459A1 (en) * | 2007-10-08 | 2011-02-24 | Willem Morkel Van Der Westhuizen | User interface |
US20090158203A1 (en) * | 2007-12-14 | 2009-06-18 | Apple Inc. | Scrolling displayed objects using a 3D remote controller in a media system |
US20100017732A1 (en) * | 2008-04-24 | 2010-01-21 | Nintendo Co., Ltd. | Computer-readable storage medium having object display order changing program stored therein and apparatus |
US20100131904A1 (en) * | 2008-11-21 | 2010-05-27 | Microsoft Corporation | Tiltable user interface |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10474250B2 (en) | 2008-10-31 | 2019-11-12 | Sony Corporation | Input device and method and program |
US9569012B2 (en) | 2008-12-24 | 2017-02-14 | Sony Corporation | Input apparatus, control apparatus, and control method for input apparatus |
US9823757B2 (en) | 2008-12-24 | 2017-11-21 | Sony Corporation | Input apparatus, control apparatus, and control method for input apparatus |
US10682577B2 (en) * | 2016-07-14 | 2020-06-16 | Yoshinori Matsumoto | Mobile terminal with call function or texting function, or a function as a game terminal, and game program and game method executed by mobile terminal |
US11110360B2 (en) | 2016-07-14 | 2021-09-07 | Yoshinori Matsumoto | Mobile terminal with call function or texting function, or a function as a game terminal, and game program and game method executed by mobile terminal |
US11439920B2 (en) | 2016-07-14 | 2022-09-13 | Yoshinori Matsumoto | Mobile terminal with call function or texting function, or a function as a game terminal, and game program and game method executed by mobile terminal |
Also Published As
Publication number | Publication date |
---|---|
US9152246B2 (en) | 2015-10-06 |
KR20100075770A (en) | 2010-07-05 |
EP2202615A3 (en) | 2013-05-01 |
CN101763182B (en) | 2014-07-30 |
US20100169824A1 (en) | 2010-07-01 |
TW201028913A (en) | 2010-08-01 |
CN101763182A (en) | 2010-06-30 |
EP2202615A2 (en) | 2010-06-30 |
JP2010152761A (en) | 2010-07-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9152246B2 (en) | Input apparatus, control apparatus, control system, electronic apparatus, and control method | |
US8413077B2 (en) | Input apparatus, handheld apparatus, and control method | |
US10747338B2 (en) | Input apparatus, control apparatus, control system, control method, and handheld apparatus | |
US8531399B2 (en) | Control apparatus, input apparatus, control system, control method, and handheld apparatus | |
US8300012B2 (en) | Input apparatus, control apparatus, control system, and control method | |
US10031595B2 (en) | Control device, input device, control system, handheld device, and control method | |
US8552977B2 (en) | Input apparatus, control apparatus, control system, handheld apparatus, and control method | |
US9829998B2 (en) | Information processing apparatus, input apparatus, information processing system, information processing method, and program | |
US8994658B2 (en) | Input apparatus, control apparatus, control method, and handheld apparatus | |
US8395583B2 (en) | Input apparatus, control apparatus, control system, control method, and handheld apparatus | |
US20090201249A1 (en) | Input apparatus, control apparatus, control system, and handheld apparatus | |
US20090309830A1 (en) | Control apparatus, input apparatus, control system, handheld information processing apparatus, control method, and program therefor | |
US8614671B2 (en) | Input apparatus, control apparatus, control system, and control method | |
US8456422B2 (en) | Input apparatus, control apparatus, control system, control method, and handheld apparatus | |
JP5412812B2 (en) | Input device, control device, control system, and handheld device | |
JP2010157106A (en) | Input device, controller, handheld device, control system, and control method | |
JP2011108270A (en) | Input device, handheld device and control method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONY CORPORATION, JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SAWAI, KUNIHITO;YAMAMOTO, KAZUYUKI;REEL/FRAME:035389/0228 Effective date: 20150113 |
|
STCV | Information on status: appeal procedure |
Free format text: ON APPEAL -- AWAITING DECISION BY THE BOARD OF APPEALS |
|
STCV | Information on status: appeal procedure |
Free format text: BOARD OF APPEALS DECISION RENDERED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- AFTER EXAMINER'S ANSWER OR BOARD OF APPEALS DECISION |