WO2011040483A1 - 表示装置、制御方法及び記録媒体 - Google Patents
表示装置、制御方法及び記録媒体 Download PDFInfo
- Publication number
- WO2011040483A1 WO2011040483A1 PCT/JP2010/066977 JP2010066977W WO2011040483A1 WO 2011040483 A1 WO2011040483 A1 WO 2011040483A1 JP 2010066977 W JP2010066977 W JP 2010066977W WO 2011040483 A1 WO2011040483 A1 WO 2011040483A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- control
- direction distance
- amount
- change
- display device
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0412—Digitisers structurally integrated in a display
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/041—Indexing scheme relating to G06F3/041 - G06F3/045
- G06F2203/04104—Multi-touch detection in digitiser, i.e. details about the simultaneous detection of a plurality of touching locations, e.g. multiple fingers or pen and finger
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04806—Zoom, i.e. interaction techniques or interactors for controlling the zooming operation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04808—Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen
Definitions
- the present invention relates to a display device, a control method, and a recording medium.
- Patent Documents 1 to 3 describe display devices that can be operated not only by a user's operation with one finger but also by touching with two or more fingers (multi-touch).
- Patent Document 1 describes a touch panel system that enlarges or reduces (zooms) display information on a screen such as a still image or a WEB page by performing an operation of expanding or narrowing two fingers.
- Patent Document 2 describes that a map displayed on a panel screen of a display device is enlarged or reduced by an operation with two fingers, and a room temperature of an in-vehicle air conditioner is adjusted by an operation with three fingers. Has been.
- Patent Document 3 in a display device in which two display units equipped with a touch panel are arranged adjacent to each other, the moving distance of the position between the first finger on the first touch panel and the second finger on the second touch panel is described. There is described a display device that calculates and moves the image data displayed on the display unit according to a change in distance when there is a difference of a certain value or more in the calculated movement distance.
- the apparatus executes one process in response to one operation with a finger. For this reason, when the apparatus executes a plurality of processes, a plurality of operations are required, and the operation is complicated.
- the present invention has been made in view of the above circumstances, a display device capable of executing a plurality of processes in response to a simple operation with a finger or the like, a control method, and for realizing these on a computer
- An object is to provide a recording medium on which a program is recorded.
- Another object of the present invention is to provide a display device capable of inputting a plurality of instructions by a simple operation, a control method, and a recording medium for realizing these on a computer.
- a display device provides: A touch panel for displaying images; A position detection unit for detecting XY coordinates of two touched positions on the touch panel; When the position detection unit detects the XY coordinates of the two contacted positions twice, each of the X direction distance and the Y direction distance between the two contacted positions from the previous coordinate detection to the subsequent coordinate detection is detected. A calculation unit for calculating the amount of change; The control of the first control object is executed according to the change amount of the X-direction distance calculated by the calculation unit, and the control of the second control object is executed according to the change amount of the Y-direction distance calculated by the calculation unit. A control execution unit for executing Is provided.
- a control method includes: Detecting XY coordinates of two contact positions on the contact surface; When the XY coordinates of the two contacted positions are detected twice in a predetermined period, each change amount of the X-direction distance and the Y-direction distance between the two contacted positions from the previous coordinate detection to the subsequent coordinate detection A step of calculating Executing control of the first control object according to the calculated amount of change in the X-direction distance, and executing control of the second control object according to the amount of change in the Y-direction distance calculated.
- a computer-readable recording medium provides: In the computer of the display device equipped with a touch panel for displaying images, Detecting XY coordinates of two touched positions on the touch panel, When the XY coordinates of the two contacted positions are detected twice in a predetermined period, each change amount of the X-direction distance and the Y-direction distance between the two contacted positions from the previous coordinate detection to the subsequent coordinate detection A step of calculating Executing control of the first control object according to the calculated amount of change in the X-direction distance, and executing control of the second control object according to the amount of change in the Y-direction distance calculated.
- the display device of the present invention can execute a plurality of processes in response to a simple operation with a finger or the like.
- FIG. 3 is a diagram showing images corresponding to steps S1 to S4 shown in FIG. (A-1), (a-2), (b-1) and (b-2) are diagrams exemplifying control when an operation of expanding or narrowing two fingers in the X direction is performed. (A-1), (a-2), (b-1), and (b-2) are diagrams exemplifying control when an operation of expanding or narrowing two fingers in the Y direction is performed.
- the display device 1 includes a touch panel 2, a position detection unit 3, a calculation unit 4, and a control execution unit 5.
- the touch panel 2 includes an LCD (Liquid Crystal Display) 2a, a capacitive touch sensor 2b arranged on the screen of the LCD 2a, a speaker 2c, and the like.
- the LCD 2a displays an image on the screen, and the touch sensor 2b supplies the position detection unit 3 with a voltage signal corresponding to the position touched by the user's finger or the like (contacted position).
- the speaker 2c outputs sound.
- the position detection unit 3 detects the XY coordinates of the position touched by the user's finger or the like on the touch panel 2 based on the voltage signal supplied from the touch sensor 2b.
- the calculation unit 4 detects the X direction distance and the Y direction distance between the two previous contacted positions, and the following 2
- Each change amount (difference) between the X-direction distance and the Y-direction distance between the two contact positions is calculated.
- the control execution unit 5 executes the control of the first control target according to the amount of change in the X-direction distance calculated by the calculation unit 4, and performs the first control according to the amount of change in the Y-direction distance calculated by the calculation unit 4.
- the control of the control target 2 is executed.
- the position detection unit 3, the calculation unit 4, and the control execution unit 5 include a CPU (Central Processing Unit) 51, a RAM (Random Access Memory) 52, and a ROM (Read Only Memory). 53, ADC (Analog to Digital Converter) 54, I / F (Interface) 55, and the like.
- the position detection unit 3, the calculation unit 4, and the control execution unit 5 are realized by the CPU 51 executing a program stored in a ROM 53, which is an example of a recording medium.
- the CPU 51 temporarily stores various data in the RAM 52.
- the ADC 54 converts an analog signal such as a voltage supplied from the touch sensor 2 b into a digital signal and supplies the digital signal to the CPU 51.
- the I / F 55 mediates data transmission / reception between the CPU 51 and other hardware.
- the display device 1 detects two contact positions on the touch panel 2 twice in succession, i) obtains an X-direction distance and a Y-direction distance between the two detected positions detected earlier, and ii) after Obtain the X-direction distance and Y-direction distance between the detected two detected positions, and iii) calculate the change amount (difference) in the X-direction distance and the change amount (difference) in the Y-direction distance, and iv) X
- the first control object is controlled according to the change amount of the direction distance
- the second control object is controlled according to the change amount of the Y direction distance.
- the display device 1 displays the X direction and the Y direction according to the amount of change in the X direction distance and the Y direction distance that accompany the continuous touch. Separate controls. Accordingly, the user can cause the display device 1 to execute a plurality of processes by a simple operation with a finger.
- the display device 1 starts the process illustrated in FIG. For ease of understanding, it is assumed that the LCD 2a of the display device 1 displays the WEB page shown in FIG. 4 (a-1). First, the position detection unit 3 of the display device 1 determines whether or not the two fingers of the user are both in contact with the touch panel 2 based on the voltage signal supplied from the touch sensor 2b (step S1).
- the position detection unit 3 waits until it is determined that both two fingers are in contact with the touch panel 2 (step S1; NO). When the position detection unit 3 determines that two fingers are both in contact with the touch panel 2 (step S1; YES), the XY coordinates of the two contact positions touched by the two fingers on the touch panel 2 are obtained. Detect each.
- the user touches the position PA, PB shown in FIG.
- the XY coordinates of the position PA are (X1, Y1)
- the XY coordinates of the position PB are (X2, Y2).
- the position detection unit 3 determines again whether or not the user's two fingers are both in contact with the touch panel 2 (step S3).
- the position detection unit 3 waits until it is determined that the two fingers of the user are both in contact with the touch panel 2 (step S3; NO), and it is determined that both of the two fingers are in contact with the touch panel 2. Then (step S3; YES), the XY coordinates on the touch panel 2 of two fingers are respectively detected.
- the user touches the position PC or PD on the touch panel 2 with a finger.
- the XY coordinates of the position PC are (X1 ′, Y1 ′) and the XY coordinates of the position PD are (X2 ′, Y2 ′).
- the calculation unit 4 calculates the X-direction distance X0 and the Y-direction distance Y0 between the contacted positions PA and PB obtained in step S2, and the X between the contacted position PC and PD obtained in step S4. Based on the direction distance X0 ′ and the Y direction distance Y0 ′, the amount of change in the X direction distance is obtained to control the enlargement / reduction of the display image on the LCD 2a, and the amount of change in the Y direction distance is obtained to determine the volume of the speaker 2c. Control (steps S5 to S12).
- step S5 the calculation unit 4 determines whether X0'-X0 is greater than 0 with respect to the amount of change in the X direction distance. That is, the calculation unit 4 determines whether the amount of change in the X direction distance is positive or negative in step S5.
- step S5 If it is determined in step S5 that the change amount in the X-direction distance is positive (step S5; YES), the control execution unit 5 determines the amount of change in the X-direction distance
- the first control object is controlled (step S6).
- the first control target is a scale of the image displayed on the LCD 2a, and the control execution unit 5 performs control to enlarge the display image according to the amount of change in the X-direction distance.
- step S5 determines the magnitude of the change amount in the X direction distance, that is,
- Step S8 the calculation unit 4 substitutes X0 'for X0 stored in the memory (Step S8). Subsequently, the control by the position detection unit 3 returns to step S3, and the position detection unit 3 enters a waiting state in preparation for the next (third) contact with the touch panel 2 by the user's two fingers.
- FIG. 4A-1 it is assumed that the display information 10 is displayed on the touch panel 2 in a normal size (1 time).
- the display information 10 is, for example, a web page image.
- the frame in the figure corresponds to the outer frame of the touch panel 2.
- the user touches the position PA with one finger, touches the position PB with the other finger, and subsequently, with one finger as shown in FIG. 4 (a-2).
- the user touches the position PC and touches the position PD with another finger.
- the control execution unit 5 generates the display information 10A by enlarging the display information 10 as shown in FIG. 4A-2, and displays the display information 10A on the LCD 2a.
- the display information 20 is displayed on the touch panel 2 in a normal size (1 time).
- the user touches the position PA with one finger, touches the position PB with the other finger, and then the position with one finger as shown in FIG. 4 (b-2).
- the control execution unit 5 reduces the display information 20 and causes the LCD 2a to display the reduced display information 20A.
- the user may appropriately set the enlargement or reduction ratio of the display information 10 and 20 in steps S6 and 7 for each application that displays the display information 10 and 20.
- the control execution unit 5 may increase or decrease the character enlargement rate by one level.
- the control execution unit 5 may increase or decrease the image enlargement ratio by 10% if the amount of change in the X-direction distance is equal to or greater than a predetermined value.
- step S9 the calculation unit 4 determines whether Y0'-Y0 is greater than 0 with respect to the amount of change in the Y-direction distance. That is, in step S9, it is determined whether the amount of change in the Y direction distance is positive or negative.
- step S9 If it is determined in step S9 that the amount of change in the Y direction distance is positive (step S9; YES), the control execution unit 5 depends on the magnitude of the amount of change in the Y direction distance, that is, the value of
- the second control target for example, the volume of a moving image or music on a WEB page can be cited.
- the control execution unit 5 performs control to increase the volume that is the second control target, in accordance with the magnitude of the amount of change in the Y-direction distance.
- step S9 when it is determined in step S9 that the change amount in the Y direction distance is negative (step S9; NO), the control execution unit 5 determines the magnitude of the change amount in the Y direction distance, that is,
- Step S12 the calculation unit 4 substitutes Y0 'for Y0 stored in the memory (Step S12). Subsequently, the control returns to step S3 by the position detector 3 again.
- the position detection unit 3 waits again until it is determined that both of the user's two fingers are in contact with the touch panel 2, and both of the two fingers are in contact with the touch panel 2. If it discriminate
- step S10 will be described in detail with reference to FIGS. 5 (a-1) and (a-2).
- display information 30 including a volume display bar 33 is displayed.
- the slider of the volume display bar 33 is at the left end, indicating that the volume is low.
- one of the fingers is brought into contact with the position PA and the other finger is brought into contact with the position PB.
- one finger is brought into contact with the position PC and the other one finger is brought into contact with the position PD.
- the user has performed an operation of spreading two fingers in the Y direction on the touch panel 2.
- control execution unit 5 controls a volume control unit (not shown) to increase the volume of the speaker 2c. Then, the control execution unit 5 displays the display information 30 ⁇ / b> A including the volume display bar 33 ⁇ / b> A on the touch panel 2. At this time, the control execution unit 5 indicates that the volume is increased by moving the slider of the volume display bar 33A to the right side.
- step S11 will be described in detail with reference to FIGS. 5B-1 and 5B-2.
- display information 40 including a volume display bar 43 is displayed.
- the slider of the volume display bar 43 is at the right end, indicating that the volume is high.
- one of the fingers is brought into contact with the position PA and the other finger is brought into contact with the position PB.
- one finger is brought into contact with the position PC and the other one finger is brought into contact with the position PD.
- the user has performed an operation of narrowing two fingers in the Y direction on the touch panel 2.
- control execution unit 5 controls the volume control unit to lower the volume of the speaker 2c. Then, the control execution unit 5 displays the display information 40 ⁇ / b> A including the volume display bar 43 ⁇ / b> A on the touch panel 2. At this time, the control execution unit 5 indicates that the volume has been reduced by moving the slider of the volume display bar 43A to the left side.
- the user may appropriately set the volume in steps S10 and S11 for each application that involves outputting audio.
- the control execution unit 5 is only one step or 10%.
- the volume may be increased or decreased.
- the amount of change in the X-direction distance and the amount of change in the Y-direction distance are calculated, and the display information is enlarged or reduced (zoomed) according to the sign of the amount of change in the X-direction distance. Further, the volume is increased or decreased according to the sign of the change amount in the Y direction distance and the magnitude thereof. That is, it is possible to perform zoom control by an operation only in the X direction and perform volume control by an operation only in the Y direction. Therefore, it is possible to perform separate control in the X direction and the Y direction, for example, it is possible to assign various different controls such as zoom control and volume control for each direction, and display by simple operation with a finger.
- the apparatus 1 can execute two processes.
- the normal size display information is converted into the XY direction.
- Control for enlarging or reducing while maintaining the ratio is performed, but the present invention is not limited to this.
- an operation of expanding or narrowing two fingers in the X direction on the touch panel 2 is performed, an operation of expanding or reducing only the X direction of the display information and expanding or narrowing two fingers in the Y direction.
- Control for enlarging or reducing only the Y direction of the display information may be assigned to the XY direction. In this way, unlike simple enlargement or reduction in which the ratio of display information in the XY direction is maintained, enlargement or reduction in which the ratio in the XY direction is changed can be performed.
- control (enlargement or reduction) of the first control target is executed in accordance with the amount of change in the X-direction distance.
- the second control target volume
- the selection unit that selects the first and second control objects from the plurality of control objects executes the control of the second control object according to the magnitude of the change amount of the X direction distance, and the Y direction.
- the control target may be appropriately assigned in the XY directions so that the control of the first control target is executed according to the magnitude of the distance change amount.
- the selection unit includes a CPU 51, a RAM 52, a ROM 53, an ADC 54, an I / F 55, and the like, and the CPU executes the program stored in the memory to execute the above-described processing. Is executed.
- the volume display bars 33, 33A, 43, and 43A are displayed on the touch panel 2 in consideration of user convenience. However, since the volume can be confirmed by hearing, the volume display bars 33, 33A, 43 and 43A may not be displayed.
- the amount of change in the distance in the X and Y directions is less than a predetermined value, it may be set not to execute the control of the first and second control objects. In this way, when the user touches the touch panel 2 with two fingers, it is possible to avoid a situation in which unintended control is executed only by moving the finger slightly.
- the control target may be changed depending on the screen display. For example, when a TV screen is displayed, enlargement or reduction of the TV screen and volume increase / decrease may be assigned in the XY directions.
- the display device 1 performs some control when the touch panel 2 is contacted at two places twice. However, if the time between the first contact and the second contact is a predetermined time or more, the process may be interrupted and the process may return to step S1.
- the control execution unit 5 includes an internal timer, and starts the internal timer when it is determined Yes in steps S1 and S3, and when it is determined No in the process of step S3, If the time measured by the timer is equal to or greater than the reference time, the process may be forcibly returned to step S1.
- the input process when two touched positions are detected has been mainly described. However, it is natural that the normal input process using the touch panel is also executed in parallel. Moreover, after detecting the contact of two fingers, when a normal input process is detected, the process may be returned to step S1.
- the capacitive sensor is exemplified as the touch sensor 2b.
- the touch sensor 2b is not limited to the capacitive sensor, and may be anything that supports multi-touch.
- the first control may be applied to the amount of change in the X-axis direction
- the second control may be applied to the amount of change in the Y-axis direction
- the third control may be applied to the amount of change in the distance in the Z-axis direction.
- the display device of the present invention is not limited to the configuration of the above embodiment, and various modifications and changes can be made from the configuration of the above embodiment. Those applied are also included in the scope of the present invention.
- Display device 2 Touch panel 2a: LCD 2b: Touch sensor 2c: Speaker 3: Position detection unit 4: Calculation unit 5: Control execution unit 10, 20, 30, 40: Display information (before control) 10A, 20A, 30A, 40A: Display information (after control)
- PA Position of first finger (before movement)
- PC Position of first finger (after movement)
- PB Second finger position (before movement)
- PD Position of second finger (after movement) 33
- 43 Volume display bar (before control) 33A
- 43A Volume display bar (after control)
Abstract
Description
また、本発明は、簡単な操作により複数の指示を入力可能な表示装置、制御方法及びこれらをコンピュータ上で実現するための記録媒体を提供することを他の目的とする。
画像を表示するタッチパネルと、
前記タッチパネル上での2つの被接触位置のXY座標をそれぞれ検出する位置検出部と、
前記位置検出部が、前記2つの被接触位置のXY座標を2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算する演算部と、
前記演算部が演算したX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算部が演算したY方向距離の変化量に応じて第2の制御対象の制御を実行する制御実行部と、
を備える。
被接触面上の2つの被接触位置のXY座標をそれぞれ検出するステップと、
前記2つの被接触位置のXY座標を所定の期間に2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算するステップと、
前記演算されたX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算されたY方向距離の変化量に応じて第2の制御対象の制御を実行するステップと、
を備える。
画像を表示するタッチパネルを備えた表示装置のコンピュータに、
前記タッチパネル上での2つの被接触位置のXY座標をそれぞれ検出するステップと、
前記2つの被接触位置のXY座標を所定の期間に2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算するステップと、
前記演算されたX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算されたY方向距離の変化量に応じて第2の制御対象の制御を実行するステップと、
を実行させるプログラムを記録する。
本実施形態に係る表示装置1は、図1(a)に示すように、タッチパネル2と、位置検出部3と、演算部4と、制御実行部5とを備える。
タッチパネル2は、LCD(Liquid Crystal Display)2aと、LCD2aの画面上に配置された静電容量方式のタッチセンサ2bと、スピーカ2c等とから構成される。LCD2aは画面に画像を表示し、タッチセンサ2bはユーザの指等で接触された位置(被接触位置)に応じた電圧信号を位置検出部3に供給する。スピーカ2cは、音を出力する。
位置検出部3は、タッチセンサ2bから供給された電圧信号に基づいて、タッチパネル2のユーザの指等により接触された位置のXY座標を検出する。
演算部4は、位置検出部3が、2つの被接触位置のXY座標を連続して2回検出すると、先の2つの被接触位置の間のX方向距離及びY方向距離と、後の2つの被接触位置の間のX方向距離及びY方向距離との各変化量(差分)を演算する。
制御実行部5は、演算部4が演算したX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、演算部4が演算したY方向距離の変化量に応じて第2の制御対象の制御を実行する。
CPU51は、RAM52に各種のデータを一時的に保存する。ADC54は、タッチセンサ2bから供給された電圧等のアナログ信号をデジタル信号に変換し、CPU51に供給する。I/F55は、CPU51と他のハードウェアとの間でデータの送受信を仲介する。
本実施形態に係る表示装置1は、電源投入後、図2に示す処理を開始する。なお、理解を容易にするため、表示装置1のLCD2aは、図4(a-1)に示すWEBページを表示しているものとする。
まず、表示装置1の位置検出部3は、ユーザの2本の指が共にタッチパネル2に接触しているか否かを、タッチセンサ2bから供給される電圧信号に基づいて判別する(ステップS1)。位置検出部3は、2本の指が共にタッチパネル2に接触していると判別するまで待ち状態となる(ステップS1;NO)。位置検出部3は2本の指が共にタッチパネル2に接触していると判別すると(ステップS1;YES)、タッチパネル2上の2本の指により接触されている2つの被接触位置のXY座標をそれぞれ検出する。
ステップS5では、演算部4が、X方向距離の変化量に関して、X0’-X0が0より大きいか否かを判定する。つまり、演算部4は、ステップS5で、X方向距離の変化量の正負を判定する。
例えば、図4(a-1)に示すように、表示情報10が通常のサイズ(1倍)でタッチパネル2に表示されているとする。表示情報10は、例えば、WEBページの画像である。なお、図中の枠は、タッチパネル2の外枠に相当する。ここで、ユーザが、1本の指で位置PAに接触し、他の1本の指で位置PBに接触し、続いて、図4(a-2)に示すように、1本の指で位置PCに接触し、他の1本の指で位置PDに接触したとする。この場合には、ユーザが、WEBページ閲覧中に2本の指をX方向に広げる操作を行ったことになる。この操作に伴い、制御実行部5は、図4(a-2)に示すように表示情報10を拡大して表示情報10Aを生成し、LCD2aに表示情報10Aを表示させる。
上記実施形態では、表示装置1は、タッチパネル2が2箇所で接触されることが、2回発生すると、何らかの制御を行った。しかし、1回目の接触から2回目の接触間での時間が所定時間以上の場合には、処理を中断し、ステップS1にもどるようにしてもよい。
この場合、例えば、制御実行部5は、内部タイマを備え、ステップS1及びS3でYesと判別された際に、内部タイマを起動し、ステップS3の処理で、Noと判別された際に、内部タイマの計時時間が基準時間以上となっている場合に、処理を強制的にステップS1にリターンするようにすればよい。
上記実施の形態では、2つの被接触位置が検出された際の入力処理を中心に説明したが、タッチパネルによる通常の入力処理も並行して実行することは当然である。また、2本の指の接触を検出した後、通常の入力処理が検出された場合に、処理をステップS1に戻すようにしてもよい。
また、上記実施形態では、タッチセンサ2bとして、静電容量方式のものを例示したが、タッチセンサ2bは静電容量方式のものに限定されず、マルチタッチに対応したものであればよい。
また、二次元で位置を検出する例を示したが、2本の指の3次元空間内の動きを検出し、X軸方向、Y軸方向、Z軸方向の距離の変化量を検出し、X軸方向の変化量に第1の制御、Y軸方向の変化量に第2の制御、Z軸方向の距離の変化量に第3の制御をわりあててもよい。
2:タッチパネル
2a:LCD
2b:タッチセンサ
2c:スピーカ
3:位置検出部
4:演算部
5:制御実行部
10,20,30,40:表示情報(制御前)
10A,20A,30A,40A:表示情報(制御後)
PA:1本目の指の位置(移動前)
PC:1本目の指の位置(移動後)
PB:2本目の指の位置(移動前)
PD:2本目の指の位置(移動後)
33,43:音量表示バー(制御前)
33A,43A:音量表示バー(制御後)
51:CPU
52:RAM
53:ROM
54:ADC
55:I/F
Claims (7)
- 画像を表示するタッチパネルと、
前記タッチパネル上での2つの被接触位置のXY座標をそれぞれ検出する位置検出部と、
前記位置検出部が、前記2つの被接触位置のXY座標を2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算する演算部と、
前記演算部が演算したX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算部が演算したY方向距離の変化量に応じて第2の制御対象の制御を実行する制御実行部と、を備える表示装置。 - 前記制御実行部は、前記変化量の大きさに対応して前記第1又は第2の制御対象の制御量を決定する、請求項1に記載の表示装置。
- 前記第1及び第2の制御対象の少なくとも一方を、複数の制御対象のうちから選択する選択部を更に備える、請求項1又は2に記載の表示装置。
- 前記制御実行部は、前記第1又は第2の制御対象が、前記画像の尺度、又は、表示装置が発生する音量である、請求項1乃至3のいずれか1項に記載の表示装置。
- 前記制御実行部は、前記X方向距離又はY方向距離の変化量の正又は負に対応して、前記第1又は第2の制御対象の制御量を一方向又は逆方向に変化させる、請求項1乃至4のいずれか1項に記載の表示装置。
- 被接触面上の2つの被接触位置のXY座標をそれぞれ検出するステップと、
前記2つの被接触位置のXY座標を所定の期間に2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算するステップと、
前記演算されたX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算されたY方向距離の変化量に応じて第2の制御対象の制御を実行するステップと、
を備える制御方法。 - 画像を表示するタッチパネルを備えた表示装置のコンピュータに、
前記タッチパネル上での2つの被接触位置のXY座標をそれぞれ検出するステップと、
前記2つの被接触位置のXY座標を所定の期間に2回検出すると、先の座標検出から後の座標検出までの、2つの被接触位置の間のX方向距離及びY方向距離の各変化量を演算するステップと、
前記演算されたX方向距離の変化量に応じて第1の制御対象の制御を実行し、且つ、前記演算されたY方向距離の変化量に応じて第2の制御対象の制御を実行するステップと、
を実行させるプログラムを記録したコンピュータ読み取り可能な記録媒体。
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020127008018A KR20120058585A (ko) | 2009-09-29 | 2010-09-29 | 표시 장치, 제어 방법 및 기록 매체 |
EP10820598.0A EP2472370B1 (en) | 2009-09-29 | 2010-09-29 | Display device, control method and recording medium |
CN201080043809.2A CN102576262B (zh) | 2009-09-29 | 2010-09-29 | 显示设备、控制方法和记录介质 |
JP2011534281A JP5686735B2 (ja) | 2009-09-29 | 2010-09-29 | 表示装置、制御方法及びプログラム |
KR1020137029638A KR101674713B1 (ko) | 2009-09-29 | 2010-09-29 | 표시 장치, 제어 방법 및 기록 매체 |
US13/499,221 US9063651B2 (en) | 2009-09-29 | 2010-09-29 | Display device, control method and recording medium |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009223583 | 2009-09-29 | ||
JP2009-223583 | 2009-09-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011040483A1 true WO2011040483A1 (ja) | 2011-04-07 |
Family
ID=43826297
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/066977 WO2011040483A1 (ja) | 2009-09-29 | 2010-09-29 | 表示装置、制御方法及び記録媒体 |
Country Status (6)
Country | Link |
---|---|
US (1) | US9063651B2 (ja) |
EP (1) | EP2472370B1 (ja) |
JP (1) | JP5686735B2 (ja) |
KR (2) | KR20120058585A (ja) |
CN (1) | CN102576262B (ja) |
WO (1) | WO2011040483A1 (ja) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130120307A1 (en) * | 2011-11-11 | 2013-05-16 | Institute For Information Industry | Handheld Device and Control Method Thereof |
JP2013200664A (ja) * | 2012-03-23 | 2013-10-03 | Denso Corp | 入力装置 |
JP2017123125A (ja) * | 2016-01-08 | 2017-07-13 | キヤノン株式会社 | 表示制御装置及びその制御方法、プログラム、並びに記憶媒体 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015078826A1 (en) | 2013-11-26 | 2015-06-04 | Koninklijke Philips N.V. | User control device, user control method, system comprising the user control device and record carrier |
KR102250856B1 (ko) * | 2014-04-30 | 2021-05-11 | 삼성전자주식회사 | 터치 입력 검출 방법, 터치 입력 감지 장치, 및 좌표 표시 장치 |
CN105468251A (zh) * | 2014-09-11 | 2016-04-06 | 深圳市麦思美汽车电子有限公司 | 一种听视觉同步反馈的音量控制方法及系统 |
CN106557243A (zh) * | 2016-11-30 | 2017-04-05 | 依偎科技(南昌)有限公司 | 智能终端调节音量的方法、装置和智能终端 |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07230352A (ja) * | 1993-09-16 | 1995-08-29 | Hitachi Ltd | タッチ位置検出装置及びタッチ指示処理装置 |
JP2006162267A (ja) | 2004-12-02 | 2006-06-22 | Denso Corp | 車両用ナビゲーション装置 |
JP3867226B2 (ja) | 2000-02-15 | 2007-01-10 | 株式会社 ニューコム | 複数の指示部位で操作可能なタッチパネルシステム |
JP2007240964A (ja) | 2006-03-09 | 2007-09-20 | Casio Comput Co Ltd | 表示装置 |
JP2008052729A (ja) * | 2006-08-22 | 2008-03-06 | Samsung Electronics Co Ltd | 多接点位置変化感知装置、方法、およびこれを用いたモバイル機器 |
JP2008217767A (ja) * | 2008-01-29 | 2008-09-18 | Fujitsu Ten Ltd | 表示装置 |
JP2009522669A (ja) * | 2005-12-30 | 2009-06-11 | アップル インコーポレイテッド | マルチタッチ入力を備えた携帯電子装置 |
JP2009223583A (ja) | 2008-03-14 | 2009-10-01 | Alpine Electronics Inc | 電圧駆動素子の異常状態検出装置 |
Family Cites Families (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5880411A (en) * | 1992-06-08 | 1999-03-09 | Synaptics, Incorporated | Object position detector with edge motion feature and gesture recognition |
US5825352A (en) * | 1996-01-04 | 1998-10-20 | Logitech, Inc. | Multiple fingers contact sensing method for emulating mouse buttons and mouse operations on a touch sensor pad |
US7663607B2 (en) * | 2004-05-06 | 2010-02-16 | Apple Inc. | Multipoint touchscreen |
JP2006277588A (ja) | 2005-03-30 | 2006-10-12 | Clarion Co Ltd | タッチパネル及びタッチパネルを備えた電子機器 |
JP5129478B2 (ja) * | 2006-03-24 | 2013-01-30 | 株式会社デンソーアイティーラボラトリ | 画面表示装置 |
CN101042300B (zh) * | 2006-03-24 | 2014-06-25 | 株式会社电装 | 画面显示装置 |
US7956847B2 (en) * | 2007-01-05 | 2011-06-07 | Apple Inc. | Gestures for controlling, manipulating, and editing of media files using touch sensitive devices |
JP5094450B2 (ja) * | 2008-02-01 | 2012-12-12 | シャープ株式会社 | 表示装置及びコンピュータプログラム |
JP4957750B2 (ja) | 2008-07-31 | 2012-06-20 | ソニー株式会社 | 情報処理装置および方法、並びにプログラム |
US8134539B2 (en) * | 2009-03-30 | 2012-03-13 | Eastman Kodak Company | Digital picture frame having near-touch and true-touch |
-
2010
- 2010-09-29 CN CN201080043809.2A patent/CN102576262B/zh not_active Expired - Fee Related
- 2010-09-29 KR KR1020127008018A patent/KR20120058585A/ko not_active Application Discontinuation
- 2010-09-29 KR KR1020137029638A patent/KR101674713B1/ko active IP Right Grant
- 2010-09-29 JP JP2011534281A patent/JP5686735B2/ja not_active Expired - Fee Related
- 2010-09-29 EP EP10820598.0A patent/EP2472370B1/en not_active Not-in-force
- 2010-09-29 US US13/499,221 patent/US9063651B2/en active Active
- 2010-09-29 WO PCT/JP2010/066977 patent/WO2011040483A1/ja active Application Filing
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07230352A (ja) * | 1993-09-16 | 1995-08-29 | Hitachi Ltd | タッチ位置検出装置及びタッチ指示処理装置 |
JP3867226B2 (ja) | 2000-02-15 | 2007-01-10 | 株式会社 ニューコム | 複数の指示部位で操作可能なタッチパネルシステム |
JP2006162267A (ja) | 2004-12-02 | 2006-06-22 | Denso Corp | 車両用ナビゲーション装置 |
JP2009522669A (ja) * | 2005-12-30 | 2009-06-11 | アップル インコーポレイテッド | マルチタッチ入力を備えた携帯電子装置 |
JP2007240964A (ja) | 2006-03-09 | 2007-09-20 | Casio Comput Co Ltd | 表示装置 |
JP2008052729A (ja) * | 2006-08-22 | 2008-03-06 | Samsung Electronics Co Ltd | 多接点位置変化感知装置、方法、およびこれを用いたモバイル機器 |
JP2008217767A (ja) * | 2008-01-29 | 2008-09-18 | Fujitsu Ten Ltd | 表示装置 |
JP2009223583A (ja) | 2008-03-14 | 2009-10-01 | Alpine Electronics Inc | 電圧駆動素子の異常状態検出装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2472370A4 |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130120307A1 (en) * | 2011-11-11 | 2013-05-16 | Institute For Information Industry | Handheld Device and Control Method Thereof |
TWI448140B (zh) * | 2011-11-11 | 2014-08-01 | Inst Information Industry | 手持裝置及其控制方法 |
JP2013200664A (ja) * | 2012-03-23 | 2013-10-03 | Denso Corp | 入力装置 |
JP2017123125A (ja) * | 2016-01-08 | 2017-07-13 | キヤノン株式会社 | 表示制御装置及びその制御方法、プログラム、並びに記憶媒体 |
Also Published As
Publication number | Publication date |
---|---|
KR20140000346A (ko) | 2014-01-02 |
US20120188193A1 (en) | 2012-07-26 |
CN102576262B (zh) | 2015-09-02 |
KR101674713B1 (ko) | 2016-11-09 |
JPWO2011040483A1 (ja) | 2013-02-28 |
EP2472370B1 (en) | 2018-02-28 |
KR20120058585A (ko) | 2012-06-07 |
EP2472370A1 (en) | 2012-07-04 |
CN102576262A (zh) | 2012-07-11 |
EP2472370A4 (en) | 2014-04-23 |
JP5686735B2 (ja) | 2015-03-18 |
US9063651B2 (en) | 2015-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5686735B2 (ja) | 表示装置、制御方法及びプログラム | |
JP5873942B2 (ja) | 携帯端末装置 | |
KR20100020311A (ko) | 터치 스크린의 정보 스크롤 제어 방법 및 장치 | |
US20140176421A1 (en) | Displaying method for flexible display device and flexible display device using the same | |
JP6171643B2 (ja) | ジェスチャ入力装置 | |
JP5999830B2 (ja) | 情報処理プログラム、情報処理装置、情報処理システム、および、情報処理方法 | |
TWI434202B (zh) | 具觸控式螢幕的電子裝置及其顯示控制方法 | |
JP2011081440A (ja) | 情報処理装置、情報処理方法及び情報処理プログラム | |
KR20100129424A (ko) | 터치스크린의 접촉위치 및 접촉힘의 세기를 이용한 사용자 인터페이스 제공 방법 및 장치 | |
JP2011034241A (ja) | タッチパネル装置及び方法並びにプログラム及び記録媒体 | |
JP2009176114A (ja) | タッチパネル装置及びユーザインタフェース装置 | |
KR20140133034A (ko) | 포터블 디바이스 및 그 제어 방법 | |
JP2010237928A (ja) | タッチパネル装置及び情報表示方法 | |
JP2010287121A (ja) | 情報処理装置、プログラム、記録媒体、及び表示制御装置 | |
JP4879933B2 (ja) | 画面表示装置、画面表示方法およびプログラム | |
JP5718433B1 (ja) | 情報処理装置 | |
EP2876540B1 (en) | Information processing device | |
JP6197559B2 (ja) | オブジェクト操作システム及びオブジェクト操作制御プログラム並びにオブジェクト操作制御方法 | |
JP5785891B2 (ja) | 表示装置 | |
KR101392739B1 (ko) | 터치 스크린 디스플레이를 통해 전자문서상에서 테이블을 생성하는 방법 및 장치 | |
JP2010205134A (ja) | 情報表示装置 | |
JP4973371B2 (ja) | 情報処理装置及びコンピュータプログラム | |
WO2017183194A1 (ja) | 表示制御装置 | |
JP2012173980A (ja) | 表示装置、表示方法、及び表示プログラム | |
KR20140072563A (ko) | 디스플레이 장치의 포인팅 동작 처리 방법 및 장치 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080043809.2 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10820598 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 20127008018 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011534281 Country of ref document: JP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13499221 Country of ref document: US Ref document number: 2900/CHENP/2012 Country of ref document: IN Ref document number: 2010820598 Country of ref document: EP |