US20210052984A1 - Program, electronic device, and method - Google Patents

Program, electronic device, and method Download PDF

Info

Publication number
US20210052984A1
US20210052984A1 US17/089,816 US202017089816A US2021052984A1 US 20210052984 A1 US20210052984 A1 US 20210052984A1 US 202017089816 A US202017089816 A US 202017089816A US 2021052984 A1 US2021052984 A1 US 2021052984A1
Authority
US
United States
Prior art keywords
axis
angle
determined
data points
slope
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US17/089,816
Other versions
US11541311B2 (en
Inventor
Shuichi Kurabayashi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Cygames Inc
Original Assignee
Cygames Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Cygames Inc filed Critical Cygames Inc
Publication of US20210052984A1 publication Critical patent/US20210052984A1/en
Assigned to CYGAMES, INC. reassignment CYGAMES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KURABAYASHI, SHUICHI
Application granted granted Critical
Publication of US11541311B2 publication Critical patent/US11541311B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04815Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/214Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads
    • A63F13/2145Input arrangements for video game devices characterised by their sensors, purposes or types for locating contacts on a surface, e.g. floor mats or touch pads the surface being also a display device, e.g. touch screens
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/218Input arrangements for video game devices characterised by their sensors, purposes or types using pressure sensors, e.g. generating a signal proportional to the pressure applied by the player
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/40Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment
    • A63F13/42Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle
    • A63F13/426Processing input control signals of video game devices, e.g. signals generated by the player or derived from the environment by mapping the input signals into game commands, e.g. mapping the displacement of a stylus on a touch screen to the steering angle of a virtual vehicle involving on-screen location information, e.g. screen coordinates of an area at which the player is aiming with a light gun
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • A63F13/525Changing parameters of virtual cameras
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • A63F13/525Changing parameters of virtual cameras
    • A63F13/5255Changing parameters of virtual cameras according to dedicated instructions from a player, e.g. using a secondary joystick to rotate the camera around a player's character
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/55Controlling game characters or game objects based on the game progress
    • A63F13/57Simulating properties, behaviour or motion of objects in the game world, e.g. computing tyre load in a car race game
    • A63F13/573Simulating properties, behaviour or motion of objects in the game world, e.g. computing tyre load in a car race game using trajectories of game objects, e.g. of a golf ball according to the point of impact
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/90Constructional details or arrangements of video game devices not provided for in groups A63F13/20 or A63F13/25, e.g. housing, wiring, connections or cabinets
    • A63F13/92Video game devices specially adapted to be hand-held while playing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/03Arrangements for converting the position or the displacement of a member into a coded form
    • G06F3/041Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
    • G06F3/044Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means
    • G06F3/0446Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means by capacitive means using a grid-like structure of electrodes in at least two directions, e.g. using row and column electrodes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04883Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/04808Several contacts: gestures triggering a specific function, e.g. scrolling, zooming, right-click, when the user establishes several contacts with the surface simultaneously; e.g. using several fingers or a combination of fingers and pen

Definitions

  • the present invention relates to programs, etc.
  • the present invention relates to a program that is executed by an electronic device having a touchscreen, etc.
  • Patent Literature 1 discloses a game device equipped with a touchscreen, as well as a program, that make it possible to set an origin in accordance with a touch operation by a user and that enables an operation simulating a joystick.
  • the game device sets reference coordinates on the basis of coordinates at the time when detection is started in the case where a touchscreen makes a transition from a state in which a touch has not been detected to a state in which a touch has been detected, and in the case where touch detection is to be continued, sets instruction coordinates on the basis of coordinates detected after the transition.
  • the game device recognizes the direction of a vector from the reference coordinates to the instruction coordinates as the direction in which a joystick is tilted and the magnitude of the vector as the degree by which the joystick is tilted, thereby realizing a virtual joystick, whereby the operation of a virtual object is realized.
  • the present invention has been made in order to solve the problem described above, and a main object thereof is to provide a program, etc. that makes it possible to improve operability in controlling an object being operated in a virtual space.
  • a program is a program that is executed on an electronic device equipped with a touchscreen, the program being characterized by causing the electronic device to execute: a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user; a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points; a step of determining the slope of a regression line on the basis of the held data points; a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
  • an amount of rotation indicating whether or not the determined slope is to be rotated by 180 degrees is determined on the basis of a displacement direction of data points temporally succeeding and following each other among the held data points.
  • the program causes the electronic device to further execute: a step of determining one of the first axis and the second axis as an axis of an independent variable and determining the other axis as an axis of a dependent variable on the basis of an amount of displacement of the values on the first axis and an amount of displacement of the values on the second axis in the held data points, and in the step of determining the slope of a regression line, the slope of the regression line is determined further on the basis of the determined axis of the independent variable and axis of the dependent variable.
  • the one of the axes is determined as the axis of the independent variable on the basis of the difference between a maximum value and a minimum value of the values on the first axis and the difference between a maximum value and a minimum value of the values on the second axis.
  • the one of the axes is determined as the axis of the independent variable by comparing the magnitude of the result of applying a weight to the difference between a maximum value and a minimum value of the values on the first axis and the magnitude of the difference between a maximum value and a minimum value of the values on the second axis.
  • an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees is determined by comparing the number of positive values and the number of negative values among the differences between values temporally succeeding and following each other on the axis of the determined independent variable in the held data points.
  • the step of determining the slope of a regression line includes: a step of determining the average of the independent variable and the average of the dependent variable in the held data points; a step of determining, by using the determined average, the deviation of the independent variable and the deviation of the dependent variable in the held data points; a step of determining, by using the determined deviation of the independent variable, the variance of the independent variable in the held data points; a step of determining, by using the determined variance of the independent variable and variance of the dependent variable, the covariance in the held data points; and a step of determining the slope of the regression line by dividing the determined covariance by the determined variance of the independent variable.
  • the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed
  • the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis
  • the slope of the regression line is determined by subtracting, from 90 degrees, an angle corresponding to the determined slope of the regression line, determined by dividing the determined covariance by the determined variance of the independent variable.
  • the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed
  • the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis.
  • an angle is determined per predefined processing period.
  • a program according to an aspect of the present invention is a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including the abovementioned program, the program suite being characterized in that the predefined processing period is a period corresponding to a frame rate for executing the game, and in that the program suite causes the execution of: a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and a step of controlling the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
  • a program suite is a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including the abovementioned program, the program being characterized in that the predefined processing period is a period corresponding to a frame rate for executing the game, and in that the program suite causes the execution of: a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and a step of controlling a virtual camera for photographing a game image, the virtual camera serving as the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
  • an electronic device is an electronic device equipped with a touchscreen, the electronic device being characterized in that: data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user, are held; the holding of a data point for which a predefined holding period has expired among the held data points is terminated; the slope of a regression line is determined on the basis of the held data points; an amount of rotation by which the determined slope of the regression line is to be rotated is determined on the basis of a displacement direction of a set of the held data points; and an angle by which the user controls an object being operated in a virtual space is determined on the basis of the determined slope of the regression line and the determined amount of rotation.
  • a method is a method that is executed on an electronic device equipped with a touchscreen, the method being characterized by including: a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user; a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points; a step of determining the slope of a regression line on the basis of the held data points; a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
  • the present invention makes it possible to improve operability in controlling an object being operated in a virtual space.
  • FIG. 1 is a block diagram showing the hardware configuration of an electronic device according to an embodiment of the present invention.
  • FIG. 2 is a functional block diagram of the electronic device according to the embodiment of the present invention.
  • FIG. 3 is an illustration showing coordinate axes consisting of a first axis and a second axis in the embodiment.
  • FIG. 4 is an illustration for explaining an example of an angle determined by an angle determination unit and a direction corresponding to that angle.
  • FIG. 5 is an illustration showing an example of data points held in a data point buffer at time t 1 .
  • FIG. 6 is an illustration showing an example of data points held in the data point buffer at time t 1 .
  • FIG. 7 is an illustration showing a regression line obtained from the set of data points shown in FIG. 5 .
  • FIG. 8 is an illustration showing an angle calculated by a function aop(x, y) by using the slope of the regression line shown in FIG. 7 .
  • FIG. 9 is an illustration showing an example of data points held in the data point buffer at time t 2 .
  • FIG. 10 is an illustration showing an angle calculated by the function aop(x, y) by using the slope of the regression line shown in FIG. 9 .
  • FIG. 11 is an illustration showing an example of data points held in the data point buffer at time t 3 .
  • FIG. 12 is an illustration showing an example of data points held in the data point buffer at time t 3 .
  • FIG. 13 is an illustration showing an angle calculated by the function aop(x, y) by using a regression line obtained from the set of data points shown in FIG. 11 and the slope of the regression line.
  • FIG. 14 is an illustration showing an example of data points held in the data point buffer at time t 4 .
  • FIG. 15 is an illustration showing an example of data points held in the data point buffer at time t 4 .
  • FIG. 16 is an illustration showing a regression line obtained from the set of data points shown in FIG. 14 .
  • FIG. 17 is an illustration showing an angle calculated by the function aop(x, y) by using the slope of the regression line shown in FIG. 16 .
  • FIG. 18 is an illustration showing how a virtual character is controlled on the basis of a set of data points by the electronic device according to the embodiment of the present invention.
  • FIG. 19 is a flowchart of information processing by the angle determination unit, according to the embodiment of the present invention.
  • FIG. 20 is an illustration showing a virtual camera, disposed in a virtual space, that captures an image displayed by the electronic device according to the embodiment of the present invention.
  • An electronic device 10 has installed therein a game application that presents a user with a virtual object disposed in a virtual space while causing the game to proceed.
  • the electronic device 10 When the game application is executed, the electronic device 10 according to this embodiment provides a virtual controller (virtual controller) for controlling an object being operated, which is a virtual object being operated by the user in the virtual space, in response to an operation by the user.
  • the virtual space is defined by the game application, and may be either a two-dimensional space or a three-dimensional space.
  • the virtual object is a character or an item disposed in the virtual space.
  • controlling the object being operated means controlling the character or the item disposed in the virtual space.
  • the electronic device 10 has the above-described game application installed therein; however, there is no limitation to the above-described game application. It suffices for the electronic device 10 to have installed therein a game application that can control an object being operated in response to an operation by a user.
  • the electronic device 10 may have installed therein an input support application or a simulation application that brings about an action of an object being operated in response to an operation by a user.
  • the virtual controller provided by the electronic device 10 in this embodiment can be used for the type of user input requiring a direction and a magnitude.
  • the term “application” refers to application programs in general, and may refer to apps that are installed on a smartphone or a tablet.
  • FIG. 1 is a block diagram showing the hardware configuration of the electronic device 10 according to the embodiment of the present invention.
  • the electronic device 10 includes a processor 11 , an input device 12 , a display device 13 , a storage device 14 , and a communication device 15 . These constituent devices are connected via a bus 16 . It is assumed that interfaces are interposed as needed between the bus 16 and the individual constituent devices.
  • the electronic device 10 is a smartphone.
  • the electronic device 10 may be a terminal such as a computer equipped with a contact-type input device, like a tablet computer or a touchpad, as long as the terminal includes the configuration described above.
  • the processor 11 controls the overall operation of the electronic device 10 .
  • the processor 11 is a CPU.
  • an electronic circuit such as an MPU may be used as the processor 11 .
  • the processor 11 executes various kinds of processing by loading programs and data stored in the storage device 14 and executing the programs.
  • the processor 11 is constituted of a plurality of processors.
  • the input device 12 is a user interface for accepting inputs to the electronic device 10 from the user; for example, the input device 12 is a touchscreen, a touchpad, a keyboard, or a mouse.
  • the display device (display) 13 displays application screens, etc. to the user of the electronic device 10 under the control of the processor 11 .
  • the electronic device 10 since the electronic device 10 is a smartphone, the electronic device 10 includes a touchscreen 17 as the input device 12 , and the touchscreen 17 also functions as the display device 13 ; that is, the input device 12 and the display device 13 have an integrated structure.
  • the touchscreen 17 in this embodiment is a projection-type capacitive touchscreen, a device having an equivalent function, if any, may be adopted.
  • the storage device 14 is a storage device included in an ordinary smartphone, including a RAM, which is a volatile memory, and a ROM, which is a non-volatile memory.
  • the storage device 14 may also include an external memory.
  • the storage device 14 stores various kinds of programs, including a game application.
  • the storage unit 14 stores an operating system (OS), middleware, application programs, various kinds of data that may be referred to when these programs are executed, etc.
  • OS operating system
  • middleware middleware
  • application programs various kinds of data that may be referred to when these programs are executed, etc.
  • the storage device 14 includes a main storage device and an auxiliary storage device.
  • the main storage device is a volatile storage medium that allows high-speed reading and writing of information, and is used as a storage area and a work area when the processor 11 processes information.
  • the auxiliary storage device stores various programs and data that are used by the processor 11 when the individual programs are executed.
  • the auxiliary storage device is, for example, a hard disk device; however, the auxiliary storage device may be any type of non-volatile storage or non-volatile memory, which may be of the removable type, that is capable of storing information.
  • the communication device 15 sends data to and receives data from other computers, such as a server, via a network.
  • the communication device 15 performs wireless communication, such as mobile communication or wireless LAN communication, to connect to the network 2 .
  • the electronic device 10 downloads a program from a server by means of the communication device 15 and stores the program in the storage device 14 .
  • the communication device 15 may perform known wired communication. In the case where data is neither sent to nor received from other computers, the electronic device 10 need not include the communication device 15 .
  • FIG. 2 is a functional block diagram of the electronic device 10 according to the embodiment of the present invention.
  • the electronic device 10 includes an input unit 21 , a display unit 22 , and a control unit 23 .
  • the control unit 23 includes an angle determination unit 24 , a state determination unit 25 , and an application unit 26 .
  • these functions are realized by the processor 11 executing a program.
  • the program that is executed is a program stored in the storage device 14 or received via the communication device 15 . Since various kinds of functions are realized by loading a program, as described above, a portion or the entirety of one part (function) may be provided in another part. Alternatively, these functions may be realized by means of hardware by configuring electronic circuits or the like for realizing the individual functions in part or in entirety.
  • the input unit 21 is configured by using the input device 12 , and accepts inputs to the electronic device 10 from the user.
  • the input unit 21 accepts a touch operation performed on the touchscreen 17 by the user and generates touch events, for which a touch detection function generally provided in a smartphone having the touchscreen 17 can be used.
  • the display unit 22 displays a game application screen on the display device 13 so as to display a screen in accordance with user operations.
  • the control unit 23 realizes a virtual controller.
  • an architecture having a three-layer structure is adopted for the control unit 23 , and the angle determination unit 24 , the state determination unit 25 , and the application unit 26 correspond to the individual layers.
  • the control unit 23 is realized by the processor 11 executing a program suite constituted of programs corresponding to the individual layers.
  • the role of the angle determination unit 24 in the virtual controller realized by the control unit 23 corresponds to a sensor IC chip inside the controller in the case of a physical controller.
  • the angle determination unit 24 determines an angle for controlling an object being operated by the user in the virtual space by mainly using touch events generated by a touch operation on the touchscreen 17 by the user, and forwards the angle to the state determination unit 25 .
  • the role of the state determination unit 25 corresponds to the entire controller that is operated in the case of a physical controller.
  • the state determination unit 25 determines a vector (angle and magnitude) corresponding to the touch operation performed on the touchscreen 17 by the user by mainly using the angle forwarded from the angle determination unit 24 .
  • the application unit 26 corresponds to a specific game application that implements actions, etc. in a game.
  • a frame rate is defined, and for example, the main loop of a main program is executed per period corresponding to the frame rate.
  • the frame rate is 30 fps (frames per second) or 60 fps.
  • the angle determination unit 24 determines an angle that is required when the user controls an object being operated in the virtual space.
  • the angle determination unit 24 is a mathematical function library for calculating an angle in real time, and is a software module implementing a statistical processing algorithm for calculating an angle from a sequence of touch events during a short period. The sequence of touch events corresponds to finger motion on the touchscreen 17 .
  • the storage device 14 includes a data point buffer.
  • the data point buffer is a buffer that enables the angle determination unit 24 to hold data points obtained on the basis of touch events and indicated by values on a first axis and values on a second axis.
  • the angle determination unit holds, in the data point buffer, data points obtained on the basis of touch events generated by user operations on the touchscreen 17 and indicated by values on the first axis and values on the second axis.
  • a touch event occurs when the user places a finger in contact with the touchscreen 17 (touchstart), when the user moves the finger while keeping the finger in contact with the touchscreen (touchmove), when the user removes the finger from the touchscreen 17 (touchend), etc.
  • the angle determination unit 24 obtains a touch event when a touch event occurs.
  • the angle determination unit 24 When obtaining a touch event, the angle determination unit 24 obtains a set of numerical values (x, y) consisting of two variables and corresponding to a position at which the electrostatic capacitance on the touchscreen 17 has changed, and also obtains time t at which (x, y) was obtained, and stores a set of numerical values (x, y, t) consisting of three variables in the data point buffer.
  • the data of the set of numerical values consisting of two variables is obtained by the angle determination unit 24 in association with a touch event, and corresponds to data points indicated by values on a value on the first axis and a value on the second axis.
  • t is a value representing a data point obtaining time, which is the time at which (x, y) was obtained, and is stored in the data point buffer in association with (x, y), as described above.
  • t is an integer value called the UNIX (registered trademark) time or a character sequence such as “2017/07/14 15:48:43.444”.
  • the holding (or termination of holding) of a data point by the angle determination unit 24 includes the holding (or termination of holding) of a data point obtaining time t associated with the data point.
  • FIG. 3 is an illustration showing coordinate axes consisting of the first axis and the second axis in this embodiment.
  • the first axis is an axis indicating the widthwise direction of the directions substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed; i.e., the first axis is a horizontal axis (x axis) substantially parallel to the shorter sides of the touchscreen 17 .
  • the second axis is an axis perpendicular to the first axis and indicating the lengthwise direction of the directions substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed; i.e., the second axis is a vertical axis (y axis) substantially parallel to the longer sides of the touchscreen 17 .
  • the first axis and the second axis are substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed, and a position on the touchscreen 17 is expressed in the form of coordinates (x, y) along the two axes.
  • the coordinates (x, y) of a data point correspond to a position on the touchscreen 17 .
  • the angle determination unit 24 holds the coordinates (x, y) as a data point in the data point buffer.
  • the coordinate setting shown in FIG. 3 is just an example, and coordinates may be set differently from the above example depending on the sensor array of the touchscreen 17 or the program implemented by the electronic device 10 .
  • the sensors are, for example, electrodes.
  • FIG. 4 is an illustration for explaining an example of an angle determined by the angle determination unit 24 and a direction corresponding to the angle in the case where coordinate axes are defined as shown in FIG. 3 .
  • An angle 31 in the first quadrant corresponds to a direction 32
  • an angle 33 in the second quadrant corresponds to a direction 34
  • an angle 35 in the fourth quadrant corresponds to a direction 36 .
  • the direction 34 and the direction 36 have the same slope, since the direction 34 and the direction 36 are opposite to each other, it is possible to confirm that the angle 33 and the angle 35 differ by 180 degrees.
  • the angle determination unit 24 terminates the holding of a data point for which a predefined holding period has expired among the data points held in the data point buffer. For example, when terminating the holding of a data point, the angle determination unit 24 may delete the data, invalidate the data, or delete the data, as appropriate, by associating a flag indicating that the holding has been terminated with the data.
  • the angle determination unit 24 defines a variable D specifying, in milliseconds, the life of the data points stored in the data point buffer. The period specified by the variable D corresponds to the predefined holding period. However, the value of the variable D is not limited to milliseconds.
  • the angle determination unit 24 monitors the time elapsed since the storage of the data point and continuously compares the time elapsed with the variable D. When the monitored time elapsed for the data point exceeds the variable D, the angle determination unit 24 terminates the holding of the data point in the data point buffer. In this manner, the angle determination unit 24 manages the life of every data point held in the data point buffer. At this time, the angle determination unit 24 can calculate the elapsed time by using the data point obtaining time t. It is assumed that the case where it is determined that the time elapsed exceeds the variable D may mean the case where it is determined that the time elapsed is greater than or equal to the variable D. Furthermore, the unit of the time elapsed for a data point, managed by the angle determination unit 24 , should preferably be the same as that of the variable D.
  • the angle determination unit 24 holds a data point stored in the data point buffer for 165 milliseconds, and after 165 milliseconds has elapsed, terminates the holding of the data point in the data point buffer.
  • the angle determination unit 24 periodically determines whether or not it is possible to calculate an angle. In the case where it is possible to calculate an angle, the angle determination unit 24 calculates an angle indicated by the set of data points held in the data point buffer by using the data points, and determines the angle as an angle by which the user controls the object being operated in the virtual space. By obtaining the angle indicated by the set of data points, as described above, the angle determination unit 24 can obtain an angle of the direction intended by the user who has performed a touch operation on the touchscreen 17 .
  • the angle determination unit 24 outputs the determined angle to the state determination unit 25 .
  • the angle determination unit 24 When outputting the determined angle to the state determination unit 25 , for example, the angle determination unit 24 outputs the determined angle together with information indicating an angle event.
  • the angle determination unit 24 may store the determined angle and the information indicating an angle event in a memory area in the storage unit 14 referred to by the state determination unit 25 .
  • the angle determination unit 24 When the obtained touch event is touchstart, the angle determination unit 24 outputs a start event to the state determination unit 25 , and does not calculate an angle. When the obtained touch event is touchend, the angle determination unit 24 outputs a stop event to the state determination unit 25 , and does not calculate an angle.
  • the angle determination unit 24 defines a variable B, and calculates and determines an angle by using the data points held in the data point buffer in the case where the number of the data points is greater than or equal to the value of the variable B. In the case where the number of the data points held in the data point buffer is less than the value of the variable B, in which case it is not possible to calculate an angle indicated by the set of data points with at least a certain level of accuracy, the angle determination unit 24 outputs a keep event to the state determination unit 25 , and does not calculate an angle. Since it is generally preferable that three or more data points are available when obtaining the slope of a regression line, the variable B should preferably be set to be 3 or greater. In this embodiment, the variable B is set to be 3.
  • the angle determination unit 24 defines a variable I specifying, in milliseconds, the time interval for determining whether or not the angle determination unit 24 can calculate an angle. In the case where the situation where the angle determination unit 24 cannot calculate an angle continues, like a situation where touch events continuously occur, the interval serves as a time interval for the angle determination unit 24 to calculate an angle.
  • the value of the variable I is not limited to milliseconds.
  • the angle determination unit 24 determines an angle per predefined processing period by determining whether or not it is possible to calculate an angle per predefined processing period, as described above.
  • the predefined processing period is a period corresponding to the frame rate for executing the game.
  • the variable I is set to be 33.
  • the angle determination unit 24 determines whether or not it is possible to calculate an angle every 33 milliseconds.
  • the angle determination unit 24 determines a calculated angle and outputs the determined angle together with an angle event to the state determination unit 25 .
  • the angle determination unit 24 outputs one of a start event, a stop event, and a keep event to the state determination unit 25 .
  • the angle event may include information indicating an angle, and in this case, the angle determination unit 24 outputs an angle event to the state determination unit 25 in the case where it is possible to calculate an angle.
  • the angle determination unit 24 determines the slope of a regression line on the basis of the data points held in the data point buffer.
  • the number of data points referred to by the angle determination unit 24 is greater than or equal to 3.
  • the angle determination unit 24 determines one of the x axis and the y axis as the axis of an independent variable on the basis of the amount of displacement in the values on the x axis and the amount of displacement in the values on the y axis in the data points held in the data point buffer.
  • the angle determination unit 24 determines the other axis as the axis of a dependent variable.
  • the angle determination unit 24 calculates the slope of a regression line by using the method of least squares.
  • the method of obtaining the slope of a regression line by using the method of least squares is known; in this method, the degree of correlation between an independent variable and a dependent variable is obtained in the form of the slope.
  • the method of least squares is generally applied by considering the x axis as the axis of an independent variable and the y axis as the axis of a dependent variable. However, in the case where the value on the x axis is fixed and only the value on the y axis changes, the dependent variable has no dependency on the independent variable.
  • the angle determination unit 24 determines which of the value on the x axis and the value on the y axis is suitable as an independent variable, and determines the slope of a regression line by using the method of least squares on the basis of the axis of the independent variable and the axis of the dependent variable thus determined.
  • the angle determination unit 24 calculates (determines) the average of the independent variable and the average of the dependent variable in the data points held in the data point buffer.
  • the angle determination unit 24 calculates (determines) the deviation of the independent variable and the deviation of the dependent variable in the data points held in the data point buffer.
  • the angle determination unit 24 calculates (determines) the variance of the independent variable in the data points held in the data point buffer.
  • the angle determination unit 24 by using the deviation of the independent variable and the deviation of the independent variable, determined in (b), calculates (determines) the covariance in the data points held in the data point buffer.
  • the angle determination unit 24 calculates (determines) the slope of the regression line by dividing the covariance, determined in (d), by the variance of the independent variable, determined in (c).
  • the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable and determines the other axis as the axis of the dependent variable on the basis of the difference between the maximum value and the minimum value of the values on the x axis and the difference between the maximum value and the minimum value of the values on the y axis in the data points held in the data point buffer.
  • the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable and determines the other axis as the axis of the dependent variable on the basis of the difference between the maximum value and the minimum value of the values on the x axis, the difference being weighted by using a weight coefficient, and the difference between the maximum value and the minimum value of the values on the y axis in the data points held in the data point buffer.
  • the angle determination unit 24 calculates (determines) the slope of a regression line by using the method of least squares, as described above.
  • the calculation of the slope of a regression line by the angle determination unit 24 means the calculation of the angle of the slope of a regression line by the angle determination unit 24 .
  • the angle determination unit 24 calculates the slope, for example, within the range of 0 to 90 degrees and 270 to 360 degrees.
  • the slope of a regression line is calculated to be 45 degrees, whether the angle indicated by the set of data points is 45 degrees or 225 degrees.
  • the angle determination unit 24 determines the amount of rotation by which the determined slope of the regression line is to be rotated. Specifically, after determining the slope of a regression line, when determining an angle by which the user controls the object being controlled, the angle determination unit 24 determines an amount of rotation indicating whether or not (the angle of) the determined slope of the regression line is to be rotated by 180 degrees.
  • the displacement direction of the set of data points indicates a direction in which the data points become displaced with time, and corresponds, for example, to a rough direction in which the user moves a finger on the touchscreen 17 .
  • the angle determination unit 24 determines the amount of rotation on the basis of the displacement direction of data points temporally succeeding and following each other among the data points held in the data point buffer. Preferably, the angle determination unit 24 determines the amount of rotation by comparing the number of positive differences and the number of negative differences between the values temporally succeeding and following each other on the axis of the determined independent variable in the data points held in the data point buffer. The amount of rotation that is determined is an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees.
  • the angle determination unit 24 stores data points so as to sequentially hold the data points in order from data points stored earlier in the data point buffer. Alternatively, when storing data points in the data point buffer, the angle determination unit 24 stores identification information that enables identification of the order of storage in association with the data points. As described above, the angle determination unit 24 holds data points in the data point buffer in such a manner that the order of storage can be identified, i.e., in such a manner that the temporal order of the stored data points can be identified. At this time, the angle determination unit 24 can use the values of data point obtaining times t.
  • the angle determination unit 24 calculates an angle indicated by the set of data points on the basis of the determined slope of the regression line and the determined amount of rotation, and determines the angle as an angle by which the user controls the object being operated in the virtual space.
  • the determination of the slope of a regression line, the determination of an amount of rotation, and the determination of an angle, performed by the angle determination unit 24 described above are realized by using an aop(x, y) function expressed in Eq. (1).
  • the function aop(x, y) calculates an angle in the form of a real value in the range of 0 to 360 degrees. It is assumed that the data point buffer holds n data points P(x, y) when the function aop(x, y) calculates an angle.
  • the function aop(x, y) uses a function rotate(x, y), a function left(x), and a function down(y).
  • the function aop(x, y) determines either one of e and y as an independent variable by using the function rotate(x, y).
  • the function rotate(x, y) determines whether or not the n data points P(x, y) are displaced mainly in the y axis direction, and returns, for example, a true or false value.
  • the case where the n data points P(x, y) are displaced mainly in the y axis direction refers to, for example, the case where a finger of the user has been moved on the touchscreen 17 mainly in the vertical direction.
  • the function rotate(x, y) determines whether the n data points P(x, y) are displaced mainly in the x axis (horizontal) direction or mainly in the y axis (vertical) direction, thereby determining which of the value on the x axis and the value on the y axis is suitable as an independent variable.
  • (max(x) ⁇ min(x)) signifies the absolute value of the difference between the maximum value and the minimum value among the x values (x 1 , x 2 , . . . x n ) of the n data points P, which indicates the amount of displacement of the n data points P in the x axis direction.
  • (max(y) ⁇ min(y)) signifies the absolute value of the difference between the maximum value and the minimum value among the y values (y 1 , y 2 , . . . y n ) of the n data points P, which indicates the amount of displacement of the n data points P in the y axis direction.
  • a variable w is a weight coefficient for applying a weight to (max (x) ⁇ min(x)).
  • the function rotate(x, y) satisfies the inequality when (max(y) ⁇ min(y)) is greater than the product of (max(x) ⁇ min(x)) and the variable w, in which case the function aop(x, y) performs coordinate transformation.
  • the function aop(x, y) determines an amount of rotation by considering the y axis as the axis of the independent variable, considering the x axis as the axis of the dependent variable, and further using the function down(y).
  • the function rotate(x, y) does not satisfy the inequality when (max(y) ⁇ min(y)) is less than or equal to the product of (max(x) ⁇ min(x)) and the variable w, in which case the function aop(x, y) does not perform coordinate transformation.
  • the function aop(x, y) determines an amount of rotation by considering the x axis as the axis of the independent variable, considering the y axis as the axis of the dependent variable, and further using the function left(x).
  • the function aop(x, y) results in performing coordinate transformation if the amount of displacement in the y axis direction is slightly greater than the amount of displacement in the x axis direction.
  • the y axis of the touchscreen 17 shown in FIG. 17 usually becomes substantially parallel to the gravitational direction.
  • the variable w should preferably be set to be a value greater than 1.
  • the variable w should preferably be set to be 2.
  • the angle determination unit 24 sets w to be a value less than 1, such as 0.5.
  • the function left(x) determines whether or not the displacement direction of the n data points P(x, y) is the ⁇ x axis direction (leftward direction), and returns, for example, a true or false value. Specifically, the function left(x) calculates the differences (x 2 ⁇ x 1 , x 3 ⁇ x 2 , . . . x n ⁇ x n-1 ) of values temporally succeeding and following each other in the x values (x 1 , x 2 , . . . x n ) of the n data points P.
  • the function left(x) determines whether or not the displacement direction of the n data points P(x, y) is the ⁇ x axis direction (leftward direction) by determining whether or not the number of negative differences is greater than the number of positive differences. In this manner, the function left(x) determines whether the displacement direction of the n data points P(x, y) is the ⁇ x axis direction (leftward direction) or the +x axis direction (rightward direction), thereby determining an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees.
  • the function aop(x, y) determines the amount of rotation to be 180 degrees in the case where the function left(x) is true, and the function aop(x, y) determines the amount of rotation to be 0 degrees in the case where the function left(x) is false.
  • the function down(y) determines whether or not the displacement direction of the n data points P(x, y) is the ⁇ y axis direction (downward direction), and returns, for example, a true or false value. Specifically, the function down(y) calculates the differences (y 2 ⁇ y 1 , y 3 ⁇ y 2 , . . . y n ⁇ y n-1 ) of values temporally succeeding and following each other in the y values (y 1 , y 2 , . . . y n ) of the n data points P.
  • the function down(y) determines whether or not the displacement direction of the n data points P(x, y) is the ⁇ y axis direction (downward direction) by determining whether or not the number of negative differences is greater than the number of positive differences. In this manner, the function down(y) determines whether the displacement direction of the n data points P(x, y) is the ⁇ y axis direction (downward direction) or the +y axis direction (upward direction), thereby determining an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees.
  • the function aop(x, y) determines the amount of rotation to be 180 degrees in the case where the function down(y) is true, and the function aop(x, y) determines the amount of rotation to be 0 degrees in the case where the function down(y) is false.
  • n ⁇ ⁇ i 1 n ⁇ ⁇ ( x i - x _ ) ⁇ ( y i - y _ )
  • ⁇ i 1 n ⁇ ⁇ ( x i - x _ ) 2
  • n ⁇ ⁇ i 1 n ⁇ ⁇ ( x i - x _ ) ⁇ ( y i - y _ )
  • ⁇ i 1 n ⁇ ⁇ ( y i - y _ ) 2
  • the angle determination unit 24 defines a variable V, and in the case where the variance of the independent variable is greater than or equal to V, calculates and determines an angle by using the function aop(x, y). In the case where the variance of the independent variable is less than V, the angle determination unit 24 outputs a keep event to the state determination unit 25 , and does not calculate an angle.
  • the case where the variance of the independent variable is less than V indicates that the n data points P(x, y) are locally concentrated.
  • the variable V is preferably set to be 0.7
  • the function aop(x, y) directly calculates an angle obtained from the slope of the regression line. In the case where the function rotate(x, y) is false and the function left(x) is true, the function aop(x, y) calculates an angle obtained by adding 180 degrees to an angle obtained from the slope of the regression line. In the case where the function rotate(x, y) is true and the function down(y) is false, the function aop(x, y) calculates an angle by subtracting an angle obtained from the slope of the regression line from 90 degrees.
  • the function aop(x, y) calculates an angle by adding 180 degrees to an angle obtained by subtracting an angle obtained from the slope of the regression line from 90 degrees.
  • the state determination unit 25 determines an angle and a magnitude per predefined processing period on the basis of the angle defined by the angle determination unit 24 per predefined processing period.
  • the state determination unit 25 outputs the determined angle and magnitude to the application unit 26 .
  • the state determination unit 25 may store information indicating the determined angle and magnitude in a memory area referred to by the application unit 26 .
  • the state determination unit 25 is a software module that converts a sequence of successively calculated angles into a function of a controller. This module is implemented as a state management model (state machine) using the angle and the time, without using touched coordinates at all.
  • the state determination unit 25 Upon receiving the information output from the angle determination unit 24 per predefined processing period, the state determination unit 25 outputs the state of the slope of the virtual controller per predefined processing period in accordance with the received event.
  • the virtual controller realizes a physical analog controller by means of software, and the slope of the virtual controller corresponds to the slope of a joystick or level of a physical controller.
  • the state determination unit 25 outputs the state of the slope of the virtual controller in the form of a vector, i.e., as a direction (angle) and a magnitude.
  • the magnitude may be acceleration information or speed information.
  • the state of the slope of the virtual controller corresponds to a quantity input to a physical analog controller.
  • the state determination unit 25 can be considered as providing a state transition model for interpreting the meaning of a generated touch event.
  • the state determination unit 25 manages the slope of the virtual controller as a state, and upon receiving the information output from the angle determination unit 24 per predefined processing period, causes a state transition to occur per predefined processing period.
  • the state determination unit 25 accumulates the information about the angle output from the angle determination unit 24 , and calculates an output to the application unit 26 in accordance with time. For example, in the case where the angle determination unit 24 outputs 45 degrees per 33 ms continuously over 165 ms, the state determination unit 25 determines acceleration information in the direction of 45 degrees for 165 ms. For example, in the case where the same angle is continuously output from the angle determination unit 24 , the state determination unit 25 outputs information indicating acceleration in the direction of that angle, and in the case where an angle opposite to the abovementioned angle by 180 degrees is subsequently input from the angle determination unit 24 , the state determination unit 25 outputs information indicating deceleration. With this configuration, it becomes possible to obtain a quantity input to the controller in the form of a period corresponding to the number of state transitions, without having to directly use actual touched coordinates.
  • the state determination unit 25 upon receiving the information output from the angle determination unit 24 , sets the state of the slope of the virtual controller in accordance with the received event, and outputs a movement event or an acceleration event to the game application together with the information indicating the angle and the magnitude.
  • the state determination unit 25 can implement acceleration processing simulating a physical analog controller by using a predetermined programming language.
  • the state determination unit 25 holds the state of the slope of the virtual controller in the form of a vector, and outputs the held vector per predefined processing period.
  • the state determination unit 25 starts state management.
  • the state determination unit 25 starts holding a vector consisting of the received angle and the magnitude of a single state transition, and outputs the vector.
  • the state determination unit 25 adds a vector consisting of the received angle and the magnitude of a single state transition to the internally held vector, and holds and outputs the resulting vector.
  • the state determination unit 25 normalizes the length of the output vector to 1.0 in the case where the length exceeds 1.0.
  • the state determination unit 25 when angle information indicating 45 degrees is received once from the angle determination unit 24 , the state determination unit 25 outputs and holds a vector consisting of an angle of 45 degrees and a magnitude of 0.2. When angle information indicating 45 degrees is received four further times from the angle determination unit 24 , the state determination unit 25 holds and outputs a vector consisting of an angle of 45 degrees and a magnitude of 1.0. Also, when angle information indicating 45 degrees is received another time from the angle determination unit 24 , the state determination unit 25 holds and outputs a vector consisting of an angle of 45 degrees and a magnitude of 1.0.
  • the state determination unit 25 When angle information indicating 30 degrees is received, the state determination unit 25 adds together a vector consisting of an angle of 30 degrees and a magnitude of 0.2 and the vector consisting of the angle of 45 degrees and the magnitude of 1.0, and holds and outputs a vector consisting of the resulting angle and the magnitude of 1.0. Also in this case, since the magnitude of the vector exceeds 1.0, the state determination unit 25 normalizes the magnitude to 1.0 when outputting the vector.
  • the state determination unit 25 When a keep event is received from the angle determination unit 24 , the state determination unit 25 continues holding the internally held vector and outputs the vector. When a stop event is received from the angle determination unit 24 , the state determination unit 25 quits holding the internally held vector, and either outputs a zero vector or outputs the stop event.
  • the application unit 26 has the function of an app activated on the electronic device 10 , and provides a service by the app.
  • the application unit 26 converts the vector output from the state determination unit 25 into motion of a specific virtual character or the like, which is a function implemented in ordinary game apps.
  • a server that communicates with the electronic device 10 has the application unit 26 in part or in entirety, and the app sends data to and receives data from the server.
  • FIG. 5 is an illustration showing an example of the data points held in the data point buffer at time t 1 .
  • a set 40 of the data points held in the data point buffer consists of six data points P 41 to P 46 .
  • the coordinates (x, y) of the x coordinate values and the y coordinate values of P 41 to P 46 are individually P 41 ( x 41 , y 41 ), P 42 ( x 42 , y 42 ), . . . P 46 ( x 46 , y 46 ), and it is assumed that the data points were stored in the data point buffer in the order P 41 , P 42 , . . . , P 46 .
  • the function aop(x, y) calculates the differences (x 42 ⁇ x 41 , x 43 ⁇ x 42 , . . . x 46 ⁇ x 45 ) between values temporally succeeding and following each other by using the function left(x). According to FIG. 5 , all the differences are positive. Thus, the function left(x) does not satisfy the inequality and thus returns a false value.
  • the function rotate(x, y) is false, and the function left(x) is false.
  • the function aop(x, y) obtains a regression line 81 , shown in FIG. 7 , from the set 40 of data points, and calculates an angle 91 from the slope of the regression line 81 , as shown in FIG. 8 .
  • FIG. 9 is an illustration showing an example of the data points held in the data point buffer at time t 2 .
  • a set 60 of the data points held in the data point buffer consists of six data points P 61 to P 66 .
  • the x coordinate values and the y coordinate values of P 61 , P 62 , . . . P 66 are respectively the same as the x coordinate values and the y coordinate values of P 46 , P 45 , . . . , P 41 , and it is assumed that the data points were stored in the order P 61 , P 62 , . . . , P 66 .
  • the function rotate(x, y) does not satisfy the inequality and thus returns a false value.
  • the function aop(x, y) calculates the differences between values temporally succeeding and following each other by using the function left(x). According to FIG. 9 , all the differences are negative. Thus, the function left(x) satisfies the inequality and returns a true value.
  • the function rotate(x, y) is false, and the function left(x) is true.
  • the function aop(x, y) obtains a regression line 82 , shown in FIG. 9 , from the set 60 of data points, and calculates an angle 92 b obtained by adding 180 degrees to an angle 92 a calculated from the slope of the regression line 82 , as shown in FIG. 10 .
  • FIG. 11 is an illustration showing an example of the data points held in the data point buffer at time t 3 .
  • a set 70 of the data points held in the data point buffer consists of six data points P 71 to P 76 .
  • the x coordinate values and the y coordinate values of P 71 to P 76 are individually P 71 ( x 71 , y 71 ), P 72 ( x 72 , y 72 ), . . . P 76 ( x 76 , y 76 ), and it is assumed that the data points were stored in the data point buffer in the order P 71 , P 72 , . . . , P 76 .
  • the function aop(x, y) calculates the differences (y 72 ⁇ y 71 , y 73 ⁇ y 72 , . . . y 76 ⁇ y 75 ) between values temporally succeeding and following each other by using the function down(y). According to FIG. 11 , all the differences are positive. Thus, the function down(y) does not satisfy the inequality and thus returns a false value.
  • the function rotate(x, y) is true, and the function down(y) is false.
  • the function aop(x, y) obtains a regression line 83 , shown in FIG. 13 , from the set 70 of data points, and calculates an angle 93 b obtained by subtracting an angle 93 a calculated from the slope of the regression line 83 from 90 degrees, as shown in FIG. 13 .
  • FIG. 14 is an illustration showing an example of the data points held in the data point buffer at time t 4 .
  • Time t 4 is the time after an elapse of time ⁇ t from time t 1 .
  • a set 40 ′ of the data points held in the data point buffer consists of eight data points P 45 to P 52 .
  • the x coordinate values and the y coordinate values of P 45 to P 52 are respectively P 45 ( x 45 , y 45 ), P 46 ( x 46 , y 46 ) . . . P 52 (x 52 , y 52 ), and it is assumed that the data points were stored in the order P 45 , P 46 , . . . , P 52 .
  • the function aop(x, y) calculates the differences (y 46 ⁇ y 45 , y 47 ⁇ y 46 , . . . y 52 ⁇ y 51 ) between values temporally succeeding and following each other by using the function down(y). According to FIG. 14 , all the differences except y 46 ⁇ y 45 are negative. Thus, the function down(y) satisfies the inequality and thus returns a true value.
  • the function rotate(x, y) is true, and the function down(y) is true.
  • the function aop(x, y) obtains a regression line 84 , shown in FIG. 16 , from the set 40 ′ of data points, and calculates an angle 94 c obtained by adding 180 degrees to an angle 94 b calculated by subtracting an angle 94 a calculated from the slope of the regression line 84 from 90 degrees, as shown in FIG. 17 .
  • FIG. 18 is an illustration showing how the electronic device 10 according to the embodiment of the present invention controls a virtual character 121 on the basis of a set 120 of data points.
  • FIG. 18 shows the set 120 of data points visualizing data points corresponding to touch events generated by a user operation on the touchscreen 17 .
  • the electronic device 10 controls motion, such as walking, running, and switching direction, of the virtual character 121 , which is the object being operated, on the basis of the angle 122 indicated by the set 120 of data points.
  • FIG. 19 is a flowchart of information processing by the angle determination unit 24 , according to the embodiment of the present invention. This flowchart starts when a touch event is generated and the angle determination unit 24 obtains a data point from the touch event.
  • step 101 the angle determination unit 24 stores the obtained data point in the data point buffer.
  • the angle determination unit 24 associates, with the data point being stored, T indicating the time, in milliseconds, elapsed after the storage thereof and D indicating the period (holding life), in milliseconds, for which the data point can be stored in the data point buffer.
  • step 102 the angle determination unit 24 outputs a start event to the state determination unit 25 .
  • step 103 the angle determination unit 24 determines whether or not a touch event has been generated.
  • Step 103 is executed after I ms from the start of the flowchart.
  • the flowchart proceeds to step 104 in the case where a touch event has been generated, and the flowchart proceeds to step 106 in the case where no touch event has been generated.
  • step 104 the angle determination unit 24 determines whether or not the generated touch event is touchend.
  • the flowchart is terminated in the case where the obtained touch event is touchend.
  • the flowchart proceeds to step 105 in the case where the obtained touch event is not touchend.
  • step 105 the angle determination unit 24 obtains a data point from the generated touch event, and stores the data point in the data point buffer. At this time, the angle determination unit 24 associates, with the data point being stored, T indicating, in milliseconds the time elapsed after the storage thereof and D indicating, in milliseconds, the period for which the data point can be stored in the data point buffer.
  • step 106 the angle determination unit 24 terminates the holding of data points for which the elapsed period T is greater than or equal to the variable D among the data points held in the data point buffer.
  • the angle determination unit 24 compares the elapsed time T with the variable D for each of the data points stored in the data point buffer, and terminates the holding of each data point for which the elapsed time is greater than or equal to the variable D.
  • step 107 the angle determination unit 24 determines whether or not the number n of the data points held in the data point buffer is greater than or equal to the value of the variable B.
  • the flowchart proceeds to step 108 in the case where the number n is greater than or equal to the variable B, and the flowchart proceeds to step 113 in the case where the number n is less than the variable B.
  • step 108 the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable on the basis of the amount of displacement of the values on the x axis and the amount of displacement of the values on the y axis in the data points held in the data point buffer. At the same time, the angle determination unit 24 determines the other axis as the axis of the dependent variable.
  • the angle determination unit 24 determines the angle of the slope of a regression line by calculating the angle of the slope by using Eq. (5). In one example, when calculating the angle of the slope by using Eq. (5), the angle determination unit 24 calculates the angle within the range of 0 to 90 degrees and 270 to 360 degrees. In the case where the y axis is determined as the axis of the independent variable in step 108 , the angle determination unit 24 calculates the angle of the slope by using Eq. (6), and subtracts the calculated angle from 90 degrees, thereby determining the angle of the slope of a regression line. In one example, when calculating the angle of the slope by using E. (6), the angle determination unit 24 calculates the angle within the range of 0 to 90 degrees and 270 to 360 degrees.
  • the angle determination unit 24 determines an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees on the basis of the displacement direction of the set of data points held in the data point buffer.
  • the angle determination unit 24 calculates the individual differences between values temporally succeeding and following each other on the x axis. The angle determination unit 24 determines the amount of rotation to be 180 degrees when the number of negative calculated differences is greater than the number of positive calculated differences, and determines the amount of rotation to be 0 degrees when the former is less than the latter.
  • the angle determination unit 24 calculates the individual differences between values temporally succeeding and following each other on the y axis.
  • the angle determination unit 24 determines the amount of rotation to be 180 degrees when the number of negative calculated differences is greater than the number of positive calculated differences, and determines the amount of rotation to be 0 degrees when the former is less than the latter.
  • the angle determination unit 24 determines an angle on the basis of the determined slope of the regression line and the determined amount of rotation. Specifically, the angle determination unit 24 determines an angle by adding the determined amount of rotation to the angle corresponding to the determined slope of the regression line. For example, in the case where the amount of rotation is 0 degrees, the angle determined by the angle determination unit 24 is the angle corresponding to the determined slope of the regression line. The angle determination unit 24 outputs an angle event together with the determined angle to the state determination unit 25 , and proceeds to step 112 .
  • step 113 the angle determination unit 24 outputs a keep event to the state determination unit 25 , and proceeds to step 112 .
  • step 112 the flowchart returns to step 103 unless terminated, for example, due to the termination of the game application.
  • the angle determination unit 24 executes the processing in steps 103 to 112 every I ms.
  • the angle determination unit 24 deletes all the data points held in the data point buffer.
  • This embodiment utilizes the characteristic of a projection-type electrostatic capacitive touchscreen, in which when the touchscreen 17 is swiped with a finger, six or more touch events are generated within an extremely short period, such as 100 ms.
  • the system architecture of the software implemented by the electronic device 10 has a three-layer structure, in which the angle determination unit 24 , the state determination unit 25 , and the application unit 26 respectively correspond to the first layer, the second layer, and the third layer, where the third layer corresponds to the game application.
  • the first layer calculates the swipe direction for a set of a plurality of generated touch events (a set of data points), in the form of a linear angle over 360 degrees.
  • the first layer Since it is possible to use touch events generated within an extremely short period, it is possible for the first layer to continue to calculate an angle at high speed. Furthermore, by obtaining an angle indicated by a set of data points held in the data point buffer just for a predefined extremely short holding period (e.g., 165 ms), it is possible for the first layer to obtain an angle of the direction intended by the user who has performed a touch operation on the touchscreen 17 , without having to use reference points.
  • a predefined extremely short holding period e.g. 165 ms
  • the electronic device 10 in this embodiment is designed on the basis of the concept that, however complex a user operation such as a swipe is, within an extremely short period, such as 100 ms, the user operation can be modeled on linear motion with constant velocity, and thus it is possible to predict and approximate an arbitrary nonlinear function ( ⁇ human operation) by continuously calculating the values of the angle of motion with constant velocity.
  • the first layer calculates an angle at a relatively low frequency is the case where the value of the variable I is set to be relatively large. In this case, even immediately after the start of generation of touch events, for example, even immediately after the user has started an operation, the first layer calculates an angle in a state where a relatively large number of data points are held in the data point buffer. As described above, by setting the value of the variable I to be relatively large, it becomes possible to calculate an angle in which a user operation is reflected relatively loosely. Meanwhile, the first layer calculates an angle at a relatively high frequency in the case where the value of the variable I is set to be relatively small.
  • the first layer calculates an angle in a state where a relatively small number of data points are held in the data point buffer. As described above, by setting the value of the variable I to be relatively small, it becomes possible to calculate an angle in which a user operation is reflected relatively immediately.
  • the first layer calculates the slope of a regression line by using the method of least squares in calculating an angle indicated by a set of data points.
  • x is fixed as the independent variable
  • y is fixed as the dependent variable
  • it is difficult to calculate the slope for example, in the case where the slope is in a direction along the y axis.
  • the first layer determines the independent variable and the dependent variable from the amount of displacement of the data points before calculating the slope of a regression line. With this configuration, it becomes possible for the first layer to stably calculate the slope of a regression line.
  • the first layer determines an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees, on the basis of the displacement direction of the set of data points, corresponding to a rough direction in which the user moves a finger on the touchscreen 17 .
  • the first layer calculates an angle corresponding to the swipe direction for a set of data points per predefined processing period, for example, per period corresponding to the frame rate of the game application as the predefined processing period.
  • the second layer determines a vector quantity (angle and magnitude) indicating the state of the slope of the virtual controller per predefined processing period by using continuously output angle information, and outputs the vector quantity to the game application.
  • the concept of reference coordinates is not involved, which makes it possible to provide higher responsiveness compared with the virtual controller with which reference coordinates are defined.
  • the concept that the current instruction coordinates approach the reference coordinates is not lacking, it is possible to switch the direction by quickly responding to the intent of the operation by the user, which is considerably advantageous.
  • a user operating a smartphone is allowed to perform operations with one hand. This makes it possible to solve the problem that it is necessary for a user to be constantly aware of reference coordinates when performing operations with the virtual joystick in the existing technology and thus there are cases where it is difficult to perform operations with one hand.
  • a virtual controller that enables quicker and more intuitive operation is realized.
  • the technology of the virtual controller provided by the electronic device 10 in this embodiment is based on a mathematical model of angle recognition from a swipe operation, the technology is applicable to a wide variety of genres, such as movement in directions over 360 degrees, acceleration and deceleration, and input of commands in a combat game or the like.
  • Another embodiment of the present invention may be a program for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention, or a computer-readable storage medium storing the program. Furthermore, another embodiment of the present invention may be a method for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention. Furthermore, another embodiment of the present invention may be a server that is capable of providing a computer with a program for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention. Furthermore, another embodiment of the present invention may be a virtual machine for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention.
  • the electronic device 10 has installed there in a game application that disposes a virtual object in a virtual space and that presents a user with a game image photographed by a virtual camera disposed in the virtual space while causing the game to proceed.
  • the electronic device 10 displays, on the touchscreen 17 , a game image obtained by photographing the virtual object, etc. disposed in the virtual space by means of the virtual camera disposed in the virtual space, and controls the virtual camera in accordance with user operations.
  • the electronic device 10 provides a virtual controller for controlling the virtual camera, which is an object being operated, in accordance with user operations.
  • the control of the virtual camera refers to controlling the motion or the visual field region of the virtual camera disposed in the virtual space.
  • FIG. 20 is an illustration showing a virtual camera 131 disposed in the virtual space, which captures an image displayed by the electronic device 10 according to the embodiment of the present invention.
  • FIG. 20 shows the position of the virtual camera 131 and a viewing direction 132 of the virtual camera 132 .
  • the viewing direction 132 is defined in the form of a three-dimensional vector with the viewpoint set at the position of the virtual camera 131 .
  • a visual field region is defined with a certain viewing angle from the virtual camera 131 into the viewing direction 132 , and a two-dimensional screen 133 , constituting a plane perpendicular to the viewing direction 132 , is defined in the visual field region.
  • the virtual object in the virtual space is projected onto the two-dimensional screen 133 , whereby a two-dimensional image is formed.
  • the input device 12 and the display device 13 have separate forms disposed at different positions.
  • the input device 12 is a touchscreen or a device having a function equivalent to the function of a projection-type electrostatic capacitive touchscreen.
  • the display device 13 it suffices for the display device 13 to be a device that displays an application screen, etc. to the user of the electronic device 10 under the control of the processor 11 .
  • the display device 13 is a liquid crystal display, a display using organic EL, or a plasma display.
  • the angle determination unit 24 calculates the slope of a regression line by using a known method other than the method of least squares. In this case, the angle determination unit 24 does not determine an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees, and thus does not determine the axis of the independent variable and the axis of the dependent variable. For example, it is possible to use an algorithm such as the Kalman filter or particle filters.
  • the angle determination unit 24 does not define the variable D, and does not terminate the holding of a data point for which it is determined that the predefined holding period has expired among the data points held in the data point buffer. In this case, the angle determination unit 24 determines an angle while referring to data points stored in a specific time band that is shifted per time defined by the variable I.
  • the angle determination unit 24 does not define the variable V. In this case, the angle determination unit 24 calculates and determines an angle by using the function aop(x, y) irrespective of the value of the variance of the independent variable.
  • the electronic device 10 is equipped with an acceleration sensor that is capable of determining the gravitational direction.
  • the angle determination unit 24 determines a weight coefficient by using information obtained from the acceleration sensor. For example, in the case where the smartphone is held in portrait orientation and thus the y axis of the touchscreen 17 is the gravitational direction, the angle determination unit 24 determines w to be 2. Meanwhile, in the case where the smartphone is held in landscape orientation and thus the x axis of the touchscreen 17 is the gravitational direction, and the game engine provided in the game application does not perform transformation between vertical and horizontal coordinates, the angle determination unit 24 determines w to be a value less than 1, such as 0.5.
  • the angle determination unit 24 when obtaining a touch event, obtains a set (x, y) of numerical values consisting of two variables, and stores the set (x, y) of numerical values consisting of the two variables in the data point buffer without associating a data point obtaining time t therewith.
  • the angle determination unit 24 can store information corresponding to a data point obtaining time t in a memory area or the like in the storage device 14 other than the data point buffer and can manage the information in association with data stored in the data point buffer.
  • the processing or operation described above may be modified freely as long as no inconsistency arises in the processing or operation, such as an inconsistency that a certain step utilizes data that may not yet be available in that step.
  • the examples described above are examples for explaining the present invention, and the present invention is not limited to those examples.
  • the present invention can be embodied in various forms as long as there is no departure from the gist thereof.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Position Input By Displaying (AREA)

Abstract

One or more embodiments of the invention is a program that is executed on an electronic device, the program causing the electronic device to execute: holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events; terminating the holding of a data point for which a predefined holding period has expired among the held data points; determining the slope of a regression line on the basis of the held data points; determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of the set of the held data points; and determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.

Description

    TECHNICAL FIELD
  • The present invention relates to programs, etc. In particular, the present invention relates to a program that is executed by an electronic device having a touchscreen, etc.
  • BACKGROUND ART
  • With recent improvements in touchscreen technology, electronic devices of the type in which user inputs are performed via a user interface on a touchscreen have become widely used. Furthermore, with games that are executed on an electronic device, instead of user inputs via a conventional physical controller, the form in which user inputs are performed via a touchscreen provided on an electronic device becoming widely used.
  • In particular, compact portable electronic devices, such as smartphones, have rapidly become popular, and a large number of games that are executed on such portable electronic devices have been released. In this situation, various technologies have been proposed concerning methods of operating a virtual object displayed on the touchscreen, such as a player character, including, for example, the technology relating to virtual pads, disclosed in Non-Patent Literature 1.
  • For example, Patent Literature 1 discloses a game device equipped with a touchscreen, as well as a program, that make it possible to set an origin in accordance with a touch operation by a user and that enables an operation simulating a joystick. The game device sets reference coordinates on the basis of coordinates at the time when detection is started in the case where a touchscreen makes a transition from a state in which a touch has not been detected to a state in which a touch has been detected, and in the case where touch detection is to be continued, sets instruction coordinates on the basis of coordinates detected after the transition. Then, the game device recognizes the direction of a vector from the reference coordinates to the instruction coordinates as the direction in which a joystick is tilted and the magnitude of the vector as the degree by which the joystick is tilted, thereby realizing a virtual joystick, whereby the operation of a virtual object is realized.
  • CITATION LIST Patent Literature {PTL 1}
  • Publication of Japanese Patent No. 3734820
  • Non-Patent Literature {NPL 1}
  • Matthias Baldauf, Peter Frohlich, Florence Adegeye, and Stefan Suette. 2015. Investigating On-Screen Gamepad Designs for Smartphone-Controlled Video Games. ACM Trans. Multimedia Comput. Commun. Appl. 12, 1s, Article 22 (October 2015), 21 pages. DOI: https://doi.org/10.1145/2808202
  • SUMMARY OF INVENTION Technical Problem
  • In the existing technology disclosed in Patent Literature 1, a user places a finger at one point on the touchscreen to cause the game device to recognize the reference coordinates, slides the finger while maintaining contact, and causes the game device to recognize the instruction coordinates with the contact position of the finger after the slide. With the thus-configured existing technology, when the user inputs a direction, it has been necessary to generate a significant distance from the reference coordinates to the instruction coordinates, which makes it difficult to realize high responsiveness. For example, in the case where the user wishes to perform an operation of considerably tilting the virtual joystick, it has been necessary to generate the magnitude of the vector from the reference coordinates to the instruction coordinates correspondingly to the degree of tilting of the considerably tilted joystick.
  • Thus, as an operation method for controlling a virtual object displayed on a touchscreen and disposed in a virtual space, for example, there is a demand for realizing a quicker and more intuitive operation method. To put it in a more generalized manner, as an operation method for controlling an object being operated, which is disposed in a virtual space and which may be operated by a user, there is a demand for realizing an operation method with improved operability.
  • The present invention has been made in order to solve the problem described above, and a main object thereof is to provide a program, etc. that makes it possible to improve operability in controlling an object being operated in a virtual space.
  • Solution to Problem
  • In order to achieve the above object, a program according to an aspect of the present invention is a program that is executed on an electronic device equipped with a touchscreen, the program being characterized by causing the electronic device to execute: a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user; a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points; a step of determining the slope of a regression line on the basis of the held data points; a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
  • Furthermore, in the present invention, preferably, in the step of determining an amount of rotation, when determining an angle for controlling the object being operated, an amount of rotation indicating whether or not the determined slope is to be rotated by 180 degrees is determined on the basis of a displacement direction of data points temporally succeeding and following each other among the held data points.
  • Furthermore, in the present invention, preferably, the program causes the electronic device to further execute: a step of determining one of the first axis and the second axis as an axis of an independent variable and determining the other axis as an axis of a dependent variable on the basis of an amount of displacement of the values on the first axis and an amount of displacement of the values on the second axis in the held data points, and in the step of determining the slope of a regression line, the slope of the regression line is determined further on the basis of the determined axis of the independent variable and axis of the dependent variable.
  • Furthermore, in the present invention, preferably, in the step of determining one of the axes, the one of the axes is determined as the axis of the independent variable on the basis of the difference between a maximum value and a minimum value of the values on the first axis and the difference between a maximum value and a minimum value of the values on the second axis.
  • Furthermore, in the present invention, preferably, in the step of determining one of the axes, the one of the axes is determined as the axis of the independent variable by comparing the magnitude of the result of applying a weight to the difference between a maximum value and a minimum value of the values on the first axis and the magnitude of the difference between a maximum value and a minimum value of the values on the second axis.
  • Furthermore, in the present invention, preferably, in the step of determining an amount of rotation, when determining an angle by which the object being controlled is controlled, an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees is determined by comparing the number of positive values and the number of negative values among the differences between values temporally succeeding and following each other on the axis of the determined independent variable in the held data points.
  • Furthermore, in the present invention, preferably, the step of determining the slope of a regression line includes: a step of determining the average of the independent variable and the average of the dependent variable in the held data points; a step of determining, by using the determined average, the deviation of the independent variable and the deviation of the dependent variable in the held data points; a step of determining, by using the determined deviation of the independent variable, the variance of the independent variable in the held data points; a step of determining, by using the determined variance of the independent variable and variance of the dependent variable, the covariance in the held data points; and a step of determining the slope of the regression line by dividing the determined covariance by the determined variance of the independent variable.
  • Furthermore, in the present invention, preferably, the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed, the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis, and in the step of determining the slope of a regression line, in the case where the determined axis of the independent variable is the second axis, the slope of the regression line is determined by subtracting, from 90 degrees, an angle corresponding to the determined slope of the regression line, determined by dividing the determined covariance by the determined variance of the independent variable.
  • Furthermore, in the present invention, preferably, the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed, and the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis.
  • Furthermore, in the present invention, preferably, in the step of determining an angle, an angle is determined per predefined processing period.
  • Furthermore, in order to achieve the above object, a program according to an aspect of the present invention is a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including the abovementioned program, the program suite being characterized in that the predefined processing period is a period corresponding to a frame rate for executing the game, and in that the program suite causes the execution of: a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and a step of controlling the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
  • Furthermore, in order to achieve the above object, a program suite according to an aspect of the present invention is a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including the abovementioned program, the program being characterized in that the predefined processing period is a period corresponding to a frame rate for executing the game, and in that the program suite causes the execution of: a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and a step of controlling a virtual camera for photographing a game image, the virtual camera serving as the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
  • Furthermore, in order to achieve the above object, an electronic device according to an aspect of the present invention is an electronic device equipped with a touchscreen, the electronic device being characterized in that: data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user, are held; the holding of a data point for which a predefined holding period has expired among the held data points is terminated; the slope of a regression line is determined on the basis of the held data points; an amount of rotation by which the determined slope of the regression line is to be rotated is determined on the basis of a displacement direction of a set of the held data points; and an angle by which the user controls an object being operated in a virtual space is determined on the basis of the determined slope of the regression line and the determined amount of rotation.
  • Furthermore, in order to achieve the above object, a method according to an aspect of the present invention is a method that is executed on an electronic device equipped with a touchscreen, the method being characterized by including: a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user; a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points; a step of determining the slope of a regression line on the basis of the held data points; a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
  • Advantageous Effects of Invention
  • The present invention makes it possible to improve operability in controlling an object being operated in a virtual space.
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a block diagram showing the hardware configuration of an electronic device according to an embodiment of the present invention.
  • FIG. 2 is a functional block diagram of the electronic device according to the embodiment of the present invention.
  • FIG. 3 is an illustration showing coordinate axes consisting of a first axis and a second axis in the embodiment.
  • FIG. 4 is an illustration for explaining an example of an angle determined by an angle determination unit and a direction corresponding to that angle.
  • FIG. 5 is an illustration showing an example of data points held in a data point buffer at time t1.
  • FIG. 6 is an illustration showing an example of data points held in the data point buffer at time t1.
  • FIG. 7 is an illustration showing a regression line obtained from the set of data points shown in FIG. 5.
  • FIG. 8 is an illustration showing an angle calculated by a function aop(x, y) by using the slope of the regression line shown in FIG. 7.
  • FIG. 9 is an illustration showing an example of data points held in the data point buffer at time t2.
  • FIG. 10 is an illustration showing an angle calculated by the function aop(x, y) by using the slope of the regression line shown in FIG. 9.
  • FIG. 11 is an illustration showing an example of data points held in the data point buffer at time t3.
  • FIG. 12 is an illustration showing an example of data points held in the data point buffer at time t3.
  • FIG. 13 is an illustration showing an angle calculated by the function aop(x, y) by using a regression line obtained from the set of data points shown in FIG. 11 and the slope of the regression line.
  • FIG. 14 is an illustration showing an example of data points held in the data point buffer at time t4.
  • FIG. 15 is an illustration showing an example of data points held in the data point buffer at time t4.
  • FIG. 16 is an illustration showing a regression line obtained from the set of data points shown in FIG. 14.
  • FIG. 17 is an illustration showing an angle calculated by the function aop(x, y) by using the slope of the regression line shown in FIG. 16.
  • FIG. 18 is an illustration showing how a virtual character is controlled on the basis of a set of data points by the electronic device according to the embodiment of the present invention.
  • FIG. 19 is a flowchart of information processing by the angle determination unit, according to the embodiment of the present invention.
  • FIG. 20 is an illustration showing a virtual camera, disposed in a virtual space, that captures an image displayed by the electronic device according to the embodiment of the present invention.
  • DESCRIPTION OF EMBODIMENTS
  • An embodiment of the present invention will be described below with reference to the drawings. The same reference signs signify the same or corresponding parts throughout the drawings unless otherwise specifically mentioned, and there are cases where the vertical to horizontal scale is shown to be different from the real scale for convenience of description. Furthermore, there are cases where descriptions that are more detailed than necessary are omitted for convenience of description. For example, there are cases where detailed descriptions of matters that are already well known and repeated descriptions of substantially the same configurations are omitted.
  • An electronic device 10 according to an embodiment of the present invention has installed therein a game application that presents a user with a virtual object disposed in a virtual space while causing the game to proceed. When the game application is executed, the electronic device 10 according to this embodiment provides a virtual controller (virtual controller) for controlling an object being operated, which is a virtual object being operated by the user in the virtual space, in response to an operation by the user. The virtual space is defined by the game application, and may be either a two-dimensional space or a three-dimensional space. For example, the virtual object is a character or an item disposed in the virtual space. For example, controlling the object being operated means controlling the character or the item disposed in the virtual space.
  • For convenience of description, in this embodiment it is assumed that the electronic device 10 has the above-described game application installed therein; however, there is no limitation to the above-described game application. It suffices for the electronic device 10 to have installed therein a game application that can control an object being operated in response to an operation by a user. For example, instead of or in addition to the game application, the electronic device 10 may have installed therein an input support application or a simulation application that brings about an action of an object being operated in response to an operation by a user. The virtual controller provided by the electronic device 10 in this embodiment can be used for the type of user input requiring a direction and a magnitude. In the following description, the term “application” refers to application programs in general, and may refer to apps that are installed on a smartphone or a tablet.
  • FIG. 1 is a block diagram showing the hardware configuration of the electronic device 10 according to the embodiment of the present invention. The electronic device 10 includes a processor 11, an input device 12, a display device 13, a storage device 14, and a communication device 15. These constituent devices are connected via a bus 16. It is assumed that interfaces are interposed as needed between the bus 16 and the individual constituent devices. In this embodiment, the electronic device 10 is a smartphone. Alternatively, the electronic device 10 may be a terminal such as a computer equipped with a contact-type input device, like a tablet computer or a touchpad, as long as the terminal includes the configuration described above.
  • The processor 11 controls the overall operation of the electronic device 10. For example, the processor 11 is a CPU. Alternatively, an electronic circuit such as an MPU may be used as the processor 11. The processor 11 executes various kinds of processing by loading programs and data stored in the storage device 14 and executing the programs. In one example, the processor 11 is constituted of a plurality of processors.
  • The input device 12 is a user interface for accepting inputs to the electronic device 10 from the user; for example, the input device 12 is a touchscreen, a touchpad, a keyboard, or a mouse. The display device (display) 13 displays application screens, etc. to the user of the electronic device 10 under the control of the processor 11. In this embodiment, since the electronic device 10 is a smartphone, the electronic device 10 includes a touchscreen 17 as the input device 12, and the touchscreen 17 also functions as the display device 13; that is, the input device 12 and the display device 13 have an integrated structure. Although the touchscreen 17 in this embodiment is a projection-type capacitive touchscreen, a device having an equivalent function, if any, may be adopted.
  • The storage device 14 is a storage device included in an ordinary smartphone, including a RAM, which is a volatile memory, and a ROM, which is a non-volatile memory. The storage device 14 may also include an external memory. The storage device 14 stores various kinds of programs, including a game application. For example, the storage unit 14 stores an operating system (OS), middleware, application programs, various kinds of data that may be referred to when these programs are executed, etc.
  • In one example, the storage device 14 includes a main storage device and an auxiliary storage device. The main storage device is a volatile storage medium that allows high-speed reading and writing of information, and is used as a storage area and a work area when the processor 11 processes information. The auxiliary storage device stores various programs and data that are used by the processor 11 when the individual programs are executed. The auxiliary storage device is, for example, a hard disk device; however, the auxiliary storage device may be any type of non-volatile storage or non-volatile memory, which may be of the removable type, that is capable of storing information.
  • The communication device 15 sends data to and receives data from other computers, such as a server, via a network. For example, the communication device 15 performs wireless communication, such as mobile communication or wireless LAN communication, to connect to the network 2. In one example, the electronic device 10 downloads a program from a server by means of the communication device 15 and stores the program in the storage device 14. Alternatively, the communication device 15 may perform known wired communication. In the case where data is neither sent to nor received from other computers, the electronic device 10 need not include the communication device 15.
  • FIG. 2 is a functional block diagram of the electronic device 10 according to the embodiment of the present invention. The electronic device 10 includes an input unit 21, a display unit 22, and a control unit 23. The control unit 23 includes an angle determination unit 24, a state determination unit 25, and an application unit 26. In this embodiment, these functions are realized by the processor 11 executing a program. For example, the program that is executed is a program stored in the storage device 14 or received via the communication device 15. Since various kinds of functions are realized by loading a program, as described above, a portion or the entirety of one part (function) may be provided in another part. Alternatively, these functions may be realized by means of hardware by configuring electronic circuits or the like for realizing the individual functions in part or in entirety.
  • The input unit 21 is configured by using the input device 12, and accepts inputs to the electronic device 10 from the user. In this embodiment, the input unit 21 accepts a touch operation performed on the touchscreen 17 by the user and generates touch events, for which a touch detection function generally provided in a smartphone having the touchscreen 17 can be used.
  • The display unit 22 displays a game application screen on the display device 13 so as to display a screen in accordance with user operations.
  • The control unit 23 realizes a virtual controller. In this embodiment, an architecture having a three-layer structure is adopted for the control unit 23, and the angle determination unit 24, the state determination unit 25, and the application unit 26 correspond to the individual layers. For example, the control unit 23 is realized by the processor 11 executing a program suite constituted of programs corresponding to the individual layers.
  • The role of the angle determination unit 24 in the virtual controller realized by the control unit 23 corresponds to a sensor IC chip inside the controller in the case of a physical controller. The angle determination unit 24 determines an angle for controlling an object being operated by the user in the virtual space by mainly using touch events generated by a touch operation on the touchscreen 17 by the user, and forwards the angle to the state determination unit 25.
  • The role of the state determination unit 25 corresponds to the entire controller that is operated in the case of a physical controller. The state determination unit 25 determines a vector (angle and magnitude) corresponding to the touch operation performed on the touchscreen 17 by the user by mainly using the angle forwarded from the angle determination unit 24.
  • The application unit 26 corresponds to a specific game application that implements actions, etc. in a game. In this game application, similarly to game applications in general, a frame rate is defined, and for example, the main loop of a main program is executed per period corresponding to the frame rate. Generally, the frame rate is 30 fps (frames per second) or 60 fps.
  • The angle determination unit 24 determines an angle that is required when the user controls an object being operated in the virtual space. In one preferred example, the angle determination unit 24 is a mathematical function library for calculating an angle in real time, and is a software module implementing a statistical processing algorithm for calculating an angle from a sequence of touch events during a short period. The sequence of touch events corresponds to finger motion on the touchscreen 17.
  • The storage device 14 includes a data point buffer. The data point buffer is a buffer that enables the angle determination unit 24 to hold data points obtained on the basis of touch events and indicated by values on a first axis and values on a second axis.
  • The angle determination unit holds, in the data point buffer, data points obtained on the basis of touch events generated by user operations on the touchscreen 17 and indicated by values on the first axis and values on the second axis. Here, a touch event occurs when the user places a finger in contact with the touchscreen 17 (touchstart), when the user moves the finger while keeping the finger in contact with the touchscreen (touchmove), when the user removes the finger from the touchscreen 17 (touchend), etc. The angle determination unit 24 obtains a touch event when a touch event occurs. When obtaining a touch event, the angle determination unit 24 obtains a set of numerical values (x, y) consisting of two variables and corresponding to a position at which the electrostatic capacitance on the touchscreen 17 has changed, and also obtains time t at which (x, y) was obtained, and stores a set of numerical values (x, y, t) consisting of three variables in the data point buffer. The data of the set of numerical values consisting of two variables is obtained by the angle determination unit 24 in association with a touch event, and corresponds to data points indicated by values on a value on the first axis and a value on the second axis. t is a value representing a data point obtaining time, which is the time at which (x, y) was obtained, and is stored in the data point buffer in association with (x, y), as described above. In one example, t is an integer value called the UNIX (registered trademark) time or a character sequence such as “2017/07/14 15:48:43.444”. Hereinafter, unless otherwise specifically mentioned, the holding (or termination of holding) of a data point by the angle determination unit 24 includes the holding (or termination of holding) of a data point obtaining time t associated with the data point.
  • In this embodiment, for convenience of description, the first axis and the second axis are defined as follows. FIG. 3 is an illustration showing coordinate axes consisting of the first axis and the second axis in this embodiment. The first axis is an axis indicating the widthwise direction of the directions substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed; i.e., the first axis is a horizontal axis (x axis) substantially parallel to the shorter sides of the touchscreen 17. The second axis is an axis perpendicular to the first axis and indicating the lengthwise direction of the directions substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed; i.e., the second axis is a vertical axis (y axis) substantially parallel to the longer sides of the touchscreen 17. As described above, the first axis and the second axis are substantially parallel to the directions in which the sensors of the touchscreen 17 are arrayed, and a position on the touchscreen 17 is expressed in the form of coordinates (x, y) along the two axes. Thus, in this embodiment, the coordinates (x, y) of a data point correspond to a position on the touchscreen 17. In this embodiment, the angle determination unit 24 holds the coordinates (x, y) as a data point in the data point buffer. The coordinate setting shown in FIG. 3 is just an example, and coordinates may be set differently from the above example depending on the sensor array of the touchscreen 17 or the program implemented by the electronic device 10. The sensors are, for example, electrodes.
  • FIG. 4 is an illustration for explaining an example of an angle determined by the angle determination unit 24 and a direction corresponding to the angle in the case where coordinate axes are defined as shown in FIG. 3. An angle 31 in the first quadrant corresponds to a direction 32, an angle 33 in the second quadrant corresponds to a direction 34, and an angle 35 in the fourth quadrant corresponds to a direction 36. Here, although the direction 34 and the direction 36 have the same slope, since the direction 34 and the direction 36 are opposite to each other, it is possible to confirm that the angle 33 and the angle 35 differ by 180 degrees.
  • The angle determination unit 24 terminates the holding of a data point for which a predefined holding period has expired among the data points held in the data point buffer. For example, when terminating the holding of a data point, the angle determination unit 24 may delete the data, invalidate the data, or delete the data, as appropriate, by associating a flag indicating that the holding has been terminated with the data. The angle determination unit 24, defines a variable D specifying, in milliseconds, the life of the data points stored in the data point buffer. The period specified by the variable D corresponds to the predefined holding period. However, the value of the variable D is not limited to milliseconds.
  • For example, after storing one data point in the data point buffer, the angle determination unit 24 monitors the time elapsed since the storage of the data point and continuously compares the time elapsed with the variable D. When the monitored time elapsed for the data point exceeds the variable D, the angle determination unit 24 terminates the holding of the data point in the data point buffer. In this manner, the angle determination unit 24 manages the life of every data point held in the data point buffer. At this time, the angle determination unit 24 can calculate the elapsed time by using the data point obtaining time t. It is assumed that the case where it is determined that the time elapsed exceeds the variable D may mean the case where it is determined that the time elapsed is greater than or equal to the variable D. Furthermore, the unit of the time elapsed for a data point, managed by the angle determination unit 24, should preferably be the same as that of the variable D.
  • For example, when 165 is set in the variable D, the angle determination unit 24 holds a data point stored in the data point buffer for 165 milliseconds, and after 165 milliseconds has elapsed, terminates the holding of the data point in the data point buffer.
  • The angle determination unit 24 periodically determines whether or not it is possible to calculate an angle. In the case where it is possible to calculate an angle, the angle determination unit 24 calculates an angle indicated by the set of data points held in the data point buffer by using the data points, and determines the angle as an angle by which the user controls the object being operated in the virtual space. By obtaining the angle indicated by the set of data points, as described above, the angle determination unit 24 can obtain an angle of the direction intended by the user who has performed a touch operation on the touchscreen 17.
  • The angle determination unit 24 outputs the determined angle to the state determination unit 25. When outputting the determined angle to the state determination unit 25, for example, the angle determination unit 24 outputs the determined angle together with information indicating an angle event. Instead of direct output to the state determination unit 25, the angle determination unit 24 may store the determined angle and the information indicating an angle event in a memory area in the storage unit 14 referred to by the state determination unit 25.
  • When the obtained touch event is touchstart, the angle determination unit 24 outputs a start event to the state determination unit 25, and does not calculate an angle. When the obtained touch event is touchend, the angle determination unit 24 outputs a stop event to the state determination unit 25, and does not calculate an angle.
  • The angle determination unit 24 defines a variable B, and calculates and determines an angle by using the data points held in the data point buffer in the case where the number of the data points is greater than or equal to the value of the variable B. In the case where the number of the data points held in the data point buffer is less than the value of the variable B, in which case it is not possible to calculate an angle indicated by the set of data points with at least a certain level of accuracy, the angle determination unit 24 outputs a keep event to the state determination unit 25, and does not calculate an angle. Since it is generally preferable that three or more data points are available when obtaining the slope of a regression line, the variable B should preferably be set to be 3 or greater. In this embodiment, the variable B is set to be 3.
  • The angle determination unit 24 defines a variable I specifying, in milliseconds, the time interval for determining whether or not the angle determination unit 24 can calculate an angle. In the case where the situation where the angle determination unit 24 cannot calculate an angle continues, like a situation where touch events continuously occur, the interval serves as a time interval for the angle determination unit 24 to calculate an angle. However, the value of the variable I is not limited to milliseconds.
  • The angle determination unit 24 determines an angle per predefined processing period by determining whether or not it is possible to calculate an angle per predefined processing period, as described above. In one preferred example, the predefined processing period is a period corresponding to the frame rate for executing the game. In the case where the frame rate is 30 fps (30 Hz), the variable I is set to be 33. When 33 is set in the variable 1, the angle determination unit 24 determines whether or not it is possible to calculate an angle every 33 milliseconds. In the case where it is possible to calculate an angle, the angle determination unit 24 determines a calculated angle and outputs the determined angle together with an angle event to the state determination unit 25. In the case where it is not possible to calculate an angle, the angle determination unit 24 outputs one of a start event, a stop event, and a keep event to the state determination unit 25. The angle event may include information indicating an angle, and in this case, the angle determination unit 24 outputs an angle event to the state determination unit 25 in the case where it is possible to calculate an angle.
  • In calculating an angle indicated by the set of data points held in the data point buffer, the angle determination unit 24 determines the slope of a regression line on the basis of the data points held in the data point buffer. Here, the number of data points referred to by the angle determination unit 24 is greater than or equal to 3. In determining the slope of a regression line, the angle determination unit 24 determines one of the x axis and the y axis as the axis of an independent variable on the basis of the amount of displacement in the values on the x axis and the amount of displacement in the values on the y axis in the data points held in the data point buffer. At the same time, the angle determination unit 24 determines the other axis as the axis of a dependent variable.
  • In this embodiment, the angle determination unit 24 calculates the slope of a regression line by using the method of least squares. The method of obtaining the slope of a regression line by using the method of least squares is known; in this method, the degree of correlation between an independent variable and a dependent variable is obtained in the form of the slope. The method of least squares is generally applied by considering the x axis as the axis of an independent variable and the y axis as the axis of a dependent variable. However, in the case where the value on the x axis is fixed and only the value on the y axis changes, the dependent variable has no dependency on the independent variable. Furthermore, in the case where a slope along the y axis is determined, the determined slope sometimes changes from a large negative value to a large positive value across the y axis, which makes it difficult to stably find the slope. Thus, the angle determination unit 24 determines which of the value on the x axis and the value on the y axis is suitable as an independent variable, and determines the slope of a regression line by using the method of least squares on the basis of the axis of the independent variable and the axis of the dependent variable thus determined.
  • The angle determination unit 24 determines the slope a of a regression line y=ax+b when the axis of the independent variable is the x axis, and determines the slope c of a regression line x=cy+d when the axis of the independent variable is the y axis. Specifically, the angle determination unit 24 obtains the slope of a regression line through (a) to (e) below.
  • (a) The angle determination unit 24 calculates (determines) the average of the independent variable and the average of the dependent variable in the data points held in the data point buffer.
    (b) By using the averages calculated in (a), the angle determination unit 24 calculates (determines) the deviation of the independent variable and the deviation of the dependent variable in the data points held in the data point buffer.
    (c) By using the deviation of the independent variable, determined in (b), the angle determination unit 24, calculates (determines) the variance of the independent variable in the data points held in the data point buffer.
    (d) The angle determination unit 24, by using the deviation of the independent variable and the deviation of the independent variable, determined in (b), calculates (determines) the covariance in the data points held in the data point buffer.
    (e) The angle determination unit 24 calculates (determines) the slope of the regression line by dividing the covariance, determined in (d), by the variance of the independent variable, determined in (c).
  • In one example, the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable and determines the other axis as the axis of the dependent variable on the basis of the difference between the maximum value and the minimum value of the values on the x axis and the difference between the maximum value and the minimum value of the values on the y axis in the data points held in the data point buffer. Preferably, the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable and determines the other axis as the axis of the dependent variable on the basis of the difference between the maximum value and the minimum value of the values on the x axis, the difference being weighted by using a weight coefficient, and the difference between the maximum value and the minimum value of the values on the y axis in the data points held in the data point buffer.
  • The angle determination unit 24 calculates (determines) the slope of a regression line by using the method of least squares, as described above. Preferably, the calculation of the slope of a regression line by the angle determination unit 24 means the calculation of the angle of the slope of a regression line by the angle determination unit 24. Here, since the calculated slope of a regression line does not have a positive or negative direction, when calculating the slope of a regression line by using the method of least squares, the angle determination unit 24 calculates the slope, for example, within the range of 0 to 90 degrees and 270 to 360 degrees. Thus, for example, the slope of a regression line is calculated to be 45 degrees, whether the angle indicated by the set of data points is 45 degrees or 225 degrees. Thus, after determining the slope of a regression line, on the basis of the displacement direction of the set of data points held in the data point buffer, the angle determination unit 24 determines the amount of rotation by which the determined slope of the regression line is to be rotated. Specifically, after determining the slope of a regression line, when determining an angle by which the user controls the object being controlled, the angle determination unit 24 determines an amount of rotation indicating whether or not (the angle of) the determined slope of the regression line is to be rotated by 180 degrees. Here, the displacement direction of the set of data points indicates a direction in which the data points become displaced with time, and corresponds, for example, to a rough direction in which the user moves a finger on the touchscreen 17.
  • In one example, the angle determination unit 24 determines the amount of rotation on the basis of the displacement direction of data points temporally succeeding and following each other among the data points held in the data point buffer. Preferably, the angle determination unit 24 determines the amount of rotation by comparing the number of positive differences and the number of negative differences between the values temporally succeeding and following each other on the axis of the determined independent variable in the data points held in the data point buffer. The amount of rotation that is determined is an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees.
  • The angle determination unit 24 stores data points so as to sequentially hold the data points in order from data points stored earlier in the data point buffer. Alternatively, when storing data points in the data point buffer, the angle determination unit 24 stores identification information that enables identification of the order of storage in association with the data points. As described above, the angle determination unit 24 holds data points in the data point buffer in such a manner that the order of storage can be identified, i.e., in such a manner that the temporal order of the stored data points can be identified. At this time, the angle determination unit 24 can use the values of data point obtaining times t.
  • The angle determination unit 24 calculates an angle indicated by the set of data points on the basis of the determined slope of the regression line and the determined amount of rotation, and determines the angle as an angle by which the user controls the object being operated in the virtual space.
  • In this embodiment, the determination of the slope of a regression line, the determination of an amount of rotation, and the determination of an angle, performed by the angle determination unit 24 described above, are realized by using an aop(x, y) function expressed in Eq. (1). The function aop(x, y) calculates an angle in the form of a real value in the range of 0 to 360 degrees. It is assumed that the data point buffer holds n data points P(x, y) when the function aop(x, y) calculates an angle. Furthermore, the value on the x axis and the value on the y axis of each of the n data points Pk (k=1 to n) are expressed by Pk(xk, yk), and it is assumed that the order of storage, in order from earlier storage, is P1(x1, y1), P2(x2, y2), . . . , Pn (xn, yn).
  • aop ( x , y ) = { 180 atan ( i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( x i - x _ ) 2 ) π , rotate ( x , y ) left ( x ) - 180 atan ( i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( y i - y _ ) 2 ) π + 90 , rotate ( x , y ) down ( y ) - 180 atan ( i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( y i - y _ ) 2 ) π + 270 , rotate ( x , y ) down ( y ) 180 atan ( i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( x i - x _ ) 2 ) π + 180 , rotate ( x , y ) left ( x ) ( 1 )
  • For the purpose of case classification, the function aop(x, y) uses a function rotate(x, y), a function left(x), and a function down(y). First, the function aop(x, y) determines either one of e and y as an independent variable by using the function rotate(x, y).
  • The function rotate(x, y) is defined by Eq. (2).

  • rotate(x,y)=w·(max(x)−min(x))<(max(y)−min(y))  (2)
  • The function rotate(x, y) determines whether or not the n data points P(x, y) are displaced mainly in the y axis direction, and returns, for example, a true or false value. The case where the n data points P(x, y) are displaced mainly in the y axis direction refers to, for example, the case where a finger of the user has been moved on the touchscreen 17 mainly in the vertical direction. In this manner, the function rotate(x, y) determines whether the n data points P(x, y) are displaced mainly in the x axis (horizontal) direction or mainly in the y axis (vertical) direction, thereby determining which of the value on the x axis and the value on the y axis is suitable as an independent variable.
  • In Eq. (2), (max(x)−min(x)) signifies the absolute value of the difference between the maximum value and the minimum value among the x values (x1, x2, . . . xn) of the n data points P, which indicates the amount of displacement of the n data points P in the x axis direction. Similarly, (max(y)−min(y)) signifies the absolute value of the difference between the maximum value and the minimum value among the y values (y1, y2, . . . yn) of the n data points P, which indicates the amount of displacement of the n data points P in the y axis direction. A variable w is a weight coefficient for applying a weight to (max (x)−min(x)).
  • The function rotate(x, y) satisfies the inequality when (max(y)−min(y)) is greater than the product of (max(x)−min(x)) and the variable w, in which case the function aop(x, y) performs coordinate transformation. In this case, the function aop(x, y) determines an amount of rotation by considering the y axis as the axis of the independent variable, considering the x axis as the axis of the dependent variable, and further using the function down(y). Meanwhile, the function rotate(x, y) does not satisfy the inequality when (max(y)−min(y)) is less than or equal to the product of (max(x)−min(x)) and the variable w, in which case the function aop(x, y) does not perform coordinate transformation. In this case, the function aop(x, y) determines an amount of rotation by considering the x axis as the axis of the independent variable, considering the y axis as the axis of the dependent variable, and further using the function left(x).
  • In the case where the function rotate(x, y) is not weighted with the variable w, the function aop(x, y) results in performing coordinate transformation if the amount of displacement in the y axis direction is slightly greater than the amount of displacement in the x axis direction. By applying a weight with the variable w, it becomes possible with the function aop(x, y) to perform coordinate transformation in the case where the n data points P(x, y) have a slope more approximate to the y axis.
  • For example, since a smartphone is usually used in portrait orientation, the y axis of the touchscreen 17 shown in FIG. 17 usually becomes substantially parallel to the gravitational direction. Considering that the dimension of the touchscreen 17 in the y axis direction is greater than that in the x axis direction, the variable w should preferably be set to be a value greater than 1. As an example, the variable w should preferably be set to be 2. In another example, in the case where the smartphone is held in landscape orientation and a game engine provided in the game app does not perform transformation between vertical and horizontal coordinates, the angle determination unit 24 sets w to be a value less than 1, such as 0.5.
  • The function left(x) is expressed by Eq. (3)

  • left(x)=|{x′|x′>0}|>|{x′|x′ i<0}|,x′←[x 1 −x 0 ,x 2 −x 1 , . . . ,x n −x n-1]  (3)
  • In the case where the function rotate(x, y) does not satisfy the inequality, the function left(x) determines whether or not the displacement direction of the n data points P(x, y) is the −x axis direction (leftward direction), and returns, for example, a true or false value. Specifically, the function left(x) calculates the differences (x2−x1, x3−x2, . . . xn−xn-1) of values temporally succeeding and following each other in the x values (x1, x2, . . . xn) of the n data points P. The function left(x) determines whether or not the displacement direction of the n data points P(x, y) is the −x axis direction (leftward direction) by determining whether or not the number of negative differences is greater than the number of positive differences. In this manner, the function left(x) determines whether the displacement direction of the n data points P(x, y) is the −x axis direction (leftward direction) or the +x axis direction (rightward direction), thereby determining an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees. For example, the function aop(x, y) determines the amount of rotation to be 180 degrees in the case where the function left(x) is true, and the function aop(x, y) determines the amount of rotation to be 0 degrees in the case where the function left(x) is false.
  • The function down(y) is expressed by Eq. (4).

  • down(y)=|{y′|y′ i>0}|<|{y′|y′ i<0}|,y′←[y−y 0 ,y 2 −y 1 , . . . ,y n −y n-1]
  • In the case where the function rotate(x, y) satisfies the inequality, the function down(y) determines whether or not the displacement direction of the n data points P(x, y) is the −y axis direction (downward direction), and returns, for example, a true or false value. Specifically, the function down(y) calculates the differences (y2−y1, y3−y2, . . . yn−yn-1) of values temporally succeeding and following each other in the y values (y1, y2, . . . yn) of the n data points P. The function down(y) determines whether or not the displacement direction of the n data points P(x, y) is the −y axis direction (downward direction) by determining whether or not the number of negative differences is greater than the number of positive differences. In this manner, the function down(y) determines whether the displacement direction of the n data points P(x, y) is the −y axis direction (downward direction) or the +y axis direction (upward direction), thereby determining an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees. For example, the function aop(x, y) determines the amount of rotation to be 180 degrees in the case where the function down(y) is true, and the function aop(x, y) determines the amount of rotation to be 0 degrees in the case where the function down(y) is false.
  • After determining the independent variable (the axis of the independent variable) and the amount of rotation in the manner described above, the function aop(x, y) calculates the slope by using the method of least squares. In the case where x is considered as the independent variable and y is considered as the dependent variable, the function aop(x, y) calculates the slope a of the regression line y=ax+b according to Eq. (5).
  • a = i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( x i - x _ ) 2 ( 5 )
  • Here,

  • x
  • signifies the average of x (independent variable),

  • y
  • signifies the average of y (dependent variable),
  • n i = 1 n ( x i - x _ ) ( y i - y _ )
  • signifies the covariance of x (independent variable) and y (dependent variable), and
  • i = 1 n ( x i - x _ ) 2
  • signifies the variance of x (independent variable).
  • In the case where y is considered as the independent variable and x is considered as the dependent variable, the function aop(x, y) calculates the slope c of the regression line x=cy+d according to Eq. (6).
  • c = i = 1 n ( x i - x _ ) ( y i - y _ ) i = 1 n ( y i - y _ ) 2 ( 6 )
  • Here,

  • x
  • signifies the average of x (dependent variable),

  • y
  • signifies the average of y (independent variable),
  • n i = 1 n ( x i - x _ ) ( y i - y _ )
  • signifies the covariance of x (dependent variable) and y (independent variable), and
  • i = 1 n ( y i - y _ ) 2
  • signifies the variance of y (independent variable).
  • The angle determination unit 24 defines a variable V, and in the case where the variance of the independent variable is greater than or equal to V, calculates and determines an angle by using the function aop(x, y). In the case where the variance of the independent variable is less than V, the angle determination unit 24 outputs a keep event to the state determination unit 25, and does not calculate an angle. The case where the variance of the independent variable is less than V indicates that the n data points P(x, y) are locally concentrated. Thus, by defining the variable V, it becomes possible for the angle determination unit 24 to calculate an angle stably while disregarding finger motion that is too small. The variable V is preferably set to be 0.7
  • In the case where the function rotate(x, y) is false and the function left(x) is false, the function aop(x, y) directly calculates an angle obtained from the slope of the regression line. In the case where the function rotate(x, y) is false and the function left(x) is true, the function aop(x, y) calculates an angle obtained by adding 180 degrees to an angle obtained from the slope of the regression line. In the case where the function rotate(x, y) is true and the function down(y) is false, the function aop(x, y) calculates an angle by subtracting an angle obtained from the slope of the regression line from 90 degrees. In the case where the function rotate(x, y) is true and the function down(y) is true, the function aop(x, y) calculates an angle by adding 180 degrees to an angle obtained by subtracting an angle obtained from the slope of the regression line from 90 degrees.
  • The state determination unit 25 determines an angle and a magnitude per predefined processing period on the basis of the angle defined by the angle determination unit 24 per predefined processing period. The state determination unit 25 outputs the determined angle and magnitude to the application unit 26. Instead of direct output to the application unit 26, the state determination unit 25 may store information indicating the determined angle and magnitude in a memory area referred to by the application unit 26. In one preferred example, the state determination unit 25 is a software module that converts a sequence of successively calculated angles into a function of a controller. This module is implemented as a state management model (state machine) using the angle and the time, without using touched coordinates at all.
  • Upon receiving the information output from the angle determination unit 24 per predefined processing period, the state determination unit 25 outputs the state of the slope of the virtual controller per predefined processing period in accordance with the received event. The virtual controller realizes a physical analog controller by means of software, and the slope of the virtual controller corresponds to the slope of a joystick or level of a physical controller. The state determination unit 25 outputs the state of the slope of the virtual controller in the form of a vector, i.e., as a direction (angle) and a magnitude. The magnitude may be acceleration information or speed information. The state of the slope of the virtual controller corresponds to a quantity input to a physical analog controller.
  • As described above, the state determination unit 25 can be considered as providing a state transition model for interpreting the meaning of a generated touch event. The state determination unit 25 manages the slope of the virtual controller as a state, and upon receiving the information output from the angle determination unit 24 per predefined processing period, causes a state transition to occur per predefined processing period.
  • In one example, the state determination unit 25 accumulates the information about the angle output from the angle determination unit 24, and calculates an output to the application unit 26 in accordance with time. For example, in the case where the angle determination unit 24 outputs 45 degrees per 33 ms continuously over 165 ms, the state determination unit 25 determines acceleration information in the direction of 45 degrees for 165 ms. For example, in the case where the same angle is continuously output from the angle determination unit 24, the state determination unit 25 outputs information indicating acceleration in the direction of that angle, and in the case where an angle opposite to the abovementioned angle by 180 degrees is subsequently input from the angle determination unit 24, the state determination unit 25 outputs information indicating deceleration. With this configuration, it becomes possible to obtain a quantity input to the controller in the form of a period corresponding to the number of state transitions, without having to directly use actual touched coordinates.
  • In one example, upon receiving the information output from the angle determination unit 24, the state determination unit 25 sets the state of the slope of the virtual controller in accordance with the received event, and outputs a movement event or an acceleration event to the game application together with the information indicating the angle and the magnitude. Specifically, the state determination unit 25 can implement acceleration processing simulating a physical analog controller by using a predetermined programming language.
  • In one preferred example, the state determination unit 25 holds the state of the slope of the virtual controller in the form of a vector, and outputs the held vector per predefined processing period. When a start event is received from the angle determination unit 24, the state determination unit 25 starts state management. When an angle event is initially received from the angle determination unit 24, the state determination unit 25 starts holding a vector consisting of the received angle and the magnitude of a single state transition, and outputs the vector. When another angle event is further received from the angle determination unit 24, the state determination unit 25 adds a vector consisting of the received angle and the magnitude of a single state transition to the internally held vector, and holds and outputs the resulting vector. The state determination unit 25 normalizes the length of the output vector to 1.0 in the case where the length exceeds 1.0.
  • For example, when angle information indicating 45 degrees is received once from the angle determination unit 24, the state determination unit 25 outputs and holds a vector consisting of an angle of 45 degrees and a magnitude of 0.2. When angle information indicating 45 degrees is received four further times from the angle determination unit 24, the state determination unit 25 holds and outputs a vector consisting of an angle of 45 degrees and a magnitude of 1.0. Also, when angle information indicating 45 degrees is received another time from the angle determination unit 24, the state determination unit 25 holds and outputs a vector consisting of an angle of 45 degrees and a magnitude of 1.0. When angle information indicating 30 degrees is received, the state determination unit 25 adds together a vector consisting of an angle of 30 degrees and a magnitude of 0.2 and the vector consisting of the angle of 45 degrees and the magnitude of 1.0, and holds and outputs a vector consisting of the resulting angle and the magnitude of 1.0. Also in this case, since the magnitude of the vector exceeds 1.0, the state determination unit 25 normalizes the magnitude to 1.0 when outputting the vector.
  • When a keep event is received from the angle determination unit 24, the state determination unit 25 continues holding the internally held vector and outputs the vector. When a stop event is received from the angle determination unit 24, the state determination unit 25 quits holding the internally held vector, and either outputs a zero vector or outputs the stop event.
  • With this configuration, it becomes possible to realize operations similar to those via a physical analog controller on a smartphone.
  • The application unit 26 has the function of an app activated on the electronic device 10, and provides a service by the app. In one preferred example, the application unit 26 converts the vector output from the state determination unit 25 into motion of a specific virtual character or the like, which is a function implemented in ordinary game apps. In the case where the app is a Web app, a server that communicates with the electronic device 10 has the application unit 26 in part or in entirety, and the app sends data to and receives data from the server.
  • An example of the angle calculation by the angle determination unit 24 using the function aop(x, y) expressed in Eq. 1 will be described below. In the following example, it is assumed that the variable w is 2 and that the variance of the independent variable is greater than or equal to V, and it is assumed that t1 to t4 each signify an arbitrary time at which the angle determination unit 24 calculates an angle.
  • FIG. 5 is an illustration showing an example of the data points held in the data point buffer at time t1. A set 40 of the data points held in the data point buffer consists of six data points P41 to P46. Furthermore, the coordinates (x, y) of the x coordinate values and the y coordinate values of P41 to P46 are individually P41(x 41, y41), P42(x 42, y42), . . . P46(x 46, y46), and it is assumed that the data points were stored in the data point buffer in the order P41, P42, . . . , P46.
  • Initially, the function aop(x, y) compares abs(max(x)−min(x)) and abs(max(y)−min(y)) with each other by using the function rotate(x, y). It is understood from FIG. 6 that the difference x46−x41 (=Δx) between the x coordinate values of P46 and P41 corresponds to abs(max(x)−min(x)) and that the difference y45−y41 (=Δy) between the y coordinate values of P45 and P41 corresponds to abs(max(y)−min(y)). According to FIG. 5, Δx>Δy. Thus, the function rotate(x, y) does not satisfy the inequality and thus returns a false value.
  • Then, the function aop(x, y) calculates the differences (x42−x41, x43−x42, . . . x46−x45) between values temporally succeeding and following each other by using the function left(x). According to FIG. 5, all the differences are positive. Thus, the function left(x) does not satisfy the inequality and thus returns a false value.
  • In the above example, the function rotate(x, y) is false, and the function left(x) is false. Thus, the function aop(x, y) obtains a regression line 81, shown in FIG. 7, from the set 40 of data points, and calculates an angle 91 from the slope of the regression line 81, as shown in FIG. 8.
  • FIG. 9 is an illustration showing an example of the data points held in the data point buffer at time t2. A set 60 of the data points held in the data point buffer consists of six data points P61 to P66. The x coordinate values and the y coordinate values of P61, P62, . . . P66 are respectively the same as the x coordinate values and the y coordinate values of P46, P45, . . . , P41, and it is assumed that the data points were stored in the order P61, P62, . . . , P66.
  • Similarly to the case of the example in FIG. 5, the function rotate(x, y) does not satisfy the inequality and thus returns a false value. Then, the function aop(x, y) calculates the differences between values temporally succeeding and following each other by using the function left(x). According to FIG. 9, all the differences are negative. Thus, the function left(x) satisfies the inequality and returns a true value.
  • In the above example, the function rotate(x, y) is false, and the function left(x) is true. Thus, the function aop(x, y) obtains a regression line 82, shown in FIG. 9, from the set 60 of data points, and calculates an angle 92 b obtained by adding 180 degrees to an angle 92 a calculated from the slope of the regression line 82, as shown in FIG. 10.
  • FIG. 11 is an illustration showing an example of the data points held in the data point buffer at time t3. A set 70 of the data points held in the data point buffer consists of six data points P71 to P76. Furthermore, the x coordinate values and the y coordinate values of P71 to P76 are individually P71(x 71, y71), P72(x 72, y72), . . . P76(x 76, y76), and it is assumed that the data points were stored in the data point buffer in the order P71, P72, . . . , P76.
  • It is understood from FIG. 12 that the difference x75−x71 (=Δx) between the x coordinate values of P75 and P71 corresponds to abs(max(x)−min(x)) and that the difference y76−y71 (=Δy) between the y coordinate values of P76 and P71 corresponds to abs(max(y)−min(y)). According to FIG. 11, 2×Δx<Δy. Thus, the function rotate(x, y) satisfies the inequality and thus returns a true value.
  • Then, the function aop(x, y) calculates the differences (y72−y71, y73−y72, . . . y76−y75) between values temporally succeeding and following each other by using the function down(y). According to FIG. 11, all the differences are positive. Thus, the function down(y) does not satisfy the inequality and thus returns a false value.
  • In the above example, the function rotate(x, y) is true, and the function down(y) is false. Thus, the function aop(x, y) obtains a regression line 83, shown in FIG. 13, from the set 70 of data points, and calculates an angle 93 b obtained by subtracting an angle 93 a calculated from the slope of the regression line 83 from 90 degrees, as shown in FIG. 13. Here, the angle 93 a shown in FIG. 13 is shown so as to correspond to the position in the case where the angle calculated from the slope of the regression line 83 in the case where y is considered as the independent variable and x is considered as the dependent variable is transformed back to the coordinates in which x is considered as the independent variable and y is considered as the dependent variable.
  • FIG. 14 is an illustration showing an example of the data points held in the data point buffer at time t4. Time t4 is the time after an elapse of time Δt from time t1. A set 40′ of the data points held in the data point buffer consists of eight data points P45 to P52. Furthermore, the x coordinate values and the y coordinate values of P45 to P52 are respectively P45(x 45, y45), P46(x 46, y46) . . . P52 (x52, y52), and it is assumed that the data points were stored in the order P45, P46, . . . , P52.
  • It is understood from FIG. 15 that the difference x52−x45 (=Δx) between the x coordinate values of P52 and P45 corresponds to abs(max(x)−min(x)) and that the difference y46−y52 (=Δy) between the y coordinate values of P46 and P52 corresponds to abs(max(y)−min(y)). According to FIG. 14, 2×Δx<Δy. Thus, the function rotate(x, y) satisfies the inequality and thus returns a true value.
  • Then, the function aop(x, y) calculates the differences (y46−y45, y47−y46, . . . y52−y51) between values temporally succeeding and following each other by using the function down(y). According to FIG. 14, all the differences except y46−y45 are negative. Thus, the function down(y) satisfies the inequality and thus returns a true value.
  • In the above example, the function rotate(x, y) is true, and the function down(y) is true. Thus, the function aop(x, y) obtains a regression line 84, shown in FIG. 16, from the set 40′ of data points, and calculates an angle 94 c obtained by adding 180 degrees to an angle 94 b calculated by subtracting an angle 94 a calculated from the slope of the regression line 84 from 90 degrees, as shown in FIG. 17.
  • FIG. 18 is an illustration showing how the electronic device 10 according to the embodiment of the present invention controls a virtual character 121 on the basis of a set 120 of data points. FIG. 18 shows the set 120 of data points visualizing data points corresponding to touch events generated by a user operation on the touchscreen 17. The electronic device 10 controls motion, such as walking, running, and switching direction, of the virtual character 121, which is the object being operated, on the basis of the angle 122 indicated by the set 120 of data points.
  • FIG. 19 is a flowchart of information processing by the angle determination unit 24, according to the embodiment of the present invention. This flowchart starts when a touch event is generated and the angle determination unit 24 obtains a data point from the touch event.
  • In step 101, the angle determination unit 24 stores the obtained data point in the data point buffer. At this time, the angle determination unit 24 associates, with the data point being stored, T indicating the time, in milliseconds, elapsed after the storage thereof and D indicating the period (holding life), in milliseconds, for which the data point can be stored in the data point buffer.
  • Then, in step 102, the angle determination unit 24 outputs a start event to the state determination unit 25.
  • Then, in step 103, the angle determination unit 24 determines whether or not a touch event has been generated. Step 103 is executed after I ms from the start of the flowchart. The flowchart proceeds to step 104 in the case where a touch event has been generated, and the flowchart proceeds to step 106 in the case where no touch event has been generated.
  • In step 104, the angle determination unit 24 determines whether or not the generated touch event is touchend. The flowchart is terminated in the case where the obtained touch event is touchend. The flowchart proceeds to step 105 in the case where the obtained touch event is not touchend.
  • In step 105, the angle determination unit 24 obtains a data point from the generated touch event, and stores the data point in the data point buffer. At this time, the angle determination unit 24 associates, with the data point being stored, T indicating, in milliseconds the time elapsed after the storage thereof and D indicating, in milliseconds, the period for which the data point can be stored in the data point buffer.
  • Then, in step 106, the angle determination unit 24 terminates the holding of data points for which the elapsed period T is greater than or equal to the variable D among the data points held in the data point buffer. The angle determination unit 24 compares the elapsed time T with the variable D for each of the data points stored in the data point buffer, and terminates the holding of each data point for which the elapsed time is greater than or equal to the variable D.
  • Then, in step 107, the angle determination unit 24 determines whether or not the number n of the data points held in the data point buffer is greater than or equal to the value of the variable B. The flowchart proceeds to step 108 in the case where the number n is greater than or equal to the variable B, and the flowchart proceeds to step 113 in the case where the number n is less than the variable B.
  • In step 108, the angle determination unit 24 determines one of the x axis and the y axis as the axis of the independent variable on the basis of the amount of displacement of the values on the x axis and the amount of displacement of the values on the y axis in the data points held in the data point buffer. At the same time, the angle determination unit 24 determines the other axis as the axis of the dependent variable.
  • Then, in step 109, in the case where the x axis is determined as the axis of the independent variable in step 108, the angle determination unit 24 determines the angle of the slope of a regression line by calculating the angle of the slope by using Eq. (5). In one example, when calculating the angle of the slope by using Eq. (5), the angle determination unit 24 calculates the angle within the range of 0 to 90 degrees and 270 to 360 degrees. In the case where the y axis is determined as the axis of the independent variable in step 108, the angle determination unit 24 calculates the angle of the slope by using Eq. (6), and subtracts the calculated angle from 90 degrees, thereby determining the angle of the slope of a regression line. In one example, when calculating the angle of the slope by using E. (6), the angle determination unit 24 calculates the angle within the range of 0 to 90 degrees and 270 to 360 degrees.
  • Then, in step 110, the angle determination unit 24 determines an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees on the basis of the displacement direction of the set of data points held in the data point buffer. In one example, in the case where the x axis is determined as the axis of the independent variable in step 108, the angle determination unit 24 calculates the individual differences between values temporally succeeding and following each other on the x axis. The angle determination unit 24 determines the amount of rotation to be 180 degrees when the number of negative calculated differences is greater than the number of positive calculated differences, and determines the amount of rotation to be 0 degrees when the former is less than the latter. In one example, in the case where the y axis is determined as the axis of the independent variable in step 108, the angle determination unit 24 calculates the individual differences between values temporally succeeding and following each other on the y axis. The angle determination unit 24 determines the amount of rotation to be 180 degrees when the number of negative calculated differences is greater than the number of positive calculated differences, and determines the amount of rotation to be 0 degrees when the former is less than the latter.
  • Then, in step 111, the angle determination unit 24 determines an angle on the basis of the determined slope of the regression line and the determined amount of rotation. Specifically, the angle determination unit 24 determines an angle by adding the determined amount of rotation to the angle corresponding to the determined slope of the regression line. For example, in the case where the amount of rotation is 0 degrees, the angle determined by the angle determination unit 24 is the angle corresponding to the determined slope of the regression line. The angle determination unit 24 outputs an angle event together with the determined angle to the state determination unit 25, and proceeds to step 112.
  • In the case where the number n is less than the variable B in step 107, in step 113, the angle determination unit 24 outputs a keep event to the state determination unit 25, and proceeds to step 112.
  • In step 112, the flowchart returns to step 103 unless terminated, for example, due to the termination of the game application. The angle determination unit 24 executes the processing in steps 103 to 112 every I ms.
  • When the flowchart is terminated, the angle determination unit 24 deletes all the data points held in the data point buffer.
  • Next, main operations and advantages of the electronic device 10 according to the embodiment of the present invention will be described. This embodiment utilizes the characteristic of a projection-type electrostatic capacitive touchscreen, in which when the touchscreen 17 is swiped with a finger, six or more touch events are generated within an extremely short period, such as 100 ms. The system architecture of the software implemented by the electronic device 10 has a three-layer structure, in which the angle determination unit 24, the state determination unit 25, and the application unit 26 respectively correspond to the first layer, the second layer, and the third layer, where the third layer corresponds to the game application. The first layer calculates the swipe direction for a set of a plurality of generated touch events (a set of data points), in the form of a linear angle over 360 degrees. Since it is possible to use touch events generated within an extremely short period, it is possible for the first layer to continue to calculate an angle at high speed. Furthermore, by obtaining an angle indicated by a set of data points held in the data point buffer just for a predefined extremely short holding period (e.g., 165 ms), it is possible for the first layer to obtain an angle of the direction intended by the user who has performed a touch operation on the touchscreen 17, without having to use reference points.
  • The electronic device 10 in this embodiment is designed on the basis of the concept that, however complex a user operation such as a swipe is, within an extremely short period, such as 100 ms, the user operation can be modeled on linear motion with constant velocity, and thus it is possible to predict and approximate an arbitrary nonlinear function (≈ human operation) by continuously calculating the values of the angle of motion with constant velocity.
  • The first layer calculates an angle at a relatively low frequency is the case where the value of the variable I is set to be relatively large. In this case, even immediately after the start of generation of touch events, for example, even immediately after the user has started an operation, the first layer calculates an angle in a state where a relatively large number of data points are held in the data point buffer. As described above, by setting the value of the variable I to be relatively large, it becomes possible to calculate an angle in which a user operation is reflected relatively loosely. Meanwhile, the first layer calculates an angle at a relatively high frequency in the case where the value of the variable I is set to be relatively small. In this case, immediately after the start of generation of touch events, for example, immediately after the user has started an operation, the first layer calculates an angle in a state where a relatively small number of data points are held in the data point buffer. As described above, by setting the value of the variable I to be relatively small, it becomes possible to calculate an angle in which a user operation is reflected relatively immediately.
  • The first layer calculates the slope of a regression line by using the method of least squares in calculating an angle indicated by a set of data points. In the case where x is fixed as the independent variable and y is fixed as the dependent variable, it is difficult to calculate the slope, for example, in the case where the slope is in a direction along the y axis. Thus, the first layer determines the independent variable and the dependent variable from the amount of displacement of the data points before calculating the slope of a regression line. With this configuration, it becomes possible for the first layer to stably calculate the slope of a regression line.
  • In calculating the angle indicated by the set of data points, the first layer determines an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees, on the basis of the displacement direction of the set of data points, corresponding to a rough direction in which the user moves a finger on the touchscreen 17. By determining the amount of rotation, which cannot be ascertained from the slope of the regression line, calculated by using the method of least squares, it is possible to obtain an angle intended by the user who has performed a touch operation on the touchscreen 17.
  • The first layer calculates an angle corresponding to the swipe direction for a set of data points per predefined processing period, for example, per period corresponding to the frame rate of the game application as the predefined processing period. The second layer determines a vector quantity (angle and magnitude) indicating the state of the slope of the virtual controller per predefined processing period by using continuously output angle information, and outputs the vector quantity to the game application. With this configuration, in this embodiment, it becomes possible to give an input to the game application on the basis of the angle calculated per predefined processing period. Furthermore, with this configuration, without requiring any layer to use past touched coordinates as reference points, it becomes possible to give an input on the basis of an accurate movement angle at high frequency at the frame rate calculated by the first layer. As described above, in this embodiment, the electronic device 10 calculates an angle without using the spatial concept of points, such as starts points (start coordinates) and end points (end coordinates), which has been used in the virtual controllers in the existing technology.
  • Furthermore, in this embodiment, as opposed to the virtual controller in the existing technology, i.e., the virtual controller that uses a vector obtained on the basis of the positional relationship between reference coordinates and current instruction coordinates, the concept of reference coordinates is not involved, which makes it possible to provide higher responsiveness compared with the virtual controller with which reference coordinates are defined. In particular, in the case of an operation for switching the direction of a virtual character or the like, even if the user performs an operation on the touchscreen in a direction considerably different from the direction before switching, since the concept that the current instruction coordinates approach the reference coordinates is not lacking, it is possible to switch the direction by quickly responding to the intent of the operation by the user, which is considerably advantageous. Furthermore, with the configuration described above, a user operating a smartphone is allowed to perform operations with one hand. This makes it possible to solve the problem that it is necessary for a user to be constantly aware of reference coordinates when performing operations with the virtual joystick in the existing technology and thus there are cases where it is difficult to perform operations with one hand. As described above, in this embodiment, a virtual controller that enables quicker and more intuitive operation is realized.
  • Furthermore, in this embodiment, as opposed to the virtual controller in the existing technology, since input is not based on the movement distance of a finger from reference coordinates, it becomes possible for the electronic device 10 to realize operations intended by the user via operations with smaller amounts of movement of a finger. Thus, as compared with the existing technology, implementation with a smaller implementation area becomes possible. For example, it becomes possible to realize the same operability irrespective of the size of the touchscreen 17.
  • Furthermore, since the technology of the virtual controller provided by the electronic device 10 in this embodiment is based on a mathematical model of angle recognition from a swipe operation, the technology is applicable to a wide variety of genres, such as movement in directions over 360 degrees, acceleration and deceleration, and input of commands in a combat game or the like.
  • The operations and advantages described above also apply similarly to other embodiments and other examples unless otherwise specifically mentioned.
  • Another embodiment of the present invention may be a program for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention, or a computer-readable storage medium storing the program. Furthermore, another embodiment of the present invention may be a method for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention. Furthermore, another embodiment of the present invention may be a server that is capable of providing a computer with a program for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention. Furthermore, another embodiment of the present invention may be a virtual machine for realizing the functions or the information processing shown in the flowchart in the above-described embodiment of the present invention.
  • Now, modifications of the embodiments of the present invention will be described. The modifications described below can be combined as appropriate and applied to any embodiment of the present invention as long as no inconsistency arises.
  • In one modification, the electronic device 10 has installed there in a game application that disposes a virtual object in a virtual space and that presents a user with a game image photographed by a virtual camera disposed in the virtual space while causing the game to proceed. When the game application is executed, the electronic device 10 displays, on the touchscreen 17, a game image obtained by photographing the virtual object, etc. disposed in the virtual space by means of the virtual camera disposed in the virtual space, and controls the virtual camera in accordance with user operations. As described above, when the game application is executed, the electronic device 10 provides a virtual controller for controlling the virtual camera, which is an object being operated, in accordance with user operations. The control of the virtual camera refers to controlling the motion or the visual field region of the virtual camera disposed in the virtual space.
  • FIG. 20 is an illustration showing a virtual camera 131 disposed in the virtual space, which captures an image displayed by the electronic device 10 according to the embodiment of the present invention. FIG. 20 shows the position of the virtual camera 131 and a viewing direction 132 of the virtual camera 132. The viewing direction 132 is defined in the form of a three-dimensional vector with the viewpoint set at the position of the virtual camera 131. A visual field region is defined with a certain viewing angle from the virtual camera 131 into the viewing direction 132, and a two-dimensional screen 133, constituting a plane perpendicular to the viewing direction 132, is defined in the visual field region. The virtual object in the virtual space is projected onto the two-dimensional screen 133, whereby a two-dimensional image is formed.
  • In one modification, the input device 12 and the display device 13 have separate forms disposed at different positions. In this case, the input device 12 is a touchscreen or a device having a function equivalent to the function of a projection-type electrostatic capacitive touchscreen. It suffices for the display device 13 to be a device that displays an application screen, etc. to the user of the electronic device 10 under the control of the processor 11. For example, the display device 13 is a liquid crystal display, a display using organic EL, or a plasma display.
  • In one modification, the angle determination unit 24 calculates the slope of a regression line by using a known method other than the method of least squares. In this case, the angle determination unit 24 does not determine an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees, and thus does not determine the axis of the independent variable and the axis of the dependent variable. For example, it is possible to use an algorithm such as the Kalman filter or particle filters.
  • In one modification, the angle determination unit 24 does not define the variable D, and does not terminate the holding of a data point for which it is determined that the predefined holding period has expired among the data points held in the data point buffer. In this case, the angle determination unit 24 determines an angle while referring to data points stored in a specific time band that is shifted per time defined by the variable I.
  • In one modification, the angle determination unit 24 does not define the variable V. In this case, the angle determination unit 24 calculates and determines an angle by using the function aop(x, y) irrespective of the value of the variance of the independent variable.
  • In one modification, the electronic device 10 is equipped with an acceleration sensor that is capable of determining the gravitational direction. The angle determination unit 24 determines a weight coefficient by using information obtained from the acceleration sensor. For example, in the case where the smartphone is held in portrait orientation and thus the y axis of the touchscreen 17 is the gravitational direction, the angle determination unit 24 determines w to be 2. Meanwhile, in the case where the smartphone is held in landscape orientation and thus the x axis of the touchscreen 17 is the gravitational direction, and the game engine provided in the game application does not perform transformation between vertical and horizontal coordinates, the angle determination unit 24 determines w to be a value less than 1, such as 0.5.
  • In one modification, when obtaining a touch event, the angle determination unit 24 obtains a set (x, y) of numerical values consisting of two variables, and stores the set (x, y) of numerical values consisting of the two variables in the data point buffer without associating a data point obtaining time t therewith. For example, the angle determination unit 24 can store information corresponding to a data point obtaining time t in a memory area or the like in the storage device 14 other than the data point buffer and can manage the information in association with data stored in the data point buffer.
  • The processing or operation described above may be modified freely as long as no inconsistency arises in the processing or operation, such as an inconsistency that a certain step utilizes data that may not yet be available in that step. Furthermore, the examples described above are examples for explaining the present invention, and the present invention is not limited to those examples. The present invention can be embodied in various forms as long as there is no departure from the gist thereof.
  • REFERENCE SIGNS LIST
    • 10 Electronic device
    • 11 Processor
    • 12 Input device
    • 13 Display device
    • 14 Storage device
    • 15 Communication device
    • 16 Bus
    • 17 Touchscreen
    • 21 Input unit
    • 22 Display unit
    • 23 Control unit
    • 24 Angle determination unit
    • 25 State determination unit
    • 26 Application unit
    • 31, 33, 35 Angle
    • 32, 34, 36 Direction
    • 81, 82, 83, 84 Regression line
    • 91, 92 a, 92 b, 93 a, 93 b, 94 a, 94 b, 94 c Angle
    • 120 Set
    • 121 Virtual character
    • 122 Angle
    • 131 Virtual camera
    • 132 Viewing direction
    • 133 Two-dimensional screen

Claims (14)

1. A non-transitory computer readable medium storing a program that is executed on an electronic device equipped with a touchscreen, the program causing the electronic device to execute:
a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user;
a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points;
a step of determining the slope of a regression line on the basis of the held data points;
a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and
a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
2. The non-transitory computer readable medium according to claim 1, wherein:
in the step of determining an amount of rotation,
when determining an angle for controlling the object being operated, an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees is determined on the basis of a displacement direction of data points temporally succeeding and following each other among the held data points.
3. The non-transitory computer readable medium according to claim 1, wherein:
the program causes the electronic device to further execute:
a step of determining one of the first axis and the second axis as an axis of an independent variable and determining the other axis as an axis of a dependent variable on the basis of an amount of displacement of the values on the first axis and an amount of displacement of the values on the second axis in the held data points, and
in the step of determining the slope of a regression line,
the slope of the regression line is determined further on the basis of the determined axis of the independent variable and axis of the dependent variable.
4. The non-transitory computer readable medium according to claim 3, wherein:
in the step of determining one of the axes,
the one of the axes is determined as the axis of the independent variable on the basis of the difference between a maximum value and a minimum value of the values on the first axis and the difference between a maximum value and a minimum value of the values on the second axis.
5. The non-transitory computer readable medium according to claim 3, wherein:
in the step of determining one of the axes,
the one of the axes is determined as the axis of the independent variable by comparing the magnitude of the result of applying a weight to the difference between a maximum value and a minimum value of the values on the first axis and the magnitude of the difference between a maximum value and a minimum value of the values on the second axis.
6. The non-transitory computer readable medium according to claim 1, wherein:
in the step of determining an amount of rotation,
when determining an angle by which the object being controlled is controlled, an amount of rotation indicating whether or not the determined slope of the regression line is to be rotated by 180 degrees is determined by comparing the number of positive values and the number of negative values among the differences between values temporally succeeding and following each other on the axis of the determined independent variable in the held data points.
7. The non-transitory computer readable medium according to claim 1, wherein:
the step of determining the slope of a regression line includes:
a step of determining the average of the independent variable and the average of the dependent variable in the held data points;
a step of determining, by using the determined average, the deviation of the independent variable and the deviation of the dependent variable in the held data points;
a step of determining, by using the determined deviation of the independent variable, the variance of the independent variable in the held data points;
a step of determining, by using the determined variance of the independent variable and variance of the dependent variable, the covariance in the held data points; and
a step of determining the slope of the regression line by dividing the determined covariance by the determined variance of the independent variable.
8. The non-transitory computer readable medium according to claim 7, wherein:
the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed,
the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis, and
in the step of determining the slope of a regression line,
in the case where the determined axis of the independent variable is the second axis, the slope of the regression line is determined by subtracting, from 90 degrees, an angle corresponding to the determined slope of the regression line, determined by dividing the determined covariance by the determined variance of the independent variable.
9. The non-transitory computer readable medium according to claim 1, wherein the first axis is an X axis indicating the widthwise direction of directions in which sensors of the touchscreen are arrayed, and the second axis is a Y axis indicating the lengthwise direction of the directions in which the sensors of the touchscreen are arrayed, which is perpendicular to the first axis.
10. The non-transitory computer readable medium according to claim 1, wherein:
in the step of determining an angle,
an angle is determined per predefined processing period.
11. A non-transitory computer readable medium storing a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including a program according to claim 10, wherein:
the predefined processing period is a period corresponding to a frame rate for executing the game,
the program suite causing the execution of:
a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and
a step of controlling the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
12. A non-transitory computer readable medium storing a program suite for a game that is executed on the electronic device equipped with the touchscreen, the program suite including a program according to claim 10, wherein:
the predefined processing period is a period corresponding to a frame rate for executing the game,
the program suite causing the execution of:
a step of determining an angle and a magnitude per the predefined processing period on the basis of the angle determined per the predefined processing period; and
a step of controlling a virtual camera for photographing a game image, the virtual camera serving as the object being controlled, which is displayed on the touchscreen, on the basis of the angle and magnitude determined per the predefined processing period.
13. An electronic device equipped with a touchscreen, wherein:
data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user, are held;
the holding of a data point for which a predefined holding period has expired among the held data points is terminated;
the slope of a regression line is determined on the basis of the held data points;
an amount of rotation by which the determined slope of the regression line is to be rotated is determined on the basis of a displacement direction of a set of the held data points; and
an angle by which the user controls an object being operated in a virtual space is determined on the basis of the determined slope of the regression line and the determined amount of rotation.
14. A method that is executed on an electronic device equipped with a touchscreen, the method comprising:
a step of holding data points indicated by values on a first axis and values on a second axis, obtained on the basis of touch events generated as a result of an operation on the touchscreen by a user;
a step of terminating the holding of a data point for which a predefined holding period has expired among the held data points;
a step of determining the slope of a regression line on the basis of the held data points;
a step of determining an amount of rotation by which the determined slope of the regression line is to be rotated, on the basis of a displacement direction of a set of the held data points; and
a step of determining an angle by which the user controls an object being operated in a virtual space, on the basis of the determined slope of the regression line and the determined amount of rotation.
US17/089,816 2018-05-16 2020-11-05 Program, electronic device, and method for simulating a joystick in a virtual space Active 2039-07-14 US11541311B2 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2018094656A JP6389581B1 (en) 2018-05-16 2018-05-16 Program, electronic apparatus, and method
JP2018-094656 2018-05-16
JPJP2018-094656 2018-05-16
PCT/JP2019/016971 WO2019220873A1 (en) 2018-05-16 2019-04-22 Program, electronic device, and method

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2019/016971 Continuation WO2019220873A1 (en) 2018-05-16 2019-04-22 Program, electronic device, and method

Publications (2)

Publication Number Publication Date
US20210052984A1 true US20210052984A1 (en) 2021-02-25
US11541311B2 US11541311B2 (en) 2023-01-03

Family

ID=63518833

Family Applications (1)

Application Number Title Priority Date Filing Date
US17/089,816 Active 2039-07-14 US11541311B2 (en) 2018-05-16 2020-11-05 Program, electronic device, and method for simulating a joystick in a virtual space

Country Status (5)

Country Link
US (1) US11541311B2 (en)
JP (1) JP6389581B1 (en)
KR (1) KR102362465B1 (en)
CN (1) CN112424739A (en)
WO (1) WO2019220873A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220008820A1 (en) * 2020-07-08 2022-01-13 Nintendo Co., Ltd. Storage medium storing information processing program, information processing apparatus, information processing system, and information processing method
US11442622B1 (en) * 2021-05-12 2022-09-13 Castles Technology Co., Ltd. Password input method
US20230256336A1 (en) * 2022-02-16 2023-08-17 Nintendo Co., Ltd. Computer-readable non-transitory storage medium having information processing program stored therein, information processing system, information processing apparatus, and information processing method
US12023588B2 (en) * 2022-02-16 2024-07-02 Nintendo Co., Ltd. Computer-readable non-transitory storage medium having information processing program stored therein, information processing system, information processing apparatus, and information processing method

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7250451B2 (en) * 2018-07-13 2023-04-03 株式会社Cygames Programs, electronic devices, and methods
JP2021067990A (en) * 2019-10-18 2021-04-30 株式会社Cygames Processing apparatus, processing method, and program
JP6748281B1 (en) * 2019-12-10 2020-08-26 株式会社Cygames Server, processing system, processing method and program
JP7143364B2 (en) * 2020-05-15 2022-09-28 株式会社Cygames Method, program and electronic device
JP6862597B1 (en) * 2020-07-14 2021-04-21 株式会社Cygames Methods, programs, and electronics

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060111182A1 (en) * 2004-11-19 2006-05-25 Nintendo Co., Ltd. Storage medium having game program stored thereon and game apparatus
US20060121985A1 (en) * 2004-11-19 2006-06-08 Nintendo Co., Ltd. Storage medium having game program stored thereon and game apparatus
US20080039201A1 (en) * 2006-05-09 2008-02-14 Nintendo Co., Ltd. Game program and game apparatus
US20120046106A1 (en) * 2010-08-20 2012-02-23 Nintendo Co., Ltd. Game system, game device, storage medium storing game program, and game process method
US20120258796A1 (en) * 2011-04-07 2012-10-11 Nintendo Co., Ltd. Input system, information processing device, storage medium storing information processing program, and three-dimensional position calculation method

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3734820A (en) 1971-09-22 1973-05-22 Calgon Corp Cationic dextran graft copolymers as dry strength additives for paper
JP3734820B1 (en) * 2004-09-03 2006-01-11 任天堂株式会社 GAME PROGRAM, GAME DEVICE, AND INPUT DEVICE
CN101498985B (en) * 2008-01-30 2012-05-30 义隆电子股份有限公司 Touch control panel for multi-object operation and its use method
JP4932010B2 (en) * 2010-01-06 2012-05-16 株式会社スクウェア・エニックス User interface processing device, user interface processing method, and user interface processing program
JP5237325B2 (en) * 2010-04-28 2013-07-17 株式会社スクウェア・エニックス Video game processing apparatus, video game processing method, and video game processing program
EP2497547B1 (en) * 2011-03-08 2018-06-27 Nintendo Co., Ltd. Information processing program, information processing apparatus, information processing system, and information processing method
JP5757118B2 (en) 2011-03-23 2015-07-29 ソニー株式会社 Information processing apparatus, information processing method, and program
JP5594847B2 (en) * 2011-05-19 2014-09-24 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, information processing apparatus control method, program, and information storage medium
US20130300704A1 (en) * 2011-09-13 2013-11-14 Tomonari Takahashi Information input device and information input method
JP2013127683A (en) * 2011-12-16 2013-06-27 Namco Bandai Games Inc Program, information storage medium, terminal, server, and network system
JP2014154055A (en) * 2013-02-13 2014-08-25 Mitsubishi Electric Corp Image processor and image processing method
KR20140105354A (en) * 2013-02-22 2014-09-01 삼성전자주식회사 Electronic device including a touch-sensitive user interface
US9772743B1 (en) * 2015-03-30 2017-09-26 Electronic Arts Inc. Implementation of a movable control pad on a touch enabled device
JP6313395B1 (en) 2016-10-17 2018-04-18 グリー株式会社 Drawing processing method, drawing processing program, and drawing processing apparatus
CN107890665A (en) * 2017-11-20 2018-04-10 上海祐胜电子科技有限公司 A kind of new game control device
JP6560801B1 (en) * 2018-09-26 2019-08-14 株式会社Cygames Program, electronic apparatus, and method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20060111182A1 (en) * 2004-11-19 2006-05-25 Nintendo Co., Ltd. Storage medium having game program stored thereon and game apparatus
US20060121985A1 (en) * 2004-11-19 2006-06-08 Nintendo Co., Ltd. Storage medium having game program stored thereon and game apparatus
US20080039201A1 (en) * 2006-05-09 2008-02-14 Nintendo Co., Ltd. Game program and game apparatus
US20120046106A1 (en) * 2010-08-20 2012-02-23 Nintendo Co., Ltd. Game system, game device, storage medium storing game program, and game process method
US20120258796A1 (en) * 2011-04-07 2012-10-11 Nintendo Co., Ltd. Input system, information processing device, storage medium storing information processing program, and three-dimensional position calculation method

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220008820A1 (en) * 2020-07-08 2022-01-13 Nintendo Co., Ltd. Storage medium storing information processing program, information processing apparatus, information processing system, and information processing method
US11577157B2 (en) * 2020-07-08 2023-02-14 Nintendo Co., Ltd. Systems and method of controlling game operations based on touch input
US11590413B2 (en) * 2020-07-08 2023-02-28 Nintendo Co., Ltd. Storage medium storing information processing program with changeable operation modes, information processing apparatus, information processing system, and information processing method
US11442622B1 (en) * 2021-05-12 2022-09-13 Castles Technology Co., Ltd. Password input method
US20230256336A1 (en) * 2022-02-16 2023-08-17 Nintendo Co., Ltd. Computer-readable non-transitory storage medium having information processing program stored therein, information processing system, information processing apparatus, and information processing method
US12023588B2 (en) * 2022-02-16 2024-07-02 Nintendo Co., Ltd. Computer-readable non-transitory storage medium having information processing program stored therein, information processing system, information processing apparatus, and information processing method

Also Published As

Publication number Publication date
KR20200138811A (en) 2020-12-10
JP2019200595A (en) 2019-11-21
CN112424739A (en) 2021-02-26
JP6389581B1 (en) 2018-09-12
KR102362465B1 (en) 2022-02-14
WO2019220873A1 (en) 2019-11-21
US11541311B2 (en) 2023-01-03

Similar Documents

Publication Publication Date Title
US11541311B2 (en) Program, electronic device, and method for simulating a joystick in a virtual space
US20210205698A1 (en) Program, electronic device, and method
US9110584B2 (en) Controlling a cursor on a touch screen
JP6522343B2 (en) Pan animation
CN104364734B (en) Remote session control using multi-touch inputs
KR101254037B1 (en) Method and mobile terminal for display processing using eyes and gesture recognition
CN110215685B (en) Method, device, equipment and storage medium for controlling virtual object in game
US9891713B2 (en) User input processing method and apparatus using vision sensor
WO2020186826A1 (en) Game control method based on intelligent wristband, intelligent wristband and storage medium
US20150153834A1 (en) Motion input apparatus and motion input method
US10073609B2 (en) Information-processing device, storage medium, information-processing method and information-processing system for controlling movement of a display area
US9665232B2 (en) Information-processing device, storage medium, information-processing method, and information-processing system for enlarging or reducing an image displayed on a display device
US9350918B1 (en) Gesture control for managing an image view display
CN108885556A (en) Control numeral input
JP7250451B2 (en) Programs, electronic devices, and methods
JP7471782B2 (en) Program, electronic device, and method
TW201621651A (en) Mouse simulation system and method
US20230117127A1 (en) Method, program, and electronic device
CN113590014B (en) Elevator calling method and device based on gesture action and computer equipment
JP7246390B2 (en) Direct manipulation of display devices using wearable computing devices
CN117112094A (en) Control interaction method and device, storage medium and electronic device
KR20210029753A (en) Method and apparatus for recognizing gesture
CN116225218A (en) Game view adjusting method and electronic device
CN112987923A (en) Method, apparatus, device and storage medium for device interaction
KR20200046764A (en) Method and apparatus for recognizing gesture

Legal Events

Date Code Title Description
FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

AS Assignment

Owner name: CYGAMES, INC., JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:KURABAYASHI, SHUICHI;REEL/FRAME:061825/0104

Effective date: 20221110

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE