US20210200373A1 - Microphone on controller with touchpad to take in audio swipe feature data - Google Patents
Microphone on controller with touchpad to take in audio swipe feature data Download PDFInfo
- Publication number
- US20210200373A1 US20210200373A1 US17/194,225 US202117194225A US2021200373A1 US 20210200373 A1 US20210200373 A1 US 20210200373A1 US 202117194225 A US202117194225 A US 202117194225A US 2021200373 A1 US2021200373 A1 US 2021200373A1
- Authority
- US
- United States
- Prior art keywords
- event
- display
- virtual keyboard
- letter
- touchpad
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
- G06F3/0416—Control or interface arrangements specially adapted for digitisers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/018—Input/output arrangements for oriental characters
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/02—Input arrangements using manually operated switches, e.g. using keyboards or dials
- G06F3/023—Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
- G06F3/0233—Character input methods
- G06F3/0237—Character input methods using prediction or retrieval techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04886—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/044—Recurrent networks, e.g. Hopfield networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/20—Input arrangements for video game devices
-
- A—HUMAN NECESSITIES
- A63—SPORTS; GAMES; AMUSEMENTS
- A63F—CARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
- A63F13/00—Video games, i.e. games using an electronically generated display having two or more dimensions
- A63F13/60—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
- A63F13/67—Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor adaptively or by learning from player actions, e.g. skill level adjustment or by storing successful combat sequences for re-use
Definitions
- the application relates generally to technically inventive, non-routine solutions that are necessarily rooted in computer technology and that produce concrete technical improvements.
- the present application relates to computer simulation controllers with touchpad input.
- Machine learning sometimes referred to as deep learning, can be used for a variety of useful applications related to data understanding, detection, and/or classification.
- Present principles thus provide a microphone on a touchpad of a computer simulation controller that can be used to receive voice signals for training a machine learning module to predict a next letter or next word, or to insert special characters/punctuations/graphics such as “smileys” during the swipe, or to indicate a tone of an Asian word character such as a Chinese character while typing with Chinese Pinyin.
- an apparatus includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to receive a touch signal from a touch surface of a computer simulation controller to identify a first alpha-numeric character.
- the instructions are executable to input the first alpha-numeric character to at least a first neural network (NN), and receive from the first NN a predicted sequence of alpha-numeric characters including at least a first predicted alpha-numeric character for presentation on at least one display.
- the instructions also are executable to receive, from at least one microphone, input indicating acceptance or rejection of at least the first predicted alpha-numeric character and provide the input from the microphone to the first NN to train the first NN.
- the first NN may include plural long short-term memory (LSTM) networks.
- the processor and microphone are embodied in the computer simulation controller.
- the processor may be embodied in a computer simulation console configured for communicating with the computer simulation controller.
- the instructions can be executable to identify at least one punctuation symbol using the input from the microphone, and responsive to identifying the punctuation symbol, present the punctuation symbol on the display.
- the instructions can be executable to identify at least one tone using the input from the microphone, and responsive to identifying the tone, identify for presentation on the display at least one Chinese Pinyin character.
- the instructions may be executable to receive from the touch surface indication of at least two Arabic letters.
- the instructions further may be executable to identify, using the Arabic letters, at least first and second candidate Chinese words, and responsive to identifying the tone, select the first Chinese word but not the second Chinese word.
- an apparatus in another aspect, includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to identify at least one tone using input from a microphone, and responsive to identifying the tone, identify for presentation on a display at least one Asian language character.
- an apparatus in another aspect, includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to identify at least one punctuation symbol using input from a microphone, and responsive to identifying the punctuation symbol, present the punctuation symbol on a display.
- FIG. 1 is a block diagram of an example system consistent with present principles
- FIG. 2 is a perspective view of a computer simulation controller with a microphone and a touch pad being used for inputting text presented on a display such as a TV or other audio video device communicating with the game controller directly or via, e.g., a computer game console;
- FIG. 3 is a schematic diagram illustrating a soft press and a hard press on the controller touch pad
- FIG. 4 is a flow chart of example logic consistent with present principles related to FIG. 3 ;
- FIG. 5 is a combination of a logic flow chart, data structures, and processing components consistent with present principles
- FIGS. 6-8B are schematic diagrams of a data structure referred to as the heat map in FIG. 5 , illustrating steps in use;
- FIG. 9 is a block diagram of an example neural network (NN) configured as plural long short-term memory (LSTM) networks for outputting a predicted next word based on current user input;
- NN neural network
- LSTM long short-term memory
- FIGS. 10-12 are schematic diagrams illustrating operation of the NN in FIG. 9 post-training
- FIG. 13 is a flow chart of example overall logic consistent with present principles
- FIG. 14 is a schematic view of a system in which the microphone is used to input ground training to the neural networks contemporaneous with operation;
- FIG. 15 is a schematic view of a system in which the microphone is used to input tones of Chinese characters to the neural networks contemporaneous with operation;
- FIG. 16 is a schematic view of a system in which the microphone is used to input punctuation or graphics contemporaneous with operation.
- FIG. 17 is a flow chart of example logic consistent with FIG. 15 .
- a system herein may include server and client components, connected over a network such that data may be exchanged between the client and server components.
- the client components may include one or more computing devices including portable televisions (e.g. smart TVs, Internet-enabled TVs), portable computers such as laptops and tablet computers, and other mobile devices including smart phones and additional examples discussed below.
- portable televisions e.g. smart TVs, Internet-enabled TVs
- portable computers such as laptops and tablet computers
- other mobile devices including smart phones and additional examples discussed below.
- These client devices may operate with a variety of operating environments.
- some of the client computers may employ, as examples, operating systems from Microsoft, or a Unix operating system, or operating systems produced by Apple Computer or Google.
- These operating environments may be used to execute one or more browsing programs, such as a browser made by Microsoft or Google or Mozilla or other browser program that can access websites hosted by the Internet servers discussed below.
- Servers and/or gateways may include one or more processors executing instructions that configure the servers to receive and transmit data over a network such as the Internet.
- a client and server can be connected over a local intranet or a virtual private network.
- a server or controller may be instantiated by a game console such as a Sony PlayStation®, a personal computer, etc.
- servers and/or clients can include firewalls, load balancers, temporary storages, and proxies, and other network infrastructure for reliability and security.
- instructions refer to computer-implemented steps for processing information in the system. Instructions can be implemented in software, firmware or hardware and include any type of programmed step undertaken by components of the system.
- a processor may be any conventional general-purpose single- or multi-chip processor that can execute logic by means of various lines such as address lines, data lines, and control lines and registers and shift registers.
- Software modules described by way of the flow charts and user interfaces herein can include various sub-routines, procedures, etc. Without limiting the disclosure, logic stated to be executed by a particular module can be redistributed to other software modules and/or combined together in a single module and/or made available in a shareable library. While flow chart format may be used, it is to be understood that software may be implemented as a state machine or other logical method.
- logical blocks, modules, and circuits described below can be implemented or performed with a general-purpose processor, a digital signal processor (DSP), a field programmable gate array (FPGA) or other programmable logic device such as an application specific integrated circuit (ASIC), discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein.
- DSP digital signal processor
- FPGA field programmable gate array
- ASIC application specific integrated circuit
- a processor can be implemented by a controller or state machine or a combination of computing devices.
- connection may establish a computer-readable medium.
- Such connections can include, as examples, hard-wired cables including fiber optics and coaxial wires and digital subscriber line (DSL) and twisted pair wires.
- a system having at least one of A, B, and C includes systems that have A alone, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc.
- FIG. 1 an example system 10 is shown, which may include one or more of the example devices mentioned above and described further below in accordance with present principles. Note that computerized devices described in all of the figures herein may include some or all of the components set forth for various devices in FIG. 1 .
- the first of the example devices included in the system 10 is a consumer electronics (CE) device configured as an example primary display device, and in the embodiment shown is an audio video display device (AVDD) 12 such as but not limited to an Internet-enabled TV with a TV tuner (equivalently, set top box controlling a TV).
- AVDD 12 may be an Android®-based system.
- the AVDD 12 alternatively may also be a computerized Internet enabled (“smart”) telephone, a tablet computer, a notebook computer, a wearable computerized device such as e.g.
- AVDD 12 and/or other computers described herein is configured to undertake present principles (e.g. communicate with other CE devices to undertake present principles, execute the logic described herein, and perform any other functions and/or operations described herein).
- the AVDD 12 can be established by some or all of the components shown in FIG. 1 .
- the AVDD 12 can include one or more displays 14 that may be implemented by a high definition or ultra-high definition “4K” or higher flat screen and that may or may not be touch-enabled for receiving user input signals via touches on the display.
- the AVDD 12 may also include one or more speakers 16 for outputting audio in accordance with present principles, and at least one additional input device 18 such as e.g. an audio receiver/microphone for e.g. entering audible commands to the AVDD 12 to control the AVDD 12 .
- the example AVDD 12 may further include one or more network interfaces 20 for communication over at least one network 22 such as the Internet, an WAN, an LAN, a PAN etc. under control of one or more processors 24 .
- the interface 20 may be, without limitation, a Wi-Fi transceiver, which is an example of a wireless computer network interface, such as but not limited to a mesh network transceiver.
- the interface 20 may be, without limitation a Bluetooth transceiver, Zigbee transceiver, IrDA transceiver, Wireless USB transceiver, wired USB, wired LAN, Powerline or MoCA.
- the processor 24 controls the AVDD 12 to undertake present principles, including the other elements of the AVDD 12 described herein such as e.g. controlling the display 14 to present images thereon and receiving input therefrom.
- the network interface 20 may be, e.g., a wired or wireless modem or router, or other appropriate interface such as, e.g., a wireless telephony transceiver, or Wi-Fi transceiver as mentioned above, etc.
- the AVDD 12 may also include one or more input ports 26 such as, e.g., a high definition multimedia interface (HDMI) port or a USB port to physically connect (e.g. using a wired connection) to another CE device and/or a headphone port to connect headphones to the AVDD 12 for presentation of audio from the AVDD 12 to a user through the headphones.
- the input port 26 may be connected via wire or wirelessly to a cable or satellite source 26 a of audio video content.
- the source 26 a may be, e.g., a separate or integrated set top box, or a satellite receiver.
- the source 26 a may be a game console or disk player.
- the AVDD 12 may further include one or more computer memories 28 such as disk-based or solid-state storage that are not transitory signals, in some cases embodied in the chassis of the AVDD as standalone devices or as a personal video recording device (PVR) or video disk player either internal or external to the chassis of the AVDD for playing back AV programs or as removable memory media.
- the AVDD 12 can include a position or location receiver such as but not limited to a cellphone receiver, GPS receiver and/or altimeter 30 that is configured to e.g. receive geographic position information from at least one satellite or cellphone tower and provide the information to the processor 24 and/or determine an altitude at which the AVDD 12 is disposed in conjunction with the processor 24 .
- a position or location receiver such as but not limited to a cellphone receiver, GPS receiver and/or altimeter 30 that is configured to e.g. receive geographic position information from at least one satellite or cellphone tower and provide the information to the processor 24 and/or determine an altitude at which the AVDD 12 is disposed in conjunction
- the AVDD 12 may include one or more cameras 32 that may be, e.g., a thermal imaging camera, a digital camera such as a webcam, and/or a camera integrated into the AVDD 12 and controllable by the processor 24 to gather pictures/images and/or video in accordance with present principles.
- a Bluetooth transceiver 34 and other Near Field Communication (NFC) element 36 for communication with other devices using Bluetooth and/or NFC technology, respectively.
- NFC element can be a radio frequency identification (RFID) element.
- the AVDD 12 may include one or more auxiliary sensors 38 (e.g., a motion sensor such as an accelerometer, gyroscope, cyclometer, or a magnetic sensor, an infrared (IR) sensor for receiving IR commands from a remote control, an optical sensor, a speed and/or cadence sensor, a gesture sensor (e.g. for sensing gesture command), etc.) providing input to the processor 24 .
- auxiliary sensors 38 e.g., a motion sensor such as an accelerometer, gyroscope, cyclometer, or a magnetic sensor, an infrared (IR) sensor for receiving IR commands from a remote control, an optical sensor, a speed and/or cadence sensor, a gesture sensor (e.g. for sensing gesture command), etc.
- the AVDD 12 may include an over-the-air TV broadcast port 40 for receiving OTA TV broadcasts providing input to the processor 24 .
- the AVDD 12 may also include an infrared (IR) transmitter and/or IR receiver and/or IR transceiver 42 such as an IR data association (IRDA) device.
- IR infrared
- IRDA IR data association
- a battery (not shown) may be provided for powering the AVDD 12 .
- the AVDD 12 may include a graphics processing unit (GPU) 44 and/or a field-programmable gate array (FPGA) 46 .
- the GPU and/or FPGA may be utilized by the AVDD 12 for, e.g., artificial intelligence processing such as training neural networks and performing the operations (e.g., inferences) of neural networks in accordance with present principles.
- the processor 24 may also be used for artificial intelligence processing such as where the processor 24 might be a central processing unit (CPU).
- the system 10 may include one or more other computer device types that may include some or all of the components shown for the AVDD 12 .
- a first device 48 and a second device 50 are shown and may include similar components as some or all of the components of the AVDD 12 . Fewer or greater devices may be used than shown.
- the system 10 also may include one or more servers 52 .
- a server 52 may include at least one server processor 54 , at least one computer memory 56 such as disk-based or solid state storage, and at least one network interface 58 that, under control of the server processor 54 , allows for communication with the other devices of FIG. 1 over the network 22 , and indeed may facilitate communication between servers, controllers, and client devices in accordance with present principles.
- the network interface 58 may be, e.g., a wired or wireless modem or router, Wi-Fi transceiver, or other appropriate interface such as, e.g., a wireless telephony transceiver.
- the server 52 may be an Internet server and may include and perform “cloud” functions such that the devices of the system 10 may access a “cloud” environment via the server 52 in example embodiments.
- the server 52 may be implemented by a game console or other computer in the same room as the other devices shown in FIG. 1 or nearby.
- the methods described herein may be implemented as software instructions executed by a processor, suitably configured application specific integrated circuits (ASIC) or field programmable gate array (FPGA) modules, or any other convenient manner as would be appreciated by those skilled in those art.
- ASIC application specific integrated circuits
- FPGA field programmable gate array
- the software instructions may be embodied in a non-transitory device such as a CD ROM or Flash drive.
- the software code instructions may alternatively be embodied in a transitory arrangement such as a radio or optical signal, or via a download over the Internet.
- FIG. 2 illustrates a system 200 the components of which may incorporate appropriate components shown in FIG. 1 .
- a computer simulation controller 202 such as a PlayStation® controller, Xbox® controller, or other controller may include a touchpad 204 that can receive touch signals from a hand 206 and communicate via wired and/or wireless paths 208 with a computer simulation console 210 and/or a display device 212 such as an Internet-enabled TV.
- the user can manipulate the touchpad 204 to generate alpha-numeric characters 214 for presentation on the display device 212 either through direct communication of signals with the display device or through the simulation console 210 .
- a user can move a screen cursor over a letter on a virtual keyboard 216 presented on the display device 212 to enter the alpha-numeric characters 214 .
- the virtual keyboard 216 may have, without limitation, a QWERTY layout.
- the controller 200 may include one or more microphones 218 communicating with the processor of the controller for purposes disclosed below.
- the microphone 218 is provided on the touchpad 204 , although it is to be understood that the microphone 218 may be provided elsewhere on the housing of the controller 200 or indeed on another component if desired.
- present principles contemplate two types of touch, namely, a “soft” press 300 (using a soft pressure on the touchpad or a hover over the touchpad with zero pressure), in which a screen cursor on the display device 212 is moved to desired locations on the virtual keyboard 216 without sending location data (i.e., a signal indicating selection of any particular virtual key) to the display device, and a “hard” press 302 of greater pressure than a soft press, in response to which a screen cursor on the display device 212 may be moved and location data sent to the display device to indicate selection of a virtual key.
- a “soft” press 300 using a soft pressure on the touchpad or a hover over the touchpad with zero pressure
- location data i.e., a signal indicating selection of any particular virtual key
- a user can look away from the touchpad 204 and view the virtual keyboard 216 while moving his or her finger across the touchpad to move a visible screen cursor to a desired letter on the virtual keyboard, and then exert a hard press to select that letter.
- an individual “next” letter may not be presented on the display, but rather the next “most possible word” may be displayed after a user has finished a “swipe”.
- the “hottest” key (based on the heatmap) may be highlighted on the virtual keyboard as well as the trace.
- a “swipe” is defined as a continuous hard-press which forms a trace.
- FIG. 4 illustrates example logic with the above description in mind.
- the logic may be executed by one or more of a processor in the simulation controller 202 , a processor in the simulation console 210 , and a processor in the display device 212 .
- a press of the touchpad 204 it is determined whether a press of the touchpad 204 has been received. This may be done by determining whether signals from one or more proximity sensors associated with the touchpad 204 indicate a hover of a finger adjacent the touchpad 204 and/or by determining whether signals from one or more pressure sensors associated with the touchpad 204 indicate a pressure of at least a first threshold pressure.
- the logic proceeds to state 402 to determine whether the touch is a soft press or hard press as indicated by, e.g., signals from a pressure sensor associated with the touchpad 204 indicating a touch of at least a threshold pressure, which is typically set to be greater than any threshold pressure used at state 400 . If the touch does not satisfy the threshold, the logic moves to block 404 to return a soft press. In some implementations the logic may proceed to state 406 to determine whether the soft press is the first soft press within, e.g., a threshold of period, for example within the last five minutes, and if so the logic can move to block 408 to enlarge an image of the virtual keyboard 216 on the display device 212 . In any case, from state 406 if the test there is negative or from block 408 , the logic moves to block 410 to move the screen cursor without sending press location information.
- a threshold pressure which is typically set to be greater than any threshold pressure used at state 400 .
- test at state 402 determines that a hard press is received, such is returned at block 412 , and the screen may be moved according to the touch with location information being sent as well indicating the location of the virtual keyboard the user has selected by means of the hard press on the touchpad 204 of the simulation controller 200 .
- FIG. 5 illustrates a combination of hardware and software blocks alluded to above.
- One or more proximity and/or pressure sensors 500 are provided in the touchpad 204 to output signals representing soft presses 502 and hard presses 504 .
- the soft presses 502 establish finger focus points 506 .
- the hard presses 504 establish points on the touchpad as detected by the sensor(s) 500 .
- a soft-press represents a cursor focus point, while “points by sensor” means “continuous points sending by the sensor”.
- a heatmap algorithm is accessed to output a sequence of letters 512 according to the hard presses 504 .
- the sequence of letters 512 is input along with a dictionary 514 to a reduction block 516 that reduces the list of candidates that might possibly form either a correction to or a completion of the sequence of letters 512 .
- the dictionary 514 is essentially a dictionary and/or thesaurus of sequences of letters that can be used to correct a mis-typed word, e.g., the dictionary 514 may correlate “thw” to “the” to return the word “the” in response to input of “thw”.
- the reduced list of candidates 516 is provided to a module 518 that outputs a predicted network or words for presentation on the screen, which a user can then select to complete his or her desired input without typing every letter of the predicted word or words.
- the module 518 may be established by one or more neural networks (NN) as described further below.
- the module 518 may receive input from a contextual user block 520 , which provides previous word strings employed by the user with the current input inferred to possibly be a repeat of a prior input, e.g., “do you” may have been followed multiple times in prior inputs by “know what I mean”, and this information can be input to help train and execute the module 518 .
- Similar training/execution aids may be input to the module 518 as shown at the right of FIG. 5 .
- queries and chat data 522 from other computer gamers may be input to a character-based NN such as a bidirectional long short-term memory (BILSTM) 524 to learn patterns of common input strings for provision to a machine learning character sequence model 526 .
- This model 526 may be input to or accessed by the module 518 in rendering a next predicted word or words.
- FIGS. 6-8B illustrate employment of the heatmap algorithm 510 in FIG. 5 .
- the “path” or “connected points” of the finger “swipe” (hard-press) and the probabilities of each letter are “discounted and accumulated” at certain time interval along the swipe. At each time interval, the letter with the highest probability is extracted, which may also have to pass a certain threshold to add to the sequence as developed further below.
- each area of the heatmap for a particular letter is divided into a three-by-three grid for nine divisions (illustrated as geometric squares) total, with the center division 600 for a particular letter indicating that the probability of that letter being desired when a cursor is in the center area being 1.
- the heatmap 510 indicates probabilities less than one but greater than zero in the border divisions 602 that surround the center division 600 of a letter, with the probabilities being associated with the letter of the center division 600 and the letter(s) immediately adjacent the border divisions 602 (or, in the case of a border division that is not adjacent another letter, only a probability less than one for the letter of the center division).
- a soft press is used to locate the starting letter of an intended input. Then, as shown at 800 in FIG. 8 , a hard press is used to indicate selection of the starting letter, in the example shown, “Q”. This causes the collection of data that “Q” is selected with a probability of one and that surrounding letters (in the example shown, “W”, “A”, and “S”) are not selected, i.e., have a probability of zero.
- FIGS. 8A and 8B illustrate the results of an ensuing swipe.
- a swipe is shown at 802 from the location starting in FIG. 8 to the location 804 indicated by the image of the hand.
- the user has moved his finger toward the letter “A”.
- This causes new heatmap statistics to be aggregated according to the path of the swipe over the border divisions 602 using the algorithm shown in FIG. 8A . Because the probability of “Q” is higher than the probabilities of “W” (which is zero), “A” (which is 0.3), and “S” (which is zero), the sequence returns “Q”.
- FIG. 8B shows at 806 that the swipe has been continued to the location 808 shown by the image of the hand. This causes further heatmap statistics to be aggregated according to the path of the swipe over the border divisions 602 using the algorithm shown in FIG. 8B . Because the probability of “A” is higher than the probabilities of “W” (which is zero), “Q” (which is 0.3), and “S” (which is zero), the sequence returns “A” to be appended after “Q” was returned in FIG. 8A , resulting in a sequence “QA”.
- the “path” or “connected points” of the finger “swipe” (hard-press) is tracked and the probabilities of each letter are discounted and accumulated at certain time intervals along the swipe.
- the letter with the highest probability is extracted, in some embodiments provided the probability of the letter satisfies a threshold probability (e.g., of 0.4) to be added to the sequence.
- FIG. 9 illustrates an example NN architecture that may be used in any of the NN-based modules of, e.g., FIG. 5 .
- a network 900 of NN may receive input letters 902 with probabilities 904 from the heatmap to output time-distributed predicted letters 906 with associated probabilities 908 .
- each letter 902 may be input to a respective recurrent NN (RNN) such as a sequence of long short-term memory (LSTM) 910 as shown.
- RNN recurrent NN
- An LSTM 910 as shown at the right in FIG. 9 may include an input gate 912 , a forget gate 914 , and an output gate 916 , all of which may execute a sigmoid function as indicated by the Greek letter ⁇ in FIG. 9 .
- the input gate 912 controls the extent to which a new value flows into the cell
- the forget gate 914 controls the extent to which a value remains in the cell
- the output gate 916 controls the extent to which the value in the cell is used to compute the output activation of the LSTM unit.
- the current value xi being input and the hidden state h t-1 from the previous iteration are input to all three gates as shown.
- the output of the sigmoid function of the input gate 912 may be combined with a hyperbolic tangent function 918 at a first combine operator 920 , which may be an element-wise product.
- the output of the first combine operator 920 is combined, as by summing if desired, with the output of s second combine operator 922 at a third combine operator 924 .
- the output of the third combine operator 924 may be fed back to the second combine operator 922 for combining with the output of the forget gate 914 .
- the output of the third combine operator 924 may be operated on if desired by a hyperbolic tangent function 926 and then combined at a fourth combine operator 928 with the output of the output gate 916 to render a hidden state vector 930 for use in the succeeding iteration.
- FIGS. 10-12 illustrate a sequence of the use of the network 900 to generate predicted text.
- the lower row of letters 1000 represents input received from hard presses on keys of the virtual keyboard 216 and/or from selection of previously predicted letters and/or words. These are input to the trained network 900 .
- a next predicted letter 1004 is generated and fed back to the model.
- the sequence shown in FIGS. 10-12 generated predicted letters for an initial input of “play” that results in the word “PlayStation”.
- FIG. 13 is a flow chart of example logic consistent with present principles.
- the NN system(s) described herein are trained at block 1300 .
- a hard press is received on the touchpad and a letter established based thereon at block 1304 using the heatmap if desired.
- the letter is input to the NN system at block 1306 , which outputs a predicted letter or words or string of words at block 1308 .
- the predicted letters/words are presented on screen at block 1310 .
- Present principles may be used in all possible deep learning-based methods for image, video and audio data processing, among others.
- a user can indicate acceptance at state 1312 by speaking into the microphone 218 illustrated in FIG. 2 .
- the user may speak “OK” to accept the predicted word or “not right” or equivalent to reject it.
- This input is provided to the neural networks described herein as ground truth data for training the neural networks.
- FIG. 14 illustrates using the components discussed above in relation to FIG. 2 .
- a predicted word 1400 has been presented on the display 212 according to description above.
- a user 1402 may speak into the microphone 218 as indicated at 1404 whether to accept or reject the predicted word 1400 contemporaneous with operation.
- the predicted word is “pony”, and the user 1402 has rejected it by speaking “no” into the microphone 218 , which is digitized and provided as training data to the neural networks.
- the user 1402 has further input a correct word, in this case, “cow”, which also is provided as ground truth for training the neural networks.
- FIG. 15 illustrates a second use case for the microphone 218 .
- the user has typed in or the neural networks have predicted one or more Asian language characters such as Chinese Pinyin characters 1500 for presentation on the display 212 .
- the characters are simply numbered 1-4 in FIG. 15 , and correspond to spoken words that may be rendered in English using identical Arabic letters, in the example shown, the letters “ma”, and thus may be indistinguishable from each other when so rendered in English.
- Below each character 1500 the word 1502 it corresponds to in English is shown and below that, a respective symbol 1504 for the corresponding tone, it being understood that in implementation neither the word 1502 nor symbol 1504 may be presented.
- a user 1506 may speak a word including an intended Asian language tone 1508 into the microphone 218 , which is input to the processor(s) herein as the correct or ground truth tone.
- the user 1506 has spoken the word using the tone corresponding to the third character 1500 , giving the character the meaning “horse”.
- the user may have input, via the controller 200 , the Arabic letters “ma” as indicated at 1510 .
- the Arabic letters 1510 may be correlated to plural candidate Chinese words, which may be presented as respective Asian language characters if desired on the display 212 . Tonal input from the microphone 218 is used to confirm and/or select which of the candidate characters words the user intended by typing in “ma”, which may then be presented on the display in lieu of the other candidate words/symbols.
- FIG. 16 illustrates that during a touchpad swipe as described above, input from the microphone 218 may be used to input special characters/punctuations/graphics such as “smileys”.
- a word 1600 in the example shown, the word “yes”, that appears on the display 212 .
- a user 1602 contemporaneously may speak the word 1600 with a tone that is detected by the neural networks to correspond to excitement as indicated at 1604 , to cause an exclamation point 1606 to appear after the word 1600 .
- the user may utter, as indicated at 1608 , the name of the desired symbol or punctuation to cause the uttered symbol to be presented on the display 212 .
- the user may utter, as indicated at 1610 , the name of a desired graphic symbol such as “smiley”, to cause the uttered graphic symbol to be presented at 1612 on the display 212 .
- FIG. 17 illustrates example logic consistent with FIG. 15 .
- input may be received, e.g., from the touch pad of the controller or from other manipulable device, indicating Arabic letters and/or Chinese character(s).
- Block 1702 indicates that such input may be ambiguous in that it may correlate to more than one candidate Asian language word with corresponding symbol, in which case the logic flows to block 1704 .
- more than one pinyin symbol may be a candidate for the user's intent.
- touch input is received on the touch pad attempting to render a Chinese character, owing to imprecision in the touch tracing more than one Chinese character might be implicated.
- the candidate pinyin symbols/corresponding Chinese words may be identified.
- handwriting recognition may be employed in the case of, e.g., an attempted trace of a Chinese character using the touch pad, to identify candidate Chinese characters that may fulfill the user's intent.
- FIG. 15 illustrates a non-limiting example screen shot showing such a prompt.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Multimedia (AREA)
- Audiology, Speech & Language Pathology (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A game controller includes a touchpad that a user, viewing a virtual keyboard on a screen, can soft-touch to move a cursor on the screen and then hard-touch to move the cursor and also send location data to a processor for inputting a letter from the virtual keyboard. A microphone on the touchpad can be used to receive voice signals for training a machine learning module to predict a next letter or next word, or to insert special characters/punctuations/graphics such as “smileys” during the swipe, or to indicate a tone of a Chinese character while typing with Chinese Pinyin.
Description
- The application relates generally to technically inventive, non-routine solutions that are necessarily rooted in computer technology and that produce concrete technical improvements. In particular, the present application relates to computer simulation controllers with touchpad input.
- Machine learning, sometimes referred to as deep learning, can be used for a variety of useful applications related to data understanding, detection, and/or classification.
- In computer simulation industries such as gaming industries, multiple data entry modes may exist that can benefit from machine learning to increase precision and robustness.
- Present principles thus provide a microphone on a touchpad of a computer simulation controller that can be used to receive voice signals for training a machine learning module to predict a next letter or next word, or to insert special characters/punctuations/graphics such as “smileys” during the swipe, or to indicate a tone of an Asian word character such as a Chinese character while typing with Chinese Pinyin.
- Accordingly, an apparatus includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to receive a touch signal from a touch surface of a computer simulation controller to identify a first alpha-numeric character. The instructions are executable to input the first alpha-numeric character to at least a first neural network (NN), and receive from the first NN a predicted sequence of alpha-numeric characters including at least a first predicted alpha-numeric character for presentation on at least one display. The instructions also are executable to receive, from at least one microphone, input indicating acceptance or rejection of at least the first predicted alpha-numeric character and provide the input from the microphone to the first NN to train the first NN. The first NN may include plural long short-term memory (LSTM) networks.
- In example embodiments, the processor and microphone are embodied in the computer simulation controller. In other embodiments the processor may be embodied in a computer simulation console configured for communicating with the computer simulation controller.
- In some implementations, the instructions can be executable to identify at least one punctuation symbol using the input from the microphone, and responsive to identifying the punctuation symbol, present the punctuation symbol on the display.
- In some implementations, the instructions can be executable to identify at least one tone using the input from the microphone, and responsive to identifying the tone, identify for presentation on the display at least one Chinese Pinyin character. In such implementations, the instructions may be executable to receive from the touch surface indication of at least two Arabic letters. The instructions further may be executable to identify, using the Arabic letters, at least first and second candidate Chinese words, and responsive to identifying the tone, select the first Chinese word but not the second Chinese word.
- In another aspect, an apparatus includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to identify at least one tone using input from a microphone, and responsive to identifying the tone, identify for presentation on a display at least one Asian language character.
- In another aspect, an apparatus includes at least one processor and at least one computer storage that is not a transitory signal and that includes instructions executable by the processor to identify at least one punctuation symbol using input from a microphone, and responsive to identifying the punctuation symbol, present the punctuation symbol on a display.
- The details of the present application, both as to its structure and operation, can best be understood in reference to the accompanying drawings, in which like reference numerals refer to like parts, and in which:
-
FIG. 1 is a block diagram of an example system consistent with present principles; -
FIG. 2 is a perspective view of a computer simulation controller with a microphone and a touch pad being used for inputting text presented on a display such as a TV or other audio video device communicating with the game controller directly or via, e.g., a computer game console; -
FIG. 3 is a schematic diagram illustrating a soft press and a hard press on the controller touch pad; -
FIG. 4 is a flow chart of example logic consistent with present principles related toFIG. 3 ; -
FIG. 5 is a combination of a logic flow chart, data structures, and processing components consistent with present principles; -
FIGS. 6-8B are schematic diagrams of a data structure referred to as the heat map inFIG. 5 , illustrating steps in use; -
FIG. 9 is a block diagram of an example neural network (NN) configured as plural long short-term memory (LSTM) networks for outputting a predicted next word based on current user input; -
FIGS. 10-12 are schematic diagrams illustrating operation of the NN inFIG. 9 post-training; -
FIG. 13 is a flow chart of example overall logic consistent with present principles; -
FIG. 14 is a schematic view of a system in which the microphone is used to input ground training to the neural networks contemporaneous with operation; -
FIG. 15 is a schematic view of a system in which the microphone is used to input tones of Chinese characters to the neural networks contemporaneous with operation; -
FIG. 16 is a schematic view of a system in which the microphone is used to input punctuation or graphics contemporaneous with operation; and -
FIG. 17 is a flow chart of example logic consistent withFIG. 15 . - Now referring to
FIG. 1 , this disclosure relates generally to computer ecosystems including aspects of computer networks that may include consumer electronics (CE) devices. A system herein may include server and client components, connected over a network such that data may be exchanged between the client and server components. The client components may include one or more computing devices including portable televisions (e.g. smart TVs, Internet-enabled TVs), portable computers such as laptops and tablet computers, and other mobile devices including smart phones and additional examples discussed below. These client devices may operate with a variety of operating environments. For example, some of the client computers may employ, as examples, operating systems from Microsoft, or a Unix operating system, or operating systems produced by Apple Computer or Google. These operating environments may be used to execute one or more browsing programs, such as a browser made by Microsoft or Google or Mozilla or other browser program that can access websites hosted by the Internet servers discussed below. - Servers and/or gateways may include one or more processors executing instructions that configure the servers to receive and transmit data over a network such as the Internet. Or, a client and server can be connected over a local intranet or a virtual private network. A server or controller may be instantiated by a game console such as a Sony PlayStation®, a personal computer, etc.
- Information may be exchanged over a network between the clients and servers. To this end and for security, servers and/or clients can include firewalls, load balancers, temporary storages, and proxies, and other network infrastructure for reliability and security.
- As used herein, instructions refer to computer-implemented steps for processing information in the system. Instructions can be implemented in software, firmware or hardware and include any type of programmed step undertaken by components of the system.
- A processor may be any conventional general-purpose single- or multi-chip processor that can execute logic by means of various lines such as address lines, data lines, and control lines and registers and shift registers.
- Software modules described by way of the flow charts and user interfaces herein can include various sub-routines, procedures, etc. Without limiting the disclosure, logic stated to be executed by a particular module can be redistributed to other software modules and/or combined together in a single module and/or made available in a shareable library. While flow chart format may be used, it is to be understood that software may be implemented as a state machine or other logical method.
- Present principles described herein can be implemented as hardware, software, firmware, or combinations thereof; hence, illustrative components, blocks, modules, circuits, and steps are set forth in terms of their functionality.
- Further to what has been alluded to above, logical blocks, modules, and circuits described below can be implemented or performed with a general-purpose processor, a digital signal processor (DSP), a field programmable gate array (FPGA) or other programmable logic device such as an application specific integrated circuit (ASIC), discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A processor can be implemented by a controller or state machine or a combination of computing devices.
- The functions and methods described below, when implemented in software, can be written in an appropriate language such as but not limited to C# or C++, and can be stored on or transmitted through a computer-readable storage medium such as a random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), compact disk read-only memory (CD-ROM) or other optical disk storage such as digital versatile disc (DVD), magnetic disk storage or other magnetic storage devices including removable thumb drives, etc. A connection may establish a computer-readable medium. Such connections can include, as examples, hard-wired cables including fiber optics and coaxial wires and digital subscriber line (DSL) and twisted pair wires.
- Components included in one embodiment can be used in other embodiments in any appropriate combination. For example, any of the various components described herein and/or depicted in the Figures may be combined, interchanged or excluded from other embodiments.
- “A system having at least one of A, B, and C” (likewise “a system having at least one of A, B, or C” and “a system having at least one of A, B, C”) includes systems that have A alone, B alone, C alone, A and B together, A and C together, B and C together, and/or A, B, and C together, etc.
- Now specifically referring to
FIG. 1 , anexample system 10 is shown, which may include one or more of the example devices mentioned above and described further below in accordance with present principles. Note that computerized devices described in all of the figures herein may include some or all of the components set forth for various devices inFIG. 1 . - The first of the example devices included in the
system 10 is a consumer electronics (CE) device configured as an example primary display device, and in the embodiment shown is an audio video display device (AVDD) 12 such as but not limited to an Internet-enabled TV with a TV tuner (equivalently, set top box controlling a TV). TheAVDD 12 may be an Android®-based system. TheAVDD 12 alternatively may also be a computerized Internet enabled (“smart”) telephone, a tablet computer, a notebook computer, a wearable computerized device such as e.g. computerized Internet-enabled watch, a computerized Internet-enabled bracelet, other computerized Internet-enabled devices, a computerized Internet-enabled music player, computerized Internet-enabled head phones, a computerized Internet-enabled implantable device such as an implantable skin device, etc. Regardless, it is to be understood that theAVDD 12 and/or other computers described herein is configured to undertake present principles (e.g. communicate with other CE devices to undertake present principles, execute the logic described herein, and perform any other functions and/or operations described herein). - Accordingly, to undertake such principles the
AVDD 12 can be established by some or all of the components shown inFIG. 1 . For example, theAVDD 12 can include one ormore displays 14 that may be implemented by a high definition or ultra-high definition “4K” or higher flat screen and that may or may not be touch-enabled for receiving user input signals via touches on the display. TheAVDD 12 may also include one ormore speakers 16 for outputting audio in accordance with present principles, and at least oneadditional input device 18 such as e.g. an audio receiver/microphone for e.g. entering audible commands to theAVDD 12 to control theAVDD 12. Theexample AVDD 12 may further include one or more network interfaces 20 for communication over at least onenetwork 22 such as the Internet, an WAN, an LAN, a PAN etc. under control of one ormore processors 24. Thus, theinterface 20 may be, without limitation, a Wi-Fi transceiver, which is an example of a wireless computer network interface, such as but not limited to a mesh network transceiver. Theinterface 20 may be, without limitation a Bluetooth transceiver, Zigbee transceiver, IrDA transceiver, Wireless USB transceiver, wired USB, wired LAN, Powerline or MoCA. It is to be understood that theprocessor 24 controls theAVDD 12 to undertake present principles, including the other elements of theAVDD 12 described herein such as e.g. controlling thedisplay 14 to present images thereon and receiving input therefrom. Furthermore, note thenetwork interface 20 may be, e.g., a wired or wireless modem or router, or other appropriate interface such as, e.g., a wireless telephony transceiver, or Wi-Fi transceiver as mentioned above, etc. - In addition to the foregoing, the
AVDD 12 may also include one ormore input ports 26 such as, e.g., a high definition multimedia interface (HDMI) port or a USB port to physically connect (e.g. using a wired connection) to another CE device and/or a headphone port to connect headphones to the AVDD 12 for presentation of audio from the AVDD 12 to a user through the headphones. For example, theinput port 26 may be connected via wire or wirelessly to a cable or satellite source 26 a of audio video content. Thus, the source 26 a may be, e.g., a separate or integrated set top box, or a satellite receiver. Or, the source 26 a may be a game console or disk player. - The
AVDD 12 may further include one ormore computer memories 28 such as disk-based or solid-state storage that are not transitory signals, in some cases embodied in the chassis of the AVDD as standalone devices or as a personal video recording device (PVR) or video disk player either internal or external to the chassis of the AVDD for playing back AV programs or as removable memory media. Also, in some embodiments, theAVDD 12 can include a position or location receiver such as but not limited to a cellphone receiver, GPS receiver and/oraltimeter 30 that is configured to e.g. receive geographic position information from at least one satellite or cellphone tower and provide the information to theprocessor 24 and/or determine an altitude at which theAVDD 12 is disposed in conjunction with theprocessor 24. However, it is to be understood that that another suitable position receiver other than a cellphone receiver, GPS receiver and/or altimeter may be used in accordance with present principles to e.g. determine the location of theAVDD 12 in e.g. all three dimensions. - Continuing the description of the
AVDD 12, in some embodiments theAVDD 12 may include one ormore cameras 32 that may be, e.g., a thermal imaging camera, a digital camera such as a webcam, and/or a camera integrated into theAVDD 12 and controllable by theprocessor 24 to gather pictures/images and/or video in accordance with present principles. Also included on theAVDD 12 may be aBluetooth transceiver 34 and other Near Field Communication (NFC)element 36 for communication with other devices using Bluetooth and/or NFC technology, respectively. An example NFC element can be a radio frequency identification (RFID) element. - Further still, the
AVDD 12 may include one or more auxiliary sensors 38 (e.g., a motion sensor such as an accelerometer, gyroscope, cyclometer, or a magnetic sensor, an infrared (IR) sensor for receiving IR commands from a remote control, an optical sensor, a speed and/or cadence sensor, a gesture sensor (e.g. for sensing gesture command), etc.) providing input to theprocessor 24. TheAVDD 12 may include an over-the-airTV broadcast port 40 for receiving OTA TV broadcasts providing input to theprocessor 24. In addition to the foregoing, it is noted that theAVDD 12 may also include an infrared (IR) transmitter and/or IR receiver and/orIR transceiver 42 such as an IR data association (IRDA) device. A battery (not shown) may be provided for powering theAVDD 12. - Still further, in some embodiments the
AVDD 12 may include a graphics processing unit (GPU) 44 and/or a field-programmable gate array (FPGA) 46. The GPU and/or FPGA may be utilized by the AVDD 12 for, e.g., artificial intelligence processing such as training neural networks and performing the operations (e.g., inferences) of neural networks in accordance with present principles. However, note that theprocessor 24 may also be used for artificial intelligence processing such as where theprocessor 24 might be a central processing unit (CPU). - Still referring to
FIG. 1 , in addition to theAVDD 12, thesystem 10 may include one or more other computer device types that may include some or all of the components shown for theAVDD 12. In one example, afirst device 48 and asecond device 50 are shown and may include similar components as some or all of the components of theAVDD 12. Fewer or greater devices may be used than shown. - The
system 10 also may include one or more servers 52. A server 52 may include at least oneserver processor 54, at least onecomputer memory 56 such as disk-based or solid state storage, and at least onenetwork interface 58 that, under control of theserver processor 54, allows for communication with the other devices ofFIG. 1 over thenetwork 22, and indeed may facilitate communication between servers, controllers, and client devices in accordance with present principles. Note that thenetwork interface 58 may be, e.g., a wired or wireless modem or router, Wi-Fi transceiver, or other appropriate interface such as, e.g., a wireless telephony transceiver. - Accordingly, in some embodiments the server 52 may be an Internet server and may include and perform “cloud” functions such that the devices of the
system 10 may access a “cloud” environment via the server 52 in example embodiments. Or, the server 52 may be implemented by a game console or other computer in the same room as the other devices shown inFIG. 1 or nearby. - The devices described below may incorporate some or all of the elements described above.
- The methods described herein may be implemented as software instructions executed by a processor, suitably configured application specific integrated circuits (ASIC) or field programmable gate array (FPGA) modules, or any other convenient manner as would be appreciated by those skilled in those art. Where employed, the software instructions may be embodied in a non-transitory device such as a CD ROM or Flash drive. The software code instructions may alternatively be embodied in a transitory arrangement such as a radio or optical signal, or via a download over the Internet.
-
FIG. 2 illustrates asystem 200 the components of which may incorporate appropriate components shown inFIG. 1 . Acomputer simulation controller 202 such as a PlayStation® controller, Xbox® controller, or other controller may include atouchpad 204 that can receive touch signals from ahand 206 and communicate via wired and/orwireless paths 208 with acomputer simulation console 210 and/or adisplay device 212 such as an Internet-enabled TV. As explained further below, the user can manipulate thetouchpad 204 to generate alpha-numeric characters 214 for presentation on thedisplay device 212 either through direct communication of signals with the display device or through thesimulation console 210. More specifically, by manipulating thetouchpad 204, a user can move a screen cursor over a letter on avirtual keyboard 216 presented on thedisplay device 212 to enter the alpha-numeric characters 214. Thevirtual keyboard 216 may have, without limitation, a QWERTY layout. - Additionally, the
controller 200 may include one ormore microphones 218 communicating with the processor of the controller for purposes disclosed below. In the example shown, themicrophone 218 is provided on thetouchpad 204, although it is to be understood that themicrophone 218 may be provided elsewhere on the housing of thecontroller 200 or indeed on another component if desired. - As shown schematically in
FIG. 3 , present principles contemplate two types of touch, namely, a “soft” press 300 (using a soft pressure on the touchpad or a hover over the touchpad with zero pressure), in which a screen cursor on thedisplay device 212 is moved to desired locations on thevirtual keyboard 216 without sending location data (i.e., a signal indicating selection of any particular virtual key) to the display device, and a “hard”press 302 of greater pressure than a soft press, in response to which a screen cursor on thedisplay device 212 may be moved and location data sent to the display device to indicate selection of a virtual key. In this way, a user can look away from thetouchpad 204 and view thevirtual keyboard 216 while moving his or her finger across the touchpad to move a visible screen cursor to a desired letter on the virtual keyboard, and then exert a hard press to select that letter. Note that an individual “next” letter may not be presented on the display, but rather the next “most possible word” may be displayed after a user has finished a “swipe”. The “hottest” key (based on the heatmap) may be highlighted on the virtual keyboard as well as the trace. In addition, a “swipe” is defined as a continuous hard-press which forms a trace. -
FIG. 4 illustrates example logic with the above description in mind. The logic may be executed by one or more of a processor in thesimulation controller 202, a processor in thesimulation console 210, and a processor in thedisplay device 212. - Commencing at
state 400 it is determined whether a press of thetouchpad 204 has been received. This may be done by determining whether signals from one or more proximity sensors associated with thetouchpad 204 indicate a hover of a finger adjacent thetouchpad 204 and/or by determining whether signals from one or more pressure sensors associated with thetouchpad 204 indicate a pressure of at least a first threshold pressure. - When it is determined that a touch has been received, the logic proceeds to
state 402 to determine whether the touch is a soft press or hard press as indicated by, e.g., signals from a pressure sensor associated with thetouchpad 204 indicating a touch of at least a threshold pressure, which is typically set to be greater than any threshold pressure used atstate 400. If the touch does not satisfy the threshold, the logic moves to block 404 to return a soft press. In some implementations the logic may proceed tostate 406 to determine whether the soft press is the first soft press within, e.g., a threshold of period, for example within the last five minutes, and if so the logic can move to block 408 to enlarge an image of thevirtual keyboard 216 on thedisplay device 212. In any case, fromstate 406 if the test there is negative or fromblock 408, the logic moves to block 410 to move the screen cursor without sending press location information. - On the other hand, if the test at
state 402 determines that a hard press is received, such is returned atblock 412, and the screen may be moved according to the touch with location information being sent as well indicating the location of the virtual keyboard the user has selected by means of the hard press on thetouchpad 204 of thesimulation controller 200. -
FIG. 5 illustrates a combination of hardware and software blocks alluded to above. One or more proximity and/orpressure sensors 500 are provided in thetouchpad 204 to output signals representingsoft presses 502 and hard presses 504. The soft presses 502 establish finger focus points 506. Thehard presses 504 establish points on the touchpad as detected by the sensor(s) 500. A soft-press represents a cursor focus point, while “points by sensor” means “continuous points sending by the sensor”. - At 510 a heatmap algorithm, discussed further below in reference to
FIGS. 6-8 , is accessed to output a sequence ofletters 512 according to the hard presses 504. The sequence ofletters 512 is input along with adictionary 514 to areduction block 516 that reduces the list of candidates that might possibly form either a correction to or a completion of the sequence ofletters 512. Thedictionary 514 is essentially a dictionary and/or thesaurus of sequences of letters that can be used to correct a mis-typed word, e.g., thedictionary 514 may correlate “thw” to “the” to return the word “the” in response to input of “thw”. - The reduced list of
candidates 516 is provided to amodule 518 that outputs a predicted network or words for presentation on the screen, which a user can then select to complete his or her desired input without typing every letter of the predicted word or words. Themodule 518 may be established by one or more neural networks (NN) as described further below. To produce a predicted word or words, themodule 518 may receive input from acontextual user block 520, which provides previous word strings employed by the user with the current input inferred to possibly be a repeat of a prior input, e.g., “do you” may have been followed multiple times in prior inputs by “know what I mean”, and this information can be input to help train and execute themodule 518. - Moreover, similar training/execution aids may be input to the
module 518 as shown at the right ofFIG. 5 . Specifically, queries and chatdata 522 from other computer gamers may be input to a character-based NN such as a bidirectional long short-term memory (BILSTM) 524 to learn patterns of common input strings for provision to a machine learningcharacter sequence model 526. Thismodel 526 may be input to or accessed by themodule 518 in rendering a next predicted word or words. -
FIGS. 6-8B illustrate employment of theheatmap algorithm 510 inFIG. 5 . Basically, the “path” or “connected points” of the finger “swipe” (hard-press) and the probabilities of each letter are “discounted and accumulated” at certain time interval along the swipe. At each time interval, the letter with the highest probability is extracted, which may also have to pass a certain threshold to add to the sequence as developed further below. - In
FIGS. 6-8B , it is to be understood that only the first four letters in the top left corner of a QWERTY keyboard (i.e., Q, next to which is “W”, and below which from left to right are “A” and “S”) are shown for clarity of disclosure, as but one example of a possible virtual keyboard layout for thevirtual keyboard 216. In theexample heatmap 510 illustrated, each area of the heatmap for a particular letter is divided into a three-by-three grid for nine divisions (illustrated as geometric squares) total, with thecenter division 600 for a particular letter indicating that the probability of that letter being desired when a cursor is in the center area being 1. In contrast, theheatmap 510 indicates probabilities less than one but greater than zero in theborder divisions 602 that surround thecenter division 600 of a letter, with the probabilities being associated with the letter of thecenter division 600 and the letter(s) immediately adjacent the border divisions 602 (or, in the case of a border division that is not adjacent another letter, only a probability less than one for the letter of the center division). - As shown in
FIG. 7 at 700, a soft press is used to locate the starting letter of an intended input. Then, as shown at 800 inFIG. 8 , a hard press is used to indicate selection of the starting letter, in the example shown, “Q”. This causes the collection of data that “Q” is selected with a probability of one and that surrounding letters (in the example shown, “W”, “A”, and “S”) are not selected, i.e., have a probability of zero. -
FIGS. 8A and 8B illustrate the results of an ensuing swipe. InFIG. 8A a swipe is shown at 802 from the location starting inFIG. 8 to thelocation 804 indicated by the image of the hand. Here, the user has moved his finger toward the letter “A”. This causes new heatmap statistics to be aggregated according to the path of the swipe over theborder divisions 602 using the algorithm shown inFIG. 8A . Because the probability of “Q” is higher than the probabilities of “W” (which is zero), “A” (which is 0.3), and “S” (which is zero), the sequence returns “Q”. -
FIG. 8B shows at 806 that the swipe has been continued to thelocation 808 shown by the image of the hand. This causes further heatmap statistics to be aggregated according to the path of the swipe over theborder divisions 602 using the algorithm shown inFIG. 8B . Because the probability of “A” is higher than the probabilities of “W” (which is zero), “Q” (which is 0.3), and “S” (which is zero), the sequence returns “A” to be appended after “Q” was returned inFIG. 8A , resulting in a sequence “QA”. - Thus, it may now be appreciated that the “path” or “connected points” of the finger “swipe” (hard-press) is tracked and the probabilities of each letter are discounted and accumulated at certain time intervals along the swipe. At each time interval, the letter with the highest probability is extracted, in some embodiments provided the probability of the letter satisfies a threshold probability (e.g., of 0.4) to be added to the sequence.
-
FIG. 9 illustrates an example NN architecture that may be used in any of the NN-based modules of, e.g.,FIG. 5 . Anetwork 900 of NN may receiveinput letters 902 withprobabilities 904 from the heatmap to output time-distributed predictedletters 906 with associatedprobabilities 908. In the example shown, eachletter 902 may be input to a respective recurrent NN (RNN) such as a sequence of long short-term memory (LSTM) 910 as shown. AnLSTM 910 as shown at the right inFIG. 9 may include aninput gate 912, a forgetgate 914, and anoutput gate 916, all of which may execute a sigmoid function as indicated by the Greek letter σ inFIG. 9 . Theinput gate 912 controls the extent to which a new value flows into the cell, the forgetgate 914 controls the extent to which a value remains in the cell and theoutput gate 916 controls the extent to which the value in the cell is used to compute the output activation of the LSTM unit. - The current value xi being input and the hidden state ht-1 from the previous iteration are input to all three gates as shown. The output of the sigmoid function of the
input gate 912 may be combined with a hyperbolictangent function 918 at afirst combine operator 920, which may be an element-wise product. The output of thefirst combine operator 920 is combined, as by summing if desired, with the output of ssecond combine operator 922 at athird combine operator 924. The output of thethird combine operator 924 may be fed back to thesecond combine operator 922 for combining with the output of the forgetgate 914. Further, the output of thethird combine operator 924 may be operated on if desired by a hyperbolictangent function 926 and then combined at afourth combine operator 928 with the output of theoutput gate 916 to render ahidden state vector 930 for use in the succeeding iteration. -
FIGS. 10-12 illustrate a sequence of the use of thenetwork 900 to generate predicted text. The lower row ofletters 1000 represents input received from hard presses on keys of thevirtual keyboard 216 and/or from selection of previously predicted letters and/or words. These are input to the trainednetwork 900. Using probabilities correlated with letters from the heatmap as illustrated at 1002, a next predictedletter 1004 is generated and fed back to the model. The sequence shown inFIGS. 10-12 generated predicted letters for an initial input of “play” that results in the word “PlayStation”. -
FIG. 13 is a flow chart of example logic consistent with present principles. The NN system(s) described herein are trained atblock 1300. Moving to block 1302, a hard press is received on the touchpad and a letter established based thereon atblock 1304 using the heatmap if desired. The letter is input to the NN system atblock 1306, which outputs a predicted letter or words or string of words atblock 1308. The predicted letters/words are presented on screen atblock 1310. - If a user does not accept the predictions at
state 1312, they may be removed from presentation atstate 1314. Otherwise, accepted predictions are confirmed atblock 1316 and presented in sequence after the letters established by the hard press. - Present principles may be used in all possible deep learning-based methods for image, video and audio data processing, among others.
- Note that a user can indicate acceptance at
state 1312 by speaking into themicrophone 218 illustrated inFIG. 2 . For example, the user may speak “OK” to accept the predicted word or “not right” or equivalent to reject it. This input is provided to the neural networks described herein as ground truth data for training the neural networks. -
FIG. 14 illustrates using the components discussed above in relation toFIG. 2 . A predictedword 1400 has been presented on thedisplay 212 according to description above. Auser 1402 may speak into themicrophone 218 as indicated at 1404 whether to accept or reject the predictedword 1400 contemporaneous with operation. In the example shown, the predicted word is “pony”, and theuser 1402 has rejected it by speaking “no” into themicrophone 218, which is digitized and provided as training data to the neural networks. Theuser 1402 has further input a correct word, in this case, “cow”, which also is provided as ground truth for training the neural networks. -
FIG. 15 illustrates a second use case for themicrophone 218. InFIG. 15 , assume that the user has typed in or the neural networks have predicted one or more Asian language characters such asChinese Pinyin characters 1500 for presentation on thedisplay 212. For illustration, the characters are simply numbered 1-4 inFIG. 15 , and correspond to spoken words that may be rendered in English using identical Arabic letters, in the example shown, the letters “ma”, and thus may be indistinguishable from each other when so rendered in English. Below eachcharacter 1500 theword 1502 it corresponds to in English is shown and below that, arespective symbol 1504 for the corresponding tone, it being understood that in implementation neither theword 1502 norsymbol 1504 may be presented. Note that while Chinese pinyin is used as an example of Asian language characters, present principles apply to other Asian languages such as Japanese in which the same Arabic letter string may be translated into two or more Asian language words differentiated from each other by tonal differences and not consonant or vowel differences. - A
user 1506 may speak a word including an intendedAsian language tone 1508 into themicrophone 218, which is input to the processor(s) herein as the correct or ground truth tone. In the example shown, theuser 1506 has spoken the word using the tone corresponding to thethird character 1500, giving the character the meaning “horse”. In this way, the user may have input, via thecontroller 200, the Arabic letters “ma” as indicated at 1510. TheArabic letters 1510 may be correlated to plural candidate Chinese words, which may be presented as respective Asian language characters if desired on thedisplay 212. Tonal input from themicrophone 218 is used to confirm and/or select which of the candidate characters words the user intended by typing in “ma”, which may then be presented on the display in lieu of the other candidate words/symbols. -
FIG. 16 illustrates that during a touchpad swipe as described above, input from themicrophone 218 may be used to input special characters/punctuations/graphics such as “smileys”. Assume that the user has input, or the neural networks have predicted, aword 1600, in the example shown, the word “yes”, that appears on thedisplay 212. A user 1602 contemporaneously may speak theword 1600 with a tone that is detected by the neural networks to correspond to excitement as indicated at 1604, to cause anexclamation point 1606 to appear after theword 1600. - Or, the user may utter, as indicated at 1608, the name of the desired symbol or punctuation to cause the uttered symbol to be presented on the
display 212. Yet again, the user may utter, as indicated at 1610, the name of a desired graphic symbol such as “smiley”, to cause the uttered graphic symbol to be presented at 1612 on thedisplay 212. -
FIG. 17 illustrates example logic consistent withFIG. 15 . Commencing atblock 1700, input may be received, e.g., from the touch pad of the controller or from other manipulable device, indicating Arabic letters and/or Chinese character(s).Block 1702 indicates that such input may be ambiguous in that it may correlate to more than one candidate Asian language word with corresponding symbol, in which case the logic flows to block 1704. - For instance, and using Chinese as an example, when input is received of Arabic letters, such as the letters “ma” discussed previously, more than one pinyin symbol (corresponding to more than one Chinese word) may be a candidate for the user's intent. Or, when touch input is received on the touch pad attempting to render a Chinese character, owing to imprecision in the touch tracing more than one Chinese character might be implicated.
- At
block 1704, using a dictionary (for example, correlating Arabic “ma” to the four pinyin characters described in reference toFIG. 15 ) the candidate pinyin symbols/corresponding Chinese words may be identified. In addition, or alternatively, handwriting recognition may be employed in the case of, e.g., an attempted trace of a Chinese character using the touch pad, to identify candidate Chinese characters that may fulfill the user's intent. - Proceeding to block 1706, a user may be prompted to speak to resolve the ambiguity and essentially select the user-preferred candidate symbol/word from
block 1704.FIG. 15 illustrates a non-limiting example screen shot showing such a prompt. - It will be appreciated that whilst present principals have been described with reference to some example embodiments, these are not intended to be limiting, and that various alternative arrangements may be used to implement the subject matter claimed herein.
Claims (13)
1. An apparatus, comprising:
at least one computer storage that is not a transitory signal and that comprises instructions executable by at least one processor to:
identify a first touch event as a soft event;
responsive to identifying the first touch event as a soft event, move a screen cursor on a display to a first location on a virtual keyboard without indicating selection of a key on the virtual keyboard to the display;
identify a second touch event as a hard event;
responsive to identifying the second touch event as a hard event, indicate selection of a key of the virtual keyboard to the display.
2. The apparatus of claim 1 , wherein a user can look away from a touchpad on which the first and second touch events are received and view the virtual keyboard on the display while moving a finger across the touchpad to move the cursor to a desired letter on the virtual keyboard, and then exert a hard event to select the letter desired.
3. The apparatus of claim 1 , wherein the instructions are executable to:
present on the display a next most possible word responsive to a swipe established by a continuous hard event forming a trace.
4. The apparatus of claim 3 , wherein the instructions are executable to:
highlight on the display a hot key based on a heatmap; and
highlight the trace.
5. The apparatus of claim 1 , wherein the instructions are executable to:
responsive to determining that the soft event is an initial soft event within a threshold of period, enlarge the virtual keyboard on the display.
6. The apparatus of claim 3 , wherein the instructions are executable to:
discount and accumulate points of the swipe and probabilities of each letter at time intervals along the swipe; and
extract for presentation on the display a letter with a highest probability.
7. The apparatus of claim 1 , comprising the at least one processor and the display.
8. A method, comprising:
identifying a first touch event as a soft event;
responsive to identifying the first touch event as a soft event, moving a screen cursor on a display to a first location on a virtual keyboard without indicating selection of a key on the virtual keyboard to the display;
identifying a second touch event as a hard event;
responsive to identifying the second touch event as a hard event, indicating selection of a key of the virtual keyboard to the display.
9. The method of claim 8 , wherein a user can look away from a touchpad on which the first and second touch events are received and view the virtual keyboard on the display while moving a finger across the touchpad to move the cursor to a desired letter on the virtual keyboard, and then exert a hard event to select the letter desired.
10. The method of claim 8 , comprising:
presenting on the display a next most possible word responsive to a swipe established by a continuous hard event forming a trace.
11. The method of claim 10 , comprising:
highlighting on the display a hot key based on a heatmap; and
highlighting the trace.
12. The method of claim 8 , comprising:
responsive to determining that the soft event is an initial soft event within a threshold of period, enlarge the virtual keyboard on the display.
13. The method of claim 10 , comprising:
discount and accumulate points of the swipe and probabilities of each letter at time intervals along the swipe; and
extract for presentation on the display a letter with a highest probability.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/194,225 US20210200373A1 (en) | 2019-07-31 | 2021-03-06 | Microphone on controller with touchpad to take in audio swipe feature data |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/527,885 US11048356B2 (en) | 2019-07-31 | 2019-07-31 | Microphone on controller with touchpad to take in audio swipe feature data |
US17/194,225 US20210200373A1 (en) | 2019-07-31 | 2021-03-06 | Microphone on controller with touchpad to take in audio swipe feature data |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/527,885 Division US11048356B2 (en) | 2019-07-31 | 2019-07-31 | Microphone on controller with touchpad to take in audio swipe feature data |
Publications (1)
Publication Number | Publication Date |
---|---|
US20210200373A1 true US20210200373A1 (en) | 2021-07-01 |
Family
ID=74230500
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/527,885 Active US11048356B2 (en) | 2019-07-31 | 2019-07-31 | Microphone on controller with touchpad to take in audio swipe feature data |
US17/194,225 Abandoned US20210200373A1 (en) | 2019-07-31 | 2021-03-06 | Microphone on controller with touchpad to take in audio swipe feature data |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/527,885 Active US11048356B2 (en) | 2019-07-31 | 2019-07-31 | Microphone on controller with touchpad to take in audio swipe feature data |
Country Status (2)
Country | Link |
---|---|
US (2) | US11048356B2 (en) |
WO (1) | WO2021021414A1 (en) |
Family Cites Families (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5680510A (en) * | 1995-01-26 | 1997-10-21 | Apple Computer, Inc. | System and method for generating and using context dependent sub-syllable models to recognize a tonal language |
US20090213081A1 (en) | 2007-01-10 | 2009-08-27 | Case Jr Charlie W | Portable Electronic Device Touchpad Input Controller |
US8542927B2 (en) * | 2008-06-26 | 2013-09-24 | Microsoft Corporation | Character auto-completion for online east asian handwriting input |
US20100138221A1 (en) * | 2008-12-02 | 2010-06-03 | Boys Donald R | Dedicated hardware/software voice-to-text system |
US8754746B2 (en) * | 2009-11-16 | 2014-06-17 | Broadcom Corporation | Hand-held gaming device that identifies user based upon input from touch sensitive panel |
US20170017393A1 (en) | 2010-04-23 | 2017-01-19 | Handscape Inc., A Delaware Corporation | Method for controlling interactive objects from a touchpad of a computerized device |
US9626354B2 (en) * | 2014-01-21 | 2017-04-18 | Lenovo (Singapore) Pte. Ltd. | Systems and methods for using tone indicator in text recognition |
KR102204919B1 (en) | 2014-06-14 | 2021-01-18 | 매직 립, 인코포레이티드 | Methods and systems for creating virtual and augmented reality |
CN108604227B (en) | 2016-01-26 | 2023-10-24 | 皇家飞利浦有限公司 | System and method for neural clinical paraphrasing generation |
US10855550B2 (en) * | 2016-11-16 | 2020-12-01 | Cisco Technology, Inc. | Network traffic prediction using long short term memory neural networks |
US11087210B2 (en) | 2017-08-18 | 2021-08-10 | MyFitnessPal, Inc. | Context and domain sensitive spelling correction in a database |
-
2019
- 2019-07-31 US US16/527,885 patent/US11048356B2/en active Active
-
2020
- 2020-07-09 WO PCT/US2020/041422 patent/WO2021021414A1/en active Application Filing
-
2021
- 2021-03-06 US US17/194,225 patent/US20210200373A1/en not_active Abandoned
Also Published As
Publication number | Publication date |
---|---|
US11048356B2 (en) | 2021-06-29 |
US20210034187A1 (en) | 2021-02-04 |
WO2021021414A1 (en) | 2021-02-04 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10275022B2 (en) | Audio-visual interaction with user devices | |
US12118683B2 (en) | Content creation in augmented reality environment | |
US20180143965A1 (en) | Trained data input system | |
KR20180102148A (en) | Search for shape symbols within the graphic keyboard | |
US11734899B2 (en) | Headset-based interface and menu system | |
US11709593B2 (en) | Electronic apparatus for providing a virtual keyboard and controlling method thereof | |
US10754441B2 (en) | Text input system using evidence from corrections | |
KR102416421B1 (en) | Game controller with touchpad input | |
JP2021531589A (en) | Motion recognition method, device and electronic device for target | |
CN111108468A (en) | System and method for determining input characters based on sliding input | |
KR102456509B1 (en) | Electronic apparatus, method for controlling thereof and the computer readable recording medium | |
KR20240124243A (en) | Electronic apparatus and control method thereof | |
US20210072890A1 (en) | Systems and methods for operating a mobile application using a communication tool | |
CN110554780A (en) | sliding input method and device | |
WO2017203764A1 (en) | Information processing device and information processing method | |
US11048356B2 (en) | Microphone on controller with touchpad to take in audio swipe feature data | |
US12020683B2 (en) | Real-time name mispronunciation detection | |
US11471765B2 (en) | Handwriting recognition using cardinal directions between sample points | |
CN115494965B (en) | Request sending method and device for sending request | |
US11317293B2 (en) | Methods for authenticating a user of an electronic device | |
CN115543099A (en) | Input method, device and device for input | |
CN115509371A (en) | Key identification method and device for identifying keys | |
KR20210086384A (en) | Method and apparatus for providing a learning service using a character containing at least one letter | |
CN117812323A (en) | Display device, voice recognition method, voice recognition device and storage medium | |
CN115512386A (en) | Gesture-based interaction method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: ADVISORY ACTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |