US9977595B2 - Keyboard with a suggested search query region - Google Patents

Keyboard with a suggested search query region Download PDF

Info

Publication number
US9977595B2
US9977595B2 US15/299,027 US201615299027A US9977595B2 US 9977595 B2 US9977595 B2 US 9977595B2 US 201615299027 A US201615299027 A US 201615299027A US 9977595 B2 US9977595 B2 US 9977595B2
Authority
US
United States
Prior art keywords
computing device
input
query suggestion
search
keys
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US15/299,027
Other versions
US20170308292A1 (en
Inventor
Min-sang Choi
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Google LLC
Original Assignee
Google LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Google LLC filed Critical Google LLC
Priority to US15/299,027 priority Critical patent/US9977595B2/en
Assigned to GOOGLE INC. reassignment GOOGLE INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHOI, MIN-SANG
Assigned to GOOGLE LLC reassignment GOOGLE LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: GOOGLE INC.
Publication of US20170308292A1 publication Critical patent/US20170308292A1/en
Application granted granted Critical
Publication of US9977595B2 publication Critical patent/US9977595B2/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/04842Selection of displayed objects or displayed text elements
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/20Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
    • G06F16/24Querying
    • G06F16/248Presentation of query results
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3322Query formulation using system suggestions
    • G06F16/3323Query formulation using system suggestions using document space presentation or visualization, e.g. category, hierarchy or range presentation and selection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/9032Query formulation
    • G06F16/90324Query formulation using system suggestions
    • G06F17/30554
    • G06F17/3097
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/02Input arrangements using manually operated switches, e.g. using keyboards or dials
    • G06F3/023Arrangements for converting discrete items of information into a coded form, e.g. arrangements for interpreting keyboard generated codes as alphanumeric codes, operand codes or instruction codes
    • G06F3/0233Character input methods
    • G06F3/0237Character input methods using prediction or retrieval techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0484Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
    • G06F3/0485Scrolling or panning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06F3/04886Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures by partitioning the display area of the touch-screen or the surface of the digitising tablet into independently controllable areas, e.g. virtual keyboards or menus
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/12Use of codes for handling textual entities
    • G06F40/134Hyperlinking
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/274Converting codes to words; Guess-ahead of partial word inputs
    • G06F17/211
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/10Text processing
    • G06F40/103Formatting, i.e. changing of presentation of documents

Definitions

  • GUI graphical user interface
  • a user of a mobile computing device may have to switch between different application GUIs. For example, a user of a mobile computing device may have to cease entering text in a messaging application and provide input to cause the device to toggle to a search application to search for a particular piece of information to use when composing a message or otherwise entering text.
  • Providing several inputs required by some computing devices to perform various tasks can be tedious, repetitive, and time consuming.
  • a method includes outputting, by a mobile computing device, for display, a graphical keyboard comprising a plurality of keys and a search element, determining, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion. The method further includes receiving, by the mobile computing device, an indication of input selecting the search element, and responsive to receiving the indication of input, outputting, by the mobile computing device, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
  • a mobile device comprises a presence-sensitive display component, at least one processor, and a memory that stores instructions that when executed cause the at least one processor to: output, for display, a graphical keyboard comprising a plurality of keys and a search element, determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion, receive an indication of input selecting the search element, and responsive to receiving the indication of input, output, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
  • a computer-readable storage medium comprises instructions that, when executed, cause at least one processor of a computing device to: output, for display, an application graphical user interface comprising an edit region, a plurality of keys, and a search element, determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion, receive an indication of input selecting the search element, and responsive to receiving the indication of input, output, for display, an updated application graphical user interface that includes a selectable link associated with the query suggestion, wherein the selectable link is positioned between the edit region and the plurality of keys.
  • FIG. 1 is a conceptual diagram illustrating an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIG. 2 is a block diagram illustrating an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIG. 3 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure.
  • FIGS. 4A-4D are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIGS. 5A-5F are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIGS. 6A-6C are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIG. 7 is a flowchart illustrating example operations of a computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • FIG. 8 is a flowchart illustrating example operations of a computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • this disclosure is directed to techniques that include enabling a computing device to display suggested search queries near the keys of a graphical keyboard while a user provides input at the keys of the graphical keyboard.
  • the computing device may output a graphical user interface (GUI) that includes a graphical keyboard.
  • GUI graphical user interface
  • the graphical keyboard may suggest a search query, for example, based on text determined from user input as the user interacts with the graphical keyboard.
  • the graphical keyboard may generate suggested queries by analyzing other information such as the user's current or recent tasks, context, or prior interaction with the graphical keyboard.
  • the suggested queries may therefore be relevant to the user's current or recent tasks in addition to the text inferred from user input, and may be updated in response to further typing or other input or actions detected by the computing device.
  • the graphical keyboard may display the suggested queries within the graphical keyboard, and in some examples, near a particular text field or other location of the graphical keyboard that has a current input focus. For example, the graphical keyboard may display a suggested query between the graphical keys and the edit region at which the user is currently typing. This arrangement may make it easier for the user to stay on task and easily view, and provide additional input to select suggested queries. In situations where the GUI has more than one text field that could have input focus, the graphical keyboard may display suggested queries above the particular text field that has the input focus, and below the other text field(s). Visual space created between the text fields in such an arrangement may reduce user confusion as to which text field has input priority.
  • the graphical keyboard may execute a search based on a suggested query and display search results within the graphical keyboard, in some examples, replacing some or all keys within the graphical keyboard with search results.
  • a graphical keyboard may allow a user to view suggested queries and obtain search results based on suggested queries without, in some cases, causing the user to lose focus, switch tasks, or switch between application user interfaces.
  • a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information.
  • information e.g., context, locations, speeds, search queries, etc.
  • the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user.
  • certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed.
  • a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined.
  • location information such as to a city, ZIP code, or state level
  • FIG. 1 is a conceptual diagram illustrating an example computing device 110 that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • Computing device 110 may represent a mobile device, such as a smart phone, a tablet computer, a laptop computer, computerized watch, or any other types of wearable and non-wearable, mobile or non-mobile computing devices that may output a graphical keyboard for display.
  • Computing device 110 includes a presence-sensitive display (PSD) 112 , user interface (UI) module 120 and keyboard module 122 .
  • Modules 120 and 122 may perform operations described using software, hardware, firmware, or a mixture of hardware, software, and firmware residing in and/or executing at computing device 110 .
  • One or more processors of computing device 110 may execute instructions that are stored at a memory or other non-transitory storage medium of computing device 110 to perform the operations of modules 120 and 122 .
  • Computing device 110 may execute modules 120 and 122 as virtual machines executing on underlying hardware.
  • Modules 120 and 122 may execute as one or more services of an operating system or computing platform.
  • Modules 120 and 122 may execute as one or more executable programs at an application layer of a computing platform.
  • PSD 112 of computing device 110 may function as respective input and/or output devices for computing device 110 .
  • PSD 112 may be implemented using various technologies. For instance, PSD 112 may function as input devices using presence-sensitive input screens, such as resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, projective capacitance touchscreens, pressure sensitive screens, acoustic pulse recognition touchscreens, or another presence-sensitive display technology.
  • presence-sensitive input screens such as resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, projective capacitance touchscreens, pressure sensitive screens, acoustic pulse recognition touchscreens, or another presence-sensitive display technology.
  • PSD 112 may also function as output (e.g., display) devices using any one or more display devices, such as liquid crystal displays (LCD), dot matrix displays, light emitting diode (LED) displays, organic light-emitting diode (OLED) displays, e-ink, or similar monochrome or color displays capable of outputting visible information to a user of computing device 110 .
  • display devices such as liquid crystal displays (LCD), dot matrix displays, light emitting diode (LED) displays, organic light-emitting diode (OLED) displays, e-ink, or similar monochrome or color displays capable of outputting visible information to a user of computing device 110 .
  • PSD 112 may detect input (e.g., touch and non-touch input) from a user of respective computing device 110 .
  • PSD 112 may detect indications of input by detecting one or more gestures from a user (e.g., the user touching, pointing, and/or swiping at or near one or more locations of PSD 112 with a finger or a stylus pen).
  • PSD 112 may output information to a user in the form of a user interface (e.g., user interface 114 ), which may be associated with functionality provided by computing device 110 .
  • PSD 112 may present user interface 114 which, as shown in FIG. 1 , is a graphical user interface of a chat application executing at computing device 110 and includes various graphical elements displayed at various locations of PSD 112 .
  • user interface 114 is a chat user interface.
  • user interface 114 may be any graphical user interface which includes (or accesses the services of) a graphical keyboard with integrated search features.
  • User interface 114 includes output region 116 A, graphical keyboard 116 B, and edit region 116 C.
  • a user of computing device 110 may provide input at graphical keyboard 116 B to produce textual characters within edit region 116 C that form the content of the electronic messages displayed within output region 116 A.
  • the messages displayed within output region 116 A form a chat conversation between a user of computing device 110 and a user of a different computing device.
  • region 116 A is a user interface for a first application that is separate and distinct from a keyboard application that provides the graphical keyboard 116 B.
  • the first application utilizes the services of the keyboard application for certain input and output functions, including text and symbol entry, text and symbol display, text and symbol entry editing, attachment selection, process invocation (such as spell checking, font and style changes, etc.), transmission and receipt controls, etc.
  • the keyboard application can implement search functionality, and in some examples provide that search functionality, or results of that search functionality, to the first application.
  • UI module 120 manages user interactions with PSD 112 and other components of computing device 110 .
  • UI module 120 may act as an intermediary between various components of computing device 110 to make determinations based on user input detected by PSD 112 and generate output at PSD 112 in response to the user input.
  • UI module 120 may receive instructions from an application, service, platform, or other module of computing device 110 to cause PSD 112 to output a user interface (e.g., user interface 114 ).
  • UI module 120 may manage inputs received by computing device 110 as a user views and interacts with the user interface presented at PSD 112 and update the user interface in response to receiving additional instructions from the application, service, platform, or other module of computing device 110 that is processing the user input.
  • Keyboard module 122 represents an application, service, or component executing at or accessible to computing device 110 that provides computing device 110 with a graphical keyboard having integrated search features. Keyboard module 122 may switch between operating in text-entry mode in which keyboard module 122 functions similar to a traditional graphical keyboard, or search mode in which keyboard module 122 performs various integrated search functions.
  • keyboard module 122 may be a stand-alone application, service, or module executing at computing device 110 and in other examples, keyboard module 122 may be a sub-component of an operating system controlling operation of device 110 .
  • keyboard module 122 may be integrated into a chat or messaging application executing at computing device 110 whereas in other examples, keyboard module 122 may be a stand-alone application or subroutine that is invoked by an application or operating platform of computing device 110 any time an application or operating platform requires graphical keyboard input functionality.
  • computing device 110 may download and install keyboard module 122 from an application repository of a service provider (e.g., via the Internet). In other examples, keyboard module 122 may be preloaded as part of the operating system of computing device 110 .
  • keyboard module 122 of computing device 110 may perform traditional, graphical keyboard operations used for text-entry, such as: generating a graphical keyboard layout for display at PSD 112 , mapping detected inputs at PSD 112 to selections of graphical keys, determining characters based on selected keys, or predicting or autocorrecting words and/or phrases based on the characters determined from selected keys.
  • Graphical keyboard 116 B includes graphical elements displayed as graphical keys 118 A.
  • Keyboard module 122 may output information to UI module 120 that specifies the layout of graphical keyboard 116 B within user interface 114 .
  • the information may include instructions that specify locations, sizes, colors, and other characteristics of graphical keys 118 A.
  • UI module 120 may cause PSD 112 to display graphical keyboard 116 B as part of user interface 114 .
  • Each key of graphical keys 118 A may be associated with a respective character (e.g., a letter, number, punctuation, or other character) displayed within the key.
  • a user of computing device 110 may provide input at locations of PSD 112 at which one or more of graphical keys 118 A are displayed to input content (e.g., characters, search results, etc.) into edit region 116 C (e.g., for composing messages that are sent and displayed within output region 116 A or for inputting a search query that computing device 110 executes from within graphical keyboard 116 B).
  • Keyboard module 122 may receive information from UI module 120 indicating locations associated with input detected by PSD 112 that are relative to the locations of each of the graphical keys. Using a spatial and/or language model, keyboard module 122 may translate the inputs to selections of keys and characters, words, and/or phrases.
  • PSD 112 may detect user inputs as a user of computing device 110 provides the user inputs at or near a location of PSD 112 where PSD 112 presents graphical keys 118 A.
  • UI module 120 may receive, from PSD 112 , an indication of the user input detected by PSD 112 and output, to keyboard module 122 , information about the user input.
  • Information about the user input may include an indication of one or more touch events (e.g., locations and other information about the input) detected by PSD 112 .
  • keyboard module 122 may map detected inputs at PSD 112 to selections of graphical keys 118 A, determine characters based on selected graphical keys 118 A, and predict or autocorrect words and/or phrases determined based on the characters associated with the selected keys 118 A.
  • keyboard module 122 may include a spatial model that may determine, based on the locations of keys 118 A and the information about the input, the most likely one or more keys 118 A being selected. Responsive to determining the most likely one or more keys 118 A being selected, keyboard module 122 may determine one or more characters, words, and/or phrases.
  • each of the one or more keys 118 A being selected from a user input at PSD 112 may represent an individual character or a keyboard operation.
  • Keyboard module 122 may determine a sequence of characters selected based on the one or more selected keys 118 A.
  • keyboard module 122 may apply a language model to the sequence of characters to determine one or more the most likely candidate letters, morphemes, words, and/or phrases that a user is trying to input based on the selection of keys 118 A.
  • Keyboard module 122 may send the sequence of characters and/or candidate words and phrases to UI module 120 and UI module 120 may cause PSD 112 to present the characters and/or candidate words determined from a selection of one or more keys 118 A as text within edit region 116 C.
  • keyboard module 122 may cause UI module 120 to display the candidate words and/or phrases as one or more selectable spelling corrections and/or selectable word or phrase suggestions within suggestion region 118 B.
  • keyboard module 122 of computing device 110 also provides integrated search capability. That is, rather than requiring a user of computing device 110 to navigate away from user interface 114 which provides graphical keyboard 116 B (e.g., to a different application or service executing at or accessible from computing device 110 ), keyboard module 122 may operate in search mode in which keyboard module 122 may invoke search operations and present search results within the same region of PSD 112 at which graphical keyboard 116 B is displayed. Keyboard module 122 may include routines for executing search functions and/or may include interfaces for communicating with a separate search engine or resource that receives search parameters from keyboard module 122 and returns search results. Search functionality invoked by keyboard module 122 may search memory of device 110 and/or external memory.
  • keyboard module 122 may execute as a stand-alone application, service, or module executing at computing device 110 or as a single, integrated sub-component thereof. Therefore, if keyboard module 122 forms part of a chat or messaging application executing at computing device 110 , keyboard module 122 may provide the chat or messaging application with text-entry capability as well as search capability. Similarly, if keyboard module 122 is a stand-alone application or subroutine that is invoked by an application or operating platform of computing device 110 any time an application or operating platform requires graphical keyboard input functionality, keyboard module 122 may provide the invoking application or operating platform with text-entry capability as well as search capability.
  • Keyboard module 122 may further operate in search mode.
  • keyboard module 122 may cause graphical keyboard 116 B to include search element 118 C.
  • Search element 118 C represents a selectable element of graphical keyboard 116 B for invoking one or more of the various search features of graphical keyboard 116 B.
  • search element 118 C e.g., by tapping or gesturing at a location or within a region of PSD 112 at which search element 118 C is displayed
  • computing device 110 can invoke the various integrated search features without having to navigate to a separate application, service, or other feature executing at or accessible from computing device 110 .
  • UI module 120 may output information to keyboard module 122 indicating that a user of computing device 110 may have selected search element 118 C. Responsive to determining that search element 118 C was selected, keyboard module 122 may transition to operating in search mode. While operating in search mode, keyboard module 122 may reconfigure graphical keyboard 116 B to execute search features as opposed to operations that are primarily attributed to text entry.
  • keyboard module 122 may configure suggestion region 118 B to present suggested content (e.g., predicted search queries, predicted emoticons or so called “emojis”, other suggested content, or other iconography symbols) as selectable elements within search suggestion region 118 B instead of predicted characters, words or phrases or other primarily linguistic information that keyboard module 122 derives from a language model, lexicon, or dictionary.
  • suggested content e.g., predicted search queries, predicted emoticons or so called “emojis”, other suggested content, or other iconography symbols
  • computing device 110 may include, within suggestion region 118 B, suggested search related content that computing device 110 determines may assist a user in providing input related to electronic communications.
  • keyboard module 122 may configure the graphical keyboard 116 B to include a query suggestion region 118 D that presents suggested content (e.g., predicted or suggested search queries, predicted emoticons, emojis, other iconographic symbols, and other suggested content) as selectable elements above the graphical keys 118 A.
  • keyboard module 122 may in some examples configure the graphical keyboard 116 B to include query suggestion region 118 D positioned between edit region 116 C and graphical keys 118 A.
  • the query suggestion region 118 D is also positioned near (e.g., adjacent to, above, below, on either side of, etc.) the suggestion region 118 B, which may have input focus.
  • the query suggestion region 118 D may be positioned between the graphical keys 118 A and the edit region 116 C, including in an example where the query suggestion region is positioned below the graphical keys and above the edit region.
  • the query suggestion region may be positioned between the suggestion region 118 B and the edit region 116 C, including an example where the query suggestion region is positioned below the suggestion region and above the edit region.
  • Other alternative arrangements of the graphical user interface are possible, including arrangements where the query suggestion region is not positioned between the graphical keys and the edit region, or between the suggestion region and the edit region.
  • input focus means that input detected by the computing device 110 may be output to one or more specific user interface component(s) that are designated as having input focus.
  • PSD 112 detects user input at or near a location of PSD 112 where PSD 112 presents graphical keyboard 116 B
  • UI module 120 may receive an indication of input and may output to the keyboard module information about the input.
  • the keyboard module 122 determines that the user is providing input at graphical keys 118 A that correspond to a string of characters
  • the keyboard module 122 may cause UI module 120 to update the user interface 114 to include that string of characters within the user interface 114 . Where that string of characters may be positioned or displayed within the user interface may be determined by which user interface component(s) have input focus.
  • User interface 114 includes several user interface components, including edit region 116 C, suggestion region 118 B, search element 118 C, and output region 116 A.
  • suggestion region 118 B has input focus
  • the computing device 110 may format and display within suggestion region 118 B the string of characters corresponding to the input.
  • the computing device 110 may determine, in some examples, which user interface component has input focus by detecting which qualifying user interface component last received touch input by the user. In other words, when PSD 112 detects user input at a location of PSD 112 corresponding to a particular user interface component, computing device 110 may give that user interface component input focus (if it is allowed or qualified to have input focus). In other examples, the computing device 110 determines which user interface component has input focus by detecting which user interface component last received input from a cursor.
  • computing device 110 may determine which user interface component has input focus based on a setting configured by an application executing on the computing device 110 .
  • a flashing cursor 119 may in some examples be displayed within the interface component to provide a visual cue indicating that it has input focus.
  • PSD 112 may detect user inputs at or near a location of PSD 112 where PSD 112 presents graphical keys 118 A, and output to the UI module 220 an indication of the user input detected by PSD 112 .
  • the UI module 120 may output to the keyboard module 122 information about the input. Responsive to the information about the input, keyboard module 122 determines a string of characters that corresponds to the input, and then causes the UI module 120 to update the user interface 114 to include that string of characters in the suggestion region 118 B, which has input focus in FIG. 1 . As shown in the example of FIG. 1 , the keyboard module 122 determines the string of characters corresponding to the input in this case to be the “meatball sho” string of characters displayed in the suggestion region 118 B.
  • Keyboard module 122 may also determine suggested queries based on the information about the input received from UI module 120 . Responsive to the information about the input received from UI module 120 , the keyboard module 122 may determine suggested queries and cause the UI module 120 to update the user interface 114 to include the suggested queries in the query suggestion region 118 D. As shown in the example of FIG. 1 , the suggested queries generated by the keyboard module 122 may include “Meatball Shoppe,” “Meatball,” and “Meatball recipe easy.” As the PSD 112 continues to detect user inputs corresponding to locations of PSD 112 that present graphical keys 118 A, information about additional input is received by the keyboard module 122 , which in response to the additional input, may determine updated suggested queries. The keyboard module 122 may cause UI module 120 to update the user interface 114 as updated suggested queries are determined.
  • the keyboard module 122 may determine suggested queries based on the information about user input received from the UI module 120 , but the suggested queries may also be determined based on an application the user is interacting with, the context of the user's activity, or some combination thereof.
  • the keyboard module 122 may also determine the suggested queries based on historical information about the user. For example, in some examples, search query suggestions may come from chat conversations on computing device 110 . The text message “do you want to get dinner?” may suggest a “Restaurants near me” query; the text message “did you watch the basketball game?” may suggest a “Team Name” query of a professional basketball team.
  • the graphical keyboard 116 B in FIG. 1 may be considered to be in search mode or suggest mode.
  • keyboard module 122 may obtain information associated with the suggested query, which may involve invoking a search operation using the content of the link as a search query, or otherwise conducting a search for information based on the suggested query.
  • keyboard module 122 may obtain information associated with the search query without invoking a search operation or performing a search, such as when the suggested query points directly an item of information, or points to an item of information (e.g., a document or contact) on the device 110 .
  • the keyboard module 122 may cause the UI module 120 to update the user interface 114 , and may cause the UI module 120 to further update the user interface 114 when search results are available. In some examples, the keyboard module 122 may cause UI module 120 to update the user interface 114 so that the search results are displayed within the graphical keyboard 116 B. In accordance with one or more aspects of the present disclosure, when search results are being displayed, the graphical keyboard 116 B may be considered to be in results mode.
  • an example computing device may perform search operations without requiring the user to leave the current application that the user is interacting with, and without the user losing focus.
  • the example computing device may present search results obtained from performing a search of a suggested query and the user may then review or interact with the search results as the example device presents them on a screen.
  • the user may provide additional input causing the computing device to incorporate the search results or information obtained from the search results into a current activity or task, or share the search results or information obtained from the search results with other devices.
  • Techniques in accordance with the present disclosure may also enable a computing device to output or present to the user suggested search queries, potentially facilitating and streamlining the process for performing a search.
  • Automatically displaying suggested search queries may result in the computing device receiving fewer inputs related to search, because fewer inputs may be required to cause a device to perform a search based on a suggested query that is automatically presented within a graphical keyboard.
  • the computing device may process fewer user inputs, execute fewer operations, and as a result, consume less electrical power.
  • FIG. 2 is a block diagram illustrating computing device 210 as an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
  • Computing device 210 of FIG. 2 is described below as an example of computing device 110 of FIG. 1 .
  • FIG. 2 illustrates only one particular example of computing device 210 , and many other examples of computing device 210 may be used in other instances and may include a subset of the components included in example computing device 210 or may include additional components not shown in FIG. 2 .
  • computing device 210 includes PSD 212 , one or more processors 240 , one or more communication units 242 , one or more input components 244 , one or more output components 246 , and one or more storage components 248 .
  • Presence-sensitive display 212 includes display component 202 and presence-sensitive input component 204 .
  • Storage components 248 of computing device 210 include UI module 220 , keyboard module 222 , and one or more application modules 224 .
  • Keyboard module 122 may include spatial model (“SM”) module 226 , language model (“LM”) module 228 , and search module 230 .
  • SM spatial model
  • LM language model
  • Communication channels 250 may interconnect each of the components 212 , 240 , 242 , 244 , 246 , and 248 for inter-component communications (physically, communicatively, and/or operatively).
  • communication channels 250 may include a system bus, a network connection, an inter-process communication data structure, or any other method for communicating data.
  • One or more communication units 242 of computing device 210 may communicate with external devices via one or more wired and/or wireless networks by transmitting and/or receiving network signals on the one or more networks.
  • Examples of communication units 242 include a network interface card (e.g. such as an Ethernet card), an optical transceiver, a radio frequency transceiver, a GPS receiver, or any other type of device that can send and/or receive information.
  • Other examples of communication units 242 may include short wave radios, cellular data radios, wireless network radios, as well as universal serial bus (USB) controllers.
  • USB universal serial bus
  • One or more input components 244 of computing device 210 may receive input. Examples of input are tactile, audio, and video input.
  • Input components 242 of computing device 210 includes a presence-sensitive input device (e.g., a touch sensitive screen, a PSD), mouse, keyboard, voice responsive system, video camera, microphone or any other type of device for detecting input from a human or machine.
  • a presence-sensitive input device e.g., a touch sensitive screen, a PSD
  • mouse e.g., keyboard, voice responsive system, video camera, microphone or any other type of device for detecting input from a human or machine.
  • input components 242 may include one or more sensor components one or more location sensors (GPS components, Wi-Fi components, cellular components), one or more temperature sensors, one or more movement sensors (e.g., accelerometers, gyros), one or more pressure sensors (e.g., barometer), one or more ambient light sensors, and one or more other sensors (e.g., microphone, camera, infrared proximity sensor, hygrometer, and the like).
  • Other sensors may include a heart rate sensor, magnetometer, glucose sensor, hygrometer sensor, olfactory sensor, compass sensor, step counter sensor, to name a few other non-limiting examples.
  • One or more output components 246 of computing device 110 may generate output. Examples of output are tactile, audio, and video output.
  • Output components 246 of computing device 210 includes a PSD, sound card, video graphics adapter card, speaker, cathode ray tube (CRT) monitor, liquid crystal display (LCD), or any other type of device for generating output to a human or machine.
  • PSD 212 of computing device 210 may be similar to PSD 112 of computing device 110 and includes display component 202 and presence-sensitive input component 204 .
  • Display component 202 may be a screen at which information is displayed by PSD 212 and presence-sensitive input component 204 may detect an object at and/or near display component 202 .
  • presence-sensitive input component 204 may detect an object, such as a finger or stylus that is within two inches or less of display component 202 .
  • Presence-sensitive input component 204 may determine a location (e.g., an [x, y] coordinate) of display component 202 at which the object was detected.
  • presence-sensitive input component 204 may detect an object six inches or less from display component 202 and other ranges are also possible.
  • Presence-sensitive input component 204 may determine the location of display component 202 selected by a user's finger using capacitive, inductive, and/or optical recognition techniques. In some examples, presence-sensitive input component 204 also provides output to a user using tactile, audio, or video stimuli as described with respect to display component 202 . In the example of FIG. 2 , PSD 212 may present a user interface (such as user interface 114 of FIG. 1 ).
  • PSD 212 may also represent an external component that shares a data path with computing device 210 for transmitting and/or receiving input and output.
  • PSD 212 represents a built-in component of computing device 210 located within and physically connected to the external packaging of computing device 210 (e.g., a screen on a mobile phone).
  • PSD 212 represents an external component of computing device 210 located outside and physically separated from the packaging or housing of computing device 210 (e.g., a monitor, a projector, etc. that shares a wired and/or wireless data path with computing device 210 ).
  • PSD 212 of computing device 210 may detect two-dimensional and/or three-dimensional gestures as input from a user of computing device 210 .
  • a sensor of PSD 212 may detect a user's movement (e.g., moving a hand, an arm, a pen, a stylus, etc.) within a threshold distance of the sensor of PSD 212 .
  • PSD 212 may determine a two or three dimensional vector representation of the movement and correlate the vector representation to a gesture input (e.g., a hand-wave, a pinch, a clap, a pen stroke, etc.) that has multiple dimensions.
  • a gesture input e.g., a hand-wave, a pinch, a clap, a pen stroke, etc.
  • PSD 212 can detect a multi-dimension gesture without requiring the user to gesture at or near a screen or surface at which PSD 212 outputs information for display. Instead, PSD 212 can detect a multi-dimensional gesture performed at or near a sensor which may or may not be located near the screen or surface at which PSD 212 outputs information for display.
  • processors 240 may implement functionality and/or execute instructions associated with computing device 210 .
  • Examples of processors 240 include application processors, display controllers, auxiliary processors, one or more sensor hubs, and any other hardware configured to function as a processor, a processing unit, or a processing device.
  • Modules 220 , 222 , 224 , 226 , 228 , and 230 may be operable by processors 240 to perform various actions, operations, or functions of computing device 210 .
  • processors 240 of computing device 210 may retrieve and execute instructions stored by storage components 248 that cause processors 240 to perform the operations modules 220 , 222 , 224 , 226 , 228 , and 230 .
  • the instructions when executed by processors 240 , may cause computing device 210 to store information within storage components 248 .
  • One or more storage components 248 within computing device 210 may store information for processing during operation of computing device 210 (e.g., computing device 210 may store data accessed by modules 220 , 222 , 224 , 226 , 228 , and 230 during execution at computing device 210 ).
  • storage component 248 is a temporary memory, meaning that a primary purpose of storage component 248 is not long-term storage.
  • Storage components 248 on computing device 210 may be configured for short-term storage of information as volatile memory and therefore not retain stored contents if powered off. Examples of volatile memories include random access memories (RAM), dynamic random access memories (DRAM), static random access memories (SRAM), and other forms of volatile memories known in the art.
  • Storage components 248 also include one or more computer-readable storage media.
  • Storage components 248 in some examples include one or more non-transitory computer-readable storage mediums.
  • Storage components 248 may be configured to store larger amounts of information than typically stored by volatile memory.
  • Storage components 248 may further be configured for long-term storage of information as non-volatile memory space and retain information after power on/off cycles. Examples of non-volatile memories include disc storage, flash memories, or forms of electrically programmable memories (EPROM) or electrically erasable and programmable (EEPROM) memories.
  • Storage components 248 may store program instructions and/or information (e.g., data) associated with modules 220 , 222 , 224 , 226 , 228 , and 230 .
  • Storage components 248 may include a memory configured to store data or other information associated with modules 220 , 222 , 224 , 226 , 228 , and 230 .
  • UI module 220 may include all functionality of UI module 120 of computing device 110 of FIG. 1 and may perform similar operations as UI module 120 for managing a user interface (e.g., user interface 114 ) that computing device 210 provides at presence-sensitive display 212 for handling input from a user.
  • UI module 220 of computing device 210 may query keyboard module 222 for a keyboard layout (e.g., an English language QWERTY keyboard, etc.).
  • UI module 220 may transmit a request for a keyboard layout over communication channels 250 to keyboard module 222 .
  • Keyboard module 222 may receive the request and reply to UI module 220 with data associated with the keyboard layout.
  • UI module 220 may receive the keyboard layout data over communication channels 250 and use the data to generate a user interface.
  • UI module 220 may transmit a display command and data over communication channels 250 to cause PSD 212 to present the user interface at PSD 212 .
  • UI module 220 may receive an indication of one or more user inputs detected at PSD 212 and may output information about the user inputs to keyboard module 222 .
  • PSD 212 may detect a user input and send data about the user input to UI module 220 .
  • UI module 220 may generate one or more touch events based on the detected input.
  • a touch event may include information that characterizes user input, such as a location component (e.g., [x,y] coordinates) of the user input, a time component (e.g., when the user input was received), a force component (e.g., an amount of pressure applied by the user input), or other data (e.g., speed, acceleration, direction, density, etc.) about the user input.
  • a location component e.g., [x,y] coordinates
  • time component e.g., when the user input was received
  • a force component e.g., an amount of pressure applied by the user input
  • other data e.g., speed,
  • UI module 220 may determine that the detected user input is associated the graphical keyboard. UI module 220 may send an indication of the one or more touch events to keyboard module 222 for further interpretation. Keyboard module 22 may determine, based on the touch events received from UI module 220 , that the detected user input represents an initial selection of one or more keys of the graphical keyboard.
  • Application modules 224 represent all the various individual applications and services executing at and accessible from computing device 210 that may rely on a graphical keyboard having integrated search features.
  • a user of computing device 210 may interact with a graphical user interface associated with one or more application modules 224 to cause computing device 210 to perform a function.
  • Numerous examples of application modules 224 may exist and include, a fitness application, a calendar application, a personal assistant or prediction engine, a search application, a map or navigation application, a transportation service application (e.g., a bus or train tracking application), a social media application, a game application, an e-mail application, a chat or messaging application, an Internet browser application, or any and all other applications that may execute at computing device 210 .
  • Keyboard module 222 may include all functionality of keyboard module 122 of computing device 110 of FIG. 1 and may perform similar operations as keyboard module 122 for providing a graphical keyboard having integrated search features. Keyboard module 222 may include various submodules, such as SM module 226 , LM module 228 , and search module 230 , which may perform the functionality of keyboard module 222 .
  • SM module 226 may receive one or more touch events as input, and output a character or sequence of characters that likely represents the one or more touch events, along with a degree of certainty or spatial model score indicative of how likely or with what accuracy the one or more characters define the touch events. In other words, SM module 226 may infer touch events as a selection of one or more keys of a keyboard and may output, based on the selection of the one or more keys, a character or sequence of characters.
  • LM module 228 may receive a character or sequence of characters as input, and output one or more candidate characters, words, or phrases that LM module 228 identifies from a lexicon as being potential replacements for a sequence of characters that LM module 228 receives as input for a given language context (e.g., a sentence in a written language).
  • Keyboard module 222 may cause UI module 220 to present one or more of the candidate words at suggestion regions 118 B and 118 D of user interface 114 .
  • the lexicon of computing device 210 may include a list of words within a written language vocabulary (e.g., a dictionary).
  • the lexicon may include a database of words (e.g., words in a standard dictionary and/or words added to a dictionary by a user or computing device 210 .
  • LM module 228 may perform a lookup in the lexicon, of a character string, to identify one or more letters, words, and/or phrases that include parts or all of the characters of the character string.
  • LM module 228 may assign a language model probability or a similarity coefficient (e.g., a Jaccard similarity coefficient) to one or more candidate words located at a lexicon of computing device 210 that include at least some of the same characters as the inputted character or sequence of characters.
  • the language model probability assigned to each of the one or more candidate words indicates a degree of certainty or a degree of likelihood that the candidate word is typically found positioned subsequent to, prior to, and/or within, a sequence of words (e.g., a sentence) generated from text input detected by presence-sensitive input component 204 prior to and/or subsequent to receiving the current sequence of characters being analyzed by LM module 228 .
  • LM module 228 may output the one or more candidate words from lexicon data stores 260 A that have the highest similarity coefficients.
  • Search module 230 of keyboard module 222 may perform integrated search functions on behalf of keyboard module 222 . That is, when invoked (e.g., automatically based on text input analysis or manually in response to a user of computing device 210 selecting selectable search element 118 C of user interface 114 ), keyboard module 222 may operate in search mode where keyboard module 222 enables computing device 210 to perform search functions from within graphical keyboard 116 B.
  • search module 230 may receive a character or sequence of characters or other information as input, and output one or more query suggestions that the search module 230 identifies based on the input.
  • the information that the search module 230 receives as input may include characters or a sequence of characters, and in some examples, may also include information about the current state of the application that the user is using, as well as other information, such as information from the application that is relying on keyboard module 122 (e.g., recent texts, pictures, geolocation data and the like) and contextual information associated with the user and computing device 210 .
  • Search module 230 may use this information to generate one or more suggested queries.
  • the keyboard module 222 may cause UI module 220 to present one or more of the suggested queries within the query suggestion region 118 D of user interface 114 .
  • search model 230 may further rely on a current context of computing device 210 to determine a suggested query.
  • a current context specifies the characteristics of the physical and/or virtual environment of a computing device, such as computing device 210 , and a user of the computing device, at a particular time.
  • contextual information is used to describe any information that can be used by a computing device to define the virtual and/or physical environmental characteristics that the computing device, and the user of the computing device, may experience at a particular time.
  • contextual information examples include: sensor information obtained by sensors (e.g., position sensors, accelerometers, gyros, barometers, ambient light sensors, proximity sensors, microphones, and any other sensor) of computing device 210 , communication information (e.g., text based communications, audible communications, video communications, etc.) sent and received by communication modules of computing device 210 , and application usage information associated with applications executing at computing device 210 (e.g., application data associated with applications, Internet search histories, text communications, voice and video communications, calendar information, social media posts and related information, etc.). Further examples of contextual information include signals and information obtained from transmitting devices that are external to computing device 210 .
  • sensors e.g., position sensors, accelerometers, gyros, barometers, ambient light sensors, proximity sensors, microphones, and any other sensor
  • communication information e.g., text based communications, audible communications, video communications, etc.
  • application usage information e.g., application data associated with applications, Internet search histories, text communications,
  • computing device 210 may initiate a search in response to different types of input.
  • keyboard module 222 may initiate a search in response to detecting user input at a location at which PSD 212 presents a suggested query within the query suggestion region 118 D or elsewhere. This input may generally correspond to the user selecting one of the displayed suggested queries.
  • keyboard module 222 while operating in search mode, may initiate a search in response to detecting user input at locations at which PSD 212 presents graphical keys 118 A followed by input at a location at which PSD 212 presents a “return” or “search” or “send” key. Such input may generally correspond to the user typing a search query using the graphical keyboard 116 B and hitting “return” without selecting a suggested search query.
  • the search module 230 within the keyboard module 222 may execute a search, which may cause the computing device 210 to send and receive information over a public or private network through communication unit(s) 242 . If search results become available, the search module 230 or keyboard module 222 may then cause the UI module 220 to present search result information to the user by, in some examples, displaying search result information in user interface 114 .
  • FIG. 3 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure.
  • Graphical content generally, may include any visual information that may be output for display, such as text, images, a group of moving images, to name only a few examples.
  • the example shown in FIG. 3 includes a computing device 310 , a PSD 312 , communication unit 342 , projector 380 , projector screen 382 , mobile device 386 , and visual display component 390 .
  • PSD 312 may be a presence-sensitive display as described in FIGS. 1-2 . Although shown for purposes of example in FIGS.
  • a computing device such as computing device 310 may, generally, be any component or system that includes a processor or other suitable computing environment for executing software instructions and, for example, need not include a presence-sensitive display.
  • computing device 310 may be a processor that includes functionality as described with respect to processors 240 in FIG. 2 .
  • computing device 310 may be operatively coupled to PSD 312 by a communication channel 362 A, which may be a system bus or other suitable connection.
  • Computing device 310 may also be operatively coupled to communication unit 342 , further described below, by a communication channel 362 B, which may also be a system bus or other suitable connection.
  • communication channel 362 B may also be a system bus or other suitable connection.
  • computing device 310 may be operatively coupled to PSD 312 and communication unit 342 by any number of one or more communication channels.
  • a computing device may refer to a portable or mobile device such as mobile phones (including smart phones), laptop computers, etc.
  • a computing device may be a desktop computer, tablet computer, smart television platform, camera, personal digital assistant (PDA), server, or mainframes.
  • PDA personal digital assistant
  • PSD 312 may include display component 302 and presence-sensitive input component 304 .
  • Display component 302 may, for example, receive data from computing device 310 and display the graphical content.
  • presence-sensitive input component 304 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at PSD 312 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input to computing device 310 using communication channel 362 A.
  • user inputs e.g., continuous gestures, multi-touch gestures, single-touch gestures
  • presence-sensitive input component 304 may be physically positioned on top of display component 302 such that, when a user positions an input unit over a graphical element displayed by display component 302 , the location at which presence-sensitive input component 304 corresponds to the location of display component 302 at which the graphical element is displayed.
  • computing device 310 may also include and/or be operatively coupled with communication unit 342 .
  • Communication unit 342 may include functionality of communication unit 242 as described in FIG. 2 .
  • Examples of communication unit 342 may include a network interface card, an Ethernet card, an optical transceiver, a radio frequency transceiver, or any other type of device that can send and receive information.
  • Other examples of such communication units may include Bluetooth, 3G, and WiFi radios, Universal Serial Bus (USB) interfaces, etc.
  • Computing device 310 may also include and/or be operatively coupled with one or more other devices (e.g., input devices, output components, memory, storage devices) that are not shown in FIG. 3 for purposes of brevity and illustration.
  • FIG. 3 also illustrates a projector 380 and projector screen 382 .
  • projection devices may include electronic whiteboards, holographic display components, and any other suitable devices for displaying graphical content.
  • Projector 380 and projector screen 382 may include one or more communication units that enable the respective devices to communicate with computing device 310 . In some examples, the one or more communication units may enable communication between projector 380 and projector screen 382 .
  • Projector 380 may receive data from computing device 310 that includes graphical content. Projector 380 , in response to receiving the data, may project the graphical content onto projector screen 382 .
  • projector 380 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen using optical recognition or other suitable techniques and send indications of such user input using one or more communication units to computing device 310 .
  • projector screen 382 may be unnecessary, and projector 380 may project graphical content on any suitable medium and detect one or more user inputs using optical recognition or other such suitable techniques.
  • Projector screen 382 may include a presence-sensitive display 384 .
  • Presence-sensitive display 384 may include a subset of functionality or all of the functionality of presence-sensitive display 112 and/or 312 as described in this disclosure.
  • presence-sensitive display 384 may include additional functionality.
  • Projector screen 382 (e.g., an electronic whiteboard), may receive data from computing device 310 and display the graphical content.
  • presence-sensitive display 384 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen 382 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units to computing device 310 .
  • FIG. 3 also illustrates mobile device 386 and visual display component 390 .
  • Mobile device 386 and visual display component 390 may each include computing and connectivity capabilities. Examples of mobile device 386 may include e-reader devices, convertible notebook devices, hybrid slate devices, etc. Examples of visual display component 390 may include other devices such as televisions, computer monitors, etc.
  • visual display component 390 may be a vehicle cockpit display or navigation display (e.g., in an automobile, aircraft, or some other vehicle). In some examples, visual display component 390 may be a home automation display or some other type of display that is separate from computing device 310 .
  • mobile device 386 may include a presence-sensitive display 388 .
  • Visual display component 390 may include a presence-sensitive display 392 .
  • Presence-sensitive displays 388 , 392 may include a subset of functionality or all of the functionality of presence-sensitive display 112 , 212 , and/or 312 as described in this disclosure.
  • presence-sensitive displays 388 , 392 may include additional functionality.
  • presence-sensitive display 392 may receive data from computing device 310 and display the graphical content.
  • presence-sensitive display 392 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units to computing device 310 .
  • user inputs e.g., continuous gestures, multi-touch gestures, single-touch gestures
  • computing device 310 may output graphical content for display at PSD 312 that is coupled to computing device 310 by a system bus or other suitable communication channel.
  • Computing device 310 may also output graphical content for display at one or more remote devices, such as projector 380 , projector screen 382 , mobile device 386 , and visual display component 390 .
  • computing device 310 may execute one or more instructions to generate and/or modify graphical content in accordance with techniques of the present disclosure.
  • Computing device 310 may output the data that includes the graphical content to a communication unit of computing device 310 , such as communication unit 342 .
  • Communication unit 342 may send the data to one or more of the remote devices, such as projector 380 , projector screen 382 , mobile device 386 , and/or visual display component 390 .
  • computing device 310 may output the graphical content for display at one or more of the remote devices.
  • one or more of the remote devices may output the graphical content at a presence-sensitive display that is included in and/or operatively coupled to the respective remote devices.
  • computing device 310 may not output graphical content at PSD 312 that is operatively coupled to computing device 310 .
  • computing device 310 may output graphical content for display at both a PSD 312 that is coupled to computing device 310 by communication channel 362 A, and at one or more remote devices.
  • the graphical content may be displayed substantially contemporaneously at each respective device. For instance, some delay may be introduced by the communication latency to send the data that includes the graphical content to the remote device.
  • graphical content generated by computing device 310 and output for display at PSD 312 may be different than graphical content display output for display at one or more remote devices.
  • Computing device 310 may send and receive data using any suitable communication techniques.
  • computing device 310 may be operatively coupled to external network 374 using network link 373 A.
  • Each of the remote devices illustrated in FIG. 3 may be operatively coupled to network external network 374 by one of respective network links 373 B, 373 C, or 373 D.
  • External network 374 may include network hubs, network switches, network routers, etc., that are operatively inter-coupled thereby providing for the exchange of information between computing device 310 and the remote devices illustrated in FIG. 3 .
  • network links 373 A- 373 D may be Ethernet, ATM or other network connections. Such connections may be wireless and/or wired connections.
  • computing device 310 may be operatively coupled to one or more of the remote devices included in FIG. 3 using direct device communication 378 .
  • Direct device communication 378 may include communications through which computing device 310 sends and receives data directly with a remote device, using wired or wireless communication. That is, in some examples of direct device communication 378 , data sent by computing device 310 may not be forwarded by one or more additional devices before being received at the remote device, and vice-versa. Examples of direct device communication 378 may include Bluetooth, Near-Field Communication, Universal Serial Bus, WiFi, infrared, etc.
  • One or more of the remote devices illustrated in FIG. 3 may be operatively coupled with computing device 310 by communication links 376 A- 376 D.
  • communication links 376 A- 376 D may be connections using Bluetooth, Near-Field Communication, Universal Serial Bus, infrared, etc. Such connections may be wireless and/or wired connections.
  • computing device 310 may in some examples be operatively coupled to visual display component 390 using external network 374 .
  • Computing device 310 may output a graphical keyboard for display at PSD 392 .
  • computing device 310 may send data that includes a representation of the graphical keyboard to communication unit 342 .
  • Communication unit 342 may send the data that includes the representation of the graphical keyboard to visual display component 390 using external network 374 .
  • Visual display component 390 in response to receiving the data using external network 374 , may cause PSD 392 to output the graphical keyboard.
  • visual display component 390 may send an indication of the user input to computing device 310 using external network 374 .
  • Communication unit 342 of may receive the indication of the user input, and send the indication to computing device 310 .
  • Computing device 310 may determine one or more candidate words or suggested search queries, and may generate an updated graphical user interface including suggested search queries, which in some examples, may be included within a query suggestion region.
  • Communication unit 342 may receive a representation of the updated graphical user interface and may send the send the representation to visual display component 390 , such that visual display component 390 may cause PSD 392 to output the updated graphical keyboard, including any suggested search queries and/or query suggestion region.
  • FIGS. 4A-4D are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • FIGS. 4A-4D illustrate example user interfaces 414 ; many other examples of graphical user interfaces may be used or presented in other instances or examples.
  • Each of the user interfaces 414 may correspond to a graphical user interface output by computing devices 110 , 210 , or 310 from FIGS. 1-3 .
  • FIGS. 4A-4D are described below within the context of computing device 210 of FIG. 2 .
  • each example user interface 414 forms part of a chat application user interface of a chat application from application module 224 .
  • User interface 414 includes output region 416 A, edit region 416 C, and graphical keyboard 416 B.
  • Output region 416 A and edit region 416 C may both be associated with the user interface of an application or service executing on or accessible from computing device 210 .
  • Included within the graphical keyboard 416 B are graphical elements displayed as keys 418 A.
  • a suggestion region 418 B which may include suggested words or other information representing selectable spelling corrections and/or selectable word suggestions to replace character strings that are included within edit region 416 C.
  • suggestion region 418 B may include suggested queries that may be of interest to the user based on information that may include the user's current activity, context, location, or situation.
  • search element 418 C Also shown included within graphical keyboard 416 B in the example of FIG. 4A is a search element 418 C.
  • computing device 210 may receive an indication of user input selecting the search element 418 C, and in response, computing device 210 may cause graphical keyboard 416 B transition to search mode in accordance with one or more aspects of the present disclosure.
  • the search element 418 C is displayed above the graphical keys 418 A, but it may be located elsewhere within the graphical keyboard 416 B. In other examples, the search element 418 C may be located elsewhere within the user interface 414 , including partially or entirely outside of the graphical keyboard 416 B.
  • the search element 418 C or a component used in a similar way could be implemented as an external input or hardware input, such as an external button, switch, or other hardware input component configured to receive user input for computing device 210 .
  • computing device 210 may respond to an indication of input selecting key 402 by transmitting to another device the string of characters in edit region 416 C as a text message.
  • Computing device 210 may also cause the string of characters in the edit region 416 C to be formatted and displayed within the output region 416 A, which in some examples may signify that the string of characters was transmitted as a text message. For instance, in response to receiving user input at locations of PSD 112 corresponding to the key 402 , computing device 210 may update the user interface 414 so that presented in output region 416 A is the text that was displayed in edit region 416 C when the computing device received the indication of input selecting key 402 .
  • the computing device 210 may respond to an indication of input selecting the return/search key 402 by performing a search using, for example, the information or text displayed within the suggestion region 418 B when the computing device 210 received the indication of input selecting the key 402 .
  • the return/search key 402 is shown as a single graphical key in the example of FIG. 4A , in other examples the key 402 could be presented or implemented as more than one key. In one example, the return/search key 402 could alternatively be implemented as a discrete graphical return key and a discrete graphical search key. In other examples, the graphical keyboard 416 B could include a discrete graphical return key and a discrete graphical search key, but may not always include a graphical representation of both keys within the graphical keyboard 416 B. In still further examples, the computing device 210 may display a discrete graphical search key only when the graphical keyboard 416 B is in search mode in accordance with one or more aspects of the present disclosure.
  • computing device 210 may receive an electronic communication or text message from a friend's device.
  • Computing device 210 may output the content of the friend's text message (“What's up?”) within the user interface 414 in output region 416 A.
  • Keyboard module 222 of computing device 210 may then receive an indication of input selecting graphical keys 418 A and determine a character string (“Din”) based on the indication of input, which the computing device 210 formats and displays within edit region 416 C as shown in FIG. 4A .
  • the keyboard module 222 may also determine, based on the indication of input, candidate words (“Dinner”, “Dine”, and “Don”).
  • the computing device 210 may format and displays these candidate words within suggestion region 418 B.
  • keyboard module 222 of computing device 210 may receive an indication of input selecting additional graphical keys 418 A, and determine, based on the input, additional text that the computing device 210 formats and displays within edit region 416 C. Keyboard module 222 may also predict one or more updated candidate words based on the further user input and display one or more updated candidate words within suggestion region 418 B.
  • the text “Dinner to” is displayed in edit region 416 C in FIG. 4B
  • candidate word “tonight” is displayed in edit region 416 B.
  • Keyboard module 222 may also determine one or more suggested queries, and as shown in FIG. 4B , computing device 210 may format and display a suggested query 406 (“Restaurants near me”) within suggestion region 418 B. In other examples, computing device 210 may cause multiple suggested queries to be displayed in suggestion region 418 B or in other areas of user interface 414 . Keyboard module 222 may generate suggested search queries based on many different sources of information, including user input, information in the edit region 416 C, information in the output region 416 A, current time, the user's context, application state, historical searches, and/or other information.
  • computing device 210 suggests search queries after the user has typed a sentence-ending punctuation or has sent a message.
  • Computing device 210 presents suggested search queries in the manner shown in FIG. 4B , or in other ways in accordance with one or more aspects of the present disclosure.
  • the computing device 210 may alert the user to availability of one or more suggested queries, such as by providing a visual, audio, tactile, or physical indication or cue.
  • the computing device 210 may alert the user to the availability of a suggested query through the search element 418 C, which may change color, change shape, animate, pulse, or otherwise change to suggest that suggested queries may be available.
  • the computing device 210 may use other aspects of the user interface 414 to perform a similar function, or in other examples computing device 210 may vibrate or provide some other physical indication.
  • Computing device 210 may determine the suggested words and suggested search queries shown in suggestion region 418 B based entirely on locally-stored information without access to information on a network or without sending information about the user's activity (including the user's current or past actions or communications) over a public or private network. In other examples, computing device 210 may determine suggested words and/or suggested search queries using information obtained through a public or private network, and possibly by sending information about the user (which could include information about the user's actions or communications) over a public or private network.
  • on-device annotators may be used to generate suggested search queries, and such annotators may use both global and local models to detect searchable text or entities.
  • Global models may include entities that may be useful to search regardless of the user's geolocation, such as a well-known person (“Barack Obama”).
  • Local models may include entities such as restaurants near the user, such as a particular restaurant near the location of the computing device 210 .
  • Suggested queries could also be based on exact language triggers, such as the phrase “let's get dinner” triggering a “Restaurants near me” suggested query.
  • a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information. For example, before a computing device or computing system can collect or may make use of information associated with a user, the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user.
  • information e.g., context, locations, speeds, search queries, etc.
  • certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed.
  • a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined.
  • location information such as to a city, ZIP code, or state level
  • a graphical symbol of a category 404 of search which may identify the category of a search with which the suggested query 406 may be associated.
  • computing device 210 may cause the graphical symbol of a category 404 to be displayed to denote a category relating to a search for restaurants or food. Symbols representing other search categories may alternatively be displayed.
  • more than one search category element may be associated with a suggested search query, and in still further examples, there may be no search category element associated with one or more suggested search queries (or associated with any suggested search queries).
  • computing device 210 may display a generic search category element, such as a magnifying glass, to designate the search query as not being associated with any particular search category.
  • the computing device 210 has updated the user interface 414 of FIG. 4B in response to an indication of input selecting graphical keys 118 A (which the keyboard module 222 may interpret as the string of text “Dinner tonight?”), followed by an indication of input selecting the “return/search” key 402 .
  • computing device 210 may compose and send a text message that includes the text “Dinner tonight?” to a friend's computing device.
  • computing device 210 may update the user interface 414 to include the “Dinner tonight?” text message within the output region 416 A.
  • the computing device 210 may update the user interface 414 in FIG.
  • computing device 210 may also remove the suggested word “tonight” from the suggestion region 418 B.
  • Computing device 210 may continue to display the suggested search query 406 within the suggestion region 418 B in the example of FIG. 4C .
  • the computing device 210 has updated the user interface 414 in response (at least partially) to detecting an indication of input selecting the suggested search query 406 (“Restaurants near me”) displayed in suggestion region 418 B in FIG. 4C (note the outline of a user's hand in FIG. 4C , which may signify a user's touch input associated with suggested query 406 ).
  • the computing device 210 may perform a search based on the suggested query 406 .
  • the suggested query 406 may be a link that, when invoked by user input, will cause computing device 210 to perform a search operation, using the content of the link as a search query.
  • a suggested query link might not require computing device to perform a search operation, such as when the link points directly to an item of information or to the location of the item, or represents the URL of a document or other resource.
  • the computing device 210 may output for display a textual and/or graphical indication of one or more search results (or other information) obtained by the computing device.
  • the computing device 210 causes the search results 418 E to be displayed within the graphical keyboard 416 B.
  • search results 418 E are presented below the suggestion region 418 B, substantially replacing the graphical keys 418 A previously displayed in FIG. 4C .
  • the graphical keyboard 416 B in FIG. 4D may be considered to be in results mode.
  • the keyboard module 222 may obtain search results based on the selected query, using the suggested query, that are or may be formatted into individual search cards 418 E that keyboard module 222 causes computing device 210 to display as overlays atop graphical keys 418 A (or as a temporary replacement for graphical keys 418 A).
  • the search results 418 E may be presented in a carousel format so that a user may swipe left or right 410 when browsing the search results 418 E.
  • the computing device 210 could alternatively output the search results of FIG. 4D in other formats, or in other locations within the graphical keyboard 416 B.
  • the computing device 210 may output the search results in other locations within the user interface 414 , including partially or entirely outside the graphical keyboard 416 B. In still further examples, the computing device 210 may output search results partially or exclusively as audio information, and/or may output some types of search result information in tactile form.
  • the computing device 210 continues to present both the search element 418 C and the suggestion region 418 B in the graphical keyboard 416 B along with, and above, the search results 418 E.
  • the computing device 210 may cause either or both to be displayed at other locations within the user interface 414 , or removed from the display.
  • FIGS. 5A-5F are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • FIGS. 5A-5F illustrate example user interfaces 514 ; many other examples of user interfaces may be used or presented in other instances or examples.
  • Each of the user interfaces 514 may correspond to a graphical user interface output by computing devices 110 , 210 , or 310 from FIGS. 1-3 .
  • FIGS. 5A-5F are described below within the context of computing device 210 of FIG. 2 .
  • each example user interface 514 forms part of a chat application user interface of a chat application from application module 224 .
  • User interface 514 includes output region 516 A, edit region 516 C, and graphical keyboard 516 B.
  • Output region 516 A and edit region 516 C may both be associated with the user interface of an application or service executing on or accessible from computing device 210 , such as the chat application described in connection FIGS. 4A-4D .
  • Included within the graphical keyboard 516 B are graphical elements displayed as keys 518 A. Also included within graphical keyboard 516 B is a suggestion region 518 B and search element 518 C.
  • the search element 518 C is displayed above the graphical keys 518 A, but may be located elsewhere within or outside the graphical keyboard 516 B. In other examples, it may be located elsewhere within the user interface 514 , or in still other examples, could be implemented as an external or hardware input to the computing device 210 .
  • computing device 210 may be exchanging electronic messages with another device, and computing device 210 may output the content of recent exchanged messages in output region 516 A.
  • the computing device 210 may also output and format a suggested query in edit region 518 B.
  • computing device 210 may receive an indication of input selecting the search element 518 C, and responsive to receiving the indication of input selecting the search element 518 C, computing device 210 may output, for display, a separate region in which suggested search queries are displayed.
  • keyboard module 222 may receive an indication of user input detected at a location at which search element 518 C is displayed within graphical keyboard 516 B, as shown in FIG. 5A . Responsive to the indication of user input, the computing device 210 may update the user interface 514 to that shown in FIG. 5B so that user interface 514 includes query suggestion region 518 D.
  • the graphical keyboard 516 B shown in the user interface 514 of FIG. 5B may be considered to be in search mode.
  • search mode may be characterized by the presence of a query suggestion region 518 D.
  • computing device 210 may change the input focus to the suggestion region 518 B.
  • This change in input focus may be signified by the cursor shown within the suggestion region 518 B, and the absence of a cursor shown within the edit region 516 C.
  • computing device 210 causes the query suggestion region 518 D to be displayed above the suggestion region 518 B, and below the edit region 516 C. This arrangement creates some visual distance between the suggestion region 518 B and edit region 516 C. This visual distance may tend to reduce user confusion in situations where both the suggestion region 518 B and the edit region 516 C are each user interface components that could potentially have input focus at various times.
  • computing device 210 could position the query suggestion region 518 D elsewhere within the graphical keyboard 516 B. In still further examples, the computing device 210 could position the query suggestion region 518 D elsewhere within the user interface 514 , partially or entirely outside the graphical keyboard 516 B.
  • the query suggestion region 518 D in some examples may include suggested search queries or other information, each of which may represent selectable search queries that could be initiated by computing device 210 .
  • suggested search queries or other information each of which may represent selectable search queries that could be initiated by computing device 210 .
  • computing device 210 may output for display within the query suggestion region 518 D suggested queries that include historical queries (previous searches performed by the device 210 ).
  • computing device 210 may output for display within the query suggestion region 518 D suggested queries that may be of interest to the user based on the user's current activity, application, context, location, situation, or other information.
  • Query suggestions in region 518 D may also be generated based, as least in part, on the contents of elements of region 516 A, such as content from a conversation “Dinner tonight?” (or similar content and context from different user interfaces that may populate the region of 516 A).
  • computing device 210 has included suggested query “Restaurants near me” in the query suggestion region 518 D.
  • This suggested query may be based at least in part on the subject of the chat shown in output region 516 A.
  • the other suggested queries, “Fast food” and “Movie times” may also be based to some extent on the subject of the chat shown in output region 516 A.
  • keyboard module 222 of computing device 210 may receive indications of taps or gestures at keys 518 A and determine, based on the user input, text that computing device 210 formats and displays within suggestion region 518 B. For example, as shown in FIG. 5C , the computing device 210 may cause the text “meatball” to be displayed within suggestion region 518 B. Keyboard module 222 of computing device 210 may also, in response to the user input, generate suggested queries that computing device 210 formats and displays within query suggestion region 518 D, as shown in FIG. 5C . The suggested queries shown in query suggestion region 518 D in FIG. 5C may be at least partially based on, and may be appropriate suggested queries for, the “meatball” text shown in the suggestion region 518 B.
  • keyboard module 222 of computing device 210 may continue to receive additional indications of taps or gestures at keys 518 A and may continue to determine, based on the additional user input, additional text that computing device 210 formats and displays within suggestion region 518 B. For example, as shown in FIG. 5D , the computing device 210 may cause the text “meatball sho” to be displayed within suggestion region 518 B. Also in response to the additional user input, keyboard module 222 of computing device 210 may generate updated suggested queries that computing device 210 formats and displays within query suggestion region 518 D, as shown in FIG. 5D .
  • the suggested queries shown in query suggestion region 518 D in FIG. 5D may be at least partially based on, and may be appropriate suggested queries for, the “meatball sho” text shown in the suggestion region 518 B.
  • a suggested query that is considered more relevant than other suggested queries may be displayed at the top of the query suggestion region 518 D.
  • computing device 210 may execute a model that assigns a score or probability to each of the suggested queries based on text input and contextual information.
  • Computing device 210 may rank the queries based on the assigned scores and probabilities.
  • Computing device 210 may display the highest scoring or highest probability queries near the top of suggestion region 518 D and the lower scoring or lower probability queries near the bottom.
  • the queries shown in query suggestion region 518 D may be listed in descending order of perceived relevance. Other arrangements could be used.
  • the determination of relevance of each suggested query may be made, at least partially, by the computing device 210 , but in other examples, this determination may be made elsewhere or by another device or other devices, such as by a search engine or other service.
  • computing device 210 may receive an indication of input selecting a suggested search query, and responsive to receiving the indication of input, computing device 210 may execute, based on the suggested query, a search for information.
  • keyboard module 222 may receive an indication of user input detected at a location at which suggested search query 509 is displayed within query suggestion region 518 D. Responsive to the indication of user input, the computing device 210 may perform a search corresponding to the “Meatball Shoppe” suggested search query 509 , and may also update the user interface 514 to indicate that a search is being performed. When search results are available, the computing device 210 may update the user interface 514 to that shown in FIG. 5F .
  • the search results 518 E are shown within the graphical keyboard 516 B, and in this example, search results 518 E are shown below the suggestion region 518 B, substantially replacing the graphical keys 518 A previously shown in FIG. 5E .
  • the search results 518 E may be presented in card-based form.
  • the query suggestion region 518 D has been removed from the user interface 514 in the example of FIG. 5F .
  • the search results could be shown in other formats, and in other locations within the graphical keyboard 516 B.
  • search results may be shown partially or entirely outside the graphical keyboard 516 B.
  • the search result information could also be provided partially or exclusively as audio information, which may enable the user to listen to at least some of the search result information.
  • the computing device 210 continues to display both the search element 518 C and the suggestion region 518 B in the user interface 514 along with, and above, the search results 518 E.
  • computing device 210 could remove either or both of these elements from the user interface 514 , or the computing device 210 could display these elements at other locations within the user interface 514 .
  • computing device 210 returns the input focus to the edit region 516 C.
  • individual cards from results 518 E may themselves be links such that a user input or gesture indicating selection thereof may link to the website indicated by the card.
  • a card from results 518 E may be “dragged” (such as by a user touching the PSD at the location of the card) and, without lifting the touching finger, move the finger to the output region 516 A. This may have the effect of copying the card into the context of the output region 516 A, such as adding it as a sendable and actionable (i.e., including a selectable link) element of a communication in a chat application.
  • cards in results 518 E may include controls, such that a user may by input or gesture invoke selected actions with respect to the cards or card content, such as copy, open, send, share, map, call, message, purchase, make reservation, and so forth. These controls may be specific to individual search results and cards (i.e., not all cards in results 518 E need provide the same controls).
  • FIGS. 6A-6C are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • FIGS. 6A-6C illustrate example user interfaces 614 ; many other examples of graphical user interfaces may be used or presented in other instances or examples.
  • Each of the user interfaces 614 may correspond to a graphical user interface output by computing devices 110 , 210 , or 310 from FIGS. 1-3 .
  • FIGS. 6A-6C are described below within the context of computing device 210 of FIG. 2 .
  • each example user interface 614 includes output region 616 A, edit region 616 C, and graphical keyboard 616 B.
  • Output region 616 A and edit region 616 C may both be associated with the user interface of an application or service executing on or accessible from computing device 210 , such as a chat application described in connection with FIGS. 4A-4D and FIGS. 5A-5F .
  • Included within the graphical keyboard 616 B are graphical keys 618 A, a suggestion region 618 B, and a search element 618 C.
  • keyboard module 222 of computing device 210 may receive indications of taps or gestures at keys 618 A and determine, based on the user input, a text string that computing device 210 formats and displays within edit region 616 C. For example, as shown in FIG. 6A , the computing device 210 may cause the text string “What should I have for dinner to” to be displayed within edit region 616 C.
  • the computing device 210 may, in some examples, format and display within suggestion region 618 B one or more suggested words, such as the suggested word “tonight” shown in suggestion region 618 B.
  • Keyboard module 222 of computing device 210 may also, in response to the user input, generate suggested queries that computing device 210 may format and display within suggestion region 618 B, such as the “Restaurants near me” suggested query shown in FIG. 6A . Any suggested queries shown in query suggestion region 618 B in FIG. 6A may be based on, and may be appropriate suggested queries for, the text shown in the edit region 616 C.
  • keyboard module 222 of computing device 210 may continue to receive additional indications of taps or gestures at keys 618 A and determine, based on the user input, text that computing device 210 formats and displays within edit region 616 C.
  • the text displayed in edit region 616 C may be a continuation of the partially-composed text message previously displayed in FIG. 6A .
  • the computing device 210 does not transmit this text message to the other person's computing device at the point shown in FIG. 6B , and does not do so until the computing device 210 detects an indication of input selecting the key 602 (or in other examples, a “send” key).
  • the computing device 210 before the computing device 210 receives an indication of input selecting key 602 , the computing device 210 receives an indication of input selecting search element 618 C.
  • the user may select the search element 618 C before selecting the key 602 for a number of possible reasons.
  • the user may have intended the text shown in edit region 616 C to be a search query, rather than a text message, but may have mistakenly entered the text in the edit region 616 C, rather than first selecting the search element 618 C and entering the text in suggestion region 618 B.
  • the user may have originally intended the text in edit region 616 C to be a text message, but may have changed his or her mind and decided to perform a search instead of (or in addition to) submitting the text as a chat message.
  • search instead of (or in addition to) submitting the text as a chat message.
  • the computing device 210 in response to the indication of input selecting search element 618 C, may output, for display, query suggestion region 618 D.
  • the computing device 210 has also changed the input focus to the suggestion region 618 B, and cleared the suggestion region 618 B of any text or suggested queries.
  • the computing device 210 formats and displays within query suggestion region 618 D a number of suggested queries.
  • computing device 210 generates the uppermost displayed suggested query based on the text in the edit region 616 C, and this suggested query contains the exact same text as that in the edit region 616 C.
  • the text may be used as a suggested query in the query suggestion region 618 D after the computing device 210 makes any stylistic or spelling corrections or other modifications.
  • the computing device also includes within the query suggestion region 618 D other suggested queries which may be relevant or pertinent to the text in edit region 616 C and possibly the information in the output region 616 A.
  • the graphical keyboard 616 B may, in some examples, be considered to be in search mode in FIG. 6C .
  • a string of characters that has already been transmitted as a text message by the computing device 210 may still be available as a suggested search query.
  • computing device 210 may, before detecting that the user has selected the search element 618 C, transmit the text in edit region 616 C in FIG. 6B as a chat message in response to user input (e.g., a selection of key 602 ). The computing device may then later detect an indication of input selecting the search element 618 C, and in response to detecting this input, the computing device 210 may cause a query suggestion region to be displayed within the user interface 614 .
  • the computing device 210 may display suggested queries within the query suggestion region that are similar to or the same as those shown in query suggestion region 618 D in FIG. 6C .
  • These suggested queries may include a suggested search query that is created with the exact text of the most recent chat message from the chat application, enabling the same text to be used as a search query.
  • the most recent and earlier text messages exchanged in the chat application may also provide enough context to generate useful other suggested search queries.
  • FIG. 7 is a flowchart illustrating an example operation of a computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • the process of FIG. 7 may be performed by one or more processors of a computing device, such as computing devices 110 , 210 , or 310 as illustrated in FIGS. 1-3 .
  • FIG. 7 is described below within the context of computing device 110 of FIG. 1 .
  • a computing device 110 may output, for display, a graphical user interface.
  • This graphical user interface may include a user interface for one or more applications.
  • the graphical user interface may also include a graphical keyboard ( 702 ).
  • the application may be a chat application, such as that described in connection with FIG. 1 , but other applications, such as a fitness, calendar, personal assistant, prediction, map, navigation, tracking, social media, email, or game application (or many others) could also or alternatively be used with a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • the graphical keyboard may include a plurality of graphical keys, a suggestion region, and a search element, such as was described, for example, in connection with FIG. 1 .
  • computing device 110 may interpret input associated with a user's selection of the search element as a request to transition the graphical keyboard into search mode in accordance with one or more aspects of the present disclosure.
  • the graphical keyboard may further include a query suggestion region, such as the query suggestion region 118 D shown in FIG. 1 .
  • the query suggestion region 118 D may be may be selectively displayed within the graphical keyboard in accordance with one or more aspects of the present disclosure.
  • the computing device 110 may determine, based on the user's interaction with one or more of the graphical keys within the graphical keyboard, one or more suggested words and suggested search queries.
  • UI module 120 may detect one or more inputs at mapped locations on PSD 112 and may generate one or more touch events corresponding to the input from the user.
  • Keyboard module 122 may include a spatial module which may select, based on the touch events, one or more keys corresponding to the user input.
  • Keyboard module 122 may also include a language module, which may identify one or more words within a lexicon that may be used to identify suggested words. The keyboard module 122 may also use this information to generate suggested queries, but keyboard module 122 may also use additional information.
  • keyboard module 122 may also base the suggested queries on the user's current activity, context, location, or situation in accordance with one or more aspects of the present disclosure.
  • Computing device 110 may update the suggested words and/or suggested queries as the computing device 110 detects input corresponding to selections of graphical keys within the graphical keyboard ( 704 ).
  • the computing device 110 may determine, based on the user's input or interaction with the computing device 110 , whether the user seeks to operate the graphical keyboard in search mode. In some examples, computing device 110 may determine that the user seeks to operate the graphical keyboard in search mode in response to detecting a selection of a search element displayed within the graphical keyboard (YES path from 706 ). Search mode may generally correspond to the sub-process 720 shown in FIG. 7 , but in other examples, search mode in accordance with one or more aspects of the present disclosure may involve more or less or different operations.
  • the computing device 110 may also determine whether it has received input selecting one of the suggested queries, if any, that may be displayed within the suggestion region ( 708 ). If the computing device does not detect input selecting a suggested query, the computing device 110 continues to update the suggested queries in the suggestion region in response to user input ( 704 ). If the computing device 110 does detect input selecting a suggested query in the suggestion region, the computing device 110 obtains information associated with the selected suggested query.
  • the suggested queries are selectable links, and in response to the computing device 110 detecting input selecting a suggested query, the computing device determines information obtained by invoking the selectable link.
  • Obtaining information corresponding to a selected suggested query may in some examples involve invoking a search operation using the content of the selected suggested query as a search query. In other examples, obtaining this information might not require a search operation. Situations where a search operation might not be performed include when the selected suggested query points directly to an item of information, or points to an item of information (e.g., a document, or a contact) on the computing device 110 .
  • the computing device 110 may output the obtained information for display (YES path from 708 ).
  • the computing device 110 may output, for display, a query suggestion region within the user interface.
  • the query suggestion region is displayed within the graphical keyboard, as shown in FIG. 1 .
  • Computing device 110 may format and display suggested search queries that may be updated as the computing device 110 continues to detect user input ( 712 ).
  • the computing device 110 may perform a search corresponding to the selected suggested query, and output, for display, the requested information ( 718 ).
  • the computing device 110 may perform a search corresponding to the string of text characters, and output, for display, the requested information ( 718 ).
  • the computing device 110 may continue to monitor user input and update the query suggestion region with suggested queries (NO path from 716 ).
  • the operations shown in FIG. 7 may be performed in a different order or presented in a different sequence, but still be in accordance with one or more aspects of the present disclosure.
  • the order of operations 706 and 708 may be reversed or done in parallel in FIG. 7 (which may involve modifications to other operations, or the order of other operations) while still remaining in accordance with one or more aspects of the present disclosure.
  • the order of operations 714 and 716 may switched or done in parallel in FIG. 7 (which may involve modifications to other operations, or the order of other operations) while still remaining in accordance with one or more aspects of the present disclosure.
  • a process in accordance with one or more aspects of the present disclosure may be implemented with less than the operations shown in FIG. 7
  • a process in accordance with one or more aspects of the present disclosure may be implemented with more or different than the operations shown in FIG. 7 .
  • FIG. 8 is a flowchart illustrating an example operation of a computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure.
  • the process of FIG. 8 may be performed by one or more processors of a computing device, such as computing devices 110 , 210 , or 310 as illustrated in FIGS. 1-3 .
  • FIG. 8 is described below within the context of computing device 110 of FIG. 1 .
  • computing device may output a graphical keyboard for display ( 802 ).
  • a chat application executing at computing device 110 may invoke keyboard module 122 (e.g., a standalone application or function of computing device 110 that is separate from the chat application) to present graphical keyboard 116 B at PSD 112 .
  • the graphical keyboard 116 B may comprise a plurality of keys and a search element.
  • Computing device 110 may determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion ( 804 ).
  • keyboard module 122 may receive information about touch inputs detected at locations of PSD 112 at which graphical keys 118 A of graphical keyboard 116 B are being displayed.
  • Keyboard module 122 may determine a suggested query based at least in part on this information.
  • Computing device 110 may receive an indication of input selecting the search element ( 806 ).
  • keyboard module 122 may receive information about a touch input detected at a location of PSD 112 at which search element 118 C is being displayed.
  • Computing device 110 may, in response to receiving the indication of input, output for display an updated graphical keyboard that includes a selectable link associated with the query suggestion.
  • keyboard module 122 may format and display a selectable link such as suggested query 406 shown in FIG. 4B or suggested query 508 in FIG. 5B .
  • a method comprising: outputting, by a mobile computing device, for display, a graphical keyboard comprising a plurality of keys and a search element; determining, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receiving, by the mobile computing device, an indication of input selecting the search element; and responsive to receiving the indication of input, outputting, by the mobile computing device, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
  • Clause 2 The method of clause 1, wherein the indication of input is a first indication of input, the method further comprising: receiving, by the mobile computing device, a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determining, based on the second indication of input, information obtained by invoking the selectable link; and outputting, by the mobile computing device, for display, the information obtained by invoking the selectable link.
  • outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, and wherein outputting the information obtained by invoking the selectable link includes outputting the information by substantially replacing the plurality of keys with the information.
  • determining information obtained by invoking the selectable link includes determining search result information by performing a search operation using content included in the selectable link as a search query.
  • outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, and with the selectable link positioned above the plurality of keys.
  • Clause 8 The method of clause 1-7, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, wherein the selectable link is a first selectable link, and wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, the method further comprising: determining, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, outputting, by the mobile computing device, for display, a further updated graphical keyboard that includes a second selectable link associated with the updated query suggestion; receiving, by the mobile computing device, a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determining, based on the second indication of input, search result information; and outputting, by the mobile computing device, for display, the search result information.
  • outputting the updated graphical keyboard includes outputting the updated graphical keyboard as part of an application graphical user interface, and wherein the application graphical user interface includes an edit region that is positioned above the selectable link associated with the query suggestion.
  • determining the query suggestion includes determining a plurality of query suggestions including a historical query suggestion based on a prior search, and wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including a selectable link associated with the historical query suggestion.
  • a mobile device comprising: a presence-sensitive display component; at least one processor; and a memory that stores instructions that when executed cause the at least one processor to: output, for display, a graphical keyboard comprising a plurality of keys and a search element; determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receive an indication of input selecting the search element; and responsive to receiving the indication of input, output, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
  • Clause 12 The mobile device of clause 11, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to: receive a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, information obtained by invoking the selectable link; and output, for display, the information obtained by invoking the selectable link.
  • Clause 13 The mobile device of clause 12, wherein the instructions, when executed, further cause the at least one processor to output the information obtained by invoking the selectable link by at least outputting, within the updated graphical keyboard, the information obtained by invoking the selectable link.
  • Clause 14 The mobile device of clause 13, wherein the updated graphical keyboard is output for display further including the plurality of keys, and wherein the information obtained by invoking the selectable link is output by substantially replacing the plurality of keys with the information obtained by invoking the selectable link.
  • Clause 15 The mobile device of clause 11-14, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the selectable link is a first selectable link, wherein the updated graphical keyboard is output for display further including the plurality of keys, and wherein the instructions, when executed, further cause the at least one processor to: determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, output, for display, a further updated graphical keyboard that includes a second selectable link associated with the updated query suggestion; receive a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, search result information; and output, for display, the search result information.
  • a computer-readable storage medium comprising instructions that, when executed, cause at least one processor of a computing device to: output, for display, an application graphical user interface comprising an edit region, a plurality of keys, and a search element; determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receive an indication of input selecting the search element; and responsive to receiving the indication of input, output, for display, an updated application graphical user interface that includes a selectable link associated with the query suggestion, wherein the selectable link is positioned between the edit region and the plurality of keys.
  • Clause 17 The computer-readable storage medium 16 , wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to: receive a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, information obtained by invoking the selectable link; and output, for display, the information obtained by invoking the selectable link.
  • Clause 18 The computer-readable storage medium of clause 17, wherein the instructions, when executed, further cause the at least one processor to output the information obtained by invoking the selectable link by at least outputting, below the edit region, the information obtained by invoking the selectable link.
  • Clause 19 The computer-readable storage medium of clause 18, wherein the updated graphical user interface is output for display further including the plurality of keys, and wherein the information obtained by invoking the selectable link is output by substantially replacing the plurality of keys with the information obtained by invoking the selectable link.
  • Clause 20 The computer-readable storage medium of clause 16-19, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the selectable link is a first selectable link, wherein the updated graphical user interface is output for display further including the plurality of keys, and wherein the instructions, when executed, further cause the at least one processor to: determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, output, for display, a further updated graphical user interface that includes a second selectable link associated with the updated query suggestion; receive a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, search result information; and output, for display, the search result information.
  • Clause 21 A system comprising means for performing any of the methods of clauses 1-10.
  • Clause 22 A computing device comprising means for performing any of the methods of clauses 1-10.
  • a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information.
  • information e.g., context, locations, speeds, search queries, etc.
  • the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user.
  • certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed.
  • a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined.
  • location information such as to a city, ZIP code, or state level
  • Computer-readable media may include computer-readable storage media, which corresponds to a tangible medium such as data storage media, or communication media including any medium that facilitates transfer of a computer program from one place to another, e.g., according to a communication protocol.
  • computer-readable media generally may correspond to (1) tangible computer-readable storage media, which is non-transitory or (2) a communication medium such as a signal or carrier wave.
  • Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for implementation of the techniques described in this disclosure.
  • a computer program product may include a computer-readable medium.
  • such computer-readable storage media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage, or other magnetic storage devices, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer.
  • any connection is properly termed a computer-readable medium.
  • a computer-readable medium For example, if instructions are transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of medium.
  • DSL digital subscriber line
  • Disk and disc includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc, where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
  • processors such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry.
  • DSPs digital signal processors
  • ASICs application specific integrated circuits
  • FPGAs field programmable logic arrays
  • processors may refer to any of the foregoing structure or any other structure suitable for implementation of the techniques described.
  • the functionality described may be provided within dedicated hardware and/or software modules. Also, the techniques could be fully implemented in one or more circuits or logic elements.
  • the techniques of this disclosure may be implemented in a wide variety of devices or apparatuses, including a wireless handset, an integrated circuit (IC) or a set of ICs (e.g., a chip set).
  • IC integrated circuit
  • a set of ICs e.g., a chip set.
  • Various components, modules, or units are described in this disclosure to emphasize functional aspects of devices configured to perform the disclosed techniques, but do not necessarily require realization by different hardware units. Rather, as described above, various units may be combined in a hardware unit or provided by a collection of interoperative hardware units, including one or more processors as described above, in conjunction with suitable software and/or firmware.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Human Computer Interaction (AREA)
  • Databases & Information Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Mathematical Physics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

A computing device is described that includes a presence-sensitive display component, at least one processor, and a memory. The memory stores instructions that when executed cause the at least one processor to output, for display, a graphical keyboard comprising a plurality of keys and a search element, and determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion. The instructions, when executed, further cause the at least one processor to receive an indication of input selecting the search element, and responsive to receiving the indication of input, output, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.

Description

RELATED APPLICATION
This application is a continuation of U.S. application Ser. No. 15/134,029, filed Apr. 20, 2016, the entire contents of which are hereby incorporated by reference.
BACKGROUND
Despite being able to simultaneously execute several applications, some mobile computing devices can only present a graphical user interface (GUI) of a single application at a time. To interact with multiple applications at once, a user of a mobile computing device may have to switch between different application GUIs. For example, a user of a mobile computing device may have to cease entering text in a messaging application and provide input to cause the device to toggle to a search application to search for a particular piece of information to use when composing a message or otherwise entering text. Providing several inputs required by some computing devices to perform various tasks can be tedious, repetitive, and time consuming.
SUMMARY
In one example, a method includes outputting, by a mobile computing device, for display, a graphical keyboard comprising a plurality of keys and a search element, determining, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion. The method further includes receiving, by the mobile computing device, an indication of input selecting the search element, and responsive to receiving the indication of input, outputting, by the mobile computing device, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
In another example, a mobile device comprises a presence-sensitive display component, at least one processor, and a memory that stores instructions that when executed cause the at least one processor to: output, for display, a graphical keyboard comprising a plurality of keys and a search element, determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion, receive an indication of input selecting the search element, and responsive to receiving the indication of input, output, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
In another example, a computer-readable storage medium comprises instructions that, when executed, cause at least one processor of a computing device to: output, for display, an application graphical user interface comprising an edit region, a plurality of keys, and a search element, determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion, receive an indication of input selecting the search element, and responsive to receiving the indication of input, output, for display, an updated application graphical user interface that includes a selectable link associated with the query suggestion, wherein the selectable link is positioned between the edit region and the plurality of keys.
The details of one or more examples are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the disclosure will be apparent from the description and drawings, and from the claims.
BRIEF DESCRIPTION OF DRAWINGS
FIG. 1 is a conceptual diagram illustrating an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIG. 2 is a block diagram illustrating an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIG. 3 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure.
FIGS. 4A-4D are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIGS. 5A-5F are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIGS. 6A-6C are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIG. 7 is a flowchart illustrating example operations of a computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
FIG. 8 is a flowchart illustrating example operations of a computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure.
DETAILED DESCRIPTION
In general, this disclosure is directed to techniques that include enabling a computing device to display suggested search queries near the keys of a graphical keyboard while a user provides input at the keys of the graphical keyboard. For example, the computing device may output a graphical user interface (GUI) that includes a graphical keyboard. The graphical keyboard may suggest a search query, for example, based on text determined from user input as the user interacts with the graphical keyboard. In some examples, in addition to analyzing text, the graphical keyboard may generate suggested queries by analyzing other information such as the user's current or recent tasks, context, or prior interaction with the graphical keyboard. The suggested queries may therefore be relevant to the user's current or recent tasks in addition to the text inferred from user input, and may be updated in response to further typing or other input or actions detected by the computing device.
The graphical keyboard may display the suggested queries within the graphical keyboard, and in some examples, near a particular text field or other location of the graphical keyboard that has a current input focus. For example, the graphical keyboard may display a suggested query between the graphical keys and the edit region at which the user is currently typing. This arrangement may make it easier for the user to stay on task and easily view, and provide additional input to select suggested queries. In situations where the GUI has more than one text field that could have input focus, the graphical keyboard may display suggested queries above the particular text field that has the input focus, and below the other text field(s). Visual space created between the text fields in such an arrangement may reduce user confusion as to which text field has input priority. If the graphical keyboard detects a user selection of a suggested query, the graphical keyboard may execute a search based on a suggested query and display search results within the graphical keyboard, in some examples, replacing some or all keys within the graphical keyboard with search results. Such a graphical keyboard may allow a user to view suggested queries and obtain search results based on suggested queries without, in some cases, causing the user to lose focus, switch tasks, or switch between application user interfaces.
Throughout the disclosure, examples are described where a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information. For example, in situations discussed below, before a computing device or computing system can collect or may make use of information associated with a user, the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user. In addition, certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed. For example, a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined. Thus, the user may have control over how information is collected about the user and used by the computing device and computing system.
FIG. 1 is a conceptual diagram illustrating an example computing device 110 that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure. Computing device 110 may represent a mobile device, such as a smart phone, a tablet computer, a laptop computer, computerized watch, or any other types of wearable and non-wearable, mobile or non-mobile computing devices that may output a graphical keyboard for display.
Computing device 110 includes a presence-sensitive display (PSD) 112, user interface (UI) module 120 and keyboard module 122. Modules 120 and 122 may perform operations described using software, hardware, firmware, or a mixture of hardware, software, and firmware residing in and/or executing at computing device 110. One or more processors of computing device 110 may execute instructions that are stored at a memory or other non-transitory storage medium of computing device 110 to perform the operations of modules 120 and 122. Computing device 110 may execute modules 120 and 122 as virtual machines executing on underlying hardware. Modules 120 and 122 may execute as one or more services of an operating system or computing platform. Modules 120 and 122 may execute as one or more executable programs at an application layer of a computing platform.
PSD 112 of computing device 110 may function as respective input and/or output devices for computing device 110. PSD 112 may be implemented using various technologies. For instance, PSD 112 may function as input devices using presence-sensitive input screens, such as resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, projective capacitance touchscreens, pressure sensitive screens, acoustic pulse recognition touchscreens, or another presence-sensitive display technology. PSD 112 may also function as output (e.g., display) devices using any one or more display devices, such as liquid crystal displays (LCD), dot matrix displays, light emitting diode (LED) displays, organic light-emitting diode (OLED) displays, e-ink, or similar monochrome or color displays capable of outputting visible information to a user of computing device 110.
PSD 112 may detect input (e.g., touch and non-touch input) from a user of respective computing device 110. PSD 112 may detect indications of input by detecting one or more gestures from a user (e.g., the user touching, pointing, and/or swiping at or near one or more locations of PSD 112 with a finger or a stylus pen). PSD 112 may output information to a user in the form of a user interface (e.g., user interface 114), which may be associated with functionality provided by computing device 110. Such user interfaces may be associated with computing platforms, operating systems, applications, and/or services executing at or accessible from computing device 110 (e.g., electronic message applications, chat applications, Internet browser applications, mobile or desktop operating systems, social media applications, electronic games, and other types of applications). For example, PSD 112 may present user interface 114 which, as shown in FIG. 1, is a graphical user interface of a chat application executing at computing device 110 and includes various graphical elements displayed at various locations of PSD 112.
As shown in FIG. 1, user interface 114 is a chat user interface. However, user interface 114 may be any graphical user interface which includes (or accesses the services of) a graphical keyboard with integrated search features. User interface 114 includes output region 116A, graphical keyboard 116B, and edit region 116C. A user of computing device 110 may provide input at graphical keyboard 116B to produce textual characters within edit region 116C that form the content of the electronic messages displayed within output region 116A. The messages displayed within output region 116A form a chat conversation between a user of computing device 110 and a user of a different computing device.
In some examples, region 116A is a user interface for a first application that is separate and distinct from a keyboard application that provides the graphical keyboard 116B. In such examples, the first application utilizes the services of the keyboard application for certain input and output functions, including text and symbol entry, text and symbol display, text and symbol entry editing, attachment selection, process invocation (such as spell checking, font and style changes, etc.), transmission and receipt controls, etc. Furthermore, the keyboard application can implement search functionality, and in some examples provide that search functionality, or results of that search functionality, to the first application.
UI module 120 manages user interactions with PSD 112 and other components of computing device 110. In other words, UI module 120 may act as an intermediary between various components of computing device 110 to make determinations based on user input detected by PSD 112 and generate output at PSD 112 in response to the user input. UI module 120 may receive instructions from an application, service, platform, or other module of computing device 110 to cause PSD 112 to output a user interface (e.g., user interface 114). UI module 120 may manage inputs received by computing device 110 as a user views and interacts with the user interface presented at PSD 112 and update the user interface in response to receiving additional instructions from the application, service, platform, or other module of computing device 110 that is processing the user input.
Keyboard module 122 represents an application, service, or component executing at or accessible to computing device 110 that provides computing device 110 with a graphical keyboard having integrated search features. Keyboard module 122 may switch between operating in text-entry mode in which keyboard module 122 functions similar to a traditional graphical keyboard, or search mode in which keyboard module 122 performs various integrated search functions.
In some examples, keyboard module 122 may be a stand-alone application, service, or module executing at computing device 110 and in other examples, keyboard module 122 may be a sub-component of an operating system controlling operation of device 110. For example, keyboard module 122 may be integrated into a chat or messaging application executing at computing device 110 whereas in other examples, keyboard module 122 may be a stand-alone application or subroutine that is invoked by an application or operating platform of computing device 110 any time an application or operating platform requires graphical keyboard input functionality. In some examples, computing device 110 may download and install keyboard module 122 from an application repository of a service provider (e.g., via the Internet). In other examples, keyboard module 122 may be preloaded as part of the operating system of computing device 110.
When operating in text-entry mode, keyboard module 122 of computing device 110 may perform traditional, graphical keyboard operations used for text-entry, such as: generating a graphical keyboard layout for display at PSD 112, mapping detected inputs at PSD 112 to selections of graphical keys, determining characters based on selected keys, or predicting or autocorrecting words and/or phrases based on the characters determined from selected keys.
Graphical keyboard 116B includes graphical elements displayed as graphical keys 118A. Keyboard module 122 may output information to UI module 120 that specifies the layout of graphical keyboard 116B within user interface 114. For example, the information may include instructions that specify locations, sizes, colors, and other characteristics of graphical keys 118A. Based on the information received from keyboard module 122, UI module 120 may cause PSD 112 to display graphical keyboard 116B as part of user interface 114.
Each key of graphical keys 118A may be associated with a respective character (e.g., a letter, number, punctuation, or other character) displayed within the key. A user of computing device 110 may provide input at locations of PSD 112 at which one or more of graphical keys 118A are displayed to input content (e.g., characters, search results, etc.) into edit region 116C (e.g., for composing messages that are sent and displayed within output region 116A or for inputting a search query that computing device 110 executes from within graphical keyboard 116B). Keyboard module 122 may receive information from UI module 120 indicating locations associated with input detected by PSD 112 that are relative to the locations of each of the graphical keys. Using a spatial and/or language model, keyboard module 122 may translate the inputs to selections of keys and characters, words, and/or phrases.
For example, PSD 112 may detect user inputs as a user of computing device 110 provides the user inputs at or near a location of PSD 112 where PSD 112 presents graphical keys 118A. UI module 120 may receive, from PSD 112, an indication of the user input detected by PSD 112 and output, to keyboard module 122, information about the user input. Information about the user input may include an indication of one or more touch events (e.g., locations and other information about the input) detected by PSD 112.
Based on the information received from UI module 120, keyboard module 122 may map detected inputs at PSD 112 to selections of graphical keys 118A, determine characters based on selected graphical keys 118A, and predict or autocorrect words and/or phrases determined based on the characters associated with the selected keys 118A. For example, keyboard module 122 may include a spatial model that may determine, based on the locations of keys 118A and the information about the input, the most likely one or more keys 118A being selected. Responsive to determining the most likely one or more keys 118A being selected, keyboard module 122 may determine one or more characters, words, and/or phrases. For example, each of the one or more keys 118A being selected from a user input at PSD 112 may represent an individual character or a keyboard operation. Keyboard module 122 may determine a sequence of characters selected based on the one or more selected keys 118A. In some examples, keyboard module 122 may apply a language model to the sequence of characters to determine one or more the most likely candidate letters, morphemes, words, and/or phrases that a user is trying to input based on the selection of keys 118A.
Keyboard module 122 may send the sequence of characters and/or candidate words and phrases to UI module 120 and UI module 120 may cause PSD 112 to present the characters and/or candidate words determined from a selection of one or more keys 118A as text within edit region 116C. In some examples, when functioning as a traditional keyboard for performing text-entry operations, and in response to receiving a user input at graphical keys 118A (e.g., as a user is typing at graphical keyboard 116B to enter text within edit region 116C), keyboard module 122 may cause UI module 120 to display the candidate words and/or phrases as one or more selectable spelling corrections and/or selectable word or phrase suggestions within suggestion region 118B.
In addition to performing traditional, graphical keyboard operations used for text-entry, keyboard module 122 of computing device 110 also provides integrated search capability. That is, rather than requiring a user of computing device 110 to navigate away from user interface 114 which provides graphical keyboard 116B (e.g., to a different application or service executing at or accessible from computing device 110), keyboard module 122 may operate in search mode in which keyboard module 122 may invoke search operations and present search results within the same region of PSD 112 at which graphical keyboard 116B is displayed. Keyboard module 122 may include routines for executing search functions and/or may include interfaces for communicating with a separate search engine or resource that receives search parameters from keyboard module 122 and returns search results. Search functionality invoked by keyboard module 122 may search memory of device 110 and/or external memory.
As indicated above, keyboard module 122 may execute as a stand-alone application, service, or module executing at computing device 110 or as a single, integrated sub-component thereof. Therefore, if keyboard module 122 forms part of a chat or messaging application executing at computing device 110, keyboard module 122 may provide the chat or messaging application with text-entry capability as well as search capability. Similarly, if keyboard module 122 is a stand-alone application or subroutine that is invoked by an application or operating platform of computing device 110 any time an application or operating platform requires graphical keyboard input functionality, keyboard module 122 may provide the invoking application or operating platform with text-entry capability as well as search capability.
Keyboard module 122 may further operate in search mode. In some examples, when operating in search mode, keyboard module 122 may cause graphical keyboard 116B to include search element 118C. Search element 118C represents a selectable element of graphical keyboard 116B for invoking one or more of the various search features of graphical keyboard 116B. By selecting search element 118C (e.g., by tapping or gesturing at a location or within a region of PSD 112 at which search element 118C is displayed), a user can cause computing device 110 to invoke the various integrated search features without having to navigate to a separate application, service, or other feature executing at or accessible from computing device 110.
For example, UI module 120 may output information to keyboard module 122 indicating that a user of computing device 110 may have selected search element 118C. Responsive to determining that search element 118C was selected, keyboard module 122 may transition to operating in search mode. While operating in search mode, keyboard module 122 may reconfigure graphical keyboard 116B to execute search features as opposed to operations that are primarily attributed to text entry.
For example, keyboard module 122 may configure suggestion region 118B to present suggested content (e.g., predicted search queries, predicted emoticons or so called “emojis”, other suggested content, or other iconography symbols) as selectable elements within search suggestion region 118B instead of predicted characters, words or phrases or other primarily linguistic information that keyboard module 122 derives from a language model, lexicon, or dictionary. In other words, rather than providing spelling or word suggestions from a dictionary within suggestion region 118B, computing device 110 may include, within suggestion region 118B, suggested search related content that computing device 110 determines may assist a user in providing input related to electronic communications.
In other examples, keyboard module 122 may configure the graphical keyboard 116B to include a query suggestion region 118D that presents suggested content (e.g., predicted or suggested search queries, predicted emoticons, emojis, other iconographic symbols, and other suggested content) as selectable elements above the graphical keys 118A. In other words, keyboard module 122 may in some examples configure the graphical keyboard 116B to include query suggestion region 118D positioned between edit region 116C and graphical keys 118A. In the example shown in FIG. 1, the query suggestion region 118D is also positioned near (e.g., adjacent to, above, below, on either side of, etc.) the suggestion region 118B, which may have input focus. In some examples where the user interface 114 may be configured differently, the query suggestion region 118D may be positioned between the graphical keys 118A and the edit region 116C, including in an example where the query suggestion region is positioned below the graphical keys and above the edit region. In still further examples, the query suggestion region may be positioned between the suggestion region 118B and the edit region 116C, including an example where the query suggestion region is positioned below the suggestion region and above the edit region. Other alternative arrangements of the graphical user interface are possible, including arrangements where the query suggestion region is not positioned between the graphical keys and the edit region, or between the suggestion region and the edit region.
In some examples, input focus means that input detected by the computing device 110 may be output to one or more specific user interface component(s) that are designated as having input focus. For instance, as PSD 112 detects user input at or near a location of PSD 112 where PSD 112 presents graphical keyboard 116B, UI module 120 may receive an indication of input and may output to the keyboard module information about the input. In some examples where the keyboard module 122 determines that the user is providing input at graphical keys 118A that correspond to a string of characters, the keyboard module 122 may cause UI module 120 to update the user interface 114 to include that string of characters within the user interface 114. Where that string of characters may be positioned or displayed within the user interface may be determined by which user interface component(s) have input focus. User interface 114, for example, includes several user interface components, including edit region 116C, suggestion region 118B, search element 118C, and output region 116A. In an example where suggestion region 118B has input focus, the computing device 110 may format and display within suggestion region 118B the string of characters corresponding to the input.
At times, in some examples, no user interface component will have input focus. In other examples, only one interface component will have input focus at a given time, and only certain types of user interface components might be allowed or be qualified to have input focus, such as text boxes or edit boxes. The computing device 110 may determine, in some examples, which user interface component has input focus by detecting which qualifying user interface component last received touch input by the user. In other words, when PSD 112 detects user input at a location of PSD 112 corresponding to a particular user interface component, computing device 110 may give that user interface component input focus (if it is allowed or qualified to have input focus). In other examples, the computing device 110 determines which user interface component has input focus by detecting which user interface component last received input from a cursor. In still other examples, computing device 110 may determine which user interface component has input focus based on a setting configured by an application executing on the computing device 110. For some user interface components, such as edit region 116C or suggestion region 118B, a flashing cursor 119 may in some examples be displayed within the interface component to provide a visual cue indicating that it has input focus.
In FIG. 1, PSD 112 may detect user inputs at or near a location of PSD 112 where PSD 112 presents graphical keys 118A, and output to the UI module 220 an indication of the user input detected by PSD 112. The UI module 120 may output to the keyboard module 122 information about the input. Responsive to the information about the input, keyboard module 122 determines a string of characters that corresponds to the input, and then causes the UI module 120 to update the user interface 114 to include that string of characters in the suggestion region 118B, which has input focus in FIG. 1. As shown in the example of FIG. 1, the keyboard module 122 determines the string of characters corresponding to the input in this case to be the “meatball sho” string of characters displayed in the suggestion region 118B.
Keyboard module 122 may also determine suggested queries based on the information about the input received from UI module 120. Responsive to the information about the input received from UI module 120, the keyboard module 122 may determine suggested queries and cause the UI module 120 to update the user interface 114 to include the suggested queries in the query suggestion region 118D. As shown in the example of FIG. 1, the suggested queries generated by the keyboard module 122 may include “Meatball Shoppe,” “Meatball,” and “Meatball recipe easy.” As the PSD 112 continues to detect user inputs corresponding to locations of PSD 112 that present graphical keys 118A, information about additional input is received by the keyboard module 122, which in response to the additional input, may determine updated suggested queries. The keyboard module 122 may cause UI module 120 to update the user interface 114 as updated suggested queries are determined.
The keyboard module 122 may determine suggested queries based on the information about user input received from the UI module 120, but the suggested queries may also be determined based on an application the user is interacting with, the context of the user's activity, or some combination thereof. The keyboard module 122 may also determine the suggested queries based on historical information about the user. For example, in some examples, search query suggestions may come from chat conversations on computing device 110. The text message “do you want to get dinner?” may suggest a “Restaurants near me” query; the text message “did you watch the basketball game?” may suggest a “Team Name” query of a professional basketball team. In accordance with one or more aspects of the present disclosure, when computing device 110 may be displaying suggested search queries, the graphical keyboard 116B in FIG. 1 may be considered to be in search mode or suggest mode.
In the example of FIG. 1, the user may optionally provide input at a location of PSD 112 at which a suggested query is displayed. In some examples, in response to receiving an indication of input selecting one of the suggested queries, keyboard module 122 may obtain information associated with the suggested query, which may involve invoking a search operation using the content of the link as a search query, or otherwise conducting a search for information based on the suggested query. In other examples, in response to receiving an indication of input selecting one of the suggested queries, keyboard module 122 may obtain information associated with the search query without invoking a search operation or performing a search, such as when the suggested query points directly an item of information, or points to an item of information (e.g., a document or contact) on the device 110.
In some examples, while a search operation is being conducted, the keyboard module 122 may cause the UI module 120 to update the user interface 114, and may cause the UI module 120 to further update the user interface 114 when search results are available. In some examples, the keyboard module 122 may cause UI module 120 to update the user interface 114 so that the search results are displayed within the graphical keyboard 116B. In accordance with one or more aspects of the present disclosure, when search results are being displayed, the graphical keyboard 116B may be considered to be in results mode.
By presenting a graphical keyboard with integrated search features, such as the graphical keyboard 116B of FIG. 1, an example computing device may perform search operations without requiring the user to leave the current application that the user is interacting with, and without the user losing focus. In some examples, the example computing device may present search results obtained from performing a search of a suggested query and the user may then review or interact with the search results as the example device presents them on a screen. For example, the user may provide additional input causing the computing device to incorporate the search results or information obtained from the search results into a current activity or task, or share the search results or information obtained from the search results with other devices. Techniques in accordance with the present disclosure may also enable a computing device to output or present to the user suggested search queries, potentially facilitating and streamlining the process for performing a search. Automatically displaying suggested search queries may result in the computing device receiving fewer inputs related to search, because fewer inputs may be required to cause a device to perform a search based on a suggested query that is automatically presented within a graphical keyboard. By receiving fewer inputs, the computing device may process fewer user inputs, execute fewer operations, and as a result, consume less electrical power.
FIG. 2 is a block diagram illustrating computing device 210 as an example computing device that is configured to present a graphical keyboard with integrated search features, in accordance with one or more aspects of the present disclosure. Computing device 210 of FIG. 2 is described below as an example of computing device 110 of FIG. 1. FIG. 2 illustrates only one particular example of computing device 210, and many other examples of computing device 210 may be used in other instances and may include a subset of the components included in example computing device 210 or may include additional components not shown in FIG. 2.
As shown in the example of FIG. 2, computing device 210 includes PSD 212, one or more processors 240, one or more communication units 242, one or more input components 244, one or more output components 246, and one or more storage components 248. Presence-sensitive display 212 includes display component 202 and presence-sensitive input component 204. Storage components 248 of computing device 210 include UI module 220, keyboard module 222, and one or more application modules 224. Keyboard module 122 may include spatial model (“SM”) module 226, language model (“LM”) module 228, and search module 230. Communication channels 250 may interconnect each of the components 212, 240, 242, 244, 246, and 248 for inter-component communications (physically, communicatively, and/or operatively). In some examples, communication channels 250 may include a system bus, a network connection, an inter-process communication data structure, or any other method for communicating data.
One or more communication units 242 of computing device 210 may communicate with external devices via one or more wired and/or wireless networks by transmitting and/or receiving network signals on the one or more networks. Examples of communication units 242 include a network interface card (e.g. such as an Ethernet card), an optical transceiver, a radio frequency transceiver, a GPS receiver, or any other type of device that can send and/or receive information. Other examples of communication units 242 may include short wave radios, cellular data radios, wireless network radios, as well as universal serial bus (USB) controllers.
One or more input components 244 of computing device 210 may receive input. Examples of input are tactile, audio, and video input. Input components 242 of computing device 210, in one example, includes a presence-sensitive input device (e.g., a touch sensitive screen, a PSD), mouse, keyboard, voice responsive system, video camera, microphone or any other type of device for detecting input from a human or machine. In some examples, input components 242 may include one or more sensor components one or more location sensors (GPS components, Wi-Fi components, cellular components), one or more temperature sensors, one or more movement sensors (e.g., accelerometers, gyros), one or more pressure sensors (e.g., barometer), one or more ambient light sensors, and one or more other sensors (e.g., microphone, camera, infrared proximity sensor, hygrometer, and the like). Other sensors may include a heart rate sensor, magnetometer, glucose sensor, hygrometer sensor, olfactory sensor, compass sensor, step counter sensor, to name a few other non-limiting examples.
One or more output components 246 of computing device 110 may generate output. Examples of output are tactile, audio, and video output. Output components 246 of computing device 210, in one example, includes a PSD, sound card, video graphics adapter card, speaker, cathode ray tube (CRT) monitor, liquid crystal display (LCD), or any other type of device for generating output to a human or machine.
PSD 212 of computing device 210 may be similar to PSD 112 of computing device 110 and includes display component 202 and presence-sensitive input component 204. Display component 202 may be a screen at which information is displayed by PSD 212 and presence-sensitive input component 204 may detect an object at and/or near display component 202. As one example range, presence-sensitive input component 204 may detect an object, such as a finger or stylus that is within two inches or less of display component 202. Presence-sensitive input component 204 may determine a location (e.g., an [x, y] coordinate) of display component 202 at which the object was detected. In another example range, presence-sensitive input component 204 may detect an object six inches or less from display component 202 and other ranges are also possible. Presence-sensitive input component 204 may determine the location of display component 202 selected by a user's finger using capacitive, inductive, and/or optical recognition techniques. In some examples, presence-sensitive input component 204 also provides output to a user using tactile, audio, or video stimuli as described with respect to display component 202. In the example of FIG. 2, PSD 212 may present a user interface (such as user interface 114 of FIG. 1).
While illustrated as an internal component of computing device 210, PSD 212 may also represent an external component that shares a data path with computing device 210 for transmitting and/or receiving input and output. For instance, in one example, PSD 212 represents a built-in component of computing device 210 located within and physically connected to the external packaging of computing device 210 (e.g., a screen on a mobile phone). In another example, PSD 212 represents an external component of computing device 210 located outside and physically separated from the packaging or housing of computing device 210 (e.g., a monitor, a projector, etc. that shares a wired and/or wireless data path with computing device 210).
PSD 212 of computing device 210 may detect two-dimensional and/or three-dimensional gestures as input from a user of computing device 210. For instance, a sensor of PSD 212 may detect a user's movement (e.g., moving a hand, an arm, a pen, a stylus, etc.) within a threshold distance of the sensor of PSD 212. PSD 212 may determine a two or three dimensional vector representation of the movement and correlate the vector representation to a gesture input (e.g., a hand-wave, a pinch, a clap, a pen stroke, etc.) that has multiple dimensions. In other words, PSD 212 can detect a multi-dimension gesture without requiring the user to gesture at or near a screen or surface at which PSD 212 outputs information for display. Instead, PSD 212 can detect a multi-dimensional gesture performed at or near a sensor which may or may not be located near the screen or surface at which PSD 212 outputs information for display.
One or more processors 240 may implement functionality and/or execute instructions associated with computing device 210. Examples of processors 240 include application processors, display controllers, auxiliary processors, one or more sensor hubs, and any other hardware configured to function as a processor, a processing unit, or a processing device. Modules 220, 222, 224, 226, 228, and 230 may be operable by processors 240 to perform various actions, operations, or functions of computing device 210. For example, processors 240 of computing device 210 may retrieve and execute instructions stored by storage components 248 that cause processors 240 to perform the operations modules 220, 222, 224, 226, 228, and 230. The instructions, when executed by processors 240, may cause computing device 210 to store information within storage components 248.
One or more storage components 248 within computing device 210 may store information for processing during operation of computing device 210 (e.g., computing device 210 may store data accessed by modules 220, 222, 224, 226, 228, and 230 during execution at computing device 210). In some examples, storage component 248 is a temporary memory, meaning that a primary purpose of storage component 248 is not long-term storage. Storage components 248 on computing device 210 may be configured for short-term storage of information as volatile memory and therefore not retain stored contents if powered off. Examples of volatile memories include random access memories (RAM), dynamic random access memories (DRAM), static random access memories (SRAM), and other forms of volatile memories known in the art.
Storage components 248, in some examples, also include one or more computer-readable storage media. Storage components 248 in some examples include one or more non-transitory computer-readable storage mediums. Storage components 248 may be configured to store larger amounts of information than typically stored by volatile memory. Storage components 248 may further be configured for long-term storage of information as non-volatile memory space and retain information after power on/off cycles. Examples of non-volatile memories include disc storage, flash memories, or forms of electrically programmable memories (EPROM) or electrically erasable and programmable (EEPROM) memories. Storage components 248 may store program instructions and/or information (e.g., data) associated with modules 220, 222, 224, 226, 228, and 230. Storage components 248 may include a memory configured to store data or other information associated with modules 220, 222, 224, 226, 228, and 230.
UI module 220 may include all functionality of UI module 120 of computing device 110 of FIG. 1 and may perform similar operations as UI module 120 for managing a user interface (e.g., user interface 114) that computing device 210 provides at presence-sensitive display 212 for handling input from a user. For example, UI module 220 of computing device 210 may query keyboard module 222 for a keyboard layout (e.g., an English language QWERTY keyboard, etc.). UI module 220 may transmit a request for a keyboard layout over communication channels 250 to keyboard module 222. Keyboard module 222 may receive the request and reply to UI module 220 with data associated with the keyboard layout. UI module 220 may receive the keyboard layout data over communication channels 250 and use the data to generate a user interface. UI module 220 may transmit a display command and data over communication channels 250 to cause PSD 212 to present the user interface at PSD 212.
In some examples, UI module 220 may receive an indication of one or more user inputs detected at PSD 212 and may output information about the user inputs to keyboard module 222. For example, PSD 212 may detect a user input and send data about the user input to UI module 220. UI module 220 may generate one or more touch events based on the detected input. A touch event may include information that characterizes user input, such as a location component (e.g., [x,y] coordinates) of the user input, a time component (e.g., when the user input was received), a force component (e.g., an amount of pressure applied by the user input), or other data (e.g., speed, acceleration, direction, density, etc.) about the user input.
Based on location information of the touch events generated from the user input, UI module 220 may determine that the detected user input is associated the graphical keyboard. UI module 220 may send an indication of the one or more touch events to keyboard module 222 for further interpretation. Keyboard module 22 may determine, based on the touch events received from UI module 220, that the detected user input represents an initial selection of one or more keys of the graphical keyboard.
Application modules 224 represent all the various individual applications and services executing at and accessible from computing device 210 that may rely on a graphical keyboard having integrated search features. A user of computing device 210 may interact with a graphical user interface associated with one or more application modules 224 to cause computing device 210 to perform a function. Numerous examples of application modules 224 may exist and include, a fitness application, a calendar application, a personal assistant or prediction engine, a search application, a map or navigation application, a transportation service application (e.g., a bus or train tracking application), a social media application, a game application, an e-mail application, a chat or messaging application, an Internet browser application, or any and all other applications that may execute at computing device 210.
Keyboard module 222 may include all functionality of keyboard module 122 of computing device 110 of FIG. 1 and may perform similar operations as keyboard module 122 for providing a graphical keyboard having integrated search features. Keyboard module 222 may include various submodules, such as SM module 226, LM module 228, and search module 230, which may perform the functionality of keyboard module 222.
SM module 226 may receive one or more touch events as input, and output a character or sequence of characters that likely represents the one or more touch events, along with a degree of certainty or spatial model score indicative of how likely or with what accuracy the one or more characters define the touch events. In other words, SM module 226 may infer touch events as a selection of one or more keys of a keyboard and may output, based on the selection of the one or more keys, a character or sequence of characters.
When keyboard module 222 operates in text-entry mode LM module 228 may receive a character or sequence of characters as input, and output one or more candidate characters, words, or phrases that LM module 228 identifies from a lexicon as being potential replacements for a sequence of characters that LM module 228 receives as input for a given language context (e.g., a sentence in a written language). Keyboard module 222 may cause UI module 220 to present one or more of the candidate words at suggestion regions 118B and 118D of user interface 114.
The lexicon of computing device 210 may include a list of words within a written language vocabulary (e.g., a dictionary). For instance, the lexicon may include a database of words (e.g., words in a standard dictionary and/or words added to a dictionary by a user or computing device 210. LM module 228 may perform a lookup in the lexicon, of a character string, to identify one or more letters, words, and/or phrases that include parts or all of the characters of the character string. For example, LM module 228 may assign a language model probability or a similarity coefficient (e.g., a Jaccard similarity coefficient) to one or more candidate words located at a lexicon of computing device 210 that include at least some of the same characters as the inputted character or sequence of characters. The language model probability assigned to each of the one or more candidate words indicates a degree of certainty or a degree of likelihood that the candidate word is typically found positioned subsequent to, prior to, and/or within, a sequence of words (e.g., a sentence) generated from text input detected by presence-sensitive input component 204 prior to and/or subsequent to receiving the current sequence of characters being analyzed by LM module 228. In response to determining the one or more candidate words, LM module 228 may output the one or more candidate words from lexicon data stores 260A that have the highest similarity coefficients.
Search module 230 of keyboard module 222 may perform integrated search functions on behalf of keyboard module 222. That is, when invoked (e.g., automatically based on text input analysis or manually in response to a user of computing device 210 selecting selectable search element 118C of user interface 114), keyboard module 222 may operate in search mode where keyboard module 222 enables computing device 210 to perform search functions from within graphical keyboard 116B.
When keyboard module 222 operates in search mode, search module 230 may receive a character or sequence of characters or other information as input, and output one or more query suggestions that the search module 230 identifies based on the input. The information that the search module 230 receives as input may include characters or a sequence of characters, and in some examples, may also include information about the current state of the application that the user is using, as well as other information, such as information from the application that is relying on keyboard module 122 (e.g., recent texts, pictures, geolocation data and the like) and contextual information associated with the user and computing device 210. Search module 230 may use this information to generate one or more suggested queries. The keyboard module 222 may cause UI module 220 to present one or more of the suggested queries within the query suggestion region 118D of user interface 114.
In some examples, search model 230 may further rely on a current context of computing device 210 to determine a suggested query. As used herein, a current context specifies the characteristics of the physical and/or virtual environment of a computing device, such as computing device 210, and a user of the computing device, at a particular time. In addition, the term “contextual information” is used to describe any information that can be used by a computing device to define the virtual and/or physical environmental characteristics that the computing device, and the user of the computing device, may experience at a particular time.
Examples of contextual information are numerous and may include: sensor information obtained by sensors (e.g., position sensors, accelerometers, gyros, barometers, ambient light sensors, proximity sensors, microphones, and any other sensor) of computing device 210, communication information (e.g., text based communications, audible communications, video communications, etc.) sent and received by communication modules of computing device 210, and application usage information associated with applications executing at computing device 210 (e.g., application data associated with applications, Internet search histories, text communications, voice and video communications, calendar information, social media posts and related information, etc.). Further examples of contextual information include signals and information obtained from transmitting devices that are external to computing device 210.
In some examples, computing device 210 may initiate a search in response to different types of input. For example, keyboard module 222 may initiate a search in response to detecting user input at a location at which PSD 212 presents a suggested query within the query suggestion region 118D or elsewhere. This input may generally correspond to the user selecting one of the displayed suggested queries. In another example, keyboard module 222, while operating in search mode, may initiate a search in response to detecting user input at locations at which PSD 212 presents graphical keys 118A followed by input at a location at which PSD 212 presents a “return” or “search” or “send” key. Such input may generally correspond to the user typing a search query using the graphical keyboard 116B and hitting “return” without selecting a suggested search query.
Responsive to detecting user input that causes computing device 210 to initiate a search, the search module 230 within the keyboard module 222 may execute a search, which may cause the computing device 210 to send and receive information over a public or private network through communication unit(s) 242. If search results become available, the search module 230 or keyboard module 222 may then cause the UI module 220 to present search result information to the user by, in some examples, displaying search result information in user interface 114.
FIG. 3 is a block diagram illustrating an example computing device that outputs graphical content for display at a remote device, in accordance with one or more techniques of the present disclosure. Graphical content, generally, may include any visual information that may be output for display, such as text, images, a group of moving images, to name only a few examples. The example shown in FIG. 3 includes a computing device 310, a PSD 312, communication unit 342, projector 380, projector screen 382, mobile device 386, and visual display component 390. In some examples, PSD 312 may be a presence-sensitive display as described in FIGS. 1-2. Although shown for purposes of example in FIGS. 1 and 2 as a stand-alone computing device 110, a computing device such as computing device 310 may, generally, be any component or system that includes a processor or other suitable computing environment for executing software instructions and, for example, need not include a presence-sensitive display.
As shown in the example of FIG. 3, computing device 310 may be a processor that includes functionality as described with respect to processors 240 in FIG. 2. In such examples, computing device 310 may be operatively coupled to PSD 312 by a communication channel 362A, which may be a system bus or other suitable connection. Computing device 310 may also be operatively coupled to communication unit 342, further described below, by a communication channel 362B, which may also be a system bus or other suitable connection. Although shown separately as an example in FIG. 3, computing device 310 may be operatively coupled to PSD 312 and communication unit 342 by any number of one or more communication channels.
In other examples, such as illustrated previously by computing device 110 in FIGS. 1-2, a computing device may refer to a portable or mobile device such as mobile phones (including smart phones), laptop computers, etc. In some examples, a computing device may be a desktop computer, tablet computer, smart television platform, camera, personal digital assistant (PDA), server, or mainframes.
PSD 312 may include display component 302 and presence-sensitive input component 304. Display component 302 may, for example, receive data from computing device 310 and display the graphical content. In some examples, presence-sensitive input component 304 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at PSD 312 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input to computing device 310 using communication channel 362A. In some examples, presence-sensitive input component 304 may be physically positioned on top of display component 302 such that, when a user positions an input unit over a graphical element displayed by display component 302, the location at which presence-sensitive input component 304 corresponds to the location of display component 302 at which the graphical element is displayed.
As shown in FIG. 3, computing device 310 may also include and/or be operatively coupled with communication unit 342. Communication unit 342 may include functionality of communication unit 242 as described in FIG. 2. Examples of communication unit 342 may include a network interface card, an Ethernet card, an optical transceiver, a radio frequency transceiver, or any other type of device that can send and receive information. Other examples of such communication units may include Bluetooth, 3G, and WiFi radios, Universal Serial Bus (USB) interfaces, etc. Computing device 310 may also include and/or be operatively coupled with one or more other devices (e.g., input devices, output components, memory, storage devices) that are not shown in FIG. 3 for purposes of brevity and illustration.
FIG. 3 also illustrates a projector 380 and projector screen 382. Other such examples of projection devices may include electronic whiteboards, holographic display components, and any other suitable devices for displaying graphical content. Projector 380 and projector screen 382 may include one or more communication units that enable the respective devices to communicate with computing device 310. In some examples, the one or more communication units may enable communication between projector 380 and projector screen 382. Projector 380 may receive data from computing device 310 that includes graphical content. Projector 380, in response to receiving the data, may project the graphical content onto projector screen 382. In some examples, projector 380 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen using optical recognition or other suitable techniques and send indications of such user input using one or more communication units to computing device 310. In such examples, projector screen 382 may be unnecessary, and projector 380 may project graphical content on any suitable medium and detect one or more user inputs using optical recognition or other such suitable techniques.
Projector screen 382, in some examples, may include a presence-sensitive display 384. Presence-sensitive display 384 may include a subset of functionality or all of the functionality of presence-sensitive display 112 and/or 312 as described in this disclosure. In some examples, presence-sensitive display 384 may include additional functionality. Projector screen 382 (e.g., an electronic whiteboard), may receive data from computing device 310 and display the graphical content. In some examples, presence-sensitive display 384 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen 382 using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units to computing device 310.
FIG. 3 also illustrates mobile device 386 and visual display component 390. Mobile device 386 and visual display component 390 may each include computing and connectivity capabilities. Examples of mobile device 386 may include e-reader devices, convertible notebook devices, hybrid slate devices, etc. Examples of visual display component 390 may include other devices such as televisions, computer monitors, etc. In some examples, visual display component 390 may be a vehicle cockpit display or navigation display (e.g., in an automobile, aircraft, or some other vehicle). In some examples, visual display component 390 may be a home automation display or some other type of display that is separate from computing device 310.
As shown in FIG. 3, mobile device 386 may include a presence-sensitive display 388. Visual display component 390 may include a presence-sensitive display 392. Presence- sensitive displays 388, 392 may include a subset of functionality or all of the functionality of presence- sensitive display 112, 212, and/or 312 as described in this disclosure. In some examples, presence- sensitive displays 388, 392 may include additional functionality. In any case, presence-sensitive display 392, for example, may receive data from computing device 310 and display the graphical content. In some examples, presence-sensitive display 392 may determine one or more user inputs (e.g., continuous gestures, multi-touch gestures, single-touch gestures) at projector screen using capacitive, inductive, and/or optical recognition techniques and send indications of such user input using one or more communication units to computing device 310.
As described above, in some examples, computing device 310 may output graphical content for display at PSD 312 that is coupled to computing device 310 by a system bus or other suitable communication channel. Computing device 310 may also output graphical content for display at one or more remote devices, such as projector 380, projector screen 382, mobile device 386, and visual display component 390. For instance, computing device 310 may execute one or more instructions to generate and/or modify graphical content in accordance with techniques of the present disclosure. Computing device 310 may output the data that includes the graphical content to a communication unit of computing device 310, such as communication unit 342. Communication unit 342 may send the data to one or more of the remote devices, such as projector 380, projector screen 382, mobile device 386, and/or visual display component 390. In this way, computing device 310 may output the graphical content for display at one or more of the remote devices. In some examples, one or more of the remote devices may output the graphical content at a presence-sensitive display that is included in and/or operatively coupled to the respective remote devices.
In some examples, computing device 310 may not output graphical content at PSD 312 that is operatively coupled to computing device 310. In other examples, computing device 310 may output graphical content for display at both a PSD 312 that is coupled to computing device 310 by communication channel 362A, and at one or more remote devices. In such examples, the graphical content may be displayed substantially contemporaneously at each respective device. For instance, some delay may be introduced by the communication latency to send the data that includes the graphical content to the remote device. In some examples, graphical content generated by computing device 310 and output for display at PSD 312 may be different than graphical content display output for display at one or more remote devices.
Computing device 310 may send and receive data using any suitable communication techniques. For example, computing device 310 may be operatively coupled to external network 374 using network link 373A. Each of the remote devices illustrated in FIG. 3 may be operatively coupled to network external network 374 by one of respective network links 373B, 373C, or 373D. External network 374 may include network hubs, network switches, network routers, etc., that are operatively inter-coupled thereby providing for the exchange of information between computing device 310 and the remote devices illustrated in FIG. 3. In some examples, network links 373A-373D may be Ethernet, ATM or other network connections. Such connections may be wireless and/or wired connections.
In some examples, computing device 310 may be operatively coupled to one or more of the remote devices included in FIG. 3 using direct device communication 378. Direct device communication 378 may include communications through which computing device 310 sends and receives data directly with a remote device, using wired or wireless communication. That is, in some examples of direct device communication 378, data sent by computing device 310 may not be forwarded by one or more additional devices before being received at the remote device, and vice-versa. Examples of direct device communication 378 may include Bluetooth, Near-Field Communication, Universal Serial Bus, WiFi, infrared, etc. One or more of the remote devices illustrated in FIG. 3 may be operatively coupled with computing device 310 by communication links 376A-376D. In some examples, communication links 376A-376D may be connections using Bluetooth, Near-Field Communication, Universal Serial Bus, infrared, etc. Such connections may be wireless and/or wired connections.
In accordance with techniques of the disclosure, computing device 310 may in some examples be operatively coupled to visual display component 390 using external network 374. Computing device 310 may output a graphical keyboard for display at PSD 392. For instance, computing device 310 may send data that includes a representation of the graphical keyboard to communication unit 342. Communication unit 342 may send the data that includes the representation of the graphical keyboard to visual display component 390 using external network 374. Visual display component 390, in response to receiving the data using external network 374, may cause PSD 392 to output the graphical keyboard. In response to receiving user input at PSD 392 to select one or more keys of the keyboard, visual display component 390 may send an indication of the user input to computing device 310 using external network 374. Communication unit 342 of may receive the indication of the user input, and send the indication to computing device 310.
Computing device 310 may determine one or more candidate words or suggested search queries, and may generate an updated graphical user interface including suggested search queries, which in some examples, may be included within a query suggestion region. Communication unit 342 may receive a representation of the updated graphical user interface and may send the send the representation to visual display component 390, such that visual display component 390 may cause PSD 392 to output the updated graphical keyboard, including any suggested search queries and/or query suggestion region.
FIGS. 4A-4D are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure. FIGS. 4A-4D illustrate example user interfaces 414; many other examples of graphical user interfaces may be used or presented in other instances or examples. Each of the user interfaces 414 may correspond to a graphical user interface output by computing devices 110, 210, or 310 from FIGS. 1-3. For purposes of illustration, FIGS. 4A-4D are described below within the context of computing device 210 of FIG. 2.
As illustrated in the examples of FIGS. 4A through 4D, each example user interface 414 forms part of a chat application user interface of a chat application from application module 224. User interface 414 includes output region 416A, edit region 416C, and graphical keyboard 416B. Output region 416A and edit region 416C may both be associated with the user interface of an application or service executing on or accessible from computing device 210. Included within the graphical keyboard 416B are graphical elements displayed as keys 418A. Also included within graphical keyboard 416B is a suggestion region 418B, which may include suggested words or other information representing selectable spelling corrections and/or selectable word suggestions to replace character strings that are included within edit region 416C. Additionally, suggestion region 418B may include suggested queries that may be of interest to the user based on information that may include the user's current activity, context, location, or situation.
Also shown included within graphical keyboard 416B in the example of FIG. 4A is a search element 418C. In some examples, computing device 210 may receive an indication of user input selecting the search element 418C, and in response, computing device 210 may cause graphical keyboard 416B transition to search mode in accordance with one or more aspects of the present disclosure. In the example shown, the search element 418C is displayed above the graphical keys 418A, but it may be located elsewhere within the graphical keyboard 416B. In other examples, the search element 418C may be located elsewhere within the user interface 414, including partially or entirely outside of the graphical keyboard 416B. In still other examples, the search element 418C or a component used in a similar way could be implemented as an external input or hardware input, such as an external button, switch, or other hardware input component configured to receive user input for computing device 210.
Included within graphical keys 418A is a return/search key 402. In some examples, computing device 210 may respond to an indication of input selecting key 402 by transmitting to another device the string of characters in edit region 416C as a text message. Computing device 210 may also cause the string of characters in the edit region 416C to be formatted and displayed within the output region 416A, which in some examples may signify that the string of characters was transmitted as a text message. For instance, in response to receiving user input at locations of PSD 112 corresponding to the key 402, computing device 210 may update the user interface 414 so that presented in output region 416A is the text that was displayed in edit region 416C when the computing device received the indication of input selecting key 402. In other situations, such as when the graphical keyboard 416B may be in search mode as described in accordance with one or more aspects of the present disclosure, the computing device 210 may respond to an indication of input selecting the return/search key 402 by performing a search using, for example, the information or text displayed within the suggestion region 418B when the computing device 210 received the indication of input selecting the key 402.
Although the return/search key 402 is shown as a single graphical key in the example of FIG. 4A, in other examples the key 402 could be presented or implemented as more than one key. In one example, the return/search key 402 could alternatively be implemented as a discrete graphical return key and a discrete graphical search key. In other examples, the graphical keyboard 416B could include a discrete graphical return key and a discrete graphical search key, but may not always include a graphical representation of both keys within the graphical keyboard 416B. In still further examples, the computing device 210 may display a discrete graphical search key only when the graphical keyboard 416B is in search mode in accordance with one or more aspects of the present disclosure.
With reference to FIG. 4A, computing device 210 may receive an electronic communication or text message from a friend's device. Computing device 210 may output the content of the friend's text message (“What's up?”) within the user interface 414 in output region 416A. Keyboard module 222 of computing device 210 may then receive an indication of input selecting graphical keys 418A and determine a character string (“Din”) based on the indication of input, which the computing device 210 formats and displays within edit region 416C as shown in FIG. 4A. The keyboard module 222 may also determine, based on the indication of input, candidate words (“Dinner”, “Dine”, and “Don”). The computing device 210 may format and displays these candidate words within suggestion region 418B.
Referring now to FIG. 4B, the computing device 210 has updated the user interface 414 in FIG. 4A in response (at least partially) to further user input. In FIG. 4B, keyboard module 222 of computing device 210 may receive an indication of input selecting additional graphical keys 418A, and determine, based on the input, additional text that the computing device 210 formats and displays within edit region 416C. Keyboard module 222 may also predict one or more updated candidate words based on the further user input and display one or more updated candidate words within suggestion region 418B. In FIG. 4B, the text “Dinner to” is displayed in edit region 416C in FIG. 4B, and candidate word “tonight” is displayed in edit region 416B.
Keyboard module 222 may also determine one or more suggested queries, and as shown in FIG. 4B, computing device 210 may format and display a suggested query 406 (“Restaurants near me”) within suggestion region 418B. In other examples, computing device 210 may cause multiple suggested queries to be displayed in suggestion region 418B or in other areas of user interface 414. Keyboard module 222 may generate suggested search queries based on many different sources of information, including user input, information in the edit region 416C, information in the output region 416A, current time, the user's context, application state, historical searches, and/or other information.
In some examples, computing device 210 suggests search queries after the user has typed a sentence-ending punctuation or has sent a message. Computing device 210 presents suggested search queries in the manner shown in FIG. 4B, or in other ways in accordance with one or more aspects of the present disclosure. In some examples, the computing device 210 may alert the user to availability of one or more suggested queries, such as by providing a visual, audio, tactile, or physical indication or cue. For example, the computing device 210 may alert the user to the availability of a suggested query through the search element 418C, which may change color, change shape, animate, pulse, or otherwise change to suggest that suggested queries may be available. In other examples, the computing device 210 may use other aspects of the user interface 414 to perform a similar function, or in other examples computing device 210 may vibrate or provide some other physical indication.
Computing device 210 may determine the suggested words and suggested search queries shown in suggestion region 418B based entirely on locally-stored information without access to information on a network or without sending information about the user's activity (including the user's current or past actions or communications) over a public or private network. In other examples, computing device 210 may determine suggested words and/or suggested search queries using information obtained through a public or private network, and possibly by sending information about the user (which could include information about the user's actions or communications) over a public or private network.
In some examples, on-device annotators may be used to generate suggested search queries, and such annotators may use both global and local models to detect searchable text or entities. Global models may include entities that may be useful to search regardless of the user's geolocation, such as a well-known person (“Barack Obama”). Local models may include entities such as restaurants near the user, such as a particular restaurant near the location of the computing device 210. Suggested queries could also be based on exact language triggers, such as the phrase “let's get dinner” triggering a “Restaurants near me” suggested query.
Throughout the disclosure, examples are described where a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information. For example, before a computing device or computing system can collect or may make use of information associated with a user, the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user. In addition, certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed. For example, a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined. Thus, the user may have control over how information is collected about the user and used by the computing device and computing system.
Shown within suggestion region 418B in the example of FIG. 4B, next to the suggested query 406, is a graphical symbol of a category 404 of search, which may identify the category of a search with which the suggested query 406 may be associated. In this example, computing device 210 may cause the graphical symbol of a category 404 to be displayed to denote a category relating to a search for restaurants or food. Symbols representing other search categories may alternatively be displayed. In other examples, more than one search category element may be associated with a suggested search query, and in still further examples, there may be no search category element associated with one or more suggested search queries (or associated with any suggested search queries). In cases where a search query remains uncategorized, computing device 210 may display a generic search category element, such as a magnifying glass, to designate the search query as not being associated with any particular search category.
In FIG. 4C, the computing device 210 has updated the user interface 414 of FIG. 4B in response to an indication of input selecting graphical keys 118A (which the keyboard module 222 may interpret as the string of text “Dinner tonight?”), followed by an indication of input selecting the “return/search” key 402. In response, computing device 210 may compose and send a text message that includes the text “Dinner tonight?” to a friend's computing device. In FIG. 4C, computing device 210 may update the user interface 414 to include the “Dinner tonight?” text message within the output region 416A. The computing device 210 may update the user interface 414 in FIG. 4C to clear the text from the edit region 416C, and the computing device 210 may also remove the suggested word “tonight” from the suggestion region 418B. Computing device 210 may continue to display the suggested search query 406 within the suggestion region 418B in the example of FIG. 4C.
In FIG. 4D, the computing device 210 has updated the user interface 414 in response (at least partially) to detecting an indication of input selecting the suggested search query 406 (“Restaurants near me”) displayed in suggestion region 418B in FIG. 4C (note the outline of a user's hand in FIG. 4C, which may signify a user's touch input associated with suggested query 406). In response to this input, the computing device 210 may perform a search based on the suggested query 406. The suggested query 406 may be a link that, when invoked by user input, will cause computing device 210 to perform a search operation, using the content of the link as a search query. (In other examples, a suggested query link might not require computing device to perform a search operation, such as when the link points directly to an item of information or to the location of the item, or represents the URL of a document or other resource.) After performing the search, the computing device 210 may output for display a textual and/or graphical indication of one or more search results (or other information) obtained by the computing device. In the example shown in FIG. 4D, the computing device 210 causes the search results 418E to be displayed within the graphical keyboard 416B. In this instance, search results 418E are presented below the suggestion region 418B, substantially replacing the graphical keys 418A previously displayed in FIG. 4C. In accordance with some aspects of present disclosure, the graphical keyboard 416B in FIG. 4D may be considered to be in results mode.
In the example of FIG. 4D, the keyboard module 222 may obtain search results based on the selected query, using the suggested query, that are or may be formatted into individual search cards 418E that keyboard module 222 causes computing device 210 to display as overlays atop graphical keys 418A (or as a temporary replacement for graphical keys 418A). In some examples, the search results 418E may be presented in a carousel format so that a user may swipe left or right 410 when browsing the search results 418E. The computing device 210 could alternatively output the search results of FIG. 4D in other formats, or in other locations within the graphical keyboard 416B. In other examples, the computing device 210 may output the search results in other locations within the user interface 414, including partially or entirely outside the graphical keyboard 416B. In still further examples, the computing device 210 may output search results partially or exclusively as audio information, and/or may output some types of search result information in tactile form.
In the example of FIG. 4D, the computing device 210 continues to present both the search element 418C and the suggestion region 418B in the graphical keyboard 416B along with, and above, the search results 418E. In other examples, the computing device 210 may cause either or both to be displayed at other locations within the user interface 414, or removed from the display.
FIGS. 5A-5F are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure. FIGS. 5A-5F illustrate example user interfaces 514; many other examples of user interfaces may be used or presented in other instances or examples. Each of the user interfaces 514 may correspond to a graphical user interface output by computing devices 110, 210, or 310 from FIGS. 1-3. For purposes of illustration, FIGS. 5A-5F are described below within the context of computing device 210 of FIG. 2.
As illustrated in the examples of FIGS. 5A through 5F, each example user interface 514 forms part of a chat application user interface of a chat application from application module 224. User interface 514 includes output region 516A, edit region 516C, and graphical keyboard 516B. Output region 516A and edit region 516C may both be associated with the user interface of an application or service executing on or accessible from computing device 210, such as the chat application described in connection FIGS. 4A-4D. Included within the graphical keyboard 516B are graphical elements displayed as keys 518A. Also included within graphical keyboard 516B is a suggestion region 518B and search element 518C. In the example shown, the search element 518C is displayed above the graphical keys 518A, but may be located elsewhere within or outside the graphical keyboard 516B. In other examples, it may be located elsewhere within the user interface 514, or in still other examples, could be implemented as an external or hardware input to the computing device 210.
Referring to FIG. 5A, computing device 210 may be exchanging electronic messages with another device, and computing device 210 may output the content of recent exchanged messages in output region 516A. The computing device 210 may also output and format a suggested query in edit region 518B.
In some examples, computing device 210 may receive an indication of input selecting the search element 518C, and responsive to receiving the indication of input selecting the search element 518C, computing device 210 may output, for display, a separate region in which suggested search queries are displayed. For example, keyboard module 222 may receive an indication of user input detected at a location at which search element 518C is displayed within graphical keyboard 516B, as shown in FIG. 5A. Responsive to the indication of user input, the computing device 210 may update the user interface 514 to that shown in FIG. 5B so that user interface 514 includes query suggestion region 518D. In accordance with one or more aspects of the present disclosure, the graphical keyboard 516B shown in the user interface 514 of FIG. 5B may be considered to be in search mode. In this example, search mode may be characterized by the presence of a query suggestion region 518D.
In the example of FIG. 5B, computing device 210 may change the input focus to the suggestion region 518B. This change in input focus may be signified by the cursor shown within the suggestion region 518B, and the absence of a cursor shown within the edit region 516C. In the example shown in FIG. 5B, computing device 210 causes the query suggestion region 518D to be displayed above the suggestion region 518B, and below the edit region 516C. This arrangement creates some visual distance between the suggestion region 518B and edit region 516C. This visual distance may tend to reduce user confusion in situations where both the suggestion region 518B and the edit region 516C are each user interface components that could potentially have input focus at various times.
In other examples, computing device 210 could position the query suggestion region 518D elsewhere within the graphical keyboard 516B. In still further examples, the computing device 210 could position the query suggestion region 518D elsewhere within the user interface 514, partially or entirely outside the graphical keyboard 516B.
The query suggestion region 518D in some examples may include suggested search queries or other information, each of which may represent selectable search queries that could be initiated by computing device 210. In some examples, such as when the user has not entered any text in edit region 516C or suggestion region 518B, there may be no suggested queries to be displayed. In other examples when the user has not entered any text in edit region 516C or suggestion region 518B, computing device 210 may output for display within the query suggestion region 518D suggested queries that include historical queries (previous searches performed by the device 210). In still other examples when the user has not entered any text in edit region 516C or suggestion region 518B, computing device 210 may output for display within the query suggestion region 518D suggested queries that may be of interest to the user based on the user's current activity, application, context, location, situation, or other information. Query suggestions in region 518D may also be generated based, as least in part, on the contents of elements of region 516A, such as content from a conversation “Dinner tonight?” (or similar content and context from different user interfaces that may populate the region of 516A). As shown in FIG. 5B, computing device 210 has included suggested query “Restaurants near me” in the query suggestion region 518D. This suggested query may be based at least in part on the subject of the chat shown in output region 516A. The other suggested queries, “Fast food” and “Movie times” may also be based to some extent on the subject of the chat shown in output region 516A.
With reference to FIG. 5C, keyboard module 222 of computing device 210 may receive indications of taps or gestures at keys 518A and determine, based on the user input, text that computing device 210 formats and displays within suggestion region 518B. For example, as shown in FIG. 5C, the computing device 210 may cause the text “meatball” to be displayed within suggestion region 518B. Keyboard module 222 of computing device 210 may also, in response to the user input, generate suggested queries that computing device 210 formats and displays within query suggestion region 518D, as shown in FIG. 5C. The suggested queries shown in query suggestion region 518D in FIG. 5C may be at least partially based on, and may be appropriate suggested queries for, the “meatball” text shown in the suggestion region 518B.
With reference to FIG. 5D, keyboard module 222 of computing device 210 may continue to receive additional indications of taps or gestures at keys 518A and may continue to determine, based on the additional user input, additional text that computing device 210 formats and displays within suggestion region 518B. For example, as shown in FIG. 5D, the computing device 210 may cause the text “meatball sho” to be displayed within suggestion region 518B. Also in response to the additional user input, keyboard module 222 of computing device 210 may generate updated suggested queries that computing device 210 formats and displays within query suggestion region 518D, as shown in FIG. 5D. The suggested queries shown in query suggestion region 518D in FIG. 5D may be at least partially based on, and may be appropriate suggested queries for, the “meatball sho” text shown in the suggestion region 518B.
In some examples, a suggested query that is considered more relevant than other suggested queries may be displayed at the top of the query suggestion region 518D. For example, computing device 210 may execute a model that assigns a score or probability to each of the suggested queries based on text input and contextual information. Computing device 210 may rank the queries based on the assigned scores and probabilities. Computing device 210 may display the highest scoring or highest probability queries near the top of suggestion region 518D and the lower scoring or lower probability queries near the bottom. In other examples, the queries shown in query suggestion region 518D may be listed in descending order of perceived relevance. Other arrangements could be used. In some examples, the determination of relevance of each suggested query may be made, at least partially, by the computing device 210, but in other examples, this determination may be made elsewhere or by another device or other devices, such as by a search engine or other service.
Referring now to FIG. 5E, computing device 210 may receive an indication of input selecting a suggested search query, and responsive to receiving the indication of input, computing device 210 may execute, based on the suggested query, a search for information. For example, as shown in FIG. 5E, keyboard module 222 may receive an indication of user input detected at a location at which suggested search query 509 is displayed within query suggestion region 518D. Responsive to the indication of user input, the computing device 210 may perform a search corresponding to the “Meatball Shoppe” suggested search query 509, and may also update the user interface 514 to indicate that a search is being performed. When search results are available, the computing device 210 may update the user interface 514 to that shown in FIG. 5F.
In FIG. 5F, the search results 518E are shown within the graphical keyboard 516B, and in this example, search results 518E are shown below the suggestion region 518B, substantially replacing the graphical keys 518A previously shown in FIG. 5E. The search results 518E may be presented in card-based form. The query suggestion region 518D has been removed from the user interface 514 in the example of FIG. 5F. In other examples, the search results could be shown in other formats, and in other locations within the graphical keyboard 516B. In still other examples, search results may be shown partially or entirely outside the graphical keyboard 516B. In yet further examples, the search result information could also be provided partially or exclusively as audio information, which may enable the user to listen to at least some of the search result information.
In the example of FIG. 5F, the computing device 210 continues to display both the search element 518C and the suggestion region 518B in the user interface 514 along with, and above, the search results 518E. In other examples, computing device 210 could remove either or both of these elements from the user interface 514, or the computing device 210 could display these elements at other locations within the user interface 514. In the example of FIG. 5F, computing device 210 returns the input focus to the edit region 516C. However, individual cards from results 518E may themselves be links such that a user input or gesture indicating selection thereof may link to the website indicated by the card. Furthermore, a card from results 518E may be “dragged” (such as by a user touching the PSD at the location of the card) and, without lifting the touching finger, move the finger to the output region 516A. This may have the effect of copying the card into the context of the output region 516A, such as adding it as a sendable and actionable (i.e., including a selectable link) element of a communication in a chat application. It should also be noted that cards in results 518E may include controls, such that a user may by input or gesture invoke selected actions with respect to the cards or card content, such as copy, open, send, share, map, call, message, purchase, make reservation, and so forth. These controls may be specific to individual search results and cards (i.e., not all cards in results 518E need provide the same controls).
FIGS. 6A-6C are conceptual diagrams illustrating example graphical user interfaces of an example computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure. FIGS. 6A-6C illustrate example user interfaces 614; many other examples of graphical user interfaces may be used or presented in other instances or examples. Each of the user interfaces 614 may correspond to a graphical user interface output by computing devices 110, 210, or 310 from FIGS. 1-3. For purposes of illustration, FIGS. 6A-6C are described below within the context of computing device 210 of FIG. 2.
As illustrated in the examples of FIGS. 6A through 6C, each example user interface 614 includes output region 616A, edit region 616C, and graphical keyboard 616B. Output region 616A and edit region 616C may both be associated with the user interface of an application or service executing on or accessible from computing device 210, such as a chat application described in connection with FIGS. 4A-4D and FIGS. 5A-5F. Included within the graphical keyboard 616B are graphical keys 618A, a suggestion region 618B, and a search element 618C.
In FIG. 6A, keyboard module 222 of computing device 210 may receive indications of taps or gestures at keys 618A and determine, based on the user input, a text string that computing device 210 formats and displays within edit region 616C. For example, as shown in FIG. 6A, the computing device 210 may cause the text string “What should I have for dinner to” to be displayed within edit region 616C. The computing device 210 may, in some examples, format and display within suggestion region 618B one or more suggested words, such as the suggested word “tonight” shown in suggestion region 618B. Keyboard module 222 of computing device 210 may also, in response to the user input, generate suggested queries that computing device 210 may format and display within suggestion region 618B, such as the “Restaurants near me” suggested query shown in FIG. 6A. Any suggested queries shown in query suggestion region 618B in FIG. 6A may be based on, and may be appropriate suggested queries for, the text shown in the edit region 616C.
Referring now to FIG. 6B, keyboard module 222 of computing device 210 may continue to receive additional indications of taps or gestures at keys 618A and determine, based on the user input, text that computing device 210 formats and displays within edit region 616C. The text displayed in edit region 616C may be a continuation of the partially-composed text message previously displayed in FIG. 6A. In accordance with the way that some chat applications operate, in some examples the computing device 210 does not transmit this text message to the other person's computing device at the point shown in FIG. 6B, and does not do so until the computing device 210 detects an indication of input selecting the key 602 (or in other examples, a “send” key).
In the example of FIG. 6B, before the computing device 210 receives an indication of input selecting key 602, the computing device 210 receives an indication of input selecting search element 618C. In this situation, the user may select the search element 618C before selecting the key 602 for a number of possible reasons. For example, the user may have intended the text shown in edit region 616C to be a search query, rather than a text message, but may have mistakenly entered the text in the edit region 616C, rather than first selecting the search element 618C and entering the text in suggestion region 618B. Alternatively, the user may have originally intended the text in edit region 616C to be a text message, but may have changed his or her mind and decided to perform a search instead of (or in addition to) submitting the text as a chat message. There may be alternative explanations.
Referring now to FIG. 6C, the computing device 210, in response to the indication of input selecting search element 618C, may output, for display, query suggestion region 618D. In the example shown, the computing device 210 has also changed the input focus to the suggestion region 618B, and cleared the suggestion region 618B of any text or suggested queries. The computing device 210 formats and displays within query suggestion region 618D a number of suggested queries. In the example of FIG. 6C, computing device 210 generates the uppermost displayed suggested query based on the text in the edit region 616C, and this suggested query contains the exact same text as that in the edit region 616C. In some examples, the text may be used as a suggested query in the query suggestion region 618D after the computing device 210 makes any stylistic or spelling corrections or other modifications. The computing device also includes within the query suggestion region 618D other suggested queries which may be relevant or pertinent to the text in edit region 616C and possibly the information in the output region 616A. In accordance with one or more aspects of the present disclosure, the graphical keyboard 616B may, in some examples, be considered to be in search mode in FIG. 6C.
In another example that may differ in some respects from examples previously described with respect to FIGS. 6A-6C, a string of characters that has already been transmitted as a text message by the computing device 210 may still be available as a suggested search query. For example, with reference to FIG. 6B, computing device 210 may, before detecting that the user has selected the search element 618C, transmit the text in edit region 616C in FIG. 6B as a chat message in response to user input (e.g., a selection of key 602). The computing device may then later detect an indication of input selecting the search element 618C, and in response to detecting this input, the computing device 210 may cause a query suggestion region to be displayed within the user interface 614. In such an example, the computing device 210 may display suggested queries within the query suggestion region that are similar to or the same as those shown in query suggestion region 618D in FIG. 6C. These suggested queries may include a suggested search query that is created with the exact text of the most recent chat message from the chat application, enabling the same text to be used as a search query. The most recent and earlier text messages exchanged in the chat application may also provide enough context to generate useful other suggested search queries.
FIG. 7 is a flowchart illustrating an example operation of a computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure. The process of FIG. 7 may be performed by one or more processors of a computing device, such as computing devices 110, 210, or 310 as illustrated in FIGS. 1-3. For purposes of illustration, FIG. 7 is described below within the context of computing device 110 of FIG. 1.
In the example of FIG. 7, a computing device 110 may output, for display, a graphical user interface. This graphical user interface may include a user interface for one or more applications. The graphical user interface may also include a graphical keyboard (702). In some examples, the application may be a chat application, such as that described in connection with FIG. 1, but other applications, such as a fitness, calendar, personal assistant, prediction, map, navigation, tracking, social media, email, or game application (or many others) could also or alternatively be used with a graphical keyboard in accordance with one or more aspects of the present disclosure.
The graphical keyboard may include a plurality of graphical keys, a suggestion region, and a search element, such as was described, for example, in connection with FIG. 1. When the graphical keyboard includes a search element, such as the search element 118C shown in FIG. 1, computing device 110 may interpret input associated with a user's selection of the search element as a request to transition the graphical keyboard into search mode in accordance with one or more aspects of the present disclosure. The graphical keyboard may further include a query suggestion region, such as the query suggestion region 118D shown in FIG. 1. The query suggestion region 118D may be may be selectively displayed within the graphical keyboard in accordance with one or more aspects of the present disclosure.
The computing device 110 may determine, based on the user's interaction with one or more of the graphical keys within the graphical keyboard, one or more suggested words and suggested search queries. For example, UI module 120 may detect one or more inputs at mapped locations on PSD 112 and may generate one or more touch events corresponding to the input from the user. Keyboard module 122 may include a spatial module which may select, based on the touch events, one or more keys corresponding to the user input. Keyboard module 122 may also include a language module, which may identify one or more words within a lexicon that may be used to identify suggested words. The keyboard module 122 may also use this information to generate suggested queries, but keyboard module 122 may also use additional information. For example, keyboard module 122 may also base the suggested queries on the user's current activity, context, location, or situation in accordance with one or more aspects of the present disclosure. Computing device 110 may update the suggested words and/or suggested queries as the computing device 110 detects input corresponding to selections of graphical keys within the graphical keyboard (704).
The computing device 110 may determine, based on the user's input or interaction with the computing device 110, whether the user seeks to operate the graphical keyboard in search mode. In some examples, computing device 110 may determine that the user seeks to operate the graphical keyboard in search mode in response to detecting a selection of a search element displayed within the graphical keyboard (YES path from 706). Search mode may generally correspond to the sub-process 720 shown in FIG. 7, but in other examples, search mode in accordance with one or more aspects of the present disclosure may involve more or less or different operations.
The computing device 110 may also determine whether it has received input selecting one of the suggested queries, if any, that may be displayed within the suggestion region (708). If the computing device does not detect input selecting a suggested query, the computing device 110 continues to update the suggested queries in the suggestion region in response to user input (704). If the computing device 110 does detect input selecting a suggested query in the suggestion region, the computing device 110 obtains information associated with the selected suggested query. In some examples, the suggested queries are selectable links, and in response to the computing device 110 detecting input selecting a suggested query, the computing device determines information obtained by invoking the selectable link.
Obtaining information corresponding to a selected suggested query may in some examples involve invoking a search operation using the content of the selected suggested query as a search query. In other examples, obtaining this information might not require a search operation. Situations where a search operation might not be performed include when the selected suggested query points directly to an item of information, or points to an item of information (e.g., a document, or a contact) on the computing device 110. When the computing device 110 determines information obtained by invoking the selectable link associated with the selected suggested query, the computing device 110 may output the obtained information for display (YES path from 708).
If the computing device 110 detects input selecting the search element (YES path from 706), the computing device 110 may output, for display, a query suggestion region within the user interface. In some examples, the query suggestion region is displayed within the graphical keyboard, as shown in FIG. 1. Computing device 110 may format and display suggested search queries that may be updated as the computing device 110 continues to detect user input (712).
If the computing device 110 detects input selecting one of the suggested queries (YES path from 714), the computing device 110 may perform a search corresponding to the selected suggested query, and output, for display, the requested information (718). Alternatively, if the computing device 110 detects input from the user requesting a search based on a string of text characters determined from user input (YES path from 716), then the computing device 110 may perform a search corresponding to the string of text characters, and output, for display, the requested information (718). The computing device 110 may continue to monitor user input and update the query suggestion region with suggested queries (NO path from 716).
In some examples, the operations shown in FIG. 7 may be performed in a different order or presented in a different sequence, but still be in accordance with one or more aspects of the present disclosure. As one example, the order of operations 706 and 708 may be reversed or done in parallel in FIG. 7 (which may involve modifications to other operations, or the order of other operations) while still remaining in accordance with one or more aspects of the present disclosure. As another example, the order of operations 714 and 716 may switched or done in parallel in FIG. 7 (which may involve modifications to other operations, or the order of other operations) while still remaining in accordance with one or more aspects of the present disclosure. Further, a process in accordance with one or more aspects of the present disclosure may be implemented with less than the operations shown in FIG. 7, and in other examples, a process in accordance with one or more aspects of the present disclosure may be implemented with more or different than the operations shown in FIG. 7.
FIG. 8 is a flowchart illustrating an example operation of a computing device that is configured to present a graphical keyboard in accordance with one or more aspects of the present disclosure. The process of FIG. 8 may be performed by one or more processors of a computing device, such as computing devices 110, 210, or 310 as illustrated in FIGS. 1-3. For purposes of illustration, FIG. 8 is described below within the context of computing device 110 of FIG. 1.
In operation, computing device may output a graphical keyboard for display (802). For example, a chat application executing at computing device 110 may invoke keyboard module 122 (e.g., a standalone application or function of computing device 110 that is separate from the chat application) to present graphical keyboard 116B at PSD 112. The graphical keyboard 116B may comprise a plurality of keys and a search element.
Computing device 110 may determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion (804). For example, keyboard module 122 may receive information about touch inputs detected at locations of PSD 112 at which graphical keys 118A of graphical keyboard 116B are being displayed. Keyboard module 122 may determine a suggested query based at least in part on this information.
Computing device 110 may receive an indication of input selecting the search element (806). For example, keyboard module 122 may receive information about a touch input detected at a location of PSD 112 at which search element 118C is being displayed.
Computing device 110 may, in response to receiving the indication of input, output for display an updated graphical keyboard that includes a selectable link associated with the query suggestion. For example, keyboard module 122 may format and display a selectable link such as suggested query 406 shown in FIG. 4B or suggested query 508 in FIG. 5B.
The following numbered clauses may illustrate one or more aspects of the disclosure:
Clause 1. A method comprising: outputting, by a mobile computing device, for display, a graphical keyboard comprising a plurality of keys and a search element; determining, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receiving, by the mobile computing device, an indication of input selecting the search element; and responsive to receiving the indication of input, outputting, by the mobile computing device, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
Clause 2. The method of clause 1, wherein the indication of input is a first indication of input, the method further comprising: receiving, by the mobile computing device, a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determining, based on the second indication of input, information obtained by invoking the selectable link; and outputting, by the mobile computing device, for display, the information obtained by invoking the selectable link.
Clause 3. The method of clause 2, wherein outputting the information obtained by invoking the selectable link includes outputting, within the updated graphical keyboard, the information obtained by invoking the selectable link.
Clause 4. The method of clause 3, wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, and wherein outputting the information obtained by invoking the selectable link includes outputting the information by substantially replacing the plurality of keys with the information.
Clause 5. The method of clause 4, wherein outputting the information obtained by invoking the selectable link includes outputting the information in a scrollable, card-based format.
Clause 6. The method of clause 2-5, wherein determining information obtained by invoking the selectable link includes determining search result information by performing a search operation using content included in the selectable link as a search query.
Clause 7. The method of clause 1-6, wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, and with the selectable link positioned above the plurality of keys.
Clause 8. The method of clause 1-7, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, wherein the selectable link is a first selectable link, and wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the plurality of keys, the method further comprising: determining, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, outputting, by the mobile computing device, for display, a further updated graphical keyboard that includes a second selectable link associated with the updated query suggestion; receiving, by the mobile computing device, a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determining, based on the second indication of input, search result information; and outputting, by the mobile computing device, for display, the search result information.
Clause 9. The method of clause 1-8, wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard as part of an application graphical user interface, and wherein the application graphical user interface includes an edit region that is positioned above the selectable link associated with the query suggestion.
Clause 10. The method of clause 1-9, wherein determining the query suggestion includes determining a plurality of query suggestions including a historical query suggestion based on a prior search, and wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including a selectable link associated with the historical query suggestion.
Clause 11. A mobile device comprising: a presence-sensitive display component; at least one processor; and a memory that stores instructions that when executed cause the at least one processor to: output, for display, a graphical keyboard comprising a plurality of keys and a search element; determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receive an indication of input selecting the search element; and responsive to receiving the indication of input, output, for display, an updated graphical keyboard that includes a selectable link associated with the query suggestion.
Clause 12. The mobile device of clause 11, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to: receive a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, information obtained by invoking the selectable link; and output, for display, the information obtained by invoking the selectable link.
Clause 13. The mobile device of clause 12, wherein the instructions, when executed, further cause the at least one processor to output the information obtained by invoking the selectable link by at least outputting, within the updated graphical keyboard, the information obtained by invoking the selectable link.
Clause 14. The mobile device of clause 13, wherein the updated graphical keyboard is output for display further including the plurality of keys, and wherein the information obtained by invoking the selectable link is output by substantially replacing the plurality of keys with the information obtained by invoking the selectable link.
Clause 15. The mobile device of clause 11-14, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the selectable link is a first selectable link, wherein the updated graphical keyboard is output for display further including the plurality of keys, and wherein the instructions, when executed, further cause the at least one processor to: determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, output, for display, a further updated graphical keyboard that includes a second selectable link associated with the updated query suggestion; receive a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, search result information; and output, for display, the search result information.
Clause 16. A computer-readable storage medium comprising instructions that, when executed, cause at least one processor of a computing device to: output, for display, an application graphical user interface comprising an edit region, a plurality of keys, and a search element; determine, based at least in part on a selection of one or more keys from the plurality of keys, a query suggestion; receive an indication of input selecting the search element; and responsive to receiving the indication of input, output, for display, an updated application graphical user interface that includes a selectable link associated with the query suggestion, wherein the selectable link is positioned between the edit region and the plurality of keys.
Clause 17. The computer-readable storage medium 16, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to: receive a second indication of input invoking the selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, information obtained by invoking the selectable link; and output, for display, the information obtained by invoking the selectable link.
Clause 18. The computer-readable storage medium of clause 17, wherein the instructions, when executed, further cause the at least one processor to output the information obtained by invoking the selectable link by at least outputting, below the edit region, the information obtained by invoking the selectable link.
Clause 19. The computer-readable storage medium of clause 18, wherein the updated graphical user interface is output for display further including the plurality of keys, and wherein the information obtained by invoking the selectable link is output by substantially replacing the plurality of keys with the information obtained by invoking the selectable link.
Clause 20. The computer-readable storage medium of clause 16-19, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the selectable link is a first selectable link, wherein the updated graphical user interface is output for display further including the plurality of keys, and wherein the instructions, when executed, further cause the at least one processor to: determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated query suggestion; responsive to determining the updated query suggestion, output, for display, a further updated graphical user interface that includes a second selectable link associated with the updated query suggestion; receive a second indication of input invoking the second selectable link; responsive to receiving the second indication of input, determine, based on the second indication of input, search result information; and output, for display, the search result information.
Clause 21. A system comprising means for performing any of the methods of clauses 1-10.
Clause 22. A computing device comprising means for performing any of the methods of clauses 1-10.
Throughout the disclosure, examples are described where a computing device and/or a computing system analyzes information (e.g., context, locations, speeds, search queries, etc.) associated with a computing device and a user of a computing device, only if the computing device receives permission from the user of the computing device to analyze the information. For example, in situations discussed below, before a computing device or computing system can collect or may make use of information associated with a user, the user may be provided with an opportunity to provide input to control whether programs or features of the computing device and/or computing system can collect and make use of user information (e.g., information about a user's current location, current speed, etc.), or to dictate whether and/or how to the device and/or system may receive content that may be relevant to the user. In addition, certain data may be treated in one or more ways before it is stored or used by the computing device and/or computing system, so that personally-identifiable information is removed. For example, a user's identity may be treated so that no personally identifiable information can be determined about the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location of a user cannot be determined. Thus, the user may have control over how information is collected about the user and used by the computing device and computing system.
In one or more examples, the functions described may be implemented in hardware, software, firmware, or any combination thereof. If implemented in software, the functions may be stored on or transmitted over, as one or more instructions or code, a computer-readable medium and executed by a hardware-based processing unit. Computer-readable media may include computer-readable storage media, which corresponds to a tangible medium such as data storage media, or communication media including any medium that facilitates transfer of a computer program from one place to another, e.g., according to a communication protocol. In this manner, computer-readable media generally may correspond to (1) tangible computer-readable storage media, which is non-transitory or (2) a communication medium such as a signal or carrier wave. Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for implementation of the techniques described in this disclosure. A computer program product may include a computer-readable medium.
By way of example, and not limitation, such computer-readable storage media can comprise RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage, or other magnetic storage devices, flash memory, or any other medium that can be used to store desired program code in the form of instructions or data structures and that can be accessed by a computer. Also, any connection is properly termed a computer-readable medium. For example, if instructions are transmitted from a website, server, or other remote source using a coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave, then the coaxial cable, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of medium. It should be understood, however, that computer-readable storage media and data storage media do not include connections, carrier waves, signals, or other transient media, but are instead directed to non-transient, tangible storage media. Disk and disc, as used, includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD), floppy disk and Blu-ray disc, where disks usually reproduce data magnetically, while discs reproduce data optically with lasers. Combinations of the above should also be included within the scope of computer-readable media.
Instructions may be executed by one or more processors, such as one or more digital signal processors (DSPs), general purpose microprocessors, application specific integrated circuits (ASICs), field programmable logic arrays (FPGAs), or other equivalent integrated or discrete logic circuitry. Accordingly, the term “processor,” as used may refer to any of the foregoing structure or any other structure suitable for implementation of the techniques described. In addition, in some aspects, the functionality described may be provided within dedicated hardware and/or software modules. Also, the techniques could be fully implemented in one or more circuits or logic elements.
The techniques of this disclosure may be implemented in a wide variety of devices or apparatuses, including a wireless handset, an integrated circuit (IC) or a set of ICs (e.g., a chip set). Various components, modules, or units are described in this disclosure to emphasize functional aspects of devices configured to perform the disclosed techniques, but do not necessarily require realization by different hardware units. Rather, as described above, various units may be combined in a hardware unit or provided by a collection of interoperative hardware units, including one or more processors as described above, in conjunction with suitable software and/or firmware.
Various examples have been described. These and other examples may be within the scope of the following claims.

Claims (20)

What is claimed is:
1. A method comprising:
outputting, by a mobile computing device, for display, a graphical keyboard comprising a plurality of keys, a cursor identifying input focus of the graphical keyboard, an edit region, a first region including a first plurality of suggested words generated based on input provided within the edit region, and a search element, the search element being a selectable icon configured to transition the graphical keyboard into a search mode, the input focus of the cursor being within the edit region;
receiving, by the mobile computing device, an indication of input selecting the search element; and
responsive to receiving the indication of the input selecting the search element:
creating a second region configured to output search results generated based on input provided within the first region;
deleting the first plurality of suggested words from the first region;
shifting input focus of the cursor to the first region to initiate the search mode;
determining, based at least in part on a selection of one or more keys from the plurality of keys provided within the first region, a first query suggestion and a second query suggestion;
and
outputting, for display, an updated graphical keyboard that includes the plurality of keys, the first region including respective characters corresponding to the one or more keys, and the second region including the first query suggestion and the second query suggestion.
2. The method of claim 1, wherein the indication of input is a first indication of input, the method further comprising:
receiving, by the mobile computing device, a second indication of input selecting the first query suggestion;
determining information associated with the first query suggestion; and
outputting, by the mobile computing device, for display, the information associated with the first query suggestion.
3. The method of claim 2, wherein outputting the information associated with the first query suggestion includes outputting, within the updated graphical keyboard, the information associated with the first query suggestion.
4. The method of claim 3, wherein outputting the information associated with the first query suggestion includes outputting the information by substantially replacing the plurality of keys with the information.
5. The method of claim 4, wherein outputting the information associated with the first query suggestion includes outputting the information in a horizontally-scrollable, card-based format.
6. The method of claim 2, wherein determining information associated with the first query suggestion includes determining search result information by performing a search operation using text included within the first query suggestion as a search query.
7. The method of claim 1, wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard with the first region positioned above the plurality of keys.
8. The method of claim 1, wherein the one or more keys is a first set of keys, and wherein the indication of input is a first indication of input, the method further comprising:
determining, based at least in part on a selection of a second set of keys from the plurality of keys, an updated first query suggestion;
responsive to determining the updated first query suggestion, outputting, by the mobile computing device, for display, a further updated graphical keyboard that includes the updated first query suggestion;
receiving, by the mobile computing device, a second indication of input selecting the updated first query suggestion;
determining search result information associated with the updated first query suggestion; and
outputting, by the mobile computing device, for display, the search result information.
9. The method of claim 1, wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard as part of an application graphical user interface, and wherein the application graphical user interface includes an edit region that is positioned above the first query suggestion and the second query suggestion.
10. The method of claim 1, wherein determining the first query suggestion and the second query suggestion further includes determining an historical query suggestion based on a prior search, and wherein outputting the updated graphical keyboard includes outputting the updated graphical keyboard further including the historical query suggestion.
11. A mobile device comprising:
a presence-sensitive display component;
at least one processor; and
a memory that stores instructions that when executed cause the at least one processor to:
output, for display, a graphical keyboard comprising a plurality of keys, a cursor identifying input focus of the graphical keyboard, an edit region, a first region including a first plurality of suggested words generated based on input provided within the edit region, and a search element, the search element being a selectable icon configured to transition the graphical keyboard into a search mode, the input focus of the cursor being within the edit region;
receive an indication of input selecting the search element; and
responsive to receiving the indication of the input selecting the search element:
create a second region configured to output search results generated based on input provided within the first region;
delete the first plurality of suggested words from the first region;
shift input focus of the cursor to the first region to initiate the search mode;
determine, based at least in part on a selection of one or more keys from the plurality of keys provided within the first region, a first query suggestion and a second query suggestion;
and
output, for display, an updated graphical keyboard that includes the plurality of keys, the first region including respective characters corresponding to the one or more keys, and the second region including the first query suggestion and the second query suggestion.
12. The mobile device of claim 11, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to:
receive a second indication of input selecting the first query suggestion;
determine information relevant to the first query suggestion; and
output, for display, the information relevant to the first query suggestion.
13. The mobile device of claim 12, wherein the instructions, when executed, further cause the at least one processor to output the information relevant to the first query suggestion by at least outputting, within the updated graphical keyboard, the information relevant to the first query suggestion.
14. The mobile device of claim 13, wherein the information relevant to the first query suggestion is output by substantially replacing the plurality of keys with the information relevant to the first query suggestion.
15. The mobile device of claim 11, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to:
determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated first query suggestion;
responsive to determining the updated first query suggestion, output, for display, a further updated graphical keyboard that includes the updated first query suggestion;
receive a second indication of input selecting the updated first query suggestion;
determine search result information relevant to the updated first query suggestion; and
output, for display, the search result information.
16. A non-transitory computer-readable storage medium comprising instructions that, when executed, cause at least one processor of a computing device to:
output, for display, an application graphical user interface comprising an edit region, a plurality of keys, a cursor identifying input focus of the graphical keyboard, a first region including a first plurality of suggested words generated based on input provided within the edit region, and a search element, the search element being a selectable icon configured to transition the graphical keyboard into a search mode, the input focus of the cursor being within the edit region;
receive an indication of input selecting the search element; and
responsive to receiving the indication of the input selecting the search element:
create a second region configured to output search results generated based on input provided within the first region;
delete the first plurality of suggested words from the first region;
shift input focus of the cursor to the first region to initiate the search mode;
determine, based at least in part on a selection of one or more keys from the plurality of keys provided within the first region, a first query suggestion and a second query suggestion; and
output, for display, an updated application graphical user interface that includes the edit region, the plurality of keys, the first region including respective characters corresponding to the one or more keys, and the second region including the first query suggestion and the second query suggestion.
17. The non-transitory computer-readable storage medium of claim 16, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to:
receive a second indication of input selecting the first query suggestion;
determine information pertaining to the first query suggestion; and
output, for display, the information pertaining to the first query suggestion.
18. The non-transitory computer-readable storage medium of claim 17, wherein the instructions, when executed, further cause the at least one processor to output the information pertaining to the first query suggestion by at least outputting, below the edit region, the information pertaining to the first query suggestion.
19. The non-transitory computer-readable storage medium of claim 18, wherein the information pertaining to the first query suggestion is output by substantially replacing the plurality of keys with the information pertaining to the first query suggestion.
20. The non-transitory computer-readable storage medium of claim 16, wherein the one or more keys is a first set of keys, wherein the indication of input is a first indication of input, and wherein the instructions, when executed, further cause the at least one processor to:
determine, based at least in part on a selection of a second set of keys from the plurality of keys, an updated first query suggestion;
responsive to determining the updated first query suggestion, output, for display, a further updated graphical user interface that includes the updated first query suggestion;
receive a second indication of input selecting the updated first query suggestion;
determine search result information pertaining to the updated first query suggestion; and
output, for display, the search result information.
US15/299,027 2016-04-20 2016-10-20 Keyboard with a suggested search query region Active US9977595B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US15/299,027 US9977595B2 (en) 2016-04-20 2016-10-20 Keyboard with a suggested search query region

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US15/134,029 US10222957B2 (en) 2016-04-20 2016-04-20 Keyboard with a suggested search query region
US15/299,027 US9977595B2 (en) 2016-04-20 2016-10-20 Keyboard with a suggested search query region

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US15/134,029 Continuation US10222957B2 (en) 2016-04-20 2016-04-20 Keyboard with a suggested search query region

Publications (2)

Publication Number Publication Date
US20170308292A1 US20170308292A1 (en) 2017-10-26
US9977595B2 true US9977595B2 (en) 2018-05-22

Family

ID=57838539

Family Applications (2)

Application Number Title Priority Date Filing Date
US15/134,029 Active 2037-05-05 US10222957B2 (en) 2016-04-20 2016-04-20 Keyboard with a suggested search query region
US15/299,027 Active US9977595B2 (en) 2016-04-20 2016-10-20 Keyboard with a suggested search query region

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US15/134,029 Active 2037-05-05 US10222957B2 (en) 2016-04-20 2016-04-20 Keyboard with a suggested search query region

Country Status (4)

Country Link
US (2) US10222957B2 (en)
EP (1) EP3408733B1 (en)
CN (2) CN108700993B (en)
WO (1) WO2017184220A1 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180239770A1 (en) * 2017-02-17 2018-08-23 Microsoft Technology Licensing, Llc Real-time personalized suggestions for communications between participants
US11216245B2 (en) 2019-03-25 2022-01-04 Samsung Electronics Co., Ltd. Electronic device and multitasking supporting method thereof
US11379529B2 (en) 2019-09-09 2022-07-05 Microsoft Technology Licensing, Llc Composing rich content messages

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10976922B2 (en) * 2013-02-17 2021-04-13 Benjamin Firooz Ghassabian Data entry systems
US9721024B2 (en) * 2014-12-19 2017-08-01 Facebook, Inc. Searching for ideograms in an online social network
USD912682S1 (en) * 2016-03-29 2021-03-09 Beijing Sogou Technology Development Co., Ltd. Display screen or portion thereof with animated graphical user interface
US11494547B2 (en) * 2016-04-13 2022-11-08 Microsoft Technology Licensing, Llc Inputting images to electronic devices
US10305828B2 (en) 2016-04-20 2019-05-28 Google Llc Search query predictions by a keyboard
US10140017B2 (en) 2016-04-20 2018-11-27 Google Llc Graphical keyboard application with integrated search
US10222957B2 (en) 2016-04-20 2019-03-05 Google Llc Keyboard with a suggested search query region
US10078673B2 (en) 2016-04-20 2018-09-18 Google Llc Determining graphical elements associated with text
US9965530B2 (en) 2016-04-20 2018-05-08 Google Llc Graphical keyboard with integrated search features
US10664157B2 (en) 2016-08-03 2020-05-26 Google Llc Image search query predictions by a keyboard
US20180101599A1 (en) * 2016-10-08 2018-04-12 Microsoft Technology Licensing, Llc Interactive context-based text completions
WO2018176012A1 (en) * 2017-03-24 2018-09-27 Inmentis, Llc Social media system with navigable, artificial-intelligence-based graphical user interface with result view
USD859453S1 (en) 2017-08-01 2019-09-10 Google Llc Display screen with an animated graphical user interface
US10482504B2 (en) * 2017-08-24 2019-11-19 William McMichael Systems and methods for analyzing input data and presenting information
USD874491S1 (en) * 2017-11-04 2020-02-04 Kika Tech (Cayman) Holdings Co., Limited Display screen with computer-generated graphical user interface
US11016957B2 (en) 2018-02-28 2021-05-25 Microsoft Technology Licensing, Llc Sensor data based query results
US12093330B2 (en) 2018-04-11 2024-09-17 Microsoft Technology Licensing, Llc IoT enhanced search results
US20190347358A1 (en) * 2018-05-10 2019-11-14 Microsoft Technology Licensing, Llc Query Formulation Using Networked Device Candidates
US10762274B2 (en) * 2018-06-18 2020-09-01 International Business Machines Corporation Execution of an application using a specifically formatted input
US10901577B2 (en) 2018-07-17 2021-01-26 Google Llc Methods and systems for input suggestion
US10956507B2 (en) * 2018-08-16 2021-03-23 Rovi Guides, Inc. Reaction compensated result selection
EP3701372B1 (en) * 2018-08-24 2023-03-29 Google LLC Dynamically configurable application control elements
US11681718B2 (en) * 2019-03-29 2023-06-20 Apple Inc. Scoping a system-wide search to a user-specified application
US10860178B1 (en) * 2019-09-05 2020-12-08 Shabu Ans Kandamkulathy Task management through soft keyboard applications
WO2021050082A1 (en) * 2019-09-13 2021-03-18 Google Llc Text entry recommendations based on stored search results
USD936078S1 (en) * 2019-10-29 2021-11-16 Optum, Inc. Display screen or portion thereof with animated graphical user interface
US11422836B1 (en) * 2021-05-07 2022-08-23 Avaya Management L.P. User guidance from gaze information during a communication session while viewing a webpage
CN113589991A (en) * 2021-08-13 2021-11-02 北京字跳网络技术有限公司 Text input method and device, electronic equipment and storage medium
JP2023101113A (en) * 2022-01-07 2023-07-20 オムロン株式会社 Character input device, character input method, and character input program
US20230252737A1 (en) * 2022-02-08 2023-08-10 Apple Inc. Devices, methods, and graphical user interfaces for interacting with virtual objects using hand gestures

Citations (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6104397A (en) 1997-06-30 2000-08-15 Sun Microsystems, Inc. Method and system for generating improved progress indicators
US20060048076A1 (en) 2004-08-31 2006-03-02 Microsoft Corporation User Interface having a carousel view
US20060294189A1 (en) 2005-06-23 2006-12-28 Microsoft Corporation Instant messaging with search
US20070088686A1 (en) 2005-10-14 2007-04-19 Microsoft Corporation Search results injected into client applications
US20070130276A1 (en) 2005-12-05 2007-06-07 Chen Zhang Facilitating retrieval of information within a messaging environment
US20070300177A1 (en) * 2006-06-23 2007-12-27 Microsoft Corporation User interface for specifying multi-valued properties
US20080201434A1 (en) 2007-02-16 2008-08-21 Microsoft Corporation Context-Sensitive Searches and Functionality for Instant Messaging Applications
US20110191364A1 (en) 2010-02-03 2011-08-04 Google Inc. Information search system with real-time feedback
US20110191321A1 (en) 2010-02-01 2011-08-04 Microsoft Corporation Contextual display advertisements for a webpage
US20110201387A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation Real-time typing assistance
US20110221678A1 (en) 2010-03-12 2011-09-15 Anton Davydov Device, Method, and Graphical User Interface for Creating and Using Duplicate Virtual Keys
US20120036469A1 (en) * 2010-07-28 2012-02-09 Daniel Suraqui Reduced keyboard with prediction solutions when input is a partial sliding trajectory
US20120102549A1 (en) 2010-10-06 2012-04-26 Citrix Systems, Inc. Mediating resource access based on a physical location of a mobile device
US20120124071A1 (en) * 2010-11-16 2012-05-17 Microsoft Corporation Extensible search term suggestion engine
US20120124519A1 (en) * 2010-11-16 2012-05-17 Microsoft Corporation Facilitating Interaction with System Level Search User Interface
US20120127083A1 (en) 2010-11-20 2012-05-24 Kushler Clifford A Systems and methods for using entered text to access and process contextual information
US20120254227A1 (en) 2011-03-31 2012-10-04 Microsoft Corporation Augmented Conversational Understanding Architecture
US20130120267A1 (en) * 2011-11-10 2013-05-16 Research In Motion Limited Methods and systems for removing or replacing on-keyboard prediction candidates
US8484573B1 (en) * 2012-05-23 2013-07-09 Google Inc. Predictive virtual keyboard
US20130246913A1 (en) * 2012-03-16 2013-09-19 Microsoft Corporation Use of touch and gestures related to tasks and business workflow
US20130285913A1 (en) * 2012-04-30 2013-10-31 Research In Motion Limited Touchscreen keyboard providing word predictions at locations in association with candidate letters
US20130297317A1 (en) 2012-04-16 2013-11-07 Htc Corporation Method for offering suggestion during conversation, electronic device using the same, and non-transitory storage medium
US20140002363A1 (en) * 2012-06-27 2014-01-02 Research In Motion Limited Touchscreen keyboard providing selection of word predictions in partitions of the touchscreen keyboard
US8650210B1 (en) 2010-02-09 2014-02-11 Google Inc. Identifying non-search actions based on a search query
US8706750B2 (en) 2010-08-19 2014-04-22 Google Inc. Predictive query completion and predictive search results
US20140115070A1 (en) * 2012-10-22 2014-04-24 Nokia Corporation Apparatus and associated methods
US8745018B1 (en) 2008-07-10 2014-06-03 Google Inc. Search application and web browser interaction
US20140172814A1 (en) * 2012-12-17 2014-06-19 Microsoft Corporation Building Long Search Queries
US20140201676A1 (en) 2011-08-25 2014-07-17 Tencent Technology (Shenzhen) Company Limited Method and apparatus for switching pages in interfaces, and computer storage medium thereof
US20140223372A1 (en) 2013-02-04 2014-08-07 602531 British Columbia Ltd. Method, system, and apparatus for executing an action related to user selection
US20140229847A1 (en) * 2011-10-13 2014-08-14 Lg Electronics Inc. Input interface controlling apparatus and method thereof
US20140282136A1 (en) 2013-03-14 2014-09-18 Microsoft Corporation Query intent expression for search in an embedded application context
US20140282203A1 (en) * 2013-03-15 2014-09-18 Research In Motion Limited System and method for predictive text input
US20140330769A1 (en) 2012-05-08 2014-11-06 24/7 Customer, Inc. Predictive 411
US20140358940A1 (en) 2013-06-03 2014-12-04 Google Inc. Query Suggestion Templates
US20150006505A1 (en) * 2013-06-28 2015-01-01 Yandex Europe Ag Method of and system for displaying a plurality of user-selectable refinements to a search query
US20150121286A1 (en) * 2013-10-30 2015-04-30 Samsung Electronics Co., Ltd. Display apparatus and user interface providing method thereof
US20150201065A1 (en) 2014-01-14 2015-07-16 Lg Electronics Inc. Mobile terminal
US20150242086A1 (en) 2014-02-21 2015-08-27 Markport Limited Drag and drop event system and method
EP2940605A1 (en) 2014-05-02 2015-11-04 Samsung Electronics Co., Ltd Information search system and method
US20150317316A1 (en) * 2014-04-30 2015-11-05 Yahoo! Inc. Method and system for user selection of query suggestions
US20150331605A1 (en) * 2014-05-16 2015-11-19 Lg Electronics Inc. Mobile terminal and method of controlling the same
US20150370434A1 (en) * 2014-06-18 2015-12-24 Lenovo Enterprise Solutions (Singapore) Pte. Ltd. Presenting search term suggestions on graphical user interfaces
US20160006856A1 (en) * 2014-07-07 2016-01-07 Verizon Patent And Licensing Inc. Messaging application with in-application search functionality
US20160034977A1 (en) 2014-08-01 2016-02-04 Yahoo! Inc. System and method for embedded search within messaging applications
US20160124926A1 (en) * 2014-10-28 2016-05-05 Idelan, Inc. Advanced methods and systems for text input error correction
US20160330150A1 (en) 2015-05-06 2016-11-10 Kakao Corp. Message service providing method for message service linked to search service and message server and user terminal to perform the method
US20160334988A1 (en) * 2014-01-03 2016-11-17 Samsung Electronics Co., Ltd. Display device and method for providing recommended characters from same
US20170102870A1 (en) 2015-10-12 2017-04-13 Microsoft Technology Licensing, Llc Multi-window keyboard

Family Cites Families (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7363295B2 (en) 2004-04-19 2008-04-22 Yahoo! Inc. Techniques for inline searching in an instant messenger environment
US7788248B2 (en) 2005-03-08 2010-08-31 Apple Inc. Immediate search feedback
US20080249764A1 (en) 2007-03-01 2008-10-09 Microsoft Corporation Smart Sentiment Classifier for Product Reviews
US20080244446A1 (en) 2007-03-29 2008-10-02 Lefevre John Disambiguation of icons and other media in text-based applications
CN101055582A (en) * 2007-05-08 2007-10-17 魏新成 Search operation method integrated in Chinese character input method
US8595252B2 (en) * 2007-09-14 2013-11-26 Google Inc. Suggesting alternative queries in query results
CN101876878A (en) 2009-04-29 2010-11-03 深圳富泰宏精密工业有限公司 Word prediction input system and method
EP2312427B1 (en) 2009-10-13 2013-09-18 BlackBerry Limited User interface for a touchscreen display
US20110099464A1 (en) * 2009-10-28 2011-04-28 Yahoo! Inc. Mechanism for adding content from a search to a document or message
US20110112824A1 (en) 2009-11-06 2011-05-12 Craig Peter Sayers Determining at least one category path for identifying input text
US20120146955A1 (en) 2010-12-10 2012-06-14 Research In Motion Limited Systems and methods for input into a portable electronic device
US8583672B1 (en) * 2011-04-14 2013-11-12 Google Inc. Displaying multiple spelling suggestions
US9176944B1 (en) 2011-08-23 2015-11-03 Google Inc. Selectively processing user input
US20130124490A1 (en) * 2011-11-10 2013-05-16 Microsoft Corporation Contextual suggestion of search queries
US9582146B2 (en) * 2012-05-29 2017-02-28 Nokia Technologies Oy Causing display of search results
US9489293B2 (en) 2012-08-17 2016-11-08 Netapp, Inc. Techniques for opportunistic data storage
US9122376B1 (en) * 2013-04-18 2015-09-01 Google Inc. System for improving autocompletion of text input
US9626960B2 (en) 2013-04-25 2017-04-18 Nuance Communications, Inc. Systems and methods for providing metadata-dependent language models
US20150100537A1 (en) 2013-10-03 2015-04-09 Microsoft Corporation Emoji for Text Predictions
US9461945B2 (en) 2013-10-18 2016-10-04 Jeffrey P. Phillips Automated messaging response
US9530416B2 (en) 2013-10-28 2016-12-27 At&T Intellectual Property I, L.P. System and method for managing models for embedded speech and language processing
US10050926B2 (en) 2014-02-05 2018-08-14 Facebook, Inc. Ideograms based on sentiment analysis
US9043196B1 (en) 2014-07-07 2015-05-26 Machine Zone, Inc. Systems and methods for identifying and suggesting emoticons
CN104156161A (en) * 2014-08-05 2014-11-19 魏新成 System and method for carrying out clicking, word capturing and searching on information equipment screen
US10824654B2 (en) 2014-09-18 2020-11-03 Snap Inc. Geolocation-based pictographs
CN104503995A (en) * 2014-12-04 2015-04-08 百度在线网络技术(北京)有限公司 Realization method and device for input method
US20160224524A1 (en) 2015-02-03 2016-08-04 Nuance Communications, Inc. User generated short phrases for auto-filling, automatically collected during normal text use
US9965530B2 (en) 2016-04-20 2018-05-08 Google Llc Graphical keyboard with integrated search features
US10078673B2 (en) 2016-04-20 2018-09-18 Google Llc Determining graphical elements associated with text
US10305828B2 (en) 2016-04-20 2019-05-28 Google Llc Search query predictions by a keyboard
US10140017B2 (en) 2016-04-20 2018-11-27 Google Llc Graphical keyboard application with integrated search
US10222957B2 (en) 2016-04-20 2019-03-05 Google Llc Keyboard with a suggested search query region

Patent Citations (51)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6104397A (en) 1997-06-30 2000-08-15 Sun Microsystems, Inc. Method and system for generating improved progress indicators
US20060048076A1 (en) 2004-08-31 2006-03-02 Microsoft Corporation User Interface having a carousel view
US20060294189A1 (en) 2005-06-23 2006-12-28 Microsoft Corporation Instant messaging with search
US20070088686A1 (en) 2005-10-14 2007-04-19 Microsoft Corporation Search results injected into client applications
US20070130276A1 (en) 2005-12-05 2007-06-07 Chen Zhang Facilitating retrieval of information within a messaging environment
US20070300177A1 (en) * 2006-06-23 2007-12-27 Microsoft Corporation User interface for specifying multi-valued properties
US20080201434A1 (en) 2007-02-16 2008-08-21 Microsoft Corporation Context-Sensitive Searches and Functionality for Instant Messaging Applications
US9086775B1 (en) * 2008-07-10 2015-07-21 Google Inc. Minimizing software based keyboard
US8745018B1 (en) 2008-07-10 2014-06-03 Google Inc. Search application and web browser interaction
US20110191321A1 (en) 2010-02-01 2011-08-04 Microsoft Corporation Contextual display advertisements for a webpage
US20110191364A1 (en) 2010-02-03 2011-08-04 Google Inc. Information search system with real-time feedback
US8650210B1 (en) 2010-02-09 2014-02-11 Google Inc. Identifying non-search actions based on a search query
US20110201387A1 (en) * 2010-02-12 2011-08-18 Microsoft Corporation Real-time typing assistance
US20110221678A1 (en) 2010-03-12 2011-09-15 Anton Davydov Device, Method, and Graphical User Interface for Creating and Using Duplicate Virtual Keys
US20120036469A1 (en) * 2010-07-28 2012-02-09 Daniel Suraqui Reduced keyboard with prediction solutions when input is a partial sliding trajectory
US8706750B2 (en) 2010-08-19 2014-04-22 Google Inc. Predictive query completion and predictive search results
US20120102549A1 (en) 2010-10-06 2012-04-26 Citrix Systems, Inc. Mediating resource access based on a physical location of a mobile device
US20120124071A1 (en) * 2010-11-16 2012-05-17 Microsoft Corporation Extensible search term suggestion engine
US20120124519A1 (en) * 2010-11-16 2012-05-17 Microsoft Corporation Facilitating Interaction with System Level Search User Interface
US20120127083A1 (en) 2010-11-20 2012-05-24 Kushler Clifford A Systems and methods for using entered text to access and process contextual information
US20120254227A1 (en) 2011-03-31 2012-10-04 Microsoft Corporation Augmented Conversational Understanding Architecture
US20140201676A1 (en) 2011-08-25 2014-07-17 Tencent Technology (Shenzhen) Company Limited Method and apparatus for switching pages in interfaces, and computer storage medium thereof
US20140229847A1 (en) * 2011-10-13 2014-08-14 Lg Electronics Inc. Input interface controlling apparatus and method thereof
US20130120267A1 (en) * 2011-11-10 2013-05-16 Research In Motion Limited Methods and systems for removing or replacing on-keyboard prediction candidates
US20130246913A1 (en) * 2012-03-16 2013-09-19 Microsoft Corporation Use of touch and gestures related to tasks and business workflow
US20130297317A1 (en) 2012-04-16 2013-11-07 Htc Corporation Method for offering suggestion during conversation, electronic device using the same, and non-transitory storage medium
US20130285913A1 (en) * 2012-04-30 2013-10-31 Research In Motion Limited Touchscreen keyboard providing word predictions at locations in association with candidate letters
US20140330769A1 (en) 2012-05-08 2014-11-06 24/7 Customer, Inc. Predictive 411
US8484573B1 (en) * 2012-05-23 2013-07-09 Google Inc. Predictive virtual keyboard
US20140002363A1 (en) * 2012-06-27 2014-01-02 Research In Motion Limited Touchscreen keyboard providing selection of word predictions in partitions of the touchscreen keyboard
US20140115070A1 (en) * 2012-10-22 2014-04-24 Nokia Corporation Apparatus and associated methods
US20140172814A1 (en) * 2012-12-17 2014-06-19 Microsoft Corporation Building Long Search Queries
US20140223372A1 (en) 2013-02-04 2014-08-07 602531 British Columbia Ltd. Method, system, and apparatus for executing an action related to user selection
US20140282136A1 (en) 2013-03-14 2014-09-18 Microsoft Corporation Query intent expression for search in an embedded application context
US20140282203A1 (en) * 2013-03-15 2014-09-18 Research In Motion Limited System and method for predictive text input
US20140358940A1 (en) 2013-06-03 2014-12-04 Google Inc. Query Suggestion Templates
US20150006505A1 (en) * 2013-06-28 2015-01-01 Yandex Europe Ag Method of and system for displaying a plurality of user-selectable refinements to a search query
US20150121286A1 (en) * 2013-10-30 2015-04-30 Samsung Electronics Co., Ltd. Display apparatus and user interface providing method thereof
US20160334988A1 (en) * 2014-01-03 2016-11-17 Samsung Electronics Co., Ltd. Display device and method for providing recommended characters from same
US20150201065A1 (en) 2014-01-14 2015-07-16 Lg Electronics Inc. Mobile terminal
US20150242086A1 (en) 2014-02-21 2015-08-27 Markport Limited Drag and drop event system and method
US20150317316A1 (en) * 2014-04-30 2015-11-05 Yahoo! Inc. Method and system for user selection of query suggestions
EP2940605A1 (en) 2014-05-02 2015-11-04 Samsung Electronics Co., Ltd Information search system and method
US20150331605A1 (en) * 2014-05-16 2015-11-19 Lg Electronics Inc. Mobile terminal and method of controlling the same
US20150370434A1 (en) * 2014-06-18 2015-12-24 Lenovo Enterprise Solutions (Singapore) Pte. Ltd. Presenting search term suggestions on graphical user interfaces
US20160006856A1 (en) * 2014-07-07 2016-01-07 Verizon Patent And Licensing Inc. Messaging application with in-application search functionality
US20160034977A1 (en) 2014-08-01 2016-02-04 Yahoo! Inc. System and method for embedded search within messaging applications
US20160124926A1 (en) * 2014-10-28 2016-05-05 Idelan, Inc. Advanced methods and systems for text input error correction
US20160330150A1 (en) 2015-05-06 2016-11-10 Kakao Corp. Message service providing method for message service linked to search service and message server and user terminal to perform the method
US20170102870A1 (en) 2015-10-12 2017-04-13 Microsoft Technology Licensing, Llc Multi-window keyboard
US20170102871A1 (en) 2015-10-12 2017-04-13 Microsoft Technology Licensing, Llc Multi-window keyboard

Non-Patent Citations (18)

* Cited by examiner, † Cited by third party
Title
Chansanchai, "Hub Keyboard app from Microsoft Garage makes it easy to multitask from one mobile screen", Microsoft, Feb. 23, 2016. Retrieved from the Internet: <https://blogs.microsoft.com/firehose/2016/02/23/hub-Keyboard-app-from-microsoft-garage-makes-it-easy-to-multitask-from-one-mobile-screen/> 6 pp.
Chansanchai, "Hub Keyboard app from Microsoft Garage makes it easy to multitask from one mobile screen", retrieved from the Internet: <https://blogs.microsoft.com/firehose/2016/02/23/hub-keyboard-app-from-microsoft-garage-makes-it-easy-to-multitask-from-one-mobile-screen/#sm.0001ekq05s8cid6jpo52ovcbjyjie>, Feb. 23, 2016, 6 pp.
International Search Report and Written Opinion of International Application No. PCT/US2016/069267, dated Apr. 3, 2017, 15 pp.
Mcalone, "Slash is the best iPhone Keyboard-Business Insider," retrieved from http://www.businessinsider.com/slash-is-the-best-iphone-keyboard-2015-9, Sep. 22, 2015, 26 pp.
Mcalone, "Slash is the best iPhone Keyboard—Business Insider," retrieved from http://www.businessinsider.com/slash-is-the-best-iphone-keyboard-2015-9, Sep. 22, 2015, 26 pp.
Response to Written Opinion dated Apr. 3, 2017, from International Application No. PCT/US2016/069267, dated Nov. 30, 2017, 34 pp.
Russell-Rose et al., "How to provide autocomplete and autosuggest on the same search box at the same time,"StackExchange, accessed from http://ux.stackexchange.com/questions/40104/how-to-provide-autocomplete-and-autosuggest-on-the-same-search-box-at-the-same-t, May 25, 2013, 2 pp.
Second Written Opinion of International Application No. PCT/US2016/069267, dated Mar. 22, 2018, 9 pp.
U.S. Appl. No. 15/133,291, by Jing Cao, filed Apr. 20, 2016.
U.S. Appl. No. 15/134,029, by Min-sang Choi, filed Apr. 20, 2016.
U.S. Appl. No. 15/134,243, by Muhammad Mohsin, filed Apr. 20, 2016.
U.S. Appl. No. 15/134,319, by Heather Luipold, filed Apr. 20, 2016.
U.S. Appl. No. 15/134,323, by Jens Nagel, filed Apr. 20, 2016.
U.S. Appl. No. 15/246,091, by Nicholas Chi-Yuen Kong, filed Aug. 24, 2016.
U.S. Appl. No. 15/289,661, by Jing Cao, filed Oct. 10, 2016.
U.S. Appl. No. 15/332,409, by Muhammad Mohsin et al., filed Oct. 24, 2016.
U.S. Appl. No. 15/332,513, by Heather Luipold, filed Oct. 24, 2016.
Whitwam, "Microsoft Hub Keyboard Comes to Android With All Things Microsoft in Tow," retrieved from http://www.androidpolice.com/2016/02/23/microsoft-hub-keyboard-comes-to-android-with-all-things-microsoft-in-tow/, Feb. 23, 2016, 5 pp.

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180239770A1 (en) * 2017-02-17 2018-08-23 Microsoft Technology Licensing, Llc Real-time personalized suggestions for communications between participants
US11907272B2 (en) * 2017-02-17 2024-02-20 Microsoft Technology Licensing, Llc Real-time personalized suggestions for communications between participants
US11216245B2 (en) 2019-03-25 2022-01-04 Samsung Electronics Co., Ltd. Electronic device and multitasking supporting method thereof
US11379529B2 (en) 2019-09-09 2022-07-05 Microsoft Technology Licensing, Llc Composing rich content messages

Also Published As

Publication number Publication date
US20170308292A1 (en) 2017-10-26
CN113391750A (en) 2021-09-14
US10222957B2 (en) 2019-03-05
EP3408733A1 (en) 2018-12-05
US20170308273A1 (en) 2017-10-26
EP3408733B1 (en) 2020-04-29
WO2017184220A1 (en) 2017-10-26
CN108700993B (en) 2021-06-01
CN108700993A (en) 2018-10-23

Similar Documents

Publication Publication Date Title
US9977595B2 (en) Keyboard with a suggested search query region
US10140017B2 (en) Graphical keyboard application with integrated search
EP3479213B1 (en) Image search query predictions by a keyboard
US9720955B1 (en) Search query predictions by a keyboard
CN108700951B (en) Iconic symbol search within a graphical keyboard
EP3400539B1 (en) Determining graphical elements associated with text
US9946773B2 (en) Graphical keyboard with integrated search features
US20170308290A1 (en) Iconographic suggestions within a keyboard
US20180173692A1 (en) Iconographic symbol predictions for a conversation
WO2017181355A1 (en) Automatic translations by keyboard

Legal Events

Date Code Title Description
AS Assignment

Owner name: GOOGLE INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:CHOI, MIN-SANG;REEL/FRAME:040082/0899

Effective date: 20160420

AS Assignment

Owner name: GOOGLE LLC, CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:GOOGLE INC.;REEL/FRAME:044567/0001

Effective date: 20170929

STCF Information on status: patent grant

Free format text: PATENTED CASE

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 4