EP2803004A1 - User interface for entertainment systems - Google Patents

User interface for entertainment systems

Info

Publication number
EP2803004A1
EP2803004A1 EP13753963.1A EP13753963A EP2803004A1 EP 2803004 A1 EP2803004 A1 EP 2803004A1 EP 13753963 A EP13753963 A EP 13753963A EP 2803004 A1 EP2803004 A1 EP 2803004A1
Authority
EP
European Patent Office
Prior art keywords
content
electronic device
search
input
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP13753963.1A
Other languages
German (de)
French (fr)
Inventor
Hoi Young
Steven Hatch
Yuen-Keen Cheong
Tapio Ilmari KOIVUNIEMI
Tuomas Antero TUONONEN
Jarkko Koivikko
Teijo Juhani KINNUNEN
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nuance Communications Inc
Original Assignee
Nuance Communications Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US13/587,353 external-priority patent/US9106957B2/en
Priority claimed from US13/587,300 external-priority patent/US9026448B2/en
Priority claimed from US13/587,280 external-priority patent/US9031848B2/en
Priority claimed from US13/587,222 external-priority patent/US8799959B2/en
Priority claimed from US13/587,260 external-priority patent/US20140052450A1/en
Priority claimed from US13/587,234 external-priority patent/US9497515B2/en
Application filed by Nuance Communications Inc filed Critical Nuance Communications Inc
Publication of EP2803004A1 publication Critical patent/EP2803004A1/en
Withdrawn legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/43Querying
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • G06F16/48Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/489Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using time information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/732Query formulation
    • G06F16/7343Query language or query format
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7844Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42204User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor
    • H04N21/42206User interfaces specially adapted for controlling a client device through a remote control device; Remote control devices therefor characterized by hardware details
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/462Content or additional data management, e.g. creating a master electronic program guide from data received from the Internet and a Head-end, controlling the complexity of a video stream by scaling the resolution or bit-rate based on the client capabilities
    • H04N21/4622Retrieving content or additional data from different sources, e.g. from a broadcast channel and the Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/478Supplemental services, e.g. displaying phone caller identification, shopping application
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/482End-user interface for program selection
    • H04N21/4828End-user interface for program selection for searching program descriptors
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04HBROADCAST COMMUNICATION
    • H04H60/00Arrangements for broadcast applications with a direct linking to broadcast information or broadcast space-time; Broadcast-related systems
    • H04H60/68Systems specially adapted for using specific information, e.g. geographical or meteorological information
    • H04H60/72Systems specially adapted for using specific information, e.g. geographical or meteorological information using electronic programme guides [EPG]

Definitions

  • EPG electronic programming guide
  • a user typically navigates through scheduled programming content entries displayed in an EPG by pressing buttons on the remote control device using a five-way navigational command system (left, right, up, down, and select). Once a user has highlighted a particular program on the EPG, the corresponding content may be shown on the electronic device by pressing the select button on the remote control device, after which the electronic device is tuned to the selected channel.
  • a five-way navigational command system left, right, up, down, and select
  • One embodiment is directed to a method of presenting search results on an electronic device, wherein the electronic device includes a tuner configured to tune the electronic device to receive scheduled programming.
  • the method comprises receiving voice input comprising a search query; searching, with at least one processor, at least one data source based, at least in part, on the search query; and presenting results of the search using a time-based axis and a time-independent axis.
  • searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
  • the results include scheduled programming content identified in the at least one scheduled programming content data source and bundled service content identified in the at least one bundled service content data source.
  • the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content.
  • time-based axis distinguishes first content that is currently available for playback from second content that is currently unavailable for playback.
  • results presented along the time-based axis include only scheduled programming content.
  • Another embodiment is directed to an electronic device, comprising: a tuner configured to tune the electronic device to receive scheduled programming; and at least one processor programmed to: receive voice input comprising a search query; search at least one data source based, at least in part, on the search query; and present results of the search using a time-based axis and a time-independent axis.
  • the time-independent axis orders the results based on relevance to one or more terms in the search query.
  • the electronic device further comprises a voice input interface configured to receive the search query as voice input.
  • searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
  • the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content.
  • the time-based axis distinguishes first content that is currently available for playback by the electronic device from second content that is currently unavailable for playback by the electronic device.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: receiving voice input comprising a search query;
  • searching at least one data source based, at least in part, on the search query ; and presenting results of the search using a time-based axis and a time-independent axis.
  • the time-independent axis orders the results based on relevance to one or more terms in the search query.
  • Another embodiment is directed to a method of presenting search results on a visual user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising: receiving a search query; searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and presenting results of the search on the user interface, wherein the results of the search combine at least one first result correspondin to scheduled programming content and at least one second result corresponding to bundled service content.
  • the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
  • the at least one bundled service data source comprises data source for a web-based service.
  • the at least one bundled service data source comprises local data source associated with the electronic device.
  • the at least one bundled service data source comprises plurality of data sources connected via a network.
  • the at least one bundled service data source comprises data source of on-demand content.
  • presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer perform a method of presenting search results on a visual user interface for a digitally- tunable electronic device configured to display scheduled programming content.
  • the method comprises receiving a search query; searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and presenting results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
  • presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
  • a digitally-tunable electronic device configured to display scheduled programming content.
  • the electronic device comprises at least one processor programmed to: receive a search query; search based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and present results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
  • the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
  • the at least one bundled service data source comprises a data source for a web-based service.
  • the at least one bundled service data source comprises a local data source associated with the electronic device.
  • the at least one bundled service data source comprises a plurality of data sources connected via a network.
  • the at least one bundled service data source comprises a data source of on-demand content.
  • Another embodiment is directed to a method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising: displaying, on the user interface, a plurality of content- independent tags, the plurality of content-independent tags comprising a first content- independent tag identifying a first content unit returned from a search; and in response to receiving first input corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input comprises first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
  • the method further comprises receiving second input, wherein a second content-independent tag of the plurality of content-independent tags is configured to be responsive to the second input; and performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input.
  • the second input is second voice input and the second content-independent tag is configured to be responsive to the second voice input.
  • receiving the first input and receiving the second input comprises receiving the second input before receiving the first input.
  • the method further comprises selecting the first content unit based on the first input.
  • performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled
  • the first content unit corresponds to a bundled service content unit comprising multiple search results
  • the performing at least one action related to the first content unit comprises displaying the multiple search results.
  • the method further comprises associating a plurality of voice commands with the user interface, the plurality of voice commands includes a first voice command and a second voice command, the first voice command invokes a first action and the second voice command invokes a second action, and the first voice command and the second voice command are recognized by the electronic device irrespective of a particular screen displayed on the user interface.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality instructions that, when executed by at least one computer, perform a method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content.
  • the method comprises displaying, on the user interface, a plurality of content-independent tags, the plurality of content-independent tags comprising a first content-independent tag identifying a first content unit returned from a search; and in response to receiving first input
  • the first content-independent tag corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
  • Another embodiment is directed to a digitally-tunable electronic device configured to display scheduled programming content.
  • the electronic device comprises at least one processor programmed to: display, on a user interface, a plurality of content- independent tags, the plurality of content-independent tags comprising a first content- independent tag identifying a first content unit returned from a search; and in response to receiving first input corresponding to the first content-independent tag, perform at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
  • the at least one processor is further programmed to: receive second input, wherein a second content-independent tag of the plurality of content- independent tags is configured to be responsive to the second input; and wherein performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input.
  • the performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled programming for the first content unit.
  • the first content unit corresponds to a bundled service content unit comprising multiple search results
  • the performing at least one action related to the first content unit comprises displaying the multiple search results
  • Another embodiment is directed to a digitally-tunable electronic device comprising: at least one processor programmed to: provide a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally- tunable electronic device; receive voice input comprising the search query; and initiate a search for content corresponding to the search query in response to receiving the search query.
  • initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
  • the at least one processor is further programmed to: display, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
  • the at least one processor is further programmed to: associate a plurality of voice commands with the user interface, the plurality of voice commands includes a first voice command and a second voice command, the first voice command invokes a first action and the second voice command invokes a second action.
  • the first action and/or the second action corresponds to initiating a search for content.
  • Another embodiment is directed to a computer-implemented method comprising: providing a user interface that enables an electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the electronic device; receiving voice input comprising the search query; and initiating a search for content corresponding to the search query in response to receiving the search query.
  • initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
  • the method further comprises displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
  • the method further comprises associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
  • the first action and/or the second action corresponds to initiating a search for content.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: providing a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally-tunable electronic device; receiving voice input comprising the search query; and initiating a search for content corresponding to the search query in response to receiving the search query.
  • initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
  • the method further comprises displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
  • the method further comprises associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
  • the first action and/or the second action corresponds to initiating a search for content.
  • Another embodiment is directed to a digitally-tunable electronic device comprising: at least one processor programmed to: receive first input; entering a listening mode in response to recognizing the first input; determine whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exit the listening mode in response to determining that second input is not received during the particular amount of time.
  • the second input is an explicit command to exit listening mode; and the at least one processor is further programmed to exit the listening mode in response to receiving the second input.
  • the at least one processor is further programmed to monitor for the first input in response to exiting the listening mode.
  • the electronic device further comprises at least one display, and the at least one processor is further programmed to display, on the least one display, an indication of whether the electronic device is currently in listening mode.
  • the indication comprises one or more selection tags and/or icons.
  • the at least one processor is further programmed to determine whether the electronic device is in listening mode; and display the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode.
  • the at least one processor is further programmed to hide the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
  • Another embodiment is directed to a computer-implemented method comprising: receiving first input; causing an electronic device to enter a listening mode in response to recognizing the first input; determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exiting the listening mode in response to determining that second input is not received during the particular amount of time.
  • the second input is an explicit command to exit listening mode; and the method further comprises exiting the listening mode in response to receiving the second input.
  • the electronic device comprises at least one display
  • the method further comprises displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
  • the indication comprises one or more selection tags and/or icons.
  • the method further comprises determining whether the electronic device is in listening mode; displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: receiving first input; causing an electronic device to enter a listening mode in response to recognizing the first input; determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exiting the listening mode in response to determining that second input is not received during the particular amount of time.
  • the electronic device comprises at least one display, and the method further comprises displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
  • the indication comprises one or more selection tags and/or icons.
  • the method further comprises determining whether the electronic device is in listening mode; displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
  • Another embodiment is directed to a method of searching for content to display and/or playback on a digitally-tunable electronic device configured to display scheduled programming content.
  • the method comprises receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device; determining, based on the search query, an action the user wants to perform; determining one or more data sources to search based, at least in part, on the action the user wants to perform; and searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
  • determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
  • determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
  • the method further comprises storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
  • searching the one or more data sources comprises searching all available data sources.
  • the method further comprises receiving search results from the one or more data sources; and presenting the search results on a user interface.
  • Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method of searching for content to display and/or playback on a digitally- tunable electronic device configured to display scheduled programming content.
  • the method comprises receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device; determining, based on the search query, an action the user wants to perform; determining one or more data sources to search based, at least in part, on the action the user wants to perform; and searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
  • determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
  • determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
  • the method further comprises storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
  • Another embodiment is directed to a digitally-tunable electronic device configured to display scheduled programming content, the electronic device comprising at least one processor programmed to: receive voice input comprising a search query from a user, wherein the search query comprises a search for content to display and/or playback on the electronic device; determine, based on the search query, an action the user wants to perform; determine one or more data sources to search based, at least in part, on the action the user wants to perform; and search based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
  • the electronic device comprising at least one processor programmed to: receive voice input comprising a search query from a user, wherein the search query comprises a search for content to display and/or playback on the electronic device; determine, based on the search query, an action the user wants to perform; determine one or more data sources to search based, at least in part, on the action the user wants to perform; and search based, at least in part, on
  • determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
  • determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
  • the at least one processor is further programmed to store one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
  • searching the one or more data sources comprises searching all available data sources.
  • FIG. 1 is an exemplary system including components of some
  • FIG. 2 is a process for presenting results of a query for content on a user interface in accordance with some embodiments of the invention
  • FIG. 3 is a schematic of a portion of a user interface for displaying results of a query for content in accordance with some embodiments of the invention.
  • FIG. 4 is a schematic of a portion of a user interface for displaying blended results of a query for content in accordance with some embodiments of the invention.
  • FIG. 5 is a schematic of portion of a user interface displaying voice tags a user may interact with using input in accordance with some embodiments of the invention
  • FIG. 6 is an exemplary process for monitoring for input using a listening mode in accordance with some embodiments of the invention.
  • FIG. 7 is an exemplary computer system that may be used to implement some embodiments of the invention.
  • the inventors have recognized and appreciated that searching for content to view on an electronic device that includes a tuner for displaying scheduled
  • EPGs Electronic program guides
  • the grid format of an EPG display is designed to receive input from a remote control device using a five- way navigation command system to move a cursor to different entries in the grid.
  • the inventors have recognized that navigation using conventional EPG displays is cumbersome.
  • a user may be required to provide a series of navigation commands (e.g., "down-down-downright-select"), which does not result in an efficient or intuitive experience for the user to select a particular program.
  • a series of navigation commands e.g., "down-down-downright-select”
  • the display of conventional EPG search systems is a "visual-first" user interface in which the main focus of the interface is to display the scheduled programming content in visually consistent boxes.
  • the inventors have appreciated that other techniques for searching for content are possible that do not require a rigid visual presentation of scheduled programming content because the user is not restricted to navigating displayed entries using a five-way navigation command system.
  • some embodiments of the invention are directed to methods and apparatus for an improved visual search interface.
  • the visual user interface is designed to encourage
  • Some embodiments encourage the use of voice input as a primary input modality, with visual aspects of the user interface being complimentary to voice input.
  • any input that can be spoken can alternatively be provided via another type of input modality (e.g., text entry in a keyboard).
  • voice e.g., text entry in a keyboard
  • other modalities can be used to allow the user to provide the same types of inputs described in the illustrated embodiments below as being provided by voice.
  • System 100 includes electronic device 102, which may be a television, a computer, a portable electronic device such as a
  • Electronic device 102 may be configured to receive input from a user to search for and select content to view on the electronic device, wherein the input is received using one or more input modalities.
  • electronic device 102 includes a voice input interface 104, such as a microphone and one or more other input interfaces 106 configured to receive the input.
  • voice input interface 104 such as a microphone
  • Other input interface(s) 106 may include, but are not limited to, a keyboard interface (e.g., a
  • QWERTY keyboard a mouse interface, a remote control interface, or any other suitable interface that enables a user to search for content to view on electronic device 102.
  • Voice input received via voice input interface 104 may be sent to automatic speech recognition (ASR) engine 108 for speech recognition to recognize the content of the received voice input.
  • ASR engine 108 is illustrated in FIG. 1 as being included as a portion of electronic device 102, ASR engine 108 may alternatively be located remotely from electronic device 102 and may be in communication with electronic device 102 using any suitable communication mediums (e.g., one or more networks).
  • electronic device 102 may be associated with one or more local ASR engines 108 and one or more remotely located ASR engines, whereby at least some received voice input is processed by the one or more local ASR engine(s) and at least some received voice input is processed by the one or more remotely-located ASR engines.
  • Voice input sent to the one or more remote ASR engine(s) for recognition using one or more communication medium(s) (e.g., one or more networks) connecting the electronic device 102 and the ASR engine(s) may be returned to the electronic device 102 from the remotely-located ASR engine(s) using the same or different communication medium(s).
  • one or more communication medium(s) e.g., one or more networks
  • ASR engine 108 may be in communication with one or more voice command data sources 110 storing one or more voice commands.
  • ASR engine 108 may match received voice input to the one or more voice commands stored in one or more voice command data sources 110 and return one or more matching voice commands to one or more processors 114 programmed to execute at least one action corresponding to the one or more matching voice commands.
  • the output of ASR engine 108 may be analyzed using a natural language understanding (NLU) engine using any suitable NLU-based technique(s), examples of which are known in the art.
  • NLU natural language understanding
  • the NLU output may be used, at least in part, to associate a voice command with the received voice input.
  • electronic device 102 may be configured to interpret voice input as a voice command in response to activation of an input signal (e.g., the user pressing a key on a remote control or a keyboard) or in response to receiving voice input corresponding to a "wake-up" word that when received indicates that the voice input following the input signal should be interpreted as a voice command.
  • voice input may be interpreted as a command without command initiation via an input signal (e.g., a key press or a wake-up word), as embodiments of the invention are not limited in this respect.
  • the set of voice commands stored by the one or more voice command data sources 110 may include any suitable commands for searching for and selecting content to view on the electronic device 102.
  • the set of voice commands may include commands to control functions of the electronic device 102 itself, examples of which include commands frequently found on a conventional remote control device, such as "mute,” "power on,” and “volume.”
  • An illustrative set of electronic device commands for controlling an electronic device 102 is shown below in Table 1, but it should be appreciated that these commands are provided merely as examples and that embodiments described herein are not limited to using these specific commands.
  • the set of voice commands may also include tuner commands used to search for and/or select scheduled programming content to display on electronic device 102, such as "show ⁇ channel name>, "next channel,” and "show ⁇ program name>.”
  • tuner commands used to search for and/or select scheduled programming content to display on electronic device 102, such as "show ⁇ channel name>, "next channel,” and "show ⁇ program name>.”
  • An illustrative set of tuner commands is shown below in Table 2, but it should be appreciated that these commands are provided merely as examples and that embodiments described herein are not limited to using these specific commands.
  • the set of voice commands may also include application commands for interacting with one or more client applications on electronic device 102 that perform any of numerous functions, including providing access to numerous services.
  • a user may use an application command to initiate a voice call session using a client application that performs calling (such as SkypeTM), search for weather information, or perform any of numerous other actions.
  • An illustrative set of application commands is shown below in Table 3.
  • Table 3 Illustrative Application Commands
  • the illustrative application commands described in Table 3 include commands to control exemplary client applications including TwitterTM, YouTubeTM, and SkypeTM.
  • embodiments of the invention may include voice commands for any other client application and are not limited in this respect.
  • voice commands for client applications including, but not limited to, Facebook®, GmailTM, NetFlix®, HuluTM, and MySpaceTM may also be included.
  • Some client applications, including many of those listed above, may be web-based services or applications that can be accessed either via a specialized application on the electronic device or via a web browser on the device and embodiments of the invention are not limited in the way in which a client application is accessed.
  • the set of voice commands may also include action commands, each for performing an action related to one or more items in a search results list.
  • action commands each for performing an action related to one or more items in a search results list.
  • An illustrative list of action commands is described below in Table 4.
  • the set of voice commands may also include EPG commands for interacting with an EPG display.
  • EPG commands for interacting with an EPG display.
  • the set of EPG commands may provide additional voice functionality to enhance the user's browsing experience.
  • An illustrative set of EPG commands is described below in Table 5.
  • electronic device 102 also includes a tuner 112 configured to enable electronic device 102 to tune to a particular channel to display scheduled programming content on electronic device 102.
  • tuner 112 may enable the electronic device to receive scheduled programming content from one or more scheduled programming content sources 124 connected to electronic device 102 via network 120.
  • Scheduled programming content sources 124 may source media content scheduled for these times including, but not limited to, stored content (e.g., television programs, and movies), and live media content.
  • Scheduled programming content sources 124 that store and/or transmit scheduled programming content to an electronic device (e.g., over a communication medium such as a cable or satellite network) is well-known.
  • Some embodiments may include a tuner configured to receive scheduled programming content from multiple content sources and/or may include multiple tuners, each of which is configured to receive scheduled programming content from a different content source.
  • Electronic device 102 may also include network interface 118 configured to communicate with one or more storage and/or processing devices via one or more network 120. As discussed above, network interface 118 may be used to transmit voice input to one or more remote ASR engines and receive speech recognition results from the one or more ASR engines. Electronic device 102 can employ a single network interface 118, or more than one network interface 118 to enable the electronic device 102 to connect to more than one type of network (e.g., to perform different functions).
  • An exemplary electronic device 102 is a television system.
  • Television systems often include a tuner to enable the television system to display scheduled programming content received by a set top box that is connected to the television.
  • a set top box is an example of a tuner 112, which receives scheduled programming content via a network interface 118 and decodes the received content to enable the electronic device to display the content on a display 122.
  • Scheduled programming content or any other content received via network 120 may be displayed on the display 122 connected to electronic device 102.
  • display 122 may be integrated as part of electronic device 102 and in other embodiments, display 122 may be connected to electronic device 102 via one or more wired or wireless connections. Display 122 can also be a single display or multiple displays and embodiments of the invention are not limited in this respect.
  • electronic device 102 may also be associated with a local content data source 116, which may be configured to store one or more television programs, one or more movies, or any other content suitable for playback on the electronic device.
  • local content data source 116 may be a digital video recorder (DVR) configured to store recorded scheduled programming content.
  • DVR digital video recorder
  • Local content data source 116 may be configured to store content related to one or more local client applications associated with electronic device 102 including, but not limited to, one or more applications that store and/or manage picture files, music files, and/or video files.
  • content when searching for content to view and/or playback on electronic device 102, content may be retrieved from local content data source 116 and/or one or more data sources (e.g., scheduled content data source 124, bundled service content data source 124) remotely located from electronic device 102 and connected to electronic device 102 by network 120.
  • the particular source(s) that are searched may depend, at least in part, on the search query.
  • the search results When results are returned from multiple sources, the search results may be ordered based, at least in part, on the source of the content corresponding to the search results.
  • a user may be able to restrict a search for content based on the source of the content and/or the type of content desired to be included in the search results, as described in more detail below.
  • Some embodiments are directed to a voice-controllable visual search interface displayed on display 122 and configured to allow the user to search for content to view and/or playback using the electronic device 102.
  • the user may be able to interact with the search interface to search for scheduled programming content, as described above.
  • Scheduled programming content includes, but is not limited to, broadcast content such as television and radio programs.
  • some embodiments may enable a user to search for bundled service content stored, for example, by one or more network-connected bundled service data sources 124 and/or local content data source 116.
  • bundled service content may be available for viewing/playback at any time.
  • Bundled service content may include, but is not limited to, content from a web-based service (e.g., Netflix®), content from a local or remote data source storing media content (e.g., DVR content, on-demand content from a service provider, etc.), and content associated with one or more client applications (e.g., iTunes®) associated with electronic device 102.
  • searching for content in accordance with some embodiments may allow a user to search for scheduled programming content, bundled service content, or both scheduled programming content and bundled service content in a single search.
  • FIG. 2 An exemplary method for processing a search query in accordance with some embodiments of the invention is illustrated in FIG. 2.
  • electronic device 102 receives a search query to search for content to view and/or playback on electronic device 102.
  • the search query may include one or more voice commands, as discussed above.
  • the voice input may be processed by one or more local and/or remote ASR engines for recognition.
  • a determination of whether to process voice input via a local and/or remote ASR engine may be made based, at least in part, on the type of voice input.
  • voice input corresponding to fixed- form device commands may be processed by a local ASR engine
  • commands including free-form input e.g., 'Search,' 'Find,' etc.
  • clouds remotely-located ASR engines.
  • fixed-form commands may be associated with a limited grammar, they can be easier to recognize and a local ASR engine may suffice for voice recognition and be faster than sending the input to a remote ASR engine.
  • free-form input may be more difficult to recognize, so sending the input to more robust ASR engine(s) may be required for acceptable recognition performance.
  • language models associated with remotely ASR engines can be updated regularly to reflect changes in popular culture, names of movies, actors, books, etc. that might otherwise be difficult for a local ASR engine to recognize.
  • at least a first portion of the voice input may be processed by a local ASR engine and at least a second portion of the voice input may be processed by a remotely-located ASR engine, and any combination of local and remote ASR engines may be used to recognize received voice input.
  • the process then proceeds to act 220, where one or more data sources are searched based, at least in part, on the recognized voice input corresponding to the search query.
  • the one or more data sources may be searched in any order, and any particular data source(s) (including less than all of the available data sources) may be searched.
  • the results that are returned from the search may be determined in any suitable way.
  • a determination of which data sources to search may be made based, at least in part, on a particular action a user wants to perform. For example, if the user says "listen Bob Seger," the system may determine that the user wants to listen to music, so only a local music data source may be searched and not a data source associated with an EPG.
  • a data source associated with an EPG and one or more data sources associated with recorded television content may be searched, but a local music data source may not be searched.
  • some embodiments of the invention may return relevant search results in a more efficient manner by reducing an amount of time to return the search results.
  • not all embodiments are limited to searching only a subset of available sources as all available sources may be searched for some searches.
  • an order in which the data sources are searched and/or an order in which the search results are presented may be determined based, at least in part, on one or more rules associating a particular command (or determined user intent) with a particular order for searching and/or presenting search results. For example, if a user says "watch Modern Family," recognition of the voice command "watch" (or an intended action if a specific command is not employed, as discussed below) may cause the electronic device to search and return results from data source(s) associated with an EPG first, search and return results from data source(s) associated with local content second, and search and return results from data source(s) associated with client applications (e.g., YouTubeTM) third. It should be appreciated that this particular order for searching and returning search results associated with a "watch" command is described merely for illustrative purposes and any other desired order may alternatively be used.
  • some embodiments may not include a fixed command grammar, but rather may determine a user' s intent from the received input in some other way, such as determining a user's intent by monitoring for particular words in free-form input, and performing an action in response to determining the user's intent. For example, if a user uses the "play" command or uses the term "play" in a free-form input, the system may recognize that the play command or use of the word "play” may suggest an intent by the user to access always accessible content, whether locally stored or available on-demand. One or more priority rules may be used to determine what media content to play.
  • the EPG may first be determined whether the EPG has any media content currently available to be played (e.g., broadcast content, on demand content, etc.) that matches the voice input. If it is determined that there is not currently available content from the EPG, it may next be determined whether any recorded media content.
  • media content currently available to be played e.g., broadcast content, on demand content, etc.
  • the electronic device may automatically playback a selected one of the matching content (e.g., the oldest unwatched (or partially watched) recorded content be selected and played back on the electronic device).
  • a selected one of the matching content e.g., the oldest unwatched (or partially watched) recorded content be selected and played back on the electronic device.
  • the process proceeds to act 222, where one or more results are displayed on a user interface associated with electronic device 102 in response to searching the one or more data sources for content that matches the search query.
  • the search result(s) may be displayed in any suitable way including using a voice-enabled search results interface, as described in more detail below.
  • displaying the results can include directly presenting a piece of content that matches user input.
  • FIG. 3 illustrates a portion of a visual search interface in accordance with some embodiments of the invention.
  • FIG. 3 shows an illustrative result screen 300 for displaying search results returned from a user-initiated search for content to view and/or play on an electronic device.
  • result screen 300 is configured to display search results along at least two axes.
  • one axis is in the foreground of the search interface and one or more other axes are in the background of the search interface, but can be brought to the foreground of the search interface in response to a command (e.g., a voice command).
  • a command e.g., a voice command
  • first axis for displaying search results may be a time-independent axis 310.
  • Search results may be ordered along time-independent axis 310 using any suitable metric not related to a time when the content corresponding to the search results is available for viewing and/or playback.
  • the search results may be ordered along the time-independent axis based, at least in part, on relevance to the search query that led to the result.
  • content displayed along the time-independent axis may be limited to content that is available at the current time, while in other embodiments it may include content not currently available.
  • relevance may be determined in any suitable way using any suitable algorithm (e.g., including giving higher relevance to content currently available for viewing), as embodiments of the invention are not limited in this respect. It should be appreciated that relevance to a search query is only one way to order results along the time-independent axis, and that any other suitable metric may be used instead of or in addition to relevance to a search query.
  • the search results may be ordered based, at least in part, on type of content, source of content, or any other factor or combination of factors.
  • search results may also be organized along a second axis that is time-based and is configured to display search results available for viewing at different times.
  • the foreground of the result screen 300 may display search results for content that is currently available, whereas content available only at a future time may be shown in the background 320 as illustrated in FIG. 3.
  • search results displayed on only two axes it should be appreciated that search results may be displayed on any number of axes, as embodiments of the invention are not limited in this respect.
  • search results for bundled service content that are available at the current time, but may require a subscription or some other fee to be paid prior to viewing the content, may be displayed along a separate axis (not shown) to indicate that the content may be viewed only after an applicable payment has been paid.
  • Result screen 300 may also include description section 330 that includes information regarding a currently selected search result.
  • Description section 330 may also include other information that enables the user to make an informed selection from the search results displayed on result screen 300.
  • description section 330 may include a description of the content associated with the currently selected search result, and this description may change based on a selection of a different search result displayed on result screen 300.
  • less than all of the search results returned from the searching process may be displayed on result screen 300 to minimize crowding of the display.
  • a user may interact with one or more voice tags, described in more detail below, to view additional search results.
  • Result screen 300 may also include discovery field 340 that enables a user to perform an additional search based on one or more of the search results displayed on result screen 300.
  • discovery field 340 when selected by a user, may initiate a search for content similar to content associated with the currently selected search result. Similarity may be determined in any suitable way using any suitable algorithm, as embodiments of the invention are not limited in this respect. For example, similarity may be determined based, at least in part, on a content source content was retrieved from, one or more content tags associated with a piece of content (e.g., parental ratings for movies), type of content, or any combination of these or other factors.
  • result screen 300 may include a toolbar 350 including information to facilitate a user interaction with search screen 300.
  • toolbar 350 may include one or more icons, text, or other indicators explaining screen- specific or global inputs (e.g., voice commands) that a user may use to interact with result screen 300.
  • Search results may comprise blended search results that include different types of content.
  • search results may include results corresponding to scheduled programming content and bundled service content.
  • FIG. 4 illustrates an illustrative result screen 400 in which the search results displayed on the result screen 400 include blended search results, i.e., at least two of the search results are associated with different types of content.
  • the blended search results may include first results returned from one or more scheduled programming content data sources and second results returned from one or more bundled service content data sources.
  • the search results displayed on result screen 400 include search results 410, 412, and 414, which may be results returned from one or more scheduled programming content data sources and search results 420 and 422, which may be results returned form one or more bundled service content data sources.
  • search results returned from scheduled programming content data sources and search results returned from bundled service content data sources may be mixed together and ordered along a time-independent axis using any suitable metric or combination of metrics including, but not limited to, relevancy to the search query, type of content, and source of content.
  • result screen 400 may include a description portion 430 that includes information about one or more of the search results displayed on result screen 400.
  • search results displayed on result screen 400 may be ordered along at least two axes including a time-independent axis and a time-based axis.
  • the search results displayed in a background portion of the search interface may include only scheduled programming content and not bundled service content. For example, because some bundled service content may always be available, in some embodiments search results corresponding to such bundled service content may always be shown in a foreground portion of the search interface to indicate that it is available for viewing and/or playing immediately. It should be appreciated, however, that this is merely one illustrative way to organize content using two axes and other organizations of content are also possible.
  • some bundled service content may only be accessed immediately upon payment of a fee.
  • a song from an online music store such as iTunes® may be accessed immediately, but only upon purchase of the song.
  • search results that correspond to content that requires payment before becoming available for viewing and/or playback on an electronic device may not be mixed with other search results, but may appear in a different portion of the search results interface.
  • a search interface in accordance with some embodiments of the invention may include one or more selection tags that facilitate interactions between a user and one or more components of the search interface.
  • FIG. 5 illustrates an exemplary result screen 500 including a plurality of selection tags.
  • one or more of the search results may be associated with a selection tag 510, which indicates input a user may provide (e.g., via spoken input, a keyboard or keypad, etc.) to select a particular search result.
  • the user may say “select 1.”
  • the user may say “select 4.”
  • Result screen 500 may also include other types of selection tags such as one or more icons 520 that enable a user to use a selection command to perform one or more actions related to the content with which the icon is associated.
  • the icons 520 may include, but are not limited to, icons that indicate to the user that they may view particular content, record particular content, or perform some other action associated with the particular content associated with the icon.
  • the user may combine an action with identification of a selection tag to perform the action on the selected content. For example, a user may view the content associated with the search result in the lower left corner of FIG. 5 by saying "view 1."
  • voice interactions with the visual search interface may be categorized into different types of voice commands depending on when the voice command is active.
  • voice commands may be categorized as screen agnostic commands, which are independent of the content displayed on any particular screen and are always active irrespective of what particular screen is currently being presented on display.
  • screen agnostic commands include, but art not limited to, commands such as volume control, channel control, screen brightness, and menu accessing commands.
  • the set of screen agnostic commands overlap substantially with
  • Screen agnostic commands may be associated with one or more labels for selective tags displayed on the search interface to alert the user to the presence of these commands, or in some embodiments the user is presumed to know that these always available commands and can be invoked so that no selection tags or icons are displayed.
  • icons for one or more of the global commands may be included in a toolbar displayed on the search interface screen and/or a list of global commands may be accessed in response to a trigger event such as asking for help, pressing a hotkey on an input device, or in some other way.
  • Some voice interactions may be associated with screen-dependent actions that enable a user to perform actions only within a particular screen of the user interface.
  • selection tags 510 may allow a user to select a particular search result as described above, and the response to selecting the particular search result, a description of the content associated with the selected search result may be updated on the search interface to provide the user with information that may help the user decide whether to perform an action (e.g., watch, purchase, record) on the content associated with the selected search result.
  • an action e.g., watch, purchase, record
  • Other screen-dependent actions may be associated with one or more navigation functions of the search interface that enable a user to navigate within a particular result screen, and/or to update the content of a result screen.
  • the search interface may display a voice-enabled label that allows a user to view additional search results not currently displayed on the search interface, for example, by speaking "next results" or some other command phrase or free-form input to update the result screen to include the additional results.
  • the search interface may be further configured with user interaction functionality associated with screen-dependent actions to allow a user to switch between search results presented at different levels of a multi-axis display.
  • user interaction functionality associated with screen-dependent actions to allow a user to switch between search results presented at different levels of a multi-axis display.
  • the user may use voice (or other forms of) input to switch which results along the time-based axis are currently in focus and can be selected.
  • voice or other forms of
  • a user may say "later" (or some other word or phrase) to select content associated with search results displayed in the background of the user interface along the time-based axis of the displayed search results.
  • This may be useful, for example, to allow the user to take an action (e.g., record) relating to scheduled programming content that is available at some later time.
  • an action e.g., record
  • the user may say "later" (or some other word or phrase), which may switch the search results displayed in the foreground and the search results displayed in the background to enable the user to select an item for which content is available only in the future.
  • Another category of user interactions may be represented by graphical icons 520 that enable a user to perform a particular action such as recording, viewing, or playback of content associated with the search result.
  • graphical icons may also correspond to additional value-added actions that enable a user to perform other "secondary" functions such as posting to Facebook®, performing an action on TwitterTM, or interacting with another client application service or website.
  • the electronic device may receive input which causes the electronic device to enter a "listening mode" during which the electronic device is expecting further input from the user to perform an action.
  • the electronic device may not interpret received input, by default, as representing a user' s intent to perform a particular action, and may be monitoring for input to trigger entry into listening mode.
  • An illustrative process for an electronic device that includes a listening mode in accordance with some embodiments of the invention is illustrated in FIG. 6. In act 610, it is determined whether a first input has been received that will cause the electronic device to enter a listening mode.
  • the first input may correspond to a voice command or some other voice input demonstrating a user's intent to perform an action, as described above.
  • the first input may be any other suitable input to cause the electronic device to enter listening mode including, but not limited to, a press of a particular key on a keyboard or keypad.
  • the process proceeds to act 620 where one or more actions corresponding to the first and/or second input are performed.
  • the process may optionally return to act 612 to enter listening mode if additional input is expected.
  • the second input may be "What's on HBO comedy listings tonight?" and the action performed in act 620 may be opening the HBO comedy listings for the current day.
  • the process may proceed to act 612 to enter listening mode in anticipation of the additional input (e.g., next, previous, exit, etc.). However, if no additional input is expected after performing the one or more actions in act 620, the process may end.
  • act 614 If it is determined in act 614 that a second input has not been received, the process proceeds to act 616 where it is determined whether a particular amount of time (e.g., 5 seconds) has elapsed since entering listening mode. If it is determined in act 616 that the particular amount of time has not yet elapsed, the process continues to monitor for a second input. Otherwise, if it is determined that the particular amount of time has elapsed, the process proceeds to act 618, where the electronic device is caused to exit listening mode, and the process returns to act 610 to monitor for a first input. In the above-described embodiment, the electronic device exits listening mode after a particular amount of time has passed after entering listening mode.
  • a particular amount of time e.g., 5 seconds
  • listening mode may alternatively be exited in other ways.
  • some embodiments may enable a user to explicitly exit listening mode by providing input instructing the electronic device to do so.
  • This explicit input may take any suitable form including, but not limited to, a voice command (or determined user intent), or a key press on a keyboard or keypad.
  • an electronic device when an electronic device is caused to enter into a listening mode, this may be noted in some way on the user interface displayed by the electronic device. For example, upon entering the listening mode, one or more selection tags and/or icons may appear on the user interface to inform a user that listening mode has been entered. It should be appreciated that changes to a user interface to reflect the mode status of the electronic device may be performed in any way and the example of displaying or hiding selection tags and/or icons is only one way to indicate the status and other ways are also possible.
  • bundled service content search results may be represented hierarchically. For example, if a particular bundled service includes one or more pieces of content that match a search query, they may initially be represented by including an identifier for the bundle service in the search results, but not the
  • an application associated with a bundled service content search result may be illustrated as an image or icon for the application, which is recognized by a user as representing that application.
  • the bundled service search result may correspond to a movie streaming service such as, Netflix® (e.g., illustrated by a Netflix® icon), and in response to selecting the movie streaming service by, for example, saying "select 2," which selects the second search result (e.g., the Netflix® search result), search results within the Netflix® application may be displayed.
  • a user may search for content corresponding to
  • search results returned in response to this search may include scheduled programming content for New York Yankees baseball games, bundled service content from a movie streaming service for movies related to the New York Yankees, music content related to the band "Damn Yankees," and electronic book content for the electronic book "A Connecticut Yankee in King Arthur's Court” by Mark Twain.
  • the search results may initially show an image for a movie streaming service which includes multiple movies related to the search term "Yankees.” Upon selecting the image for the movie streaming service using, for example, a selection voice tag associated with the image, one or more movies available from that streaming service and corresponding to the search term "Yankees" may be displayed in the search results.
  • Hierarchically- presented search results may have any number of hierarchical levels, as embodiments of the invention are not limited in this respect.
  • FIG. 1 Another illustrative example of content that can be represented hierarchically is songs in a music library.
  • a user may want to play a song by Metallica, but may not remember the name of the song.
  • the user may initiate a search for "Metallica,” and the search results that are returned may include multiple albums for Metallica stored locally on a local data source associated with electronic device and/or one or more network connected music data sources (e.g., databases) that include music from Metallica.
  • the search results may be presented as a series of images for album covers for the different Metallica albums and the user may be prompted to select one of the albums from the search results (e.g., by using selection voice tags associated with each album image).
  • the selected album may be expanded to reveal search results for one or more of the songs associated with album, and the user may select one of the songs on the album for playback.
  • search results may be presented in a hierarchical manner such that sequential selections of search results may allow the user to navigate through the search results to identify the content the user wants to playback on the electronic device.
  • voice tags may be employed that instruct the user how to interact with different components of the user interface.
  • Content-neutral voice tags are labels that are independent of the content that the voice tag identifies.
  • An example of a content-neutral voice tag is the selection voice tags (1, 2, 3, 4) illustrated in FIG. 5, which allow the user to select one of the search results without having to speak the name of the content identifying the search result.
  • content-neutral voice tags do not identify the content itself with which they are associated, but merely identify a slot or position of the search result to which the content is assigned.
  • the user is provided with a set of voice tags that facilitates a consistent user experience when interacting with the visual search interface.
  • content-neutral voice tags allows the user to use less speech to invoke commands that would be required if the voice tags were created based, at least in part, on the content the tags identified.
  • Use of content-neutral voice tags also allow the use of a fixed grammar (e.g., numbers, letters) that can result in more accurate ASR for content that may have unique names, etc.
  • the user may be able to provide input (e.g., via voice) that results in the activation of more than one type of command or interaction. For example, a user may say, "watch 1.” In response, two interactions take place: (1) the first item in the search results may be selected and (2) the selected item may be played. In this way a single input may cause multiple commands or interactions to be executed without the user having to issue separate commands.
  • input e.g., via voice
  • a user may say, "watch 1.”
  • two interactions take place: (1) the first item in the search results may be selected and (2) the selected item may be played.
  • a single input may cause multiple commands or interactions to be executed without the user having to issue separate commands.
  • Such a system enables the user to use natural input (e.g., speech) to control the electronic device.
  • Searches may be performed for any type of content has described above, including searches for video clips on YouTubeTM, HuluTM, or any other video-sharing website or web service. Additionally, the user may search for music content, as described above, in one music libraries stored locally or accessible via a network (e.g., the Internet). The user may have the ability to restrict a search to a particular type of content such as music, videos, or scheduled programming content, such that the search results only include a particular type of content for which the user desires to search. To this end, the search interface may be associated with advanced searching capabilities that may be configured using voice input and/or other input modalities.
  • a piece of content displayed in the search results may include multiple tags (e.g., voice tags), with one tag identifying the content, and another tag indicating an action to perform with the content (e.g., watch, record, or purchase the content).
  • tags e.g., voice tags
  • Another tag indicating an action to perform with the content (e.g., watch, record, or purchase the content).
  • the number of tags associated with search results for a search interface is not a limitation of embodiments of the invention.
  • FIG. 7 An illustrative implementation of a computer system 700 that may be used in connection with any of the embodiments of the invention described herein is shown in FIG. 7.
  • the computer system 700 may include one or more processors 710 and one or more computer-readable non-transitory storage media (e.g., memory 720 and one or more non-volatile storage media 730).
  • the processor 710 may control writing data to and reading data from the memory 720 and the non-volatile storage device 730 in any suitable manner, as the aspects of the present invention described herein are not limited in this respect.
  • the processor 710 may execute one or more instructions stored in one or more computer-readable storage media (e.g., the memory 720), which may serve as non-transitory computer-readable storage media storing instructions for execution by the processor 710.
  • computer-readable storage media e.g., the memory 720
  • the processor 710 may execute one or more instructions stored in one or more computer-readable storage media (e.g., the memory 720), which may serve as non-transitory computer-readable storage media storing instructions for execution by the processor 710.
  • the above-described embodiments of the present invention can be implemented in any of numerous ways.
  • the embodiments may be implemented using hardware, software or a combination thereof.
  • the software code can be executed on any suitable processor or collection of processors, whether provided in a single computer or distributed among multiple computers.
  • any component or collection of components that perform the functions described above can be generically considered as one or more controllers that control the above-discussed functions.
  • the one or more controllers can be implemented in numerous ways, such as with dedicated hardware, or with general purpose hardware (e.g., one or more processors) that is programmed using microcode or software to perform the functions recited above.
  • one implementation of the embodiments of the present invention comprises at least one non-transitory computer- readable storage medium (e.g., a computer memory, a floppy disk, a compact disk, a tape, etc.) encoded with a computer program (i.e., a plurality of instructions), which, when executed on a processor, performs the above-discussed functions of the
  • the computer-readable storage medium can be transportable such that the program stored thereon can be loaded onto any computer resource to implement the aspects of the present invention discussed herein.
  • the reference to a computer program which, when executed, performs the above-discussed functions is not limited to an application program running on a host computer. Rather, the term computer program is used herein in a generic sense to reference any type of computer code (e.g., software or microcode) that can be employed to program a processor to implement the above-discussed aspects of the present invention.
  • embodiments of the invention may be implemented as one or more methods, of which an example has been provided.
  • the acts performed as part of the method(s) may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.

Abstract

Methods and apparatus (100) for providing a search interface for an electronic device including a tuner (112) configured to tune the electronic device to receive scheduled programming content (124). Voice input (104)comprising a search query is received and one or more data sources (124) including information about media content are searched based, at least in part, on the search query. The results of the search are presented on a user interface (122) of the electronic device.

Description

USER INTERFACE FOR ENTERTAINMENT SYSTEMS
BACKGROUND
[0001] Some electronic devices that can be used to access entertainment content
(e.g., audio and/or video content) include, or are connected to, a tuner that enables the electronic device to receive and display scheduled programming content such as a television, radio, or other media program. To select a particular program for viewing on the electronic device, a user typically uses a remote control device to input the number of a channel associated with, e.g., broadcasting, the particular program. Some tunable electronic devices also allow a user to use a remote control device to search for and select content using an electronic programming guide (EPG), which is a user interface that displays scheduling information for current and future scheduled programming content in a time-based grid format. A user typically navigates through scheduled programming content entries displayed in an EPG by pressing buttons on the remote control device using a five-way navigational command system (left, right, up, down, and select). Once a user has highlighted a particular program on the EPG, the corresponding content may be shown on the electronic device by pressing the select button on the remote control device, after which the electronic device is tuned to the selected channel.
SUMMARY
[0002] One embodiment is directed to a method of presenting search results on an electronic device, wherein the electronic device includes a tuner configured to tune the electronic device to receive scheduled programming. The method comprises receiving voice input comprising a search query; searching, with at least one processor, at least one data source based, at least in part, on the search query; and presenting results of the search using a time-based axis and a time-independent axis.
[0003] In one aspect, the time-independent axis orders the results based on relevance to one or more terms in the search query. [0004] In another aspect, searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
[0005] In another aspect, the results include scheduled programming content identified in the at least one scheduled programming content data source and bundled service content identified in the at least one bundled service content data source.
[0006] In another aspect, the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content.
[0007] In another aspect, the time-based axis distinguishes first content that is currently available for playback from second content that is currently unavailable for playback.
[0008] In another aspect, the results presented along the time-based axis include only scheduled programming content.
[0009] Another embodiment is directed to an electronic device, comprising: a tuner configured to tune the electronic device to receive scheduled programming; and at least one processor programmed to: receive voice input comprising a search query; search at least one data source based, at least in part, on the search query; and present results of the search using a time-based axis and a time-independent axis.
[0010] In one aspect, the time-independent axis orders the results based on relevance to one or more terms in the search query.
[0011] In another aspect, the electronic device further comprises a voice input interface configured to receive the search query as voice input.
[0012] In another aspect, searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
[0013] In another aspect, the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content. [0014] In another aspect, the time-based axis distinguishes first content that is currently available for playback by the electronic device from second content that is currently unavailable for playback by the electronic device.
[0015] Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: receiving voice input comprising a search query;
searching at least one data source based, at least in part, on the search query; and presenting results of the search using a time-based axis and a time-independent axis.
[0016] In one aspect, the time-independent axis orders the results based on relevance to one or more terms in the search query.
[0017] Another embodiment is directed to a method of presenting search results on a visual user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising: receiving a search query; searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and presenting results of the search on the user interface, wherein the results of the search combine at least one first result correspondin to scheduled programming content and at least one second result corresponding to bundled service content.
[0018] In one aspect, the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
[0019] In another aspect, the at least one bundled service data source comprises data source for a web-based service.
[0020] In another aspect, the at least one bundled service data source comprises local data source associated with the electronic device.
[0021] In another aspect, the at least one bundled service data source comprises plurality of data sources connected via a network.
[0022] In another aspect, the at least one bundled service data source comprises data source of on-demand content. [0023] In another aspect, presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
[0024] Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer perform a method of presenting search results on a visual user interface for a digitally- tunable electronic device configured to display scheduled programming content. The method comprises receiving a search query; searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and presenting results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
[0025] In one aspect, presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
[0026] Another embodiment is directed to A digitally-tunable electronic device configured to display scheduled programming content. The electronic device comprises at least one processor programmed to: receive a search query; search based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and present results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
[0027] In one aspect, the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
[0028] In another aspect, the at least one bundled service data source comprises a data source for a web-based service.
[0029] In another aspect, the at least one bundled service data source comprises a local data source associated with the electronic device. [0030] In another aspect, the at least one bundled service data source comprises a plurality of data sources connected via a network.
[0031] In another aspect, the at least one bundled service data source comprises a data source of on-demand content.
[0032] Another embodiment is directed to a method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising: displaying, on the user interface, a plurality of content- independent tags, the plurality of content-independent tags comprising a first content- independent tag identifying a first content unit returned from a search; and in response to receiving first input corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input comprises first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
[0033] In one aspect, the method further comprises receiving second input, wherein a second content-independent tag of the plurality of content-independent tags is configured to be responsive to the second input; and performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input.
[0034] In another aspect, the second input is second voice input and the second content-independent tag is configured to be responsive to the second voice input.
[0035] In another aspect, receiving the first input and receiving the second input comprises receiving the second input before receiving the first input.
[0036] In another aspect, the method further comprises selecting the first content unit based on the first input.
[0037] In another aspect, performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled
programming for the first content unit.
[0038] In another aspect, the first content unit corresponds to a bundled service content unit comprising multiple search results, and the performing at least one action related to the first content unit comprises displaying the multiple search results. [0039] In another aspect, the method further comprises associating a plurality of voice commands with the user interface, the plurality of voice commands includes a first voice command and a second voice command, the first voice command invokes a first action and the second voice command invokes a second action, and the first voice command and the second voice command are recognized by the electronic device irrespective of a particular screen displayed on the user interface.
[0040] Another embodiment is directed to a computer-readable storage medium encoded with a plurality instructions that, when executed by at least one computer, perform a method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content. The method comprises displaying, on the user interface, a plurality of content-independent tags, the plurality of content-independent tags comprising a first content-independent tag identifying a first content unit returned from a search; and in response to receiving first input
corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
[0041] Another embodiment is directed to a digitally-tunable electronic device configured to display scheduled programming content. The electronic device comprises at least one processor programmed to: display, on a user interface, a plurality of content- independent tags, the plurality of content-independent tags comprising a first content- independent tag identifying a first content unit returned from a search; and in response to receiving first input corresponding to the first content-independent tag, perform at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
[0042] In one aspect, the at least one processor is further programmed to: receive second input, wherein a second content-independent tag of the plurality of content- independent tags is configured to be responsive to the second input; and wherein performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input. [0043] In another aspect, the performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled programming for the first content unit.
[0044] In another aspect, the first content unit corresponds to a bundled service content unit comprising multiple search results, and the performing at least one action related to the first content unit comprises displaying the multiple search results.
[0045] Another embodiment is directed to a digitally-tunable electronic device comprising: at least one processor programmed to: provide a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally- tunable electronic device; receive voice input comprising the search query; and initiate a search for content corresponding to the search query in response to receiving the search query.
[0046] In one aspect, initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
[0047] In another aspect, the at least one processor is further programmed to: display, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
[0048] In another aspect, the at least one processor is further programmed to: associate a plurality of voice commands with the user interface, the plurality of voice commands includes a first voice command and a second voice command, the first voice command invokes a first action and the second voice command invokes a second action.
[0049] In another aspect, the first action and/or the second action corresponds to initiating a search for content.
[0050] Another embodiment is directed to a computer-implemented method comprising: providing a user interface that enables an electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the electronic device; receiving voice input comprising the search query; and initiating a search for content corresponding to the search query in response to receiving the search query.
[0051] In one aspect, initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
[0052] In another aspect, the method further comprises displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
[0053] In another aspect, the method further comprises associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
[0054] In another aspect, the first action and/or the second action corresponds to initiating a search for content.
[0055] Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: providing a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally-tunable electronic device; receiving voice input comprising the search query; and initiating a search for content corresponding to the search query in response to receiving the search query.
[0056] In one aspect, initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
[0057] In another aspect, the method further comprises displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content. [0058] In another aspect, the method further comprises associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
[0059] In another aspect, the first action and/or the second action corresponds to initiating a search for content.
[0060] Another embodiment is directed to a digitally-tunable electronic device comprising: at least one processor programmed to: receive first input; entering a listening mode in response to recognizing the first input; determine whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exit the listening mode in response to determining that second input is not received during the particular amount of time.
[0061] In one aspect, the second input is an explicit command to exit listening mode; and the at least one processor is further programmed to exit the listening mode in response to receiving the second input.
[0062] In another aspect, the at least one processor is further programmed to monitor for the first input in response to exiting the listening mode.
[0063] In another aspect, the electronic device further comprises at least one display, and the at least one processor is further programmed to display, on the least one display, an indication of whether the electronic device is currently in listening mode.
[0064] In another aspect, the indication comprises one or more selection tags and/or icons.
[0065] In another aspect, the at least one processor is further programmed to determine whether the electronic device is in listening mode; and display the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode.
[0066] In another aspect, the at least one processor is further programmed to hide the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode. [0067] Another embodiment is directed to a computer-implemented method comprising: receiving first input; causing an electronic device to enter a listening mode in response to recognizing the first input; determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exiting the listening mode in response to determining that second input is not received during the particular amount of time.
[0068] In one aspect, the second input is an explicit command to exit listening mode; and the method further comprises exiting the listening mode in response to receiving the second input.
[0069] In another aspect, the electronic device comprises at least one display, and the method further comprises displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
[0070] In another aspect, the indication comprises one or more selection tags and/or icons.
[0071] In another aspect, the method further comprises determining whether the electronic device is in listening mode; displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
[0072] Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising: receiving first input; causing an electronic device to enter a listening mode in response to recognizing the first input; determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and exiting the listening mode in response to determining that second input is not received during the particular amount of time. [0073] In one aspect, the electronic device comprises at least one display, and the method further comprises displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
[0074] In another aspect, the indication comprises one or more selection tags and/or icons.
[0075] In another aspect, the method further comprises determining whether the electronic device is in listening mode; displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
[0076] Another embodiment is directed to a method of searching for content to display and/or playback on a digitally-tunable electronic device configured to display scheduled programming content. The method comprises receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device; determining, based on the search query, an action the user wants to perform; determining one or more data sources to search based, at least in part, on the action the user wants to perform; and searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
[0077] In one aspect, determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
[0078] In another aspect, determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
[0079] In another aspect, the method further comprises storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
[0080] In another aspect, searching the one or more data sources comprises searching all available data sources.
[0081] In another aspect, the method further comprises receiving search results from the one or more data sources; and presenting the search results on a user interface.
[0082] Another embodiment is directed to a computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method of searching for content to display and/or playback on a digitally- tunable electronic device configured to display scheduled programming content. The method comprises receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device; determining, based on the search query, an action the user wants to perform; determining one or more data sources to search based, at least in part, on the action the user wants to perform; and searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
[0083] In one aspect, determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
[0084] In another aspect, determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
[0085] In another aspect, the method further comprises storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules. [0086] Another embodiment is directed to a digitally-tunable electronic device configured to display scheduled programming content, the electronic device comprising at least one processor programmed to: receive voice input comprising a search query from a user, wherein the search query comprises a search for content to display and/or playback on the electronic device; determine, based on the search query, an action the user wants to perform; determine one or more data sources to search based, at least in part, on the action the user wants to perform; and search based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
[0087] In one aspect, determining the action the user wants to perform comprises determining the user wants to listen to media content, and determining one or more data sources to search comprises determining the one or more data sources as a music data source.
[0088] In another aspect, determining the action the user wants to perform comprises determining the user wants to watch media content, and determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
[0089] In another aspect, the at least one processor is further programmed to store one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
[0090] In another aspect, searching the one or more data sources comprises searching all available data sources.
[0091] It should be appreciated that all combinations of the foregoing concepts and additional concepts discussed in greater detail below (provided that such concepts are not mutually inconsistent) are contemplated as being part of the inventive subject matter disclosed herein. BRIEF DESCRIPTION OF DRAWINGS
[0092] The accompanying drawings are not intended to be drawn to scale. In the drawings, each identical or nearly identical component that is illustrated in various figures is represented by a like numeral. For purposes of clarity, not every component may be labeled in every drawing. In the drawings:
[0093] FIG. 1 is an exemplary system including components of some
embodiments of the invention that relate to a search interface for entertainment systems;
[0094] FIG. 2 is a process for presenting results of a query for content on a user interface in accordance with some embodiments of the invention;
[0095] FIG. 3 is a schematic of a portion of a user interface for displaying results of a query for content in accordance with some embodiments of the invention;
[0096] FIG. 4 is a schematic of a portion of a user interface for displaying blended results of a query for content in accordance with some embodiments of the invention;
[0097] FIG. 5 is a schematic of portion of a user interface displaying voice tags a user may interact with using input in accordance with some embodiments of the invention;
[0098] FIG. 6 is an exemplary process for monitoring for input using a listening mode in accordance with some embodiments of the invention; and
[0099] FIG. 7 is an exemplary computer system that may be used to implement some embodiments of the invention.
DETAILED DESCRIPTION
[00100] The inventors have recognized and appreciated that searching for content to view on an electronic device that includes a tuner for displaying scheduled
programming content may be improved. Electronic program guides (EPGs) of the type discussed above are frequently used to search for and select scheduled programming content to view on a tunable electronic device, and are configured using a grid-like display format corresponding to the times when the scheduled programming content is available for viewing. The grid format of an EPG display is designed to receive input from a remote control device using a five- way navigation command system to move a cursor to different entries in the grid. The inventors have recognized that navigation using conventional EPG displays is cumbersome. To navigate through entries in a conventional grid-like EPG display using a five- way navigation command system, a user may be required to provide a series of navigation commands (e.g., "down-down-downright-select"), which does not result in an efficient or intuitive experience for the user to select a particular program.
[00101] The inventors have recognized that the display of conventional EPG search systems is a "visual-first" user interface in which the main focus of the interface is to display the scheduled programming content in visually consistent boxes. The inventors have appreciated that other techniques for searching for content are possible that do not require a rigid visual presentation of scheduled programming content because the user is not restricted to navigating displayed entries using a five-way navigation command system. To this end, some embodiments of the invention are directed to methods and apparatus for an improved visual search interface. In particular, in some embodiments of the invention, the visual user interface is designed to encourage
"natural-language first" input such that the interface is configured to encourage users to speak or otherwise provide input to the search system in a natural way.
[00102] Some embodiments encourage the use of voice input as a primary input modality, with visual aspects of the user interface being complimentary to voice input. However, it should be appreciated that any input that can be spoken can alternatively be provided via another type of input modality (e.g., text entry in a keyboard). Thus, while illustrated embodiments described below describe the use of voice as the input modality, other modalities can be used to allow the user to provide the same types of inputs described in the illustrated embodiments below as being provided by voice.
[00103] An exemplary system 100 for use in accordance with some embodiments of the invention is illustrated in FIG. 1. System 100 includes electronic device 102, which may be a television, a computer, a portable electronic device such as a
smartphone, or any other tunable electronic device configured to receive and display scheduled programming content. Electronic device 102 may be configured to receive input from a user to search for and select content to view on the electronic device, wherein the input is received using one or more input modalities. In some embodiments, electronic device 102 includes a voice input interface 104, such as a microphone and one or more other input interfaces 106 configured to receive the input. Other input interface(s) 106 may include, but are not limited to, a keyboard interface (e.g., a
QWERTY keyboard), a mouse interface, a remote control interface, or any other suitable interface that enables a user to search for content to view on electronic device 102.
[00104] Voice input received via voice input interface 104 may be sent to automatic speech recognition (ASR) engine 108 for speech recognition to recognize the content of the received voice input. Although ASR engine 108 is illustrated in FIG. 1 as being included as a portion of electronic device 102, ASR engine 108 may alternatively be located remotely from electronic device 102 and may be in communication with electronic device 102 using any suitable communication mediums (e.g., one or more networks). In some embodiments, electronic device 102 may be associated with one or more local ASR engines 108 and one or more remotely located ASR engines, whereby at least some received voice input is processed by the one or more local ASR engine(s) and at least some received voice input is processed by the one or more remotely-located ASR engines. Examples of such distributed ASR systems are known by those of skill in the art. Voice input sent to the one or more remote ASR engine(s) for recognition using one or more communication medium(s) (e.g., one or more networks) connecting the electronic device 102 and the ASR engine(s) may be returned to the electronic device 102 from the remotely-located ASR engine(s) using the same or different communication medium(s).
[00105] In some embodiments, ASR engine 108 may be in communication with one or more voice command data sources 110 storing one or more voice commands.
ASR engine 108 may match received voice input to the one or more voice commands stored in one or more voice command data sources 110 and return one or more matching voice commands to one or more processors 114 programmed to execute at least one action corresponding to the one or more matching voice commands. In some
embodiments, the output of ASR engine 108 may be analyzed using a natural language understanding (NLU) engine using any suitable NLU-based technique(s), examples of which are known in the art. The NLU output may be used, at least in part, to associate a voice command with the received voice input. In some embodiments, electronic device 102 may be configured to interpret voice input as a voice command in response to activation of an input signal (e.g., the user pressing a key on a remote control or a keyboard) or in response to receiving voice input corresponding to a "wake-up" word that when received indicates that the voice input following the input signal should be interpreted as a voice command. Alternatively, voice input may be interpreted as a command without command initiation via an input signal (e.g., a key press or a wake-up word), as embodiments of the invention are not limited in this respect.
[00106] The set of voice commands stored by the one or more voice command data sources 110 may include any suitable commands for searching for and selecting content to view on the electronic device 102. In some embodiments, the set of voice commands may include commands to control functions of the electronic device 102 itself, examples of which include commands frequently found on a conventional remote control device, such as "mute," "power on," and "volume." An illustrative set of electronic device commands for controlling an electronic device 102 is shown below in Table 1, but it should be appreciated that these commands are provided merely as examples and that embodiments described herein are not limited to using these specific commands.
Voice Command Example(s) Description
Dragon 'Dragon' Wake-up-word will start speech recording or when device is on standby, will start device
Mute (on/off), mute, 'Mute on,' 'Unmute' Mutes/unmutes volume unmute
Shutdown, standby, 'Shutdown,' 'Power off,' Sets device to standby (switch/turn/power) off 'Turn off mode
Volume <num_upto_100>, 'Set volume to 50,' Sets the device volume Set volume to 'Volume 70' from 0-100
<num_upto_ 100> Volume (up/down) 'Volume up' Set volume up/down
Turn it (up/down), louder, 'Turn it up,' 'Louder' Set volume up/down softer
(Enable/start/disable/stop) 'Enable read out,' 'Read Turn on/off text-to-speech (read out/screen reader), out off,' 'Enable screen (TTS)
(Read out/screen reader) reader,' 'Screen reader on'
(on/off)
Exit, close 'Exit,' 'Close' Exit the current
screen/window
Cancel 'Cancel' Cancel a request
Help/What can I say 'Help,' 'What can I say?' Show help
Table 1: Illustrative Electronic Device Commands
[00107] The set of voice commands may also include tuner commands used to search for and/or select scheduled programming content to display on electronic device 102, such as "show <channel name>, "next channel," and "show <program name>." An illustrative set of tuner commands is shown below in Table 2, but it should be appreciated that these commands are provided merely as examples and that embodiments described herein are not limited to using these specific commands.
[show] channel <channel 'show channel HBO' Switch to specified channel name> name show <channel name> 'show HBO' Switch to specified channel name switch to <channel name> 'switch to HBO' Switch to specified channel name show listings 'show listings' Show electronic program guide show EPG 'show EPG' Show electronic program guide show [the] [program] 'show program Show electronic program guide guide' guide show channels 'show channels' Show channels list show channel list 'show channel list' Show channels list what's on [<channel 'what's on HBO at 6' List programs on channel at name>] [at <time>] specified time what's on channel 'what' s on channel List programs on channel at <number> [at <time>] 5' specified time what's on <channel 'what's on HBO on List programs on channel at name> <day> [at <time>] Monday,' 'what's on specified time
HBO tonight at 9,'
what's on HBO
tomorrow at 6 pm' what's on channel 'what' s on channel 5 List programs on channel at <number> <day> [at on Monday morning' specified time
<time>] show <program name> 'show Frasier' Switch to channel that is currently playing program or has program in schedule
within 30 minutes what am I (looking at / 'what am I Shows current channel
watching), [show] watching,' 'show information
information information,'
'information'
Find/Search 'find movies with Advanced searching using
<program/genre> Harrison Ford,' program guide information
[<actor>] [<channel>] [<at 'search for comedies
time>] tonight,' 'find news
on PBS,' 'find action
movies'
Table 2: Illustrative Tuner Commands
[00108] The set of voice commands may also include application commands for interacting with one or more client applications on electronic device 102 that perform any of numerous functions, including providing access to numerous services. For example, a user may use an application command to initiate a voice call session using a client application that performs calling (such as Skype™), search for weather information, or perform any of numerous other actions. An illustrative set of application commands is shown below in Table 3.
Find/search/show/play 'show Michael Jackson on Start YouTube™ client and <video> on video sharing YouTube,' 'Find ice hockey use <video> part to search service (e.g., YouTube) goal on YouTube'
[show/open] YouTube 'show YouTube,' Open YouTube™ client
'YouTube'
Find/search [for] <video> 'find Miami Vice' Search videos on on YouTube YouTube™ (YouTube™ client already open)
Play/show/watch [full 'play,' 'watch full screen' Select video from results screen] list and show it on small screen [or full screen] . If a video is already playing in small screen, open the full screen playback
Play/show/watch/select 'play Rihanna Man Down' Select a video to play from <title> [full screen] results list [open full screen playback]
Play/show/watch/select 'play number 6 full screen' Select a video to play from number <item number> results list [open full screen [full screen] playback]
Page up/ previous page, 'page down,' 'previous Switch to the next/previous page down/ next page page' page
[go to the] (first item / 'first item' Browse to the beginning of beginning) results list
[go the the] (last item / end) 'go to the last item' Browse to the end of results list
Back 'back' If video playback is on, go back to results list. If showing results list, close YouTube™ client.
Find weather in <city> 'find weather in Seattle' Update weather information on toolbar
Search [for] weather in 'search weather in Las Update weather information <city> Vegas,' Search for weather on toolbar
in Stockholm'
What' s the weather in 'what' s the weather in New Update weather information <city> York' on toolbar
Share on [the] Twitter 'share on Twitter' Open Twitter™ client update window, start dictation automatically
Send to [the] Twitter 'send to the Twitter' Open Twitter™ client update window, start dictation automatically
Update [the] Twitter 'update Twitter' Open Twitter™ client update window, start dictation automatically
Open/show/go to Twitter 'open Twitter' Open Twitter™ client update window
Close 'close' Cancel Twitter™ update
Dictate 'dictate' Starts dictation within open
Twitter™ client application
Send 'send' Update Twitter™ account with text shown on 'edit box'
Close 'close' Close Twitter™ client and cancel update
Table 3: Illustrative Application Commands [00109] The illustrative application commands described in Table 3 include commands to control exemplary client applications including Twitter™, YouTube™, and Skype™. However, it should be appreciated that embodiments of the invention may include voice commands for any other client application and are not limited in this respect. For example, voice commands for client applications including, but not limited to, Facebook®, Gmail™, NetFlix®, Hulu™, and MySpace™ may also be included. Some client applications, including many of those listed above, may be web-based services or applications that can be accessed either via a specialized application on the electronic device or via a web browser on the device and embodiments of the invention are not limited in the way in which a client application is accessed.
[00110] The set of voice commands may also include action commands, each for performing an action related to one or more items in a search results list. An illustrative list of action commands is described below in Table 4.
beginning) the results list
[go to the] (last item / end) 'go to the last item' Browse to the end of the results list
Watch 'watch' Change the channel and starts playing the selected program
DVR 'DVR' Display 'recording set' notification'
Cancel 'cancel' Cancel a selection
Table 4: Illustrative Action Commands
[00111] In some embodiments, the set of voice commands may also include EPG commands for interacting with an EPG display. Rather than restricting the user' s navigation through EPG entries using a conventional five-way navigation command system used with remote control devices, the set of EPG commands may provide additional voice functionality to enhance the user's browsing experience. An illustrative set of EPG commands is described below in Table 5.
<channel name> in EPG
[select/play/show/watch] 'show channel 11,' channel Change to channel visible channel [number] <channel number 6' in EPG
number>
Page up / previous page, 'previous page,' 'page up,' Switch to another page in page down / next page, 'page left,' 'page right' the EPG view
page left, page right
Previous channel / step up, 'next channel,' 'previous Step up/down in the next channel / step down channel' channel list
[go to the] (first item / 'first item' Browse to the beginning of beginning) the channels list
[go to the] (last item / end) 'go to the last item' Browse to the end of the channels list
Table 5: Illustrative EPG Commands
[00112] In some embodiments, electronic device 102 also includes a tuner 112 configured to enable electronic device 102 to tune to a particular channel to display scheduled programming content on electronic device 102. For example, tuner 112 may enable the electronic device to receive scheduled programming content from one or more scheduled programming content sources 124 connected to electronic device 102 via network 120. Scheduled programming content sources 124 may source media content scheduled for these times including, but not limited to, stored content (e.g., television programs, and movies), and live media content. Scheduled programming content sources 124 that store and/or transmit scheduled programming content to an electronic device (e.g., over a communication medium such as a cable or satellite network) is well-known. Some embodiments may include a tuner configured to receive scheduled programming content from multiple content sources and/or may include multiple tuners, each of which is configured to receive scheduled programming content from a different content source.
[00113] Electronic device 102 may also include network interface 118 configured to communicate with one or more storage and/or processing devices via one or more network 120. As discussed above, network interface 118 may be used to transmit voice input to one or more remote ASR engines and receive speech recognition results from the one or more ASR engines. Electronic device 102 can employ a single network interface 118, or more than one network interface 118 to enable the electronic device 102 to connect to more than one type of network (e.g., to perform different functions).
[00114] An exemplary electronic device 102 is a television system. Television systems often include a tuner to enable the television system to display scheduled programming content received by a set top box that is connected to the television. A set top box is an example of a tuner 112, which receives scheduled programming content via a network interface 118 and decodes the received content to enable the electronic device to display the content on a display 122. Scheduled programming content or any other content received via network 120 may be displayed on the display 122 connected to electronic device 102. In some embodiments, display 122 may be integrated as part of electronic device 102 and in other embodiments, display 122 may be connected to electronic device 102 via one or more wired or wireless connections. Display 122 can also be a single display or multiple displays and embodiments of the invention are not limited in this respect.
[00115] In some embodiments, electronic device 102 may also be associated with a local content data source 116, which may be configured to store one or more television programs, one or more movies, or any other content suitable for playback on the electronic device. For example, in some embodiments, local content data source 116 may be a digital video recorder (DVR) configured to store recorded scheduled programming content. Local content data source 116 may be configured to store content related to one or more local client applications associated with electronic device 102 including, but not limited to, one or more applications that store and/or manage picture files, music files, and/or video files.
[00116] In accordance with some embodiments, when searching for content to view and/or playback on electronic device 102, content may be retrieved from local content data source 116 and/or one or more data sources (e.g., scheduled content data source 124, bundled service content data source 124) remotely located from electronic device 102 and connected to electronic device 102 by network 120. The particular source(s) that are searched may depend, at least in part, on the search query. When results are returned from multiple sources, the search results may be ordered based, at least in part, on the source of the content corresponding to the search results. In some embodiments, a user may be able to restrict a search for content based on the source of the content and/or the type of content desired to be included in the search results, as described in more detail below.
[00117] Some embodiments are directed to a voice-controllable visual search interface displayed on display 122 and configured to allow the user to search for content to view and/or playback using the electronic device 102. The user may be able to interact with the search interface to search for scheduled programming content, as described above. Scheduled programming content includes, but is not limited to, broadcast content such as television and radio programs.
[00118] Additionally, some embodiments may enable a user to search for bundled service content stored, for example, by one or more network-connected bundled service data sources 124 and/or local content data source 116. Unlike scheduled programming content which may only be available for viewing at particular times (e.g., when the content is scheduled for broadcast), bundled service content may be available for viewing/playback at any time. Bundled service content may include, but is not limited to, content from a web-based service (e.g., Netflix®), content from a local or remote data source storing media content (e.g., DVR content, on-demand content from a service provider, etc.), and content associated with one or more client applications (e.g., iTunes®) associated with electronic device 102. As described in more detail below, searching for content in accordance with some embodiments may allow a user to search for scheduled programming content, bundled service content, or both scheduled programming content and bundled service content in a single search.
[00119] An exemplary method for processing a search query in accordance with some embodiments of the invention is illustrated in FIG. 2. In act 210, electronic device 102 receives a search query to search for content to view and/or playback on electronic device 102. For example, the search query may include one or more voice commands, as discussed above. To interpret received voice input as one or more voice commands, the voice input may be processed by one or more local and/or remote ASR engines for recognition. In some embodiments, a determination of whether to process voice input via a local and/or remote ASR engine may be made based, at least in part, on the type of voice input. For example, in one implementation, voice input corresponding to fixed- form device commands (e.g., "mute," "power off," "volume up") may be processed by a local ASR engine, whereas commands including free-form input (e.g., 'Search,' 'Find,' etc.) may be processed by one or more remotely-located (e.g., "cloud") ASR engines. For example, since fixed-form commands may be associated with a limited grammar, they can be easier to recognize and a local ASR engine may suffice for voice recognition and be faster than sending the input to a remote ASR engine. However, free-form input may be more difficult to recognize, so sending the input to more robust ASR engine(s) may be required for acceptable recognition performance. Also, language models associated with remotely ASR engines can be updated regularly to reflect changes in popular culture, names of movies, actors, books, etc. that might otherwise be difficult for a local ASR engine to recognize. In some embodiments, at least a first portion of the voice input may be processed by a local ASR engine and at least a second portion of the voice input may be processed by a remotely-located ASR engine, and any combination of local and remote ASR engines may be used to recognize received voice input.
[00120] The process then proceeds to act 220, where one or more data sources are searched based, at least in part, on the recognized voice input corresponding to the search query. The one or more data sources may be searched in any order, and any particular data source(s) (including less than all of the available data sources) may be searched. The results that are returned from the search may be determined in any suitable way. In some embodiments, a determination of which data sources to search may be made based, at least in part, on a particular action a user wants to perform. For example, if the user says "listen Bob Seger," the system may determine that the user wants to listen to music, so only a local music data source may be searched and not a data source associated with an EPG. Alternatively, if the user says "watch Modern Family," it may be determined that the user wants to watch a television show. Accordingly, a data source associated with an EPG and one or more data sources associated with recorded television content may be searched, but a local music data source may not be searched. By selectively searching less than all available data sources, some embodiments of the invention may return relevant search results in a more efficient manner by reducing an amount of time to return the search results. However, not all embodiments are limited to searching only a subset of available sources as all available sources may be searched for some searches.
[00121] In some embodiments, an order in which the data sources are searched and/or an order in which the search results are presented may be determined based, at least in part, on one or more rules associating a particular command (or determined user intent) with a particular order for searching and/or presenting search results. For example, if a user says "watch Modern Family," recognition of the voice command "watch" (or an intended action if a specific command is not employed, as discussed below) may cause the electronic device to search and return results from data source(s) associated with an EPG first, search and return results from data source(s) associated with local content second, and search and return results from data source(s) associated with client applications (e.g., YouTube™) third. It should be appreciated that this particular order for searching and returning search results associated with a "watch" command is described merely for illustrative purposes and any other desired order may alternatively be used.
[00122] Additionally, some embodiments may not include a fixed command grammar, but rather may determine a user' s intent from the received input in some other way, such as determining a user's intent by monitoring for particular words in free-form input, and performing an action in response to determining the user's intent. For example, if a user uses the "play" command or uses the term "play" in a free-form input, the system may recognize that the play command or use of the word "play" may suggest an intent by the user to access always accessible content, whether locally stored or available on-demand. One or more priority rules may be used to determine what media content to play. For example, it may first be determined whether the EPG has any media content currently available to be played (e.g., broadcast content, on demand content, etc.) that matches the voice input. If it is determined that there is not currently available content from the EPG, it may next be determined whether any recorded media content
(e.g., DVR content) matches the voice input. If more than one piece of recorded media content match the voice input, the one or more priority rules may specify that all the matching content be displayed so that the user may select among them. Alternatively, in some embodiments, rather than displaying a plurality of search results to a user, the electronic device may automatically playback a selected one of the matching content (e.g., the oldest unwatched (or partially watched) recorded content be selected and played back on the electronic device). Although the above-described example relates to playing media content using a particular set of priority rules, it should be appreciated that these rules are merely illustrative, and that any suitable set of priority rules may be used for any voice input as embodiments of the invention are not limited in this respect.
[00123] Returning to the process of FIG. 2, after the one or more data sources has been searched in act 220, the process proceeds to act 222, where one or more results are displayed on a user interface associated with electronic device 102 in response to searching the one or more data sources for content that matches the search query. The search result(s) may be displayed in any suitable way including using a voice-enabled search results interface, as described in more detail below. As discussed above, displaying the results can include directly presenting a piece of content that matches user input.
[00124] FIG. 3 illustrates a portion of a visual search interface in accordance with some embodiments of the invention. In particular, FIG. 3 shows an illustrative result screen 300 for displaying search results returned from a user-initiated search for content to view and/or play on an electronic device. In some embodiments, result screen 300 is configured to display search results along at least two axes. In one illustrative embodiment, one axis is in the foreground of the search interface and one or more other axes are in the background of the search interface, but can be brought to the foreground of the search interface in response to a command (e.g., a voice command). In yet some other embodiments, only the content displayed on the axis in the foreground is fully focused, and other axes are partially out of focus to draw the user's attention to the foreground axis. However, not all embodiments using foreground and background axes are limited in this respect, as some embodiments can present two or more axes fully in focus.
[00125] In some embodiments, first axis for displaying search results may be a time-independent axis 310. Search results may be ordered along time-independent axis 310 using any suitable metric not related to a time when the content corresponding to the search results is available for viewing and/or playback. For example, in one non-limiting example, the search results may be ordered along the time-independent axis based, at least in part, on relevance to the search query that led to the result. In some
embodiments, content displayed along the time-independent axis may be limited to content that is available at the current time, while in other embodiments it may include content not currently available. For embodiments that order the returned search results based on relevance to a search query, relevance may be determined in any suitable way using any suitable algorithm (e.g., including giving higher relevance to content currently available for viewing), as embodiments of the invention are not limited in this respect. It should be appreciated that relevance to a search query is only one way to order results along the time-independent axis, and that any other suitable metric may be used instead of or in addition to relevance to a search query. For example, the search results may be ordered based, at least in part, on type of content, source of content, or any other factor or combination of factors.
[00126] In some embodiments, search results may also be organized along a second axis that is time-based and is configured to display search results available for viewing at different times. For example, the foreground of the result screen 300 may display search results for content that is currently available, whereas content available only at a future time may be shown in the background 320 as illustrated in FIG. 3.
[00127] Although the result screen 300 includes search results displayed on only two axes it should be appreciated that search results may be displayed on any number of axes, as embodiments of the invention are not limited in this respect. For example, in some embodiments search results for bundled service content that are available at the current time, but may require a subscription or some other fee to be paid prior to viewing the content, may be displayed along a separate axis (not shown) to indicate that the content may be viewed only after an applicable payment has been paid.
[00128] Result screen 300 may also include description section 330 that includes information regarding a currently selected search result. Description section 330 may also include other information that enables the user to make an informed selection from the search results displayed on result screen 300. For example, description section 330 may include a description of the content associated with the currently selected search result, and this description may change based on a selection of a different search result displayed on result screen 300. [00129] In some embodiments, less than all of the search results returned from the searching process may be displayed on result screen 300 to minimize crowding of the display. In some embodiments, a user may interact with one or more voice tags, described in more detail below, to view additional search results. Result screen 300 may also include discovery field 340 that enables a user to perform an additional search based on one or more of the search results displayed on result screen 300. For example, discovery field 340, when selected by a user, may initiate a search for content similar to content associated with the currently selected search result. Similarity may be determined in any suitable way using any suitable algorithm, as embodiments of the invention are not limited in this respect. For example, similarity may be determined based, at least in part, on a content source content was retrieved from, one or more content tags associated with a piece of content (e.g., parental ratings for movies), type of content, or any combination of these or other factors.
[00130] In some embodiments, result screen 300 may include a toolbar 350 including information to facilitate a user interaction with search screen 300. For example, toolbar 350 may include one or more icons, text, or other indicators explaining screen- specific or global inputs (e.g., voice commands) that a user may use to interact with result screen 300.
[00131] Search results, in accordance with some embodiments of the invention, may comprise blended search results that include different types of content. For example, as discussed above, search results may include results corresponding to scheduled programming content and bundled service content. FIG. 4 illustrates an illustrative result screen 400 in which the search results displayed on the result screen 400 include blended search results, i.e., at least two of the search results are associated with different types of content. The blended search results may include first results returned from one or more scheduled programming content data sources and second results returned from one or more bundled service content data sources. The search results displayed on result screen 400 include search results 410, 412, and 414, which may be results returned from one or more scheduled programming content data sources and search results 420 and 422, which may be results returned form one or more bundled service content data sources. [00132] As illustrated in FIG. 4, search results returned from scheduled programming content data sources and search results returned from bundled service content data sources may be mixed together and ordered along a time-independent axis using any suitable metric or combination of metrics including, but not limited to, relevancy to the search query, type of content, and source of content. As with result screen 300, result screen 400 may include a description portion 430 that includes information about one or more of the search results displayed on result screen 400.
[00133] In some embodiments, search results displayed on result screen 400 may be ordered along at least two axes including a time-independent axis and a time-based axis. In some embodiments, the search results displayed in a background portion of the search interface may include only scheduled programming content and not bundled service content. For example, because some bundled service content may always be available, in some embodiments search results corresponding to such bundled service content may always be shown in a foreground portion of the search interface to indicate that it is available for viewing and/or playing immediately. It should be appreciated, however, that this is merely one illustrative way to organize content using two axes and other organizations of content are also possible.
[00134] As discussed above, some bundled service content may only be accessed immediately upon payment of a fee. For example, a song from an online music store such as iTunes® may be accessed immediately, but only upon purchase of the song. In some embodiments, although not shown in FIG. 4, search results that correspond to content that requires payment before becoming available for viewing and/or playback on an electronic device may not be mixed with other search results, but may appear in a different portion of the search results interface.
[00135] In some embodiments, a search interface in accordance with some embodiments of the invention may include one or more selection tags that facilitate interactions between a user and one or more components of the search interface. FIG. 5 illustrates an exemplary result screen 500 including a plurality of selection tags. As illustrated in FIG. 5, one or more of the search results may be associated with a selection tag 510, which indicates input a user may provide (e.g., via spoken input, a keyboard or keypad, etc.) to select a particular search result. In the exemplary result screen 500, for embodiments employing a voice interface, if the user wanted to select the first search result in the lower left corner, the user may say "select 1." Alternatively, if the user wanted to select the search result displayed on the lower right corner, the user may say "select 4."
[00136] Result screen 500 may also include other types of selection tags such as one or more icons 520 that enable a user to use a selection command to perform one or more actions related to the content with which the icon is associated. For example, the icons 520 may include, but are not limited to, icons that indicate to the user that they may view particular content, record particular content, or perform some other action associated with the particular content associated with the icon. The user may combine an action with identification of a selection tag to perform the action on the selected content. For example, a user may view the content associated with the search result in the lower left corner of FIG. 5 by saying "view 1."
[00137] In some embodiments, that employ a voice interface, voice interactions with the visual search interface may be categorized into different types of voice commands depending on when the voice command is active. For example, some voice commands may be categorized as screen agnostic commands, which are independent of the content displayed on any particular screen and are always active irrespective of what particular screen is currently being presented on display. Some non-limiting examples of screen agnostic commands include, but art not limited to, commands such as volume control, channel control, screen brightness, and menu accessing commands. In some embodiments, the set of screen agnostic commands overlap substantially with
functionality conventionally controlled by a remote control device. Screen agnostic commands may be associated with one or more labels for selective tags displayed on the search interface to alert the user to the presence of these commands, or in some embodiments the user is presumed to know that these always available commands and can be invoked so that no selection tags or icons are displayed. In some embodiments, icons for one or more of the global commands may be included in a toolbar displayed on the search interface screen and/or a list of global commands may be accessed in response to a trigger event such as asking for help, pressing a hotkey on an input device, or in some other way. [00138] Some voice interactions may be associated with screen-dependent actions that enable a user to perform actions only within a particular screen of the user interface. For example, in the result screen 500 illustrated in FIG. 5, selection tags 510 may allow a user to select a particular search result as described above, and the response to selecting the particular search result, a description of the content associated with the selected search result may be updated on the search interface to provide the user with information that may help the user decide whether to perform an action (e.g., watch, purchase, record) on the content associated with the selected search result.
[00139] Other screen-dependent actions may be associated with one or more navigation functions of the search interface that enable a user to navigate within a particular result screen, and/or to update the content of a result screen. For example, the search interface may display a voice-enabled label that allows a user to view additional search results not currently displayed on the search interface, for example, by speaking "next results" or some other command phrase or free-form input to update the result screen to include the additional results.
[00140] The search interface may be further configured with user interaction functionality associated with screen-dependent actions to allow a user to switch between search results presented at different levels of a multi-axis display. For example, in embodiments in which search results are displayed along a time-independent axis and a time-based axis, the user may use voice (or other forms of) input to switch which results along the time-based axis are currently in focus and can be selected. For example, a user may say "later" (or some other word or phrase) to select content associated with search results displayed in the background of the user interface along the time-based axis of the displayed search results. This may be useful, for example, to allow the user to take an action (e.g., record) relating to scheduled programming content that is available at some later time. To perform this action, the user may say "later" (or some other word or phrase), which may switch the search results displayed in the foreground and the search results displayed in the background to enable the user to select an item for which content is available only in the future.
[00141] Another category of user interactions may be represented by graphical icons 520 that enable a user to perform a particular action such as recording, viewing, or playback of content associated with the search result. In some embodiments, graphical icons may also correspond to additional value-added actions that enable a user to perform other "secondary" functions such as posting to Facebook®, performing an action on Twitter™, or interacting with another client application service or website.
[00142] Although the illustrative examples above discuss three different types of user interactions, (i.e., screen- agnostic commands, screen-dependent actions, and graphical icon actions) it should be appreciated that any number of types of user interactions may be used, as embodiments of the invention are not limited in this respect.
[00143] In some embodiments, that support a voice interface, the presence or absence of selection tags and/or other icons displayed on a search interface may be determined based, at least in part, on whether voice interaction has been activated for the search interface. Determining whether voice interaction is active for the search interface may be performed in any suitable way including, for example, determining whether a microphone is active, determining whether voice recognition software is executing on the electronic device, determining whether a user has issued a voice command to display the voice tags on the search interface, or other suitable ways.
[00144] In some embodiments, the electronic device may receive input which causes the electronic device to enter a "listening mode" during which the electronic device is expecting further input from the user to perform an action. In contrast, when the electronic device is not in listening mode, the electronic device may not interpret received input, by default, as representing a user' s intent to perform a particular action, and may be monitoring for input to trigger entry into listening mode. An illustrative process for an electronic device that includes a listening mode in accordance with some embodiments of the invention is illustrated in FIG. 6. In act 610, it is determined whether a first input has been received that will cause the electronic device to enter a listening mode. For example, in embodiments that include a voice interface, the first input may correspond to a voice command or some other voice input demonstrating a user's intent to perform an action, as described above. Alternatively, the first input may be any other suitable input to cause the electronic device to enter listening mode including, but not limited to, a press of a particular key on a keyboard or keypad. [00145] If it is determined in act 610 that a first input has not been received, the process continues to monitor for a received first input. Otherwise, if a first input has been received, the process proceeds to act 612, where the electronic device is caused to enter a listening mode in response to recognizing the first input. While in listening mode, the process proceeds to act 614, where it is determined whether a second input has been received. If it is determined that a second input has been received while in listening mode, the process proceeds to act 620 where one or more actions corresponding to the first and/or second input are performed. After performing the one or more actions in act 620, the process may optionally return to act 612 to enter listening mode if additional input is expected. For example, the second input may be "What's on HBO comedy listings tonight?" and the action performed in act 620 may be opening the HBO comedy listings for the current day. Because additional input is expected, the process may proceed to act 612 to enter listening mode in anticipation of the additional input (e.g., next, previous, exit, etc.). However, if no additional input is expected after performing the one or more actions in act 620, the process may end.
[00146] If it is determined in act 614 that a second input has not been received, the process proceeds to act 616 where it is determined whether a particular amount of time (e.g., 5 seconds) has elapsed since entering listening mode. If it is determined in act 616 that the particular amount of time has not yet elapsed, the process continues to monitor for a second input. Otherwise, if it is determined that the particular amount of time has elapsed, the process proceeds to act 618, where the electronic device is caused to exit listening mode, and the process returns to act 610 to monitor for a first input. In the above-described embodiment, the electronic device exits listening mode after a particular amount of time has passed after entering listening mode. However, it should be appreciated that in some embodiments, listening mode may alternatively be exited in other ways. For example, some embodiments may enable a user to explicitly exit listening mode by providing input instructing the electronic device to do so. This explicit input may take any suitable form including, but not limited to, a voice command (or determined user intent), or a key press on a keyboard or keypad.
[00147] In some embodiments, when an electronic device is caused to enter into a listening mode, this may be noted in some way on the user interface displayed by the electronic device. For example, upon entering the listening mode, one or more selection tags and/or icons may appear on the user interface to inform a user that listening mode has been entered. It should be appreciated that changes to a user interface to reflect the mode status of the electronic device may be performed in any way and the example of displaying or hiding selection tags and/or icons is only one way to indicate the status and other ways are also possible.
[00148] In some embodiments, bundled service content search results may be represented hierarchically. For example, if a particular bundled service includes one or more pieces of content that match a search query, they may initially be represented by including an identifier for the bundle service in the search results, but not the
corresponding pieces it content themselves, and then selection of the bundled service from among the search results may cause the corresponding pieces of content to be shown on the user interface. For example, an application associated with a bundled service content search result may be illustrated as an image or icon for the application, which is recognized by a user as representing that application. For example, the bundled service search result may correspond to a movie streaming service such as, Netflix® (e.g., illustrated by a Netflix® icon), and in response to selecting the movie streaming service by, for example, saying "select 2," which selects the second search result (e.g., the Netflix® search result), search results within the Netflix® application may be displayed.
[00149] In an illustrative example, a user may search for content corresponding to
"Yankees," and search results returned in response to this search may include scheduled programming content for New York Yankees baseball games, bundled service content from a movie streaming service for movies related to the New York Yankees, music content related to the band "Damn Yankees," and electronic book content for the electronic book "A Connecticut Yankee in King Arthur's Court" by Mark Twain. Rather than showing all movies the bundled service includes related to the Yankees, the search results may initially show an image for a movie streaming service which includes multiple movies related to the search term "Yankees." Upon selecting the image for the movie streaming service using, for example, a selection voice tag associated with the image, one or more movies available from that streaming service and corresponding to the search term "Yankees" may be displayed in the search results. Hierarchically- presented search results may have any number of hierarchical levels, as embodiments of the invention are not limited in this respect.
[00150] Another illustrative example of content that can be represented hierarchically is songs in a music library. For example, a user may want to play a song by Metallica, but may not remember the name of the song. To search for the song, the user may initiate a search for "Metallica," and the search results that are returned may include multiple albums for Metallica stored locally on a local data source associated with electronic device and/or one or more network connected music data sources (e.g., databases) that include music from Metallica. In some embodiments, the search results may be presented as a series of images for album covers for the different Metallica albums and the user may be prompted to select one of the albums from the search results (e.g., by using selection voice tags associated with each album image). Upon selection of one of the albums, the selected album may be expanded to reveal search results for one or more of the songs associated with album, and the user may select one of the songs on the album for playback. Accordingly, in this respect, search results may be presented in a hierarchical manner such that sequential selections of search results may allow the user to navigate through the search results to identify the content the user wants to playback on the electronic device.
[00151] In some embodiments, voice tags may be employed that instruct the user how to interact with different components of the user interface. Content-neutral voice tags are labels that are independent of the content that the voice tag identifies. An example of a content-neutral voice tag is the selection voice tags (1, 2, 3, 4) illustrated in FIG. 5, which allow the user to select one of the search results without having to speak the name of the content identifying the search result. Thus, content-neutral voice tags do not identify the content itself with which they are associated, but merely identify a slot or position of the search result to which the content is assigned. By using content-neutral voice tags, the user is provided with a set of voice tags that facilitates a consistent user experience when interacting with the visual search interface. Additionally, using content-neutral voice tags allows the user to use less speech to invoke commands that would be required if the voice tags were created based, at least in part, on the content the tags identified. Use of content-neutral voice tags also allow the use of a fixed grammar (e.g., numbers, letters) that can result in more accurate ASR for content that may have unique names, etc.
[00152] In some embodiments, the user may be able to provide input (e.g., via voice) that results in the activation of more than one type of command or interaction. For example, a user may say, "watch 1." In response, two interactions take place: (1) the first item in the search results may be selected and (2) the selected item may be played. In this way a single input may cause multiple commands or interactions to be executed without the user having to issue separate commands. Such a system enables the user to use natural input (e.g., speech) to control the electronic device.
[00153] Searches may be performed for any type of content has described above, including searches for video clips on YouTube™, Hulu™, or any other video-sharing website or web service. Additionally, the user may search for music content, as described above, in one music libraries stored locally or accessible via a network (e.g., the Internet). The user may have the ability to restrict a search to a particular type of content such as music, videos, or scheduled programming content, such that the search results only include a particular type of content for which the user desires to search. To this end, the search interface may be associated with advanced searching capabilities that may be configured using voice input and/or other input modalities.
[00154] In some embodiments, a piece of content displayed in the search results may include multiple tags (e.g., voice tags), with one tag identifying the content, and another tag indicating an action to perform with the content (e.g., watch, record, or purchase the content). The number of tags associated with search results for a search interface is not a limitation of embodiments of the invention.
[00155] An illustrative implementation of a computer system 700 that may be used in connection with any of the embodiments of the invention described herein is shown in FIG. 7. The computer system 700 may include one or more processors 710 and one or more computer-readable non-transitory storage media (e.g., memory 720 and one or more non-volatile storage media 730). The processor 710 may control writing data to and reading data from the memory 720 and the non-volatile storage device 730 in any suitable manner, as the aspects of the present invention described herein are not limited in this respect. To perform any of the functionality described herein, the processor 710 may execute one or more instructions stored in one or more computer-readable storage media (e.g., the memory 720), which may serve as non-transitory computer-readable storage media storing instructions for execution by the processor 710.
[00156] The above-described embodiments of the present invention can be implemented in any of numerous ways. For example, the embodiments may be implemented using hardware, software or a combination thereof. When implemented in software, the software code can be executed on any suitable processor or collection of processors, whether provided in a single computer or distributed among multiple computers. It should be appreciated that any component or collection of components that perform the functions described above can be generically considered as one or more controllers that control the above-discussed functions. The one or more controllers can be implemented in numerous ways, such as with dedicated hardware, or with general purpose hardware (e.g., one or more processors) that is programmed using microcode or software to perform the functions recited above.
[00157] In this respect, it should be appreciated that one implementation of the embodiments of the present invention comprises at least one non-transitory computer- readable storage medium (e.g., a computer memory, a floppy disk, a compact disk, a tape, etc.) encoded with a computer program (i.e., a plurality of instructions), which, when executed on a processor, performs the above-discussed functions of the
embodiments of the present invention. The computer-readable storage medium can be transportable such that the program stored thereon can be loaded onto any computer resource to implement the aspects of the present invention discussed herein. In addition, it should be appreciated that the reference to a computer program which, when executed, performs the above-discussed functions, is not limited to an application program running on a host computer. Rather, the term computer program is used herein in a generic sense to reference any type of computer code (e.g., software or microcode) that can be employed to program a processor to implement the above-discussed aspects of the present invention.
[00158] Various aspects of the present invention may be used alone, in
combination, or in a variety of arrangements not specifically discussed in the
embodiments described in the foregoing and are therefore not limited in their application to the details and arrangement of components set forth in the foregoing description or illustrated in the drawings. For example, aspects described in one embodiment may be combined in any manner with aspects described in other embodiments.
[00159] Also, embodiments of the invention may be implemented as one or more methods, of which an example has been provided. The acts performed as part of the method(s) may be ordered in any suitable way. Accordingly, embodiments may be constructed in which acts are performed in an order different than illustrated, which may include performing some acts simultaneously, even though shown as sequential acts in illustrative embodiments.
[00160] Use of ordinal terms such as "first," "second," "third," etc., in the claims to modify a claim element does not by itself connote any priority, precedence, or order of one claim element over another or the temporal order in which acts of a method are performed. Such terms are used merely as labels to distinguish one claim element having a certain name from another element having a same name (but for use of the ordinal term).
[00161] The phraseology and terminology used herein is for the purpose of description and should not be regarded as limiting. The use of "including,"
"comprising," "having," "containing", "involving", and variations thereof, is meant to encompass the items listed thereafter and additional items.
[00162] Having described several embodiments of the invention in detail, various modifications and improvements will readily occur to those skilled in the art. Such modifications and improvements are intended to be within the spirit and scope of the invention. Accordingly, the foregoing description is by way of example only, and is not intended as limiting. The invention is limited only as defined by the following claims and the equivalents thereto.
[00163] What is claimed is:

Claims

1. A method of presenting search results on an electronic device, wherein the electronic device includes a tuner configured to tune the electronic device to receive scheduled programming, the method comprising:
receiving voice input comprising a search query;
searching, with at least one processor, at least one data source based, at least in part, on the search query; and
presenting results of the search using a time-based axis and a time-independent axis.
2. The method of claim 1, wherein the time-independent axis orders the results based on relevance to one or more terms in the search query.
3. The method of claim 1, wherein searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
4. The method of claim 3, wherein the results include scheduled programming content identified in the at least one scheduled programming content data source and bundled service content identified in the at least one bundled service content data source.
5. The method of claim 1, wherein the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content.
6. The method of claim 1, wherein the time-based axis distinguishes first content that is currently available for playback from second content that is currently unavailable for playback.
7. The method of claim 1, wherein the results presented along the time-based axis include only scheduled programming content.
8. An electronic device, comprising:
a tuner configured to tune the electronic device to receive scheduled
programming; and
at least one processor programmed to:
receive voice input comprising a search query;
search at least one data source based, at least in part, on the search query; and
present results of the search using a time-based axis and a time- independent axis.
9. The electronic device of claim 8, wherein the time-independent axis orders the results based on relevance to one or more terms in the search query.
10. The electronic device of claim 8, further comprising:
a voice input interface configured to receive the search query as voice input.
11. The electronic device of claim 8, wherein searching at least one data source comprises searching at least one scheduled programming content data source and at least one bundled service content data source.
12. The electronic device of claim 8, wherein the results include bundled service content that, when selected, displays multiple results from at least one bundled service content data source associated with the selected bundled service content.
13. The electronic device of claim 8, wherein the time-based axis distinguishes first content that is currently available for playback by the electronic device from second content that is currently unavailable for playback by the electronic device.
14. A computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method, comprising:
receiving voice input comprising a search query;
searching at least one data source based, at least in part, on the search query; and presenting results of the search using a time-based axis and a time-independent axis.
15. The computer-readable storage medium of claim 14, wherein the time- independent axis orders the results based on relevance to one or more terms in the search query.
16. A method of presenting search results on a visual user interface for a digitally- tunable electronic device configured to display scheduled programming content, the method comprising:
receiving a search query;
searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and
presenting results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
17. The method of claim 16, wherein the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
18. The method of claim 16, wherein the at least one bundled service data source comprises a data source for a web-based service.
19. The method of claim 16, wherein the at least one bundled service data source comprises a local data source associated with the electronic device.
20. The method of claim 16, wherein the at least one bundled service data source comprises a plurality of data sources connected via a network.
21. The method of claim 16, wherein the at least one bundled service data source comprises a data source of on-demand content.
22. The method of claim 16, wherein presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
23. A computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer perform a method of presenting search results on a visual user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising:
receiving a search query;
searching based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and
presenting results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
24. The computer-readable storage medium of claim 23, wherein presenting the results of the search comprises presenting an image corresponding to at least one of the results of the search.
25. A digitally- tunable electronic device configured to display scheduled
programming content, the electronic device comprising:
at least one processor programmed to:
receive a search query;
search based, at least in part, on a natural language understanding analysis of the search query, at least one scheduled programming content data source and at least one bundled service content data source; and
present results of the search on the user interface, wherein the results of the search combine at least one first result corresponding to scheduled programming content and at least one second result corresponding to bundled service content.
26. The digitally- tunable electronic device of claim 25, wherein the at least one second result includes at least one bundled service result that, when selected, displays multiple results from the corresponding bundled service data source.
27. The digitally- tunable electronic device of claim 25, wherein the at least one bundled service data source comprises a data source for a web-based service.
28. The digitally- tunable electronic device of claim 25, wherein the at least one bundled service data source comprises a local data source associated with the electronic device.
29. The digitally- tunable electronic device of claim 25, wherein the at least one bundled service data source comprises a plurality of data sources connected via a network.
30. The digitally-tunable electronic device of claim 25, wherein the at least one bundled service data source comprises a data source of on-demand content.
31. A method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising:
displaying, on the user interface, a plurality of content-independent tags, the plurality of content-independent tags comprising a first content-independent tag identifying a first content unit returned from a search; and
in response to receiving first input corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input comprises first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
32. The method of claim 31, further comprising:
receiving second input, wherein a second content-independent tag of the plurality of content-independent tags is configured to be responsive to the second input; and wherein performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input.
33. The method of claim 32, wherein the second input is second voice input and wherein the second content-independent tag is configured to be responsive to the second voice input.
34. The method of claim 32, wherein receiving the first input and receiving the second input comprises receiving the second input before receiving the first input.
35. The method of claim 31, further comprising:
selecting the first content unit based on the first input.
36. The method of claim 31, wherein the performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled programming for the first content unit.
37. The method of claim 31, wherein the first content unit corresponds to a bundled service content unit comprising multiple search results, and wherein the performing at least one action related to the first content unit comprises displaying the multiple search results.
38. The method of claim 31, further comprising:
associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action, and wherein the first voice command and the second voice command are recognized by the electronic device irrespective of a particular screen displayed on the user interface.
39. A computer-readable storage medium encoded with a plurality instructions that, when executed by at least one computer, perform a method of providing a user interface for a digitally-tunable electronic device configured to display scheduled programming content, the method comprising:
displaying, on the user interface, a plurality of content-independent tags, the plurality of content-independent tags comprising a first content-independent tag identifying a first content unit returned from a search; and
in response to receiving first input corresponding to the first content-independent tag, performing at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
40. A digitally- tunable electronic device configured to display scheduled
programming content, the electronic device comprising:
at least one processor programmed to:
display, on a user interface, a plurality of content-independent tags, the plurality of content-independent tags comprising a first content-independent tag identifying a first content unit returned from a search; and
in response to receiving first input corresponding to the first content-independent tag, perform at least one action related to the first content unit, wherein the first input is first voice input and wherein the first content-independent tag is configured to be responsive to the first voice input.
41. The digitally- tunable electronic device of claim 40, wherein the at least one processor is further programmed to:
receive second input, wherein a second content-independent tag of the plurality of content-independent tags is configured to be responsive to the second input; and
wherein performing the at least one action related to the first content unit comprises performing the at least one action based, at least in part, on the second input.
42. The digitally-tunable electronic device of claim 40, wherein the performing at least one action related to the first content unit comprises tuning the electronic device to a channel to display scheduled programming for the first content unit.
43. The digitally- tunable electronic device of claim 40, wherein the first content unit corresponds to a bundled service content unit comprising multiple search results, and wherein the performing at least one action related to the first content unit comprises displaying the multiple search results.
44. A digitally-tunable electronic device comprising:
at least one processor programmed to:
provide a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally-tunable electronic device;
receive voice input comprising the search query; and
initiate a search for content corresponding to the search query in response to receiving the search query.
45. The digitally- tunable electronic device of claim 44, wherein initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
46 The digitally-tunable electronic device of claim 44, wherein the at least one processor is further programmed to:
display, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
47. The digitally- tunable electronic device of claim 44, wherein the at least one processor is further programmed to:
associate a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
48. The digitally- tunable electronic device of claim 47, wherein the first action and/or the second action corresponds to initiating a search for content.
49. A computer-implemented method, comprising:
providing a user interface that enables an electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the electronic device;
receiving voice input comprising the search query; and
initiating a search for content corresponding to the search query in response to receiving the search query.
50. The computer-implemented method of claim 49, wherein initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
51. The computer- implemented method of claim 49, further comprises:
displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
52. The computer-implemented method of claim 49, further comprising:
associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
53. The computer- implemented method of claim 52, wherein the first action and/or the second action corresponds to initiating a search for content.
54. A computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method comprising: providing a user interface that enables the electronic device to receive a search query regardless of which of a plurality of screens is displayed on the user interface when the search query is received, wherein the search query comprises a search for content to display and/or playback on the digitally-tunable electronic device;
receiving voice input comprising the search query; and
initiating a search for content corresponding to the search query in response to receiving the search query.
55. The computer-readable storage medium of claim 54, wherein initiating a search for content comprises sending a query for content to at least one data source connected to the electronic device by one or more networks.
56. The computer-readable storage medium of claim 54, wherein the method further comprises:
displaying, on the user interface, one or more icons indicating input a user may use to initiate a search for content.
57. The computer-readable storage medium of claim 54, wherein the method further comprises:
associating a plurality of voice commands with the user interface, wherein the plurality of voice commands includes a first voice command and a second voice command, wherein the first voice command invokes a first action and the second voice command invokes a second action.
58. The computer-readable storage medium of claim 57, wherein the first action and/or the second action corresponds to initiating a search for content.
59. A digitally-tunable electronic device comprising:
at least one processor programmed to:
receive first input;
entering a listening mode in response to recognizing the first input; determine whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and
exit the listening mode in response to determining that second input is not received during the particular amount of time.
60. The digitally-tunable electronic device of claim 59, wherein the second input is an explicit command to exit listening mode; and wherein the at least one processor is further programmed to:
exit the listening mode in response to receiving the second input.
61. The digitally- tunable electronic device of claim 59, wherein the at least one processor is further programmed to:
monitor for the first input in response to exiting the listening mode.
62. The digitally-tunable electronic device of claim 59, wherein the electronic device further comprises at least one display, and wherein the at least one processor is further programmed to:
display, on the least one display, an indication of whether the electronic device is currently in listening mode.
63. The digitally- tunable electronic device of claim 62, wherein the indication comprises one or more selection tags and/or icons.
64. The digitally-tunable electronic device of claim 59, wherein the at least one processor is further programmed to:
determine whether the electronic device is in listening mode; and
display the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode.
65. The digitally- tunable electronic device of claim 64, wherein the at least one processor is further programmed to: hide the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
66. A computer-implemented method comprising:
receiving first input;
causing an electronic device to enter a listening mode in response to recognizing the first input
determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and
exiting the listening mode in response to determining that second input is not received during the particular amount of time.
67. The computer- implemented method of claim 66, wherein the second input is an explicit command to exit listening mode; and wherein the method further comprises: exiting the listening mode in response to receiving the second input.
68. The computer- implemented method of claim 66, wherein the electronic device comprises at least one display, and wherein the method further comprises:
displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
69. The computer-implemented method of claim 68, wherein the indication comprises one or more selection tags and/or icons.
70. The computer-implemented method of claim 66, further comprising:
determining whether the electronic device is in listening mode;
displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and
hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
71. A computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method comprising:
receiving first input;
causing an electronic device to enter a listening mode in response to recognizing the first input
determining whether second input is received during a particular amount of time while the electronic device is in the listening mode, wherein the first input and/or the second input is voice input; and
exiting the listening mode in response to determining that second input is not received during the particular amount of time.
72. The computer-readable storage medium of claim 71, wherein the electronic device comprises at least one display, and wherein the method further comprises:
displaying, on the least one display, an indication of whether the electronic device is currently in listening mode.
73. The computer-readable storage medium of claim 72, wherein the indication comprises one or more selection tags and/or icons.
74. The computer-readable storage medium of claim 71, further comprising:
determining whether the electronic device is in listening mode;
displaying the plurality of content-independent voice tags in response to determining that the electronic device is in listening mode; and
hiding the plurality of content-independent voice tags in response to determining that the electronic device is not in listening mode.
75. A method of searching for content to display and/or playback on a digitally- tunable electronic device configured to display scheduled programming content, the method comprising:
receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device; determining, based on the search query, an action the user wants to perform;
determining one or more data sources to search based, at least in part, on the action the user wants to perform; and
searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
76. The method of claim 75, wherein determining the action the user wants to perform comprises determining the user wants to listen to media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as a music data source.
77. The method of claim 75, wherein determining the action the user wants to perform comprises determining the user wants to watch media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
78. The method of claim 75, further comprising:
storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and
wherein searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
79. The method of claim 75, wherein searching the one or more data sources comprises searching all available data sources.
80. The method of claim 75, further comprising:
receiving search results from the one or more data sources; and
presenting the search results on a user interface.
81. A computer-readable storage medium encoded with a plurality of instructions that, when executed by at least one computer, perform a method of searching for content to display and/or playback on a digitally-tunable electronic device configured to display scheduled programming content, the method comprising:
receiving voice input comprising a search query from a user, wherein the search query comprises a search for the content to display and/or playback on the electronic device;
determining, based on the search query, an action the user wants to perform; determining one or more data sources to search based, at least in part, on the action the user wants to perform; and
searching based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
82. The computer-readable storage medium of claim 81, wherein determining the action the user wants to perform comprises determining the user wants to listen to media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as a music data source.
83. The computer-readable storage medium of claim 81, wherein determining the action the user wants to perform comprises determining the user wants to watch media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
84. The computer-readable storage medium of claim 81, wherein the method further comprises:
storing one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and
wherein searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
85. A digitally-tunable electronic device configured to display scheduled
programming content, the electronic device comprising:
at least one processor programmed to: receive voice input comprising a search query from a user, wherein the search query comprises a search for content to display and/or playback on the electronic device;
determine, based on the search query, an action the user wants to perform; determine one or more data sources to search based, at least in part, on the action the user wants to perform; and
search based, at least in part, on the search query, the one or more data sources for the content to display and/or playback on the electronic device.
86. The digitally-tunable electronic device of claim 85, wherein determining the action the user wants to perform comprises determining the user wants to listen to media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as a music data source.
87. The digitally-tunable electronic device of claim 85, wherein determining the action the user wants to perform comprises determining the user wants to watch media content, and wherein determining one or more data sources to search comprises determining the one or more data sources as at least one scheduled programming content source.
88. The digitally-tunable electronic device of claim 85, wherein the at least one processor is further programmed to:
store one or more rules associating a particular type of input with a particular order for searching the one or more data sources; and
wherein searching the one or more data sources comprises searching the one or more data sources in accordance with the one or more rules.
89. The digitally-tunable electronic device of claim 85, wherein searching the one or more data sources comprises searching all available data sources.
EP13753963.1A 2012-08-16 2013-08-16 User interface for entertainment systems Withdrawn EP2803004A1 (en)

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US13/587,353 US9106957B2 (en) 2012-08-16 2012-08-16 Method and apparatus for searching data sources for entertainment systems
US13/587,300 US9026448B2 (en) 2012-08-16 2012-08-16 User interface for entertainment systems
US13/587,280 US9031848B2 (en) 2012-08-16 2012-08-16 User interface for searching a bundled service content data source
US13/587,222 US8799959B2 (en) 2012-08-16 2012-08-16 User interface for entertainment systems
US13/587,260 US20140052450A1 (en) 2012-08-16 2012-08-16 User interface for entertainment systems
US13/587,234 US9497515B2 (en) 2012-08-16 2012-08-16 User interface for entertainment systems
PCT/US2013/055259 WO2014028797A1 (en) 2012-08-16 2013-08-16 User interface for entertainment systems

Publications (1)

Publication Number Publication Date
EP2803004A1 true EP2803004A1 (en) 2014-11-19

Family

ID=49083777

Family Applications (1)

Application Number Title Priority Date Filing Date
EP13753963.1A Withdrawn EP2803004A1 (en) 2012-08-16 2013-08-16 User interface for entertainment systems

Country Status (4)

Country Link
EP (1) EP2803004A1 (en)
KR (1) KR20150045404A (en)
CN (2) CN104170397B (en)
WO (1) WO2014028797A1 (en)

Families Citing this family (168)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8645137B2 (en) 2000-03-16 2014-02-04 Apple Inc. Fast, language-independent method for user authentication by voice
US8677377B2 (en) 2005-09-08 2014-03-18 Apple Inc. Method and apparatus for building an intelligent automated assistant
US9318108B2 (en) 2010-01-18 2016-04-19 Apple Inc. Intelligent automated assistant
US8977255B2 (en) 2007-04-03 2015-03-10 Apple Inc. Method and system for operating a multi-function portable electronic device using voice-activation
US10002189B2 (en) 2007-12-20 2018-06-19 Apple Inc. Method and apparatus for searching using an active ontology
US9330720B2 (en) 2008-01-03 2016-05-03 Apple Inc. Methods and apparatus for altering audio output signals
US8996376B2 (en) 2008-04-05 2015-03-31 Apple Inc. Intelligent text-to-speech conversion
US20100030549A1 (en) 2008-07-31 2010-02-04 Lee Michael M Mobile device having human language translation capability with positional feedback
US8676904B2 (en) 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
US10706373B2 (en) 2011-06-03 2020-07-07 Apple Inc. Performing actions associated with task items that represent tasks to perform
US10241752B2 (en) 2011-09-30 2019-03-26 Apple Inc. Interface for a virtual digital assistant
US10241644B2 (en) 2011-06-03 2019-03-26 Apple Inc. Actionable reminder entries
US9858925B2 (en) 2009-06-05 2018-01-02 Apple Inc. Using context information to facilitate processing of commands in a virtual assistant
US9431006B2 (en) 2009-07-02 2016-08-30 Apple Inc. Methods and apparatuses for automatic speech recognition
US10276170B2 (en) 2010-01-18 2019-04-30 Apple Inc. Intelligent automated assistant
US8682667B2 (en) 2010-02-25 2014-03-25 Apple Inc. User profiling for selecting user specific voice input processing information
US9262612B2 (en) 2011-03-21 2016-02-16 Apple Inc. Device access using voice authentication
US10057736B2 (en) 2011-06-03 2018-08-21 Apple Inc. Active transport based notifications
US8994660B2 (en) 2011-08-29 2015-03-31 Apple Inc. Text correction processing
US10134385B2 (en) 2012-03-02 2018-11-20 Apple Inc. Systems and methods for name pronunciation
US9280610B2 (en) 2012-05-14 2016-03-08 Apple Inc. Crowd sourcing information to fulfill user requests
US10417037B2 (en) 2012-05-15 2019-09-17 Apple Inc. Systems and methods for integrating third party services with a digital assistant
US9721563B2 (en) 2012-06-08 2017-08-01 Apple Inc. Name recognition system
US9547647B2 (en) 2012-09-19 2017-01-17 Apple Inc. Voice-based media searching
JP2016508007A (en) 2013-02-07 2016-03-10 アップル インコーポレイテッド Voice trigger for digital assistant
US10652394B2 (en) 2013-03-14 2020-05-12 Apple Inc. System and method for processing voicemail
US10748529B1 (en) 2013-03-15 2020-08-18 Apple Inc. Voice activated device for use with a voice-based digital assistant
WO2014197334A2 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for user-specified pronunciation of words for speech synthesis and recognition
WO2014197336A1 (en) 2013-06-07 2014-12-11 Apple Inc. System and method for detecting errors in interactions with a voice-based digital assistant
US9582608B2 (en) 2013-06-07 2017-02-28 Apple Inc. Unified ranking with entropy-weighted information for phrase-based semantic auto-completion
WO2014197335A1 (en) 2013-06-08 2014-12-11 Apple Inc. Interpreting and acting upon commands that involve sharing information with remote devices
US10176167B2 (en) 2013-06-09 2019-01-08 Apple Inc. System and method for inferring user intent from speech inputs
CN110442699A (en) 2013-06-09 2019-11-12 苹果公司 Operate method, computer-readable medium, electronic equipment and the system of digital assistants
US10296160B2 (en) 2013-12-06 2019-05-21 Apple Inc. Method for extracting salient dialog usage from live data
US9715875B2 (en) 2014-05-30 2017-07-25 Apple Inc. Reducing the need for manual start/end-pointing and trigger phrases
US9966065B2 (en) 2014-05-30 2018-05-08 Apple Inc. Multi-command single utterance input method
US10170123B2 (en) 2014-05-30 2019-01-01 Apple Inc. Intelligent assistant for home automation
US9633004B2 (en) 2014-05-30 2017-04-25 Apple Inc. Better resolution when referencing to concepts
US9430463B2 (en) 2014-05-30 2016-08-30 Apple Inc. Exemplar-based natural language processing
US9842101B2 (en) 2014-05-30 2017-12-12 Apple Inc. Predictive conversion of language input
US10659851B2 (en) 2014-06-30 2020-05-19 Apple Inc. Real-time digital assistant knowledge updates
US9338493B2 (en) 2014-06-30 2016-05-10 Apple Inc. Intelligent automated assistant for TV user interactions
US9818400B2 (en) 2014-09-11 2017-11-14 Apple Inc. Method and apparatus for discovering trending terms in speech requests
US10789041B2 (en) 2014-09-12 2020-09-29 Apple Inc. Dynamic thresholds for always listening speech trigger
US9830321B2 (en) 2014-09-30 2017-11-28 Rovi Guides, Inc. Systems and methods for searching for a media asset
US10074360B2 (en) 2014-09-30 2018-09-11 Apple Inc. Providing an indication of the suitability of speech recognition
US9646609B2 (en) 2014-09-30 2017-05-09 Apple Inc. Caching apparatus for serving phonetic pronunciations
US9886432B2 (en) 2014-09-30 2018-02-06 Apple Inc. Parsimonious handling of word inflection via categorical stem + suffix N-gram language models
US10127911B2 (en) 2014-09-30 2018-11-13 Apple Inc. Speaker identification and unsupervised speaker adaptation techniques
US9668121B2 (en) 2014-09-30 2017-05-30 Apple Inc. Social reminders
US10460720B2 (en) 2015-01-03 2019-10-29 Microsoft Technology Licensing, Llc. Generation of language understanding systems and methods
US10152299B2 (en) 2015-03-06 2018-12-11 Apple Inc. Reducing response latency of intelligent automated assistants
US9865280B2 (en) 2015-03-06 2018-01-09 Apple Inc. Structured dictation using intelligent automated assistants
US9886953B2 (en) 2015-03-08 2018-02-06 Apple Inc. Virtual assistant activation
US10567477B2 (en) 2015-03-08 2020-02-18 Apple Inc. Virtual assistant continuity
US9721566B2 (en) 2015-03-08 2017-08-01 Apple Inc. Competing devices responding to voice triggers
US9899019B2 (en) 2015-03-18 2018-02-20 Apple Inc. Systems and methods for structured stem and suffix language models
US9842105B2 (en) 2015-04-16 2017-12-12 Apple Inc. Parsimonious continuous-space phrase representations for natural language processing
US10460227B2 (en) 2015-05-15 2019-10-29 Apple Inc. Virtual assistant in a communication session
US10083688B2 (en) 2015-05-27 2018-09-25 Apple Inc. Device voice control for selecting a displayed affordance
US10200824B2 (en) 2015-05-27 2019-02-05 Apple Inc. Systems and methods for proactively identifying and surfacing relevant content on a touch-sensitive device
US10127220B2 (en) 2015-06-04 2018-11-13 Apple Inc. Language identification from short strings
US10101822B2 (en) 2015-06-05 2018-10-16 Apple Inc. Language input correction
US9578173B2 (en) 2015-06-05 2017-02-21 Apple Inc. Virtual assistant aided communication with 3rd party service in a communication session
US11025565B2 (en) 2015-06-07 2021-06-01 Apple Inc. Personalized prediction of responses for instant messaging
US10255907B2 (en) 2015-06-07 2019-04-09 Apple Inc. Automatic accent detection using acoustic models
US10186254B2 (en) 2015-06-07 2019-01-22 Apple Inc. Context-based endpoint detection
US20160378747A1 (en) 2015-06-29 2016-12-29 Apple Inc. Virtual assistant for media playback
US11170060B2 (en) 2015-08-18 2021-11-09 Lg Electronics Inc. Mobile terminal and control method thereof
US10671428B2 (en) 2015-09-08 2020-06-02 Apple Inc. Distributed personal assistant
DK179690B1 (en) * 2015-09-08 2019-03-26 Apple Inc. Intelligent automated media search and playback assistant
US10740384B2 (en) 2015-09-08 2020-08-11 Apple Inc. Intelligent automated assistant for media search and playback
US10747498B2 (en) 2015-09-08 2020-08-18 Apple Inc. Zero latency digital assistant
US10331312B2 (en) * 2015-09-08 2019-06-25 Apple Inc. Intelligent automated assistant in a media environment
US9697820B2 (en) 2015-09-24 2017-07-04 Apple Inc. Unit-selection text-to-speech synthesis using concatenation-sensitive neural networks
US10366158B2 (en) 2015-09-29 2019-07-30 Apple Inc. Efficient word encoding for recurrent neural network language models
US11010550B2 (en) 2015-09-29 2021-05-18 Apple Inc. Unified language modeling framework for word prediction, auto-completion and auto-correction
US11587559B2 (en) 2015-09-30 2023-02-21 Apple Inc. Intelligent device identification
US10691473B2 (en) 2015-11-06 2020-06-23 Apple Inc. Intelligent automated assistant in a messaging environment
US10956666B2 (en) 2015-11-09 2021-03-23 Apple Inc. Unconventional virtual assistant interactions
US10049668B2 (en) 2015-12-02 2018-08-14 Apple Inc. Applying neural network language models to weighted finite state transducers for automatic speech recognition
US10223066B2 (en) 2015-12-23 2019-03-05 Apple Inc. Proactive assistance based on dialog communication between devices
US10446143B2 (en) 2016-03-14 2019-10-15 Apple Inc. Identification of voice inputs providing credentials
US9934775B2 (en) 2016-05-26 2018-04-03 Apple Inc. Unit-selection text-to-speech synthesis based on predicted concatenation parameters
US9972304B2 (en) 2016-06-03 2018-05-15 Apple Inc. Privacy preserving distributed evaluation framework for embedded personalized systems
US10249300B2 (en) 2016-06-06 2019-04-02 Apple Inc. Intelligent list reading
US11227589B2 (en) 2016-06-06 2022-01-18 Apple Inc. Intelligent list reading
US10049663B2 (en) 2016-06-08 2018-08-14 Apple, Inc. Intelligent automated assistant for media exploration
DK179309B1 (en) 2016-06-09 2018-04-23 Apple Inc Intelligent automated assistant in a home environment
US10490187B2 (en) 2016-06-10 2019-11-26 Apple Inc. Digital assistant providing automated status report
US10509862B2 (en) 2016-06-10 2019-12-17 Apple Inc. Dynamic phrase expansion of language input
US10586535B2 (en) 2016-06-10 2020-03-10 Apple Inc. Intelligent digital assistant in a multi-tasking environment
US10192552B2 (en) 2016-06-10 2019-01-29 Apple Inc. Digital assistant providing whispered speech
US10067938B2 (en) 2016-06-10 2018-09-04 Apple Inc. Multilingual word prediction
DK179415B1 (en) 2016-06-11 2018-06-14 Apple Inc Intelligent device arbitration and control
DK179049B1 (en) 2016-06-11 2017-09-18 Apple Inc Data driven natural language event detection and classification
DK179343B1 (en) 2016-06-11 2018-05-14 Apple Inc Intelligent task discovery
DK201670540A1 (en) 2016-06-11 2018-01-08 Apple Inc Application integration with a digital assistant
JP6834256B2 (en) * 2016-08-30 2021-02-24 株式会社リコー Information processing equipment, information processing systems, and programs
US10474753B2 (en) 2016-09-07 2019-11-12 Apple Inc. Language identification using recurrent neural networks
US10043516B2 (en) 2016-09-23 2018-08-07 Apple Inc. Intelligent automated assistant
US11281993B2 (en) 2016-12-05 2022-03-22 Apple Inc. Model and ensemble compression for metric learning
US10593346B2 (en) 2016-12-22 2020-03-17 Apple Inc. Rank-reduced token representation for automatic speech recognition
US11204787B2 (en) 2017-01-09 2021-12-21 Apple Inc. Application integration with a digital assistant
CN107093426A (en) * 2017-04-26 2017-08-25 医惠科技有限公司 The input method of voice, apparatus and system
US10417266B2 (en) 2017-05-09 2019-09-17 Apple Inc. Context-aware ranking of intelligent response suggestions
DK201770383A1 (en) 2017-05-09 2018-12-14 Apple Inc. User interface for correcting recognition errors
US10726832B2 (en) 2017-05-11 2020-07-28 Apple Inc. Maintaining privacy of personal information
US10395654B2 (en) 2017-05-11 2019-08-27 Apple Inc. Text normalization based on a data-driven learning network
DK201770439A1 (en) 2017-05-11 2018-12-13 Apple Inc. Offline personal assistant
DK180048B1 (en) 2017-05-11 2020-02-04 Apple Inc. MAINTAINING THE DATA PROTECTION OF PERSONAL INFORMATION
DK179496B1 (en) 2017-05-12 2019-01-15 Apple Inc. USER-SPECIFIC Acoustic Models
DK201770429A1 (en) 2017-05-12 2018-12-14 Apple Inc. Low-latency intelligent automated assistant
US11301477B2 (en) 2017-05-12 2022-04-12 Apple Inc. Feedback analysis of a digital assistant
DK179745B1 (en) 2017-05-12 2019-05-01 Apple Inc. SYNCHRONIZATION AND TASK DELEGATION OF A DIGITAL ASSISTANT
DK201770432A1 (en) 2017-05-15 2018-12-21 Apple Inc. Hierarchical belief states for digital assistants
DK201770431A1 (en) 2017-05-15 2018-12-20 Apple Inc. Optimizing dialogue policy decisions for digital assistants using implicit feedback
US20180336892A1 (en) 2017-05-16 2018-11-22 Apple Inc. Detecting a trigger of a digital assistant
US10303715B2 (en) 2017-05-16 2019-05-28 Apple Inc. Intelligent automated assistant for media exploration
US10403278B2 (en) 2017-05-16 2019-09-03 Apple Inc. Methods and systems for phonetic matching in digital assistant services
US10311144B2 (en) 2017-05-16 2019-06-04 Apple Inc. Emoji word sense disambiguation
DK179549B1 (en) 2017-05-16 2019-02-12 Apple Inc. Far-field extension for digital assistant services
US10657328B2 (en) 2017-06-02 2020-05-19 Apple Inc. Multi-task recurrent neural network architecture for efficient morphology handling in neural language modeling
US10445429B2 (en) 2017-09-21 2019-10-15 Apple Inc. Natural language understanding using vocabularies with compressed serialized tries
US10755051B2 (en) 2017-09-29 2020-08-25 Apple Inc. Rule-based natural language processing
US10636424B2 (en) 2017-11-30 2020-04-28 Apple Inc. Multi-turn canned dialog
US10733982B2 (en) 2018-01-08 2020-08-04 Apple Inc. Multi-directional dialog
US10733375B2 (en) 2018-01-31 2020-08-04 Apple Inc. Knowledge-based framework for improving natural language understanding
US10789959B2 (en) 2018-03-02 2020-09-29 Apple Inc. Training speaker recognition models for digital assistants
US10592604B2 (en) 2018-03-12 2020-03-17 Apple Inc. Inverse text normalization for automatic speech recognition
US10818288B2 (en) 2018-03-26 2020-10-27 Apple Inc. Natural assistant interaction
US10909331B2 (en) 2018-03-30 2021-02-02 Apple Inc. Implicit identification of translation payload with neural machine translation
US10928918B2 (en) 2018-05-07 2021-02-23 Apple Inc. Raise to speak
US11145294B2 (en) 2018-05-07 2021-10-12 Apple Inc. Intelligent automated assistant for delivering content from user experiences
US10984780B2 (en) 2018-05-21 2021-04-20 Apple Inc. Global semantic word embeddings using bi-directional recurrent neural networks
DK201870355A1 (en) 2018-06-01 2019-12-16 Apple Inc. Virtual assistant operation in multi-device environments
US11386266B2 (en) 2018-06-01 2022-07-12 Apple Inc. Text correction
DK180639B1 (en) 2018-06-01 2021-11-04 Apple Inc DISABILITY OF ATTENTION-ATTENTIVE VIRTUAL ASSISTANT
DK179822B1 (en) 2018-06-01 2019-07-12 Apple Inc. Voice interaction at a primary device to access call functionality of a companion device
US10892996B2 (en) 2018-06-01 2021-01-12 Apple Inc. Variable latency device coordination
US10496705B1 (en) 2018-06-03 2019-12-03 Apple Inc. Accelerated task performance
CN109104631B (en) * 2018-09-06 2021-09-28 网易传媒科技(北京)有限公司 Video display method and device
US11010561B2 (en) 2018-09-27 2021-05-18 Apple Inc. Sentiment prediction from textual data
US10839159B2 (en) 2018-09-28 2020-11-17 Apple Inc. Named entity normalization in a spoken dialog system
US11462215B2 (en) 2018-09-28 2022-10-04 Apple Inc. Multi-modal inputs for voice commands
US11170166B2 (en) 2018-09-28 2021-11-09 Apple Inc. Neural typographical error modeling via generative adversarial networks
CN109147801B (en) * 2018-09-30 2021-08-10 深圳市元征科技股份有限公司 Voice interaction method, system, terminal and storage medium
US11475898B2 (en) 2018-10-26 2022-10-18 Apple Inc. Low-latency multi-speaker speech recognition
US11638059B2 (en) 2019-01-04 2023-04-25 Apple Inc. Content playback on multiple devices
US11348573B2 (en) 2019-03-18 2022-05-31 Apple Inc. Multimodality in digital assistant systems
US11475884B2 (en) 2019-05-06 2022-10-18 Apple Inc. Reducing digital assistant latency when a language is incorrectly determined
US11423908B2 (en) 2019-05-06 2022-08-23 Apple Inc. Interpreting spoken requests
DK201970509A1 (en) 2019-05-06 2021-01-15 Apple Inc Spoken notifications
US11307752B2 (en) 2019-05-06 2022-04-19 Apple Inc. User configurable task triggers
US11140099B2 (en) 2019-05-21 2021-10-05 Apple Inc. Providing message response suggestions
US11289073B2 (en) 2019-05-31 2022-03-29 Apple Inc. Device text to speech
US11496600B2 (en) 2019-05-31 2022-11-08 Apple Inc. Remote execution of machine-learned models
DK180129B1 (en) 2019-05-31 2020-06-02 Apple Inc. User activity shortcut suggestions
DK201970511A1 (en) 2019-05-31 2021-02-15 Apple Inc Voice identification in digital assistant systems
US11360641B2 (en) 2019-06-01 2022-06-14 Apple Inc. Increasing the relevance of new available information
US11468890B2 (en) 2019-06-01 2022-10-11 Apple Inc. Methods and user interfaces for voice-based control of electronic devices
US11488406B2 (en) 2019-09-25 2022-11-01 Apple Inc. Text detection using global geometry estimators
US11061543B1 (en) 2020-05-11 2021-07-13 Apple Inc. Providing relevant data items based on context
US11183193B1 (en) 2020-05-11 2021-11-23 Apple Inc. Digital assistant hardware abstraction
US11755276B2 (en) 2020-05-12 2023-09-12 Apple Inc. Reducing description length based on confidence
US11490204B2 (en) 2020-07-20 2022-11-01 Apple Inc. Multi-device audio adjustment coordination
US11438683B2 (en) 2020-07-21 2022-09-06 Apple Inc. User identification using headphones
CN113301395B (en) * 2021-04-30 2023-07-07 当趣网络科技(杭州)有限公司 Voice searching method combined with user grade in video playing state

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2001309256A (en) * 2000-04-26 2001-11-02 Sanyo Electric Co Ltd Receiver of digital tv broadcasting
JP4923604B2 (en) * 2006-02-13 2012-04-25 ソニー株式会社 Information processing apparatus and method, and program
US8200688B2 (en) * 2006-03-07 2012-06-12 Samsung Electronics Co., Ltd. Method and system for facilitating information searching on electronic devices
JP4892601B2 (en) * 2009-10-30 2012-03-07 株式会社ソニー・コンピュータエンタテインメント Information processing apparatus, tuner, and information processing method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
None *
See also references of WO2014028797A1 *

Also Published As

Publication number Publication date
CN104170397A (en) 2014-11-26
CN107613353B (en) 2020-10-16
WO2014028797A1 (en) 2014-02-20
CN104170397B (en) 2017-10-24
CN107613353A (en) 2018-01-19
KR20150045404A (en) 2015-04-28

Similar Documents

Publication Publication Date Title
US9066150B2 (en) User interface for entertainment systems
US9031848B2 (en) User interface for searching a bundled service content data source
US9497515B2 (en) User interface for entertainment systems
US9026448B2 (en) User interface for entertainment systems
CN107613353B (en) Method for presenting search results on electronic device, electronic device and computer storage medium
US9106957B2 (en) Method and apparatus for searching data sources for entertainment systems
US20140052450A1 (en) User interface for entertainment systems
US11838579B2 (en) Intelligent automated assistant for TV user interactions
US8634944B2 (en) Auto-station tuning
US20140195244A1 (en) Display apparatus and method of controlling display apparatus
US20140006022A1 (en) Display apparatus, method for controlling display apparatus, and interactive system
KR20190019041A (en) Approximate Template Matching for Natural Language Queries
US9544528B2 (en) Matrix search of video using closed caption information
JP2013085232A (en) Method and device for providing integrated epg information of plurality of input sources
TWI587253B (en) Method and apparatus for providing notice of availability of audio description
KR20160003714A (en) Providing correlated programming information for broadcast media content and streaming media content
US10555044B2 (en) Systems and methods for control of channel surfing

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20140815

AK Designated contracting states

Kind code of ref document: A1

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

DAX Request for extension of the european patent (deleted)
STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: EXAMINATION IS IN PROGRESS

17Q First examination report despatched

Effective date: 20190125

GRAP Despatch of communication of intention to grant a patent

Free format text: ORIGINAL CODE: EPIDOSNIGR1

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: GRANT OF PATENT IS INTENDED

RIC1 Information provided on ipc code assigned before grant

Ipc: H04N 21/422 20110101ALN20200122BHEP

Ipc: G06F 16/48 20190101ALN20200122BHEP

Ipc: G06F 16/43 20190101ALN20200122BHEP

Ipc: G10L 15/18 20130101ALI20200122BHEP

Ipc: G10L 15/22 20060101ALN20200122BHEP

Ipc: H04N 5/44 20110101ALN20200122BHEP

Ipc: H04N 21/482 20110101ALN20200122BHEP

Ipc: H04N 5/445 20110101AFI20200122BHEP

Ipc: G10L 15/08 20060101ALI20200122BHEP

Ipc: G06F 16/783 20190101ALN20200122BHEP

Ipc: H04N 21/462 20110101ALN20200122BHEP

Ipc: G06F 16/732 20190101ALN20200122BHEP

Ipc: G10L 15/26 20060101ALN20200122BHEP

Ipc: H04H 60/72 20080101ALN20200122BHEP

Ipc: G10L 15/10 20060101ALI20200122BHEP

RIC1 Information provided on ipc code assigned before grant

Ipc: H04N 5/44 20110101ALN20200128BHEP

Ipc: H04H 60/72 20080101ALN20200128BHEP

Ipc: H04N 21/462 20110101ALN20200128BHEP

Ipc: G06F 16/48 20190101ALN20200128BHEP

Ipc: G10L 15/10 20060101ALI20200128BHEP

Ipc: G10L 15/08 20060101ALI20200128BHEP

Ipc: G10L 15/22 20060101ALN20200128BHEP

Ipc: H04N 21/422 20110101ALN20200128BHEP

Ipc: G06F 16/732 20190101ALN20200128BHEP

Ipc: G06F 16/43 20190101ALN20200128BHEP

Ipc: G10L 15/26 20060101ALN20200128BHEP

Ipc: G10L 15/18 20130101ALI20200128BHEP

Ipc: H04N 21/482 20110101ALN20200128BHEP

Ipc: G06F 16/783 20190101ALN20200128BHEP

Ipc: H04N 5/445 20110101AFI20200128BHEP

INTG Intention to grant announced

Effective date: 20200219

RIC1 Information provided on ipc code assigned before grant

Ipc: H04N 5/445 20110101AFI20200211BHEP

Ipc: H04N 5/44 20110101ALN20200211BHEP

Ipc: H04H 60/72 20080101ALN20200211BHEP

Ipc: G10L 15/10 20060101ALI20200211BHEP

Ipc: G06F 16/783 20190101ALN20200211BHEP

Ipc: G06F 16/48 20190101ALN20200211BHEP

Ipc: G06F 16/43 20190101ALN20200211BHEP

Ipc: H04N 21/462 20110101ALN20200211BHEP

Ipc: H04N 21/482 20110101ALN20200211BHEP

Ipc: G10L 15/26 20060101ALN20200211BHEP

Ipc: G06F 16/732 20190101ALN20200211BHEP

Ipc: G10L 15/08 20060101ALI20200211BHEP

Ipc: H04N 21/422 20110101ALN20200211BHEP

Ipc: G10L 15/22 20060101ALN20200211BHEP

Ipc: G10L 15/18 20130101ALI20200211BHEP

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20200630