WO2014103568A1 - 情報処理装置、情報処理方法、およびプログラム - Google Patents
情報処理装置、情報処理方法、およびプログラム Download PDFInfo
- Publication number
- WO2014103568A1 WO2014103568A1 PCT/JP2013/081312 JP2013081312W WO2014103568A1 WO 2014103568 A1 WO2014103568 A1 WO 2014103568A1 JP 2013081312 W JP2013081312 W JP 2013081312W WO 2014103568 A1 WO2014103568 A1 WO 2014103568A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- information processing
- utterance content
- processing apparatus
- character string
- present
- Prior art date
Links
- 230000010365 information processing Effects 0.000 title claims abstract description 372
- 238000003672 processing method Methods 0.000 title claims description 64
- 238000004458 analytical method Methods 0.000 claims abstract description 217
- 238000012545 processing Methods 0.000 claims description 187
- 238000000034 method Methods 0.000 description 259
- 230000008569 process Effects 0.000 description 247
- 238000004891 communication Methods 0.000 description 33
- 230000000877 morphologic effect Effects 0.000 description 18
- 230000006870 function Effects 0.000 description 16
- 230000009471 action Effects 0.000 description 12
- 230000005540 biological transmission Effects 0.000 description 12
- 238000010586 diagram Methods 0.000 description 10
- 230000006399 behavior Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000007619 statistical method Methods 0.000 description 3
- 230000000007 visual effect Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 239000000470 constituent Substances 0.000 description 2
- 238000004590 computer program Methods 0.000 description 1
- 238000012937 correction Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005401 electroluminescence Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/166—Editing, e.g. inserting or deleting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/237—Lexical tools
- G06F40/242—Dictionaries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/268—Morphological analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/30—Semantic analysis
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/221—Announcement of recognition results
Definitions
- the present disclosure relates to an information processing apparatus, an information processing method, and a program.
- a device that can recognize and process voice has appeared.
- a technique using voice input has also been developed.
- a technique for using voice input and other input in combination for example, a technique described in Patent Document 1 below can be cited.
- existing device capable of performing processing by recognizing voice
- existing device including the case where the technique described in Patent Document 1 is used
- utterance is performed.
- the user cannot know which part of the content has been recognized and processed. Therefore, when an existing device is used, there may occur a situation where the convenience of the user may be lowered, for example, “the subsequent behavior due to voice utterance in the device is not sufficiently transmitted to the user”.
- This disclosure proposes a new and improved information processing apparatus, information processing method, and program capable of improving user convenience.
- an analysis unit that analyzes a character string indicating utterance content obtained as a result of speech recognition, and a display control unit that displays the character string indicating the utterance content and the analysis result on a display screen.
- An information processing apparatus is provided.
- this indication has the step which analyzes the character string which shows the utterance content obtained as a result of speech recognition, and the step which displays the character string which shows the said utterance content, and the result of an analysis on a display screen An information processing method is provided.
- the computer executes a step of analyzing a character string indicating the utterance content obtained as a result of the speech recognition, and a step of displaying the character string indicating the utterance content and the analysis result on a display screen.
- a program is provided.
- the information processing apparatus analyzes a character string of the utterance content obtained as a result of the speech recognition (hereinafter, simply referred to as “character string of the utterance content”) (analysis process). Then, the information processing apparatus according to the present embodiment displays the character string of the utterance content and the analysis result on the display screen (display control process).
- the information processing apparatus obtains a character string of utterance content to be processed by, for example, performing processing related to speech recognition on voice data indicating the utterance content.
- the acquisition method of the character string of the utterance content to be processed is not limited to the above.
- the information processing apparatus performs processing by acquiring data indicating a character string of the utterance content obtained as a result of processing related to speech recognition in an external apparatus such as a server from the external apparatus. You may obtain the character string of the utterance content used as object.
- examples of the process related to speech recognition according to the present embodiment include a process using a statistical method, a process using a hidden Markov model, and the process related to the speech recognition according to the present embodiment. It is not limited to the above.
- the information processing apparatus according to the present embodiment or an external device such as a server can use any process that can obtain a character string of the utterance content from voice data indicating the utterance content.
- the display screen on which the information processing apparatus according to the present embodiment displays the character string of the utterance content and the analysis result for example, a display unit (described later) included in the own apparatus (the information processing apparatus according to the present embodiment)
- the display screen on which the information processing apparatus according to the present embodiment displays the character string of the utterance content and the analysis result is not limited to the above.
- the information processing apparatus according to the present embodiment may display the character string of the utterance content and the analysis result on the display screen of a display device provided in an external apparatus connected by wire or wirelessly.
- the information processing apparatus displays the character string of the utterance content obtained as a result of the voice recognition on the display screen, so that the user who made the utterance visually determines how the utterance has been voice-recognized. I can grasp it.
- the information processing apparatus displays the result of the analysis on the character string of the utterance content together with the character string of the utterance content, so that the user who performed the utterance is analyzed how the utterance is analyzed. Can be visually grasped. Therefore, the user who made the utterance processes the device (for example, the information processing apparatus according to the present embodiment or the apparatus whose operation is controlled by the information processing apparatus according to the present embodiment) performed after the utterance. However, it is possible to grasp which part of the utterance content is recognized and performed.
- the information processing apparatus performs (1) analysis processing and (2) display control processing as processing related to the information processing method according to the present embodiment, thereby improving user convenience. Can be achieved.
- FIG. 1 is an explanatory diagram for explaining an overview of processing related to the information processing method according to the present embodiment, and illustrates an example of a screen displayed on the display screen by the information processing apparatus according to the present embodiment.
- FIG. 1 shows an example of an application selection screen for a television receiver
- A1 shown in FIG. 1 shows an example of an icon for selecting an application for the user to perform voice recognition. Is shown.
- FIG. 1A for example, when the icon shown in A1 in FIG. 1 is selected by a user operation, execution of an application for allowing the user to perform voice recognition, for example, as shown in FIG. A screen is displayed.
- the information processing apparatus When the user utters a voice input device such as a microphone on the screen shown in FIG. 1B, the information processing apparatus according to the present embodiment performs the present implementation based on data indicating the character string of the utterance content.
- the process (1) analysis process
- the information processing apparatus performs the process (2) (display control process) related to the information processing method according to the present embodiment, thereby, for example, the utterance content as illustrated in C of FIG. A screen that displays the character string and the analysis result together is displayed.
- FIG. 1C shows an example of a screen displayed when the user speaks “I want to see tomorrow's XXX”.
- the information processing apparatus displays, for example, a character string indicating the utterance content “I want to see tomorrow's XXX”, and further displays the analysis result.
- FIG. 1C shows an example in which the information processing apparatus according to the present embodiment highlights the result of the analysis of the character string indicating the utterance content with quotation marks, but the result of the analysis according to the present embodiment.
- the display method is not limited to the above.
- the information processing apparatus according to the present embodiment can change the utterance content by any method capable of giving a visual change to the character, such as underlining the character or changing the color or font of the character.
- the analysis result of the character string to be shown may be highlighted.
- the information processing apparatus performs the process (2) (display control process) according to the information processing method according to the present embodiment, for example, as illustrated by C2 in FIG.
- An application corresponding to the utterance content and / or a service corresponding to the utterance content may be displayed on the display screen.
- an application corresponding to the utterance content for example, an application for displaying a television broadcast program guide, an application for viewing a television broadcast, or a list of recorded contents is displayed.
- applications such as an application, an application related to reproduction of content stored in a recording medium, and an application for connecting to a network such as a home network.
- a service corresponding to the utterance content according to the present embodiment there are various services such as a video service, a music service, a service provided on an image posting website, a search service provided on a search site, and the like. Can be mentioned.
- FIG. 1D a screen corresponding to the selected application or the selected service is displayed as shown in FIG. 1D, for example.
- the information processing apparatus displays a character string indicating the utterance content and the analysis result on the display screen. Therefore, for example, the user who has seen the screen shown in FIG. 1C can visually grasp how the speech is recognized by the character string indicating the utterance content. By highlighting, it is possible to visually grasp how the utterance was analyzed.
- the information processing apparatus can improve user convenience.
- the screen displayed by the information processing apparatus according to the present embodiment by the processing according to the information processing method according to the present embodiment is not limited to the example illustrated in FIG. 1C, and the information processing apparatus according to the present embodiment. Needless to say, the screen displayed by is not limited to the examples shown in A of FIG. 1, B of FIG. 1, and D of FIG.
- the information processing apparatus includes, for example, the process (1) (analysis process) and the process (2) (display control) as the process related to the information processing method according to the present embodiment. Process).
- the information processing apparatus analyzes a character string indicating the utterance content obtained as a result of speech recognition.
- the information processing apparatus performs a morphological analysis using, for example, a syntax analysis dictionary stored in a storage unit (described later) or an external recording medium, thereby generating a character string indicating the utterance content. Divide into words such as nouns and verbs.
- the dictionary according to the present embodiment may be, for example, database format data or table format data.
- the information processing apparatus further uses, for example, a semantic analysis dictionary stored in a storage unit (to be described later) or an external recording medium to make a meaning for a character string indicating the utterance content. Analysis (or intention analysis) may be performed.
- the information processing apparatus indicates the utterance content based on the result of the syntax analysis for the character string indicating the utterance content or the result of the syntax analysis and the result of the semantic analysis for the character string indicating the utterance content.
- a score may be set for the character string.
- the information processing apparatus sets a score for each word divided by, for example, morphological analysis. More specifically, the information processing apparatus according to the present embodiment is divided by, for example, setting a reference score for each divided word and adjusting the score set for each divided word. Set a score corresponding to each word. An example of processing related to the score setting according to the present embodiment will be described later.
- the score set in the process (1) (analysis process) is used, for example, in an execution process according to this embodiment described later.
- An example of processing related to the score setting according to the present embodiment and an example of processing using the set score will be described later.
- the information processing apparatus may analyze the operation on a character string indicating the utterance content, for example.
- the information processing apparatus uses, for example, a motion analysis dictionary stored in a storage unit (described later) or an external recording medium to analyze motion in words divided by morphological analysis.
- the action is analyzed for the character string indicating the utterance content by determining whether or not the word indicating the action registered in the dictionary is included.
- the analysis result of the operation in the process (1) is used in the process (2) (display control process), for example.
- An example of a process in which the analysis result of the operation in the process (2) (display control process) is used will be described later.
- the information processing apparatus may analyze the tense with respect to a character string indicating the utterance content, for example.
- the information processing apparatus according to the present embodiment uses, for example, a tense analysis in words divided by morphological analysis using a tense analysis dictionary stored in a storage unit (described later) or an external recording medium.
- the tense is analyzed with respect to the character string indicating the utterance content by determining whether or not the word indicating the tense registered in the dictionary is included.
- the tense analysis result in the process (1) (analysis process) is used, for example, in an execution process according to the present embodiment described later.
- An example of processing using the tense analysis result in the execution processing according to the present embodiment to be described later will be described later.
- the fluctuation of speech recognition according to the present embodiment is, for example, divided when “a word expressed in a specific kanji is expressed in hiragana” or “uttered by an abbreviation rather than an official name”.
- the abbreviation should be the abbreviation.
- the information processing apparatus may correct fluctuations in voice recognition in order to prevent the normal processing from being performed as described above, for example.
- the information processing apparatus according to the present embodiment uses, for example, a fluctuation absorbing dictionary stored in a storage unit (described later) or an external recording medium, and replaces words divided by morphological analysis or the like. Correct fluctuations in voice recognition.
- the information processing apparatus corrects fluctuations in voice recognition in the process (analysis process) (1), the screen displayed in the process (display control process) (2) A character string indicating the utterance content in which the fluctuation of recognition is corrected is displayed.
- the information processing apparatus performs, for example, one or more of the processes as described above as the analysis process according to the present embodiment. Needless to say, the analysis process according to the present embodiment is not limited to the example of the process described above.
- the information processing apparatus displays a character string indicating the utterance content and the result of analysis in the process (1) (analysis process) on the display screen.
- the information processing apparatus displays a character string indicating the utterance content and the analysis result so that the analysis result is highlighted, for example, as indicated by C1 in FIG. .
- the information processing apparatus may display, for example, a character string indicating the utterance content displayed on the display screen in a state where it can be edited.
- the information processing apparatus makes it possible to edit a character string, for example, for each word divided by morphological analysis or the like in the entire character string indicating the utterance content or the character string indicating the utterance content.
- the information processing apparatus can edit a character string indicating the utterance content by the user, for example, by an input operation using a virtual input device such as a keyboard or an input device such as a software keyboard.
- a virtual input device such as a keyboard
- an input device such as a software keyboard.
- the editing method of the character string which shows the speech content based on this embodiment is not restricted above.
- the information processing apparatus according to the present embodiment may be able to edit a character string indicating the content of the utterance by the user by the utterance by the user.
- the information processing apparatus displays the character string indicating the utterance content on the display screen, so that the user who made the utterance visually recognizes the erroneous recognition in the voice recognition. I can grasp it.
- the information processing apparatus displays the character string indicating the utterance content to be displayed on the display screen in a state where the character string can be edited. Even if there is an error, the user can correct the error part and cause the information processing apparatus (or external apparatus) according to the present embodiment to perform, for example, an execution process described later.
- the information processing apparatus can prompt the user to correct the utterance content, for example.
- High processing can be performed by the own apparatus (information processing apparatus according to the present embodiment) or an external apparatus.
- the user is prevented from, for example, not realizing an operation desired by the user unless the user utters the utterance many times. Therefore, user convenience can be further improved.
- the information processing apparatus includes an application corresponding to the utterance content based on the analysis result in the process (1) (analysis process), for example, as indicated by C2 in FIG.
- a service corresponding to the utterance content may be further displayed on the display screen.
- the information processing apparatus determines, for example, application candidates to be displayed on the display screen and / or service candidates to be displayed on the display screen based on capability information.
- the information processing apparatus for example, selects an application and a service corresponding to the analysis result in the process (analysis process) of (1) from the determined application candidates and / or service candidates. To display the selected application and service on the display screen.
- the capability information according to the present embodiment includes, for example, data indicating the capability of the device (for example, data indicating the function of the device, data indicating the capability of each function, etc.).
- the capability information used by the information processing apparatus according to the present embodiment for processing for example, the capability information of the own apparatus (the information processing apparatus according to the present embodiment) or the information processing apparatus according to the present embodiment functions.
- An external device that performs control may be mentioned.
- a device corresponding to capability information used for processing by the information processing apparatus according to the present embodiment, that is, a target device for which the information processing apparatus according to the present embodiment controls the function is referred to as a “control target device”. There is a case.
- the information processing apparatus displays, for example, a display screen based on capability information from a list of applications and / or services corresponding to the analysis result in the process (1) (analysis process). Candidates for applications to be displayed on the screen and / or candidates for services to be displayed on the display screen may be determined. In the above case, the information processing apparatus according to the present embodiment displays the determined application candidates and service candidates on the display screen.
- the information processing apparatus when displaying an application corresponding to the utterance content and / or a service corresponding to the utterance content, the information processing apparatus according to the present embodiment includes an application, The service may be highlighted.
- the information processing apparatus assigns an icon (for example, a microphone icon) indicating an application or service related to the analysis result of the utterance content, attaches an icon indicating a recommendation,
- the application or service may be highlighted and displayed by any method capable of giving a visual change to the application or service, such as changing the color or font.
- the information processing apparatus when an operation analysis is performed on a character string indicating the utterance content in the process (1) (analysis process), the information processing apparatus according to the present embodiment performs an operation included in the analysis result. Based on the analysis results, the applications and services to be displayed on the display screen are highlighted.
- the information processing apparatus when the analysis result of the operation on the character string indicating the utterance content indicates a viewing operation such as “I want to see”, the information processing apparatus according to the present embodiment, for example, a television broadcast program In an application for displaying a table, an application for viewing a television broadcast, an application for displaying a list of recorded content, an application for playing back content stored in a recording medium, a video service, or an image posting website Highlight applications and services related to viewing, such as services provided.
- the analysis result of the operation on the character string indicating the utterance content indicates a listening operation such as “I want to hear”
- the information processing apparatus stores the content stored in the recording medium.
- Applications and services related to reproduction of sound such as applications related to reproduction and music services are highlighted.
- the information processing apparatus sets, for example, a search service provided on a search site or the like You may highlight highlighted applications and services.
- the information processing apparatus highlights and displays applications and services to be displayed on the display screen.
- processing related to highlighting of applications and services according to the present embodiment is not limited to the above.
- the information processing apparatus according to the present embodiment has a higher priority among the applications and services to be highlighted.
- Applications and services may be displayed with more emphasis.
- the information processing apparatus performs, for example, the above process as the display control process according to the present embodiment.
- the display control process according to the present embodiment is not limited to the example of the process described above.
- the information processing apparatus performs, for example, the above-described processing (1) (analysis processing) and (2) processing (display control processing) as processing related to the information processing method according to the present embodiment. I do.
- the processing according to the information processing method according to the present embodiment is not limited to the processing (1) (analysis processing) and the processing (2) (display control processing).
- the information processing apparatus includes an application corresponding to the utterance content selected on a screen on which an application corresponding to the utterance content or a service corresponding to the utterance content is displayed, such as the screen illustrated in FIG.
- processing related to the service corresponding to the utterance content may be executed (execution processing).
- the information processing apparatus performs, for example, the following process (a), process (b), or process (c) shown in FIG. Do as.
- process (a), process (b), or process (c) shown in FIG. Do A specific example of the execution process according to the present embodiment will be described later.
- (A) First example of execution process according to this embodiment For example, when a score is set for a character string indicating the utterance content in the process (1) (analysis process) described above, Based on the comparison result between the set reference value and the set score, the information processing apparatus according to the present invention executes part or all of the character string indicating the utterance content, execution of an application corresponding to the utterance content, It is determined as a character string used for execution of processing related to the service corresponding to the utterance content. The information processing apparatus according to the present embodiment uses the determined character string to execute processing related to the application corresponding to the utterance content and the service corresponding to the utterance content.
- the reference value according to the present embodiment may be, for example, a fixed value set in advance, or a variable value that can be changed based on a user operation or the like.
- the information when there is no score larger than the reference value (or a score equal to or higher than the reference value; hereinafter the same) in the character string indicating the utterance content, the information according to the present embodiment.
- the processing device uses the entire character string indicating the utterance content for processing. Further, for example, when a character string indicating the utterance content has a score larger than the reference value, the information processing apparatus according to the present embodiment has 1 or 2 having a large score among the character strings indicating the utterance content.
- the partial character string described above is used for processing.
- the information processing apparatus is, for example, a process that combines the processing according to the first example and the processing according to the second example. It is also possible to perform.
- the information processing apparatus performs, for example, the process (a), the process (b), or the process (c) described above as the execution process according to the present embodiment.
- processing according to the information processing method according to the present embodiment includes “the processing (1) (analysis processing) and (2) (display control processing)” and “the processing (1) (analysis)”. Process), the process (2) (display control process), and the execution process ".
- the information processing apparatus may record the analysis result in the process (1) (analysis process) as a log (log recording control process).
- the information to be recorded by the information processing apparatus includes, for example, a storage unit (described later), an external recording medium connected to the information processing apparatus according to the present embodiment, and a network. And (or directly) an external device capable of communication.
- a network for example, a wired network such as a LAN (Local Area Network) or a WAN (Wide Area Network), a wireless LAN (WLAN: Wireless Local Area Network) or a wireless WAN via a base station (WWAN: Examples include a wireless network such as Wireless Wide Area Network, or the Internet using a communication protocol such as TCP / IP (Transmission Control Protocol / Internet Protocol).
- An apparatus for example, an information processing apparatus according to the present embodiment or an external apparatus such as a server
- An apparatus that performs processing related to speech recognition by causing the information processing apparatus according to the present embodiment to record an analysis result as a log
- the log can be used, for example, for improving the accuracy of voice recognition or improving the behavior of an application.
- FIG. 2 is a flowchart illustrating a first example of processing according to the information processing method according to this embodiment.
- the processing of steps S104 and S106 shown in FIG. 2 corresponds to the processing (analysis processing) of (1) above, and the processing of step S108 shown in FIG. Processing).
- FIG. 2 illustrates an example in which the information processing apparatus according to the present embodiment performs processing related to speech recognition for speech data indicating the utterance content. Note that, as described above, processing related to speech recognition for speech data indicating the utterance content may be performed in an external device such as a server.
- the information processing apparatus determines whether or not voice is detected (S100).
- the information processing apparatus determines that a voice has been detected when, for example, voice data indicating the utterance content transmitted from a voice input device such as a microphone is received.
- step S100 If it is not determined in step S100 that sound has been detected, the information processing apparatus according to the present embodiment does not proceed until it is determined that sound has been detected.
- the information processing apparatus performs processing related to voice recognition on the voice data (S102).
- the information processing apparatus according to the present embodiment performs, for example, a process using a statistical method or a process using a hidden Markov model as a process related to speech recognition.
- the information processing apparatus analyzes a character string indicating the utterance content obtained by the process of step S102 (S104).
- the information processing apparatus according to the present embodiment performs, for example, syntax analysis and semantic analysis on a character string indicating the utterance content.
- the information processing apparatus corrects fluctuations in voice recognition (S106).
- the information processing apparatus according to the present embodiment uses, for example, a fluctuation absorbing dictionary stored in a storage unit (described later) or the like to replace a word divided by morphological analysis or the like, thereby performing speech recognition. Correct fluctuations.
- the information processing apparatus displays the character string indicating the utterance content and the analysis result in step S104 on the display screen (S108).
- the information processing apparatus emphasizes and displays the analysis result of the character string indicating the utterance content with respect to the character string indicating the utterance content, for example, as indicated by C1 in FIG.
- the information processing apparatus performs, for example, the processing illustrated in FIG. 2 as processing related to the information processing method according to the present embodiment. For example, by performing the process shown in FIG. 2, the process (1) (analysis process) and the process (2) (display control process) according to the information processing method according to the present embodiment are realized.
- the information processing apparatus according to the present embodiment can improve user convenience.
- processing related to the information processing method according to the first example according to the present embodiment is not limited to the example shown in FIG.
- the information processing apparatus according to the present embodiment may not perform the process of step S106 illustrated in FIG.
- the information processing apparatus according to the present embodiment relates to the process (1) (analysis process) and (2) according to the information processing method according to the present embodiment. (Display control processing) can be performed. Therefore, for example, even when the process of step S106 is not performed, the information processing apparatus according to the present embodiment can improve user convenience.
- the process according to the information processing method according to this embodiment is not limited to the process according to the first example.
- the information processing apparatus according to the present embodiment includes the process (1) (analysis) in the process (2) (display control process) according to the present embodiment as indicated by C2 in FIG. Based on the result of the analysis in the processing), the application corresponding to the utterance content and / or the service corresponding to the utterance content can be further displayed on the display screen.
- FIG. 3 is a flowchart showing a second example of processing related to the information processing method according to the present embodiment.
- FIG. 3 shows an example of processing when the information processing apparatus according to the present embodiment further displays an application corresponding to the utterance content and / or a service corresponding to the utterance content on the display screen.
- the process of step S208 shown in FIG. 3 corresponds to the process (1) (analysis process)
- the process of step S210 shown in FIG. 3 is the process (2) (display control process). It corresponds to.
- the process according to the second example illustrated in FIG. 3 may be performed in parallel with the process according to the first example illustrated in FIG. 2, or the process according to the first example illustrated in FIG. It may be done after it has been done.
- the information processing apparatus determines whether or not the capability information of the control target apparatus has been acquired (S200). If it is determined in step S200 that the capability information of the control target device has been acquired, the information processing apparatus according to the present embodiment performs the process of step S204 described later.
- the information processing device acquires capability information from the control target device (S202). For example, when the control target device according to the present embodiment is its own device (information processing device according to the present embodiment), the information processing device according to the present embodiment generates capability information, thereby generating capability information. To get. Further, for example, when the control target device according to the present embodiment is an external device, the capability information is acquired by transmitting a capability information transmission request including a transmission command for transmitting capability information to the external device. To do.
- the information processing apparatus When it is determined in step S200 that the capability information of the control target device has been acquired, or when the capability information of the control target device is acquired by the processing in step S202, the information processing apparatus according to the present embodiment is For example, the application and service corresponding to the control target device are specified from the list of applications and services based on the capability information (S204).
- the information processing apparatus determines whether or not a character string indicating the utterance content has been acquired (S206). If it is not determined in step S206 that the character string indicating the utterance content is acquired, the information processing apparatus according to the present embodiment, for example, until it is determined that the character string indicating the utterance content is acquired, Cannot proceed.
- the information processing apparatus analyzes the operation on the character string indicating the utterance content (S208). ).
- the information processing apparatus according to the present embodiment uses, for example, a dictionary for motion analysis stored in a storage unit (described later) or the like, and uses the morphological analysis in the process of step S104 in FIG. By determining whether or not the word indicating the motion registered in the motion analysis dictionary is included therein, the motion is analyzed with respect to the character string indicating the utterance content.
- FIG. 4 is an explanatory diagram showing an example of a dictionary for motion analysis according to the present embodiment.
- the word is indicated as “Keyword” and the operation is indicated as “Type”.
- 4A shows an example of a table-type motion analysis dictionary in which English words and actions are associated with each other, and B shown in FIG. 4 shows Japanese words and actions.
- An example of a dictionary for motion analysis in the associated table format is shown.
- FIG. 4 shows an example in which information indicating the part of speech is associated with each word in the dictionary for motion analysis.
- the information processing apparatus determines whether or not a word indicating an action registered in the action analysis dictionary illustrated in FIG. 4 is included in words divided by, for example, morphological analysis. To do. When it is determined that a word is included, the operation corresponding to the word is specified as the operation indicated by the character string indicating the utterance content, thereby analyzing the operation on the character string indicating the utterance content. .
- processing related to the analysis of the action on the character string indicating the utterance content according to the present embodiment is not limited to the example shown above, and the action analysis dictionary according to the present embodiment is shown in FIG. It goes without saying that it is not limited to examples.
- step S208 the information processing apparatus according to the present embodiment highlights and displays the corresponding application or service based on the analysis result of the operation in the process of step S208 (S210).
- the information processing apparatus when the analysis result of the operation on the character string indicating the utterance content in step S208 indicates a viewing operation such as “I want to see”, the information processing apparatus according to the present embodiment, for example, a television broadcast program guide Highlight the application or service related to viewing, such as the application that displays. If the analysis result of the operation on the character string indicating the utterance content in step S208 indicates a listening operation such as “I want to hear”, the information processing apparatus according to the present embodiment is stored in, for example, a recording medium. An application or service related to reproduction of sound (including music) such as an application related to reproduction of existing content is highlighted.
- the information processing apparatus when the analysis result of the operation on the character string indicating the utterance content in step S208 does not indicate a specific operation, the information processing apparatus according to the present embodiment, for example, provides a search service provided at a search site or the like For example, a set application or service may be highlighted.
- a preferentially displayed application or service may be set.
- FIG. 5 is an explanatory diagram showing an example of applications and services corresponding to the analysis results according to the present embodiment.
- the analysis result is indicated as “Tense or Action Type”, and the corresponding application or service is indicated as “High Priority Category”.
- FIG. 5 shows an example of a table in which analysis results are associated with applications and services.
- a shown in FIG. 5 shows an example of the tense analysis result and an application or service corresponding to the tense analysis result
- B shown in FIG. 5 corresponds to the operation analysis result or the operation analysis result.
- An example of an application or service to be performed is shown.
- C shown in FIG. 5 shows an example of a corresponding application or service when the tense and the action are not included in the character string indicating the utterance content as a result of the tense analysis and the action analysis.
- the information processing apparatus identifies an application or service corresponding to the analysis result of the operation on the character string indicating the utterance content in step S208 by referring to the table shown in FIG. Then, for example, the information processing apparatus according to the present embodiment preferentially displays the identified application or service on the display screen.
- the information processing apparatus performs, for example, the process illustrated in FIG. 3 as the process related to the information processing method according to the present embodiment in parallel with the process according to the first example illustrated in FIG. This is performed after the processing according to the first example shown in FIG.
- the information processing apparatus uses the analysis result in the process (1) (analysis process) as shown in C2 of C of FIG. Based on this, an application corresponding to the utterance content and / or a service corresponding to the utterance content can be further displayed on the display screen.
- the processing related to the information processing method according to the second example according to the present embodiment is not limited to the example illustrated in FIG. 3.
- FIG. 6 is a flowchart showing a third example of processing related to the information processing method according to the present embodiment.
- FIG. 6 shows an example in which the information processing apparatus according to the present embodiment performs a process related to a content search using a character string indicating the utterance content as an execution process according to the present embodiment.
- content data such as image data and audio data stored in a recording medium such as an optical disk or a magnetic disk, a television broadcast program
- Examples include various information and data obtained through a Web search service.
- the processes in steps S308 to S318 shown in FIG. 6 correspond to the process (1) (analysis process), and the processes in steps S314 and S320 to S324 shown in FIG. 6 correspond to the execution process.
- the process according to the third example illustrated in FIG. 6 may be performed in parallel with the process according to the first example illustrated in FIG. 2, or the process according to the first example illustrated in FIG. It may be done after it has been done.
- the process according to the third example illustrated in FIG. 6 may be performed in parallel with the process according to the second example illustrated in FIG.
- the information processing apparatus determines whether or not the capability information of the control target apparatus has been acquired (S300). When it is determined in step S300 that the capability information of the control target device has been acquired, the information processing apparatus according to the present embodiment performs the process of step S304 described later.
- step S300 If it is not determined in step S300 that the capability information of the control target device has been acquired, the information processing device according to the present embodiment acquires capability information from the control target device in the same manner as in step S202 of FIG. (S302).
- step S300 When it is determined in step S300 that the capability information of the control target device is acquired, or when the capability information of the control target device is acquired by the processing in step S302, the information processing apparatus according to the present embodiment is Based on the capability information, the application and service corresponding to the control target device are identified from the list of applications and services (S304).
- the information processing apparatus determines whether or not a character string indicating the utterance content has been acquired (S306). If it is not determined in step S306 that the character string indicating the utterance content is acquired, the information processing apparatus according to the present embodiment, for example, until it is determined that the character string indicating the utterance content is acquired, Cannot proceed.
- the information processing apparatus sets a reference score for the character string indicating the utterance content ( S308).
- the information processing apparatus sets the reference score “1” for each word divided by morphological analysis in the process of step S104 of FIG. For example, when the character string indicating the utterance content (or the character string in which the fluctuation of the speech recognition is corrected) is “I want to see tomorrow's XXX” shown in C of FIG.
- the processing apparatus sets a reference score “1” for each of “tomorrow” / “no” / “XXX” / “ga” / “see” / “ta” / “yes”. Needless to say, the reference score according to the present embodiment is not limited to “1”.
- the information processing apparatus adjusts the set reference score (S310).
- the information processing apparatus according to the present embodiment adjusts the set reference score using, for example, a score adjustment dictionary stored in a storage unit (described later) or an external recording medium.
- FIG. 7 is an explanatory diagram showing an example of a score adjustment dictionary according to the present embodiment.
- the word is indicated as “Keyword”
- the adjustment value for adjusting the score is indicated as “Additional Value”.
- a shown in FIG. 7 shows an example of a table format score adjustment dictionary in which English words and adjustment values are associated
- B shown in FIG. 7 shows Japanese words and adjustment values. Shows an example of a dictionary for score adjustment in the form of a table in which and are associated with each other.
- the information processing apparatus refers to, for example, a score adjustment dictionary as illustrated in FIG. 7, and adjustment values corresponding to the words divided by morphological analysis in the process of step S104 in FIG. Is identified.
- the information processing apparatus adjusts the set reference score, for example, by adding an adjustment value corresponding to each identified word to the reference score of the corresponding word.
- the processing device adjusts the scores of “tomorrow” / “no” / “XXX” / “ga” / “see” / “ta” / “yes”. Examples of the score after adjustment include the following examples. Needless to say, the adjusted score is not limited to the example shown below.
- processing related to the adjustment of the score set in the character string indicating the utterance content according to the present embodiment is not limited to the example shown above, and the score adjustment dictionary according to the present embodiment is illustrated in FIG. Needless to say, the example is not limited to the example shown in FIG.
- step S310 determines whether or not a character string indicating the utterance content has a score larger than the set reference value (S314).
- the information processing apparatus searches for content using, for example, the entire character string indicating the utterance content ( S314).
- step S314 When it is determined in step S314 that there is a score larger than the reference value set, the information processing apparatus according to the present embodiment determines a search word based on the set score, for example. (S316).
- the information processing apparatus determines, for example, a part of a character string having the highest score as a search word among character strings indicating utterance contents. For example, the adjusted score shown above when the character string indicating the utterance content (or the character string in which the fluctuation of the speech recognition is corrected) is “I want to see tomorrow's XXX” shown in FIG. In the example, the score of “XXX” is the largest among the adjusted scores. Therefore, in the above case, the information processing apparatus according to the present embodiment determines, for example, “XXX” as a search word.
- XXX is determined as the search word in step S316
- an example of a screen displayed on the display screen when the process from S318 described later is performed is, for example, the screen shown in FIG. is there.
- the screen displayed on the display screen when the processing from S318 described later is performed is not limited to the screen illustrated in FIG. 1C. Needless to say.
- the information processing apparatus may determine, for example, a plurality of character strings having a large score among character strings indicating utterance contents as search words.
- the information processing apparatus searches for content by, for example, AND search of a plurality of character strings or OR search of a plurality of character strings.
- the information processing apparatus analyzes the tense with respect to the character string indicating the utterance content (S318).
- the information processing apparatus according to the present embodiment uses, for example, a tense-analysis dictionary among words divided by morphological analysis using a tense-analysis dictionary stored in a storage unit (described later).
- the tense is analyzed with respect to the character string indicating the utterance content by determining whether or not the word indicating the tense registered in is included.
- FIG. 8 is an explanatory diagram showing an example of a tense analysis dictionary according to the present embodiment.
- the word is indicated as “Keyword”
- the corresponding tense is indicated as “Type”.
- 8 shows an example of a table format tense analysis dictionary in which English words and corresponding tenses are associated with each other, and B shown in FIG. 8 corresponds to Japanese words.
- 3 shows an example of a table format tense analysis dictionary in which the tense to be associated is associated with.
- the information processing apparatus refers to, for example, a tense analysis dictionary as shown in FIG. 8, and the tense analysis is performed in the words divided by the morphological analysis in the process of step S104 in FIG. It is determined whether or not a word indicating a tense registered in the dictionary is included.
- the information processing apparatus uses, for example, a character string indicating the utterance content of the tense corresponding to the word. It is determined as a tense included.
- the information processing apparatus when a word indicating tense is not included in words divided by morphological analysis or the like, the information processing apparatus according to the present embodiment, for example, a character string indicating tense in a character string indicating utterance content Is determined not to be included.
- processing related to the tense analysis on the character string indicating the utterance content according to the present embodiment is not limited to the example shown above, and the tense analysis dictionary according to the present embodiment is shown in FIG. It goes without saying that it is not limited to examples.
- the information processing apparatus determines whether or not there is a character string indicating tense in the character string indicating the utterance content (S320). For example, when the tense included in the character string indicating the utterance content is determined in step S318, the information processing apparatus according to the present embodiment determines that there is a character string indicating the tense in the character string indicating the utterance content.
- step S320 If it is not determined in step S320 that there is a character string indicating tense in the character string indicating utterance content, the information processing apparatus according to the present embodiment searches for the content using the search word determined in step S316. (S322).
- step S320 If it is determined in step S320 that there is a character string indicating tense in the character string indicating the utterance content, the information processing apparatus according to the present embodiment uses the tense determined in step S318 as a parameter, step S316. The content is searched using the search word determined in (S324).
- the information processing apparatus may execute content data (an example of content) such as image data and audio data stored in a recording medium such as an optical disk or a magnetic disk. )
- content data an example of content
- the information processing apparatus searches for a television broadcast program (an example of content).
- the information processing apparatus performs, for example, the process illustrated in FIG. 6 as the process related to the information processing method according to the present embodiment in parallel with the process according to the first example illustrated in FIG. 3 is performed in parallel with the processing according to the second example shown in FIG. 3 or after the processing according to the first example shown in FIG.
- the information processing apparatus displays a screen on which an application corresponding to the utterance content and a service corresponding to the utterance content are displayed, such as the screen shown in FIG. It is possible to execute processing related to the application corresponding to the utterance content or the service corresponding to the utterance content selected in.
- processing related to the information processing method according to the third example according to the present embodiment is not limited to the example shown in FIG.
- the information processing apparatus according to the present embodiment may not perform the tense processing shown in steps S318, S312 and S324 shown in FIG.
- the information processing apparatus is, for example, the process according to the first example shown in (A) or the second example shown in (B) as the process related to the information processing method according to the present embodiment.
- the process according to the third example shown in (C) above is performed.
- the processing related to the information processing method according to the present embodiment is not limited to the processing described above.
- FIG. 9 is a block diagram illustrating an example of the configuration of the information processing apparatus 100 according to the present embodiment.
- the information processing apparatus 100 includes, for example, a communication unit 102 and a control unit 104.
- the information processing apparatus 100 includes, for example, a ROM (Read Only Memory, not shown), a RAM (Random Access Memory, not shown), a storage unit (not shown), and a user-operable operation unit (see FIG. And a display unit (not shown) for displaying various screens on the display screen.
- the information processing apparatus 100 connects the above-described constituent elements by, for example, a bus as a data transmission path.
- the ROM (not shown) stores control data such as programs and calculation parameters used by the control unit 104.
- a RAM (not shown) temporarily stores a program executed by the control unit 104.
- the storage unit (not shown) is a storage unit included in the information processing apparatus 100, and stores various data such as various dictionaries and applications.
- examples of the storage unit (not shown) include a magnetic recording medium such as a hard disk, and a non-volatile memory such as a flash memory. Further, the storage unit (not shown) may be detachable from the information processing apparatus 100.
- examples of the operation unit include an operation input device described later
- examples of the display unit include a display device described later.
- FIG. 10 is an explanatory diagram illustrating an example of a hardware configuration of the information processing apparatus 100 according to the present embodiment.
- the information processing apparatus 100 includes, for example, an MPU 150, a ROM 152, a RAM 154, a recording medium 156, an input / output interface 158, an operation input device 160, a display device 162, and a communication interface 164.
- the information processing apparatus 100 connects each component with a bus 166 as a data transmission path, for example.
- the MPU 150 includes, for example, an MPU (Micro Processing Unit) and various processing circuits, and functions as the control unit 104 that controls the entire information processing apparatus 100.
- the MPU 150 serves as an analysis unit 110, a display control unit 112, a log recording control unit 114, and a processing unit 116, which will be described later.
- the ROM 152 stores programs used by the MPU 150, control data such as calculation parameters, and the like.
- the RAM 154 temporarily stores a program executed by the MPU 150, for example.
- the recording medium 156 functions as a storage unit (not shown), and stores various data such as various dictionaries and applications.
- examples of the recording medium 156 include a magnetic recording medium such as a hard disk and a non-volatile memory such as a flash memory. Further, the recording medium 156 may be detachable from the information processing apparatus 100.
- the input / output interface 158 connects, for example, the operation input device 160 and the display device 162.
- the operation input device 160 functions as an operation unit (not shown), and the display device 162 functions as a display unit (not shown).
- Examples of the input / output interface 158 include a USB (Universal Serial Bus) terminal, a DVI (Digital Visual Interface) terminal, an HDMI (High-Definition Multimedia Interface) (registered trademark) terminal, and various processing circuits.
- the operation input device 160 is provided on the information processing apparatus 100 and is connected to the input / output interface 158 inside the information processing apparatus 100.
- Examples of the operation input device 160 include a rotary selector such as a button, a direction key, and a jog dial, or a combination thereof.
- the display device 162 is provided on the information processing apparatus 100 and is connected to the input / output interface 158 inside the information processing apparatus 100.
- Examples of the display device 162 include a liquid crystal display (LCD), an organic EL display (Organic ElectroLuminescence display, or an OLED display (Organic Light Emitting Diode display)), and the like.
- the input / output interface 158 can be connected to an external device such as an operation input device (for example, a keyboard or a mouse) or a display device as an external device of the information processing apparatus 100.
- the display device 162 may be a device capable of display and user operation, such as a touch screen.
- the communication interface 164 is a communication unit included in the information processing apparatus 100, and a communication unit 102 for performing wireless / wired communication with an external apparatus such as a control target apparatus or a server via a network (or directly).
- Examples of the communication interface 164 include a communication antenna and an RF (Radio Frequency) circuit (wireless communication), an IEEE 802.15.1 port and a transmission / reception circuit (wireless communication), an IEEE 802.11b port and a transmission / reception circuit (wireless communication). ), Or a LAN terminal and a transmission / reception circuit (wired communication).
- RF Radio Frequency
- the information processing apparatus 100 performs processing related to the information processing method according to the present embodiment, for example, with the configuration illustrated in FIG. Note that the hardware configuration of the information processing apparatus 100 according to the present embodiment is not limited to the configuration illustrated in FIG.
- the information processing apparatus 100 communicates with an external apparatus such as a control target apparatus or a server via an external communication device connected via, for example, the input / output interface 158, the information processing apparatus 100
- the communication interface 164 may not be provided.
- the information processing apparatus 100 can be configured not to include the operation device 160 and the display device 162, for example.
- the information processing apparatus 100 may further include, for example, an audio input device such as a microphone, an amplifier circuit that amplifies an audio signal transmitted from the audio input device, and the like.
- an audio input device such as a microphone
- an amplifier circuit that amplifies an audio signal transmitted from the audio input device, and the like.
- the communication unit 102 is a communication unit included in the information processing apparatus 100, and performs wireless / wired communication with an external apparatus such as a control target apparatus or a server via a network (or directly).
- the communication of the communication unit 102 is controlled by the control unit 104, for example.
- examples of the communication unit 102 include a communication antenna and an RF circuit, a LAN terminal, and a transmission / reception circuit, but the configuration of the communication unit 102 is not limited to the above.
- the communication unit 102 can take a configuration corresponding to an arbitrary standard capable of performing communication, such as a USB terminal and a transmission / reception circuit, or an arbitrary configuration capable of communicating with an external device via a network.
- the control unit 104 is configured by, for example, an MPU and plays a role of controlling the entire information processing apparatus 100.
- the control unit 104 includes, for example, an analysis unit 110, a display control unit 112, a log recording control unit 114, and a processing unit 116, and performs processing related to the information processing method according to the present embodiment. Play a role.
- the analysis unit 110 plays a leading role in performing the process (1) (analysis process), and analyzes a character string indicating the utterance content obtained as a result of speech recognition.
- the character string indicating the utterance content to be processed according to the present embodiment is, for example, a process related to speech recognition performed on speech data indicating the utterance content by a component included in the information processing apparatus 100 such as the control unit 104.
- the method for acquiring the character string of the utterance content to be processed according to the present embodiment is not limited to the above.
- the information processing apparatus 100 acquires, from the external apparatus, data indicating a character string of the utterance content obtained as a result of processing related to speech recognition performed in an external apparatus such as a server. You may obtain the character string of the speech content used as a process target.
- the analysis unit 110 uses the fluctuation absorbing dictionary stored in the storage unit (not shown) or the like, for example, to replace the words divided by the morphological analysis, thereby reducing the fluctuation of the voice recognition. It may be corrected.
- the analysis unit 110 may analyze the operation and tense by using, for example, a dictionary for the character string indicating the utterance content.
- the analysis unit 110 can set a score for a character string indicating the utterance content based on the analysis result, for example.
- the analysis unit 110 sets the score by performing the processes of steps S308 and S310 illustrated in FIG.
- the display control unit 112 plays a leading role in performing the process (2) (display control process), and displays the character string indicating the utterance content and the analysis result in the analysis unit 110 on the display screen.
- the display control unit 112 edits the character string indicating the utterance content, for example, for the entire character string indicating the utterance content or for each word divided by morphological analysis in the character string indicating the utterance content. It may be displayed in a possible state.
- the display control unit 112 displays, for example, a character string indicating the utterance content in which the fluctuation of the voice recognition is corrected.
- the display control unit 112 determines an application corresponding to the utterance content and / or a service corresponding to the utterance content based on the analysis result in the analysis unit 110. It may be displayed highlighted on the display screen.
- the display control unit 112 performs the operation included in the analysis result in the analysis unit 110, for example. Based on the analysis results, the applications and services to be displayed on the display screen are highlighted.
- the display control unit 112 When displaying an application corresponding to the utterance content and / or a service corresponding to the utterance content on the display screen, the display control unit 112 displays the display screen based on the capability information of the control target device, for example. Candidates for applications to be displayed on the screen and / or candidates for services to be displayed on the display screen may be determined.
- the log recording control unit 114 plays a role of leading the log recording control process, and records the analysis result in the analysis unit 110 as a log.
- Examples of targets for which the log recording control unit 114 records logs include a storage unit (not shown), an external recording medium, and an external device that can communicate (or directly) via a network.
- the log recording control unit 114 causes the communication unit 102 to transmit log data to the external device, for example.
- the processing unit 116 plays a role of leading the execution process, and executes a process related to an application corresponding to the selected utterance content or a service corresponding to the selected utterance content.
- the processing unit 116 executes the application corresponding to the utterance content based on the comparison result between the set reference value and the set score.
- the processing may be performed by determining a character string used to execute processing related to the service corresponding to the utterance content. More specifically, for example, when there is no score larger than the reference value, the processing unit 116 uses the entire character string indicating the utterance content for processing, and when there is a score larger than the reference value, Among the character strings indicating the utterance contents, one or more partial character strings having a high score are used for the processing.
- the processing unit 116 when the tense is analyzed in the analysis unit 110, the processing unit 116, for example, when the analysis result of the tense in the analysis unit 110 indicates that the character string indicating the utterance content includes a word indicating the tense. As shown in step S324 of FIG. 6, a word indicating tense included in the character string indicating the utterance content may be used for the processing.
- the control unit 104 includes, for example, an analysis unit 110, a display control unit 112, a log recording control unit 114, and a processing unit 116, thereby leading the processing related to the information processing method according to the present embodiment.
- the information processing apparatus 100 has, for example, the configuration shown in FIG. 9 and processes (for example, the process (1) (analysis process) and the process (2) (display control process) according to the information processing method according to the present embodiment. , The execution process, and the log recording control process).
- processes for example, the process (1) (analysis process) and the process (2) (display control process) according to the information processing method according to the present embodiment. , The execution process, and the log recording control process).
- the information processing apparatus 100 can improve user convenience, for example, by the configuration shown in FIG.
- the information processing apparatus 100 can achieve an effect by performing the processing related to the information processing method according to the present embodiment described above, for example, with the configuration illustrated in FIG. 9.
- the configuration of the information processing apparatus 100 according to the present embodiment is not limited to the configuration illustrated in FIG. 9.
- the information processing apparatus 100 according to the present embodiment may not include the log recording control unit 114 and / or the processing unit 116. Even in a configuration that does not include the log recording control unit 114 and / or the processing unit 116, the information processing apparatus 100 according to the present embodiment performs the process (1) according to the information processing method according to the present embodiment. (Analysis process) and the process (2) (display control process) can be led.
- the information processing apparatus 100 can improve the convenience for the user.
- the information processing apparatus 100 includes one or more of the analysis unit 110, the display control unit 112, the log recording control unit 114, and the processing unit 116 illustrated in FIG. Can be provided separately (for example, realized by a separate processing circuit).
- the information processing apparatus 100 when the information processing apparatus 100 according to the present embodiment communicates with an external apparatus such as a control target apparatus or a server via an external communication device, the information processing apparatus 100 includes the communication unit 102. It does not have to be.
- the information processing apparatus includes, for example, the process (1) (analysis process) and the process (2) (display) as the processes related to the information processing method according to the present embodiment. Control process).
- the information processing apparatus performed the speech by displaying the character string of the speech content obtained as a result of the speech recognition on the display screen in the process (2) (display control process).
- the user can visually grasp how the utterance is recognized as voice.
- the information processing apparatus analyzes the result of analysis on the character string of the utterance content in the process (analysis process) of (1). By displaying together with the character string, the user who made the utterance can visually understand how the utterance was analyzed.
- the information processing apparatus can improve user convenience.
- the information processing apparatus displays, for example, a character string indicating the utterance content to be displayed on the display screen in an editable state in the process (2) (display control process). .
- the information processing apparatus displays, for example, the utterance content to the user by displaying the character string indicating the utterance content in an editable state. Correction can be promoted, and processing with higher accuracy can be performed by the own device (information processing device according to the present embodiment) or an external device.
- the process (2) display control process
- the information processing apparatus according to the present embodiment can improve the convenience of the user. Further improvement can be achieved.
- the information processing apparatus performs the analysis in the process (1) (analysis process) in the process (2) (display control process), for example, as indicated by C2 in FIG. Based on the result, the application corresponding to the utterance content and / or the service corresponding to the utterance content is displayed on the display screen. Also, the information processing apparatus according to the present embodiment displays, for example, an application corresponding to the utterance content and / or a service corresponding to the utterance content in a highlighted manner on the display screen.
- the application corresponding to the utterance content and / or the service corresponding to the utterance content is displayed on the display screen based on the analysis result, for example, a display more suitable for the control target device. Can be realized.
- the information processing apparatus displays the analysis result of the utterance content in a more suitable form instead of one, thereby absorbing the ambiguity of the utterance and discovering new content by the user. It becomes possible.
- the process (2) display control process
- the application corresponding to the utterance content and / or the service corresponding to the utterance content is displayed on the display screen.
- the information processing apparatus according to the present embodiment can further improve user convenience.
- the information processing apparatus has been described as the present embodiment, but the present embodiment is not limited to such a form.
- a tablet-type device for example, a communication device such as a mobile phone or a smartphone, a video / music playback device (or video / music recording / playback device), a game machine, a computer such as a PC (Personal Computer), a remote -It can be applied to various devices such as controllers.
- the present embodiment can be applied to, for example, a processing IC (Integrated Circuit) that can be incorporated in the above devices.
- a processing IC Integrated Circuit
- the information processing apparatus can process a character string indicating the utterance content obtained as a result of the speech recognition process performed by an external apparatus such as a server. Therefore, next, the information processing apparatus according to the present embodiment can process the character string indicating the utterance content obtained as a result of the voice recognition process performed by the external apparatus.
- an example of the system will be described.
- FIG. 11 is an explanatory diagram showing an example of the information processing system 1000 according to the present embodiment.
- the information processing system 1000 includes, for example, the information processing apparatus 100, the speech recognition processing apparatus 200, and control target apparatuses 300A, 300B, which are target apparatuses on which the information processing apparatus 100 performs control of functions (hereinafter collectively referred to as "collection"). It may be indicated as “control target device 300”). Further, the information processing apparatus 100 and the speech recognition processing apparatus 200 communicate via the network 400, for example.
- the network 400 include a wired network such as a LAN or WAN, a wireless network such as a wireless WAN via a wireless LAN or a base station, or the Internet using a communication protocol such as TCP / IP.
- the information processing system 1000 has shown the example which has several control object apparatus 300, the information processing system which concerns on this embodiment is not restricted above.
- the information processing system according to the present embodiment may be configured to include one control target device 300.
- the information processing system according to the present embodiment includes the control target apparatuses 300A and 300B illustrated in FIG. The structure which does not have ... may be sufficient.
- the speech recognition processing apparatus 200 performs speech recognition processing such as processing using a statistical method and processing using a hidden Markov model on speech data indicating the utterance content.
- the speech recognition processing device 200 transmits data indicating the character string of the utterance content obtained as a result of processing related to speech recognition on the speech data to the information processing device 100.
- the speech recognition processing apparatus 200 manages, for example, data related to the dictionary according to the present embodiment (hereinafter referred to as “dictionary data”).
- dictionary data data related to the dictionary according to the present embodiment
- other devices such as a server may manage the dictionary data.
- the dictionary data By managing the dictionary in an external device such as the speech recognition processing device 200, for example, the behavior in the processing (1) (analysis processing) in the information processing device 100 can be changed more easily.
- the speech recognition processing apparatus 200 manages dictionary data will be described as an example.
- the voice recognition processing device 200 manages log data, for example.
- another device such as a server may manage log data, or the information processing device 100 may manage log data.
- the voice recognition processing device 200 manages log data will be described as an example.
- the speech recognition processing device 200 has, for example, the hardware configuration shown in FIG. 10, but the hardware configuration of the speech recognition processing device 200 is not limited to the above.
- the information processing apparatus 100 performs processing related to the information processing method according to the above-described embodiment on the data indicating the character string of the utterance content transmitted from the speech recognition processing apparatus 200.
- the control target device 300 performs processing based on, for example, a control command transmitted from the information processing device 100, and the function is controlled by the information processing device 100.
- the control target device 300 has, for example, the hardware configuration illustrated in FIG. 10, but the hardware configuration of the control target device 300 is not limited to the above.
- the information processing apparatus 100 can process a character string indicating the utterance content obtained as a result of the voice recognition processing performed by the voice recognition processing apparatus 200 (an example of an external apparatus).
- An information processing system is realized.
- FIG. 12 is an explanatory diagram illustrating an example of processing in the information processing system 1000 according to the present embodiment.
- the process of step S414 shown in FIG. 12 corresponds to the above process (1) (analysis process)
- the process of step S418 shown in FIG. 12 corresponds to the above process (2) (display control process).
- steps S416 and S420 shown in FIG. 12 show an example of the execution process according to the present embodiment
- step S422 shown in FIG. 12 corresponds to the log recording control process according to the present embodiment.
- the information processing apparatus 100 and the control target apparatus 300 perform connection processing (S400).
- the process of step S400 is performed, for example, when the information processing apparatus 100 transmits a connection request to the control target apparatus 300 based on a user's connection operation.
- the information processing apparatus 100 specifies the capability of the control target apparatus 300 (S402). For example, the information processing apparatus 100 transmits a capability information transmission request for transmitting capability information to the control target device 300, and controls the control target based on the capability information transmitted from the control target device 300 in response to the capability information transmission request.
- the capability of the device 300 is specified.
- the information processing apparatus 100 transmits a dictionary update confirmation request to the speech recognition processing apparatus 200, for example (S404).
- the speech recognition processing apparatus 200 that has received the dictionary update confirmation request transmitted from the information processing apparatus 100 in step S404 confirms the dictionary data stored in the recording medium 156 or the like, and if the dictionary data has been updated. Then, the dictionary data is transmitted to the information processing apparatus 100 (S406). Through the processing in step S406, for example, the dictionary according to the present embodiment stored in a storage unit (not shown) included in the information processing apparatus 100 is updated.
- the speech recognition processing apparatus 200 may transmit dictionary data indicating the entire dictionary, or may transmit only data corresponding to the updated portion in the dictionary as dictionary data.
- the information processing apparatus 100 transmits the voice data to the voice recognition processing apparatus 200 (S408).
- step S408 the voice recognition processing apparatus 200 that has received the voice data transmitted from the information processing apparatus 100 performs a process related to voice recognition on the voice data (S410). Then, the speech recognition processing device 200 transmits data indicating the character string of the utterance content obtained as a result of the processing related to speech recognition to the information processing device 100 (S412).
- step S412 the information processing apparatus 100 that has received the data indicating the character string of the utterance content transmitted from the speech recognition processing device 200 performs analysis processing on the data indicating the character string of the utterance content (S414).
- the information processing apparatus 100 searches for content based on the analysis result in step S414 (S416).
- examples of the information processing device 100 to search for content include a storage unit (not shown), an external recording medium such as an optical disk, an external device such as the voice recognition processing device 200, and the like.
- the information processing apparatus 100 searches for a program in a connected television receiver (an example of the control target apparatus 300) in cooperation with a content search engine in an external apparatus such as the voice recognition processing apparatus 200, for example. Also, search for videos and music from internet-linked services.
- the information processing apparatus 100 communicates with, for example, a connected television receiver (an example of the control target device 300) to search for a recorded title content in the television receiver, an application of the television receiver, or the like.
- the information processing apparatus 100 communicates with, for example, a PC (an example of the control target apparatus 300) in the home network and stores image data (data indicating moving images and still images) stored in a recording medium included in the PC. Search music data.
- the information processing apparatus 100 performs web search, for example. Further, for example, the information processing apparatus 100 may search for a disk inserted in a disk player (an example of the control target apparatus 300).
- the information processing apparatus 100 displays the character string of the utterance content and the analysis result in step S414 on the display screen (S418). Further, the information processing apparatus 100, for example, based on the analysis result in step S414 and the search result in step S416, as shown in C2 of FIG. A service corresponding to the content may be displayed on the display screen.
- step S4108 when an operation for an application or service (an operation for causing the control target apparatus 300 to perform an operation) is detected on the screen displayed in step S418, the information processing apparatus 100 performs control according to the operation.
- the command is transmitted to the control target device 300 (S420).
- the information processing apparatus 100 transmits, for example, the log of the analysis result in step S414 and the log data indicating the log in each process to the voice recognition processing apparatus 200 (S422).
- the processing shown in FIG. 12 is performed.
- the information processing apparatus 100 can perform the process related to the information processing method according to the present embodiment described above. Therefore, for example, the information processing system capable of improving the convenience for the user is realized by performing the processing shown in FIG. Needless to say, the processing in the information processing system 1000 according to the present embodiment is not limited to the processing shown in FIG.
- the speech recognition processing apparatus 200 is described as a constituent element of the information processing system according to the present embodiment, but the present embodiment is not limited to such a form.
- the present embodiment can be applied to various devices such as computers such as PCs and servers.
- the present embodiment can be applied to a processing IC that can be incorporated in the above-described device, for example.
- the processing related to the speech recognition processing device 200 is performed by a system composed of a plurality of devices on the premise of connection to a network (or communication between devices) such as cloud computing. It may be realized.
- control target device 300 is described as a component of the information processing system according to the present embodiment.
- the present embodiment is not limited to such a form.
- the present embodiment can be applied to various devices such as a television receiver, a computer such as a PC or a server, a video / music playback device (or video / music recording / playback device), and a game machine.
- the present embodiment can be applied to a processing IC that can be incorporated in the above-described device, for example.
- a program for causing a computer to function as the information processing apparatus according to the present embodiment for example, “the process (1) (analysis process) and (2) process (display control process)” or “( 1) Process (Analysis Process), Process (2) (Display Control Process), and Execution Process "," Process (1) (Analysis Process), Process (2) (Display Control Process) " And the above-mentioned log recording control process ”,“ the process (1) (analysis process), (2) the process (display control process), the execution process, and the log recording control process ”.
- a program for causing a computer to function as the information processing apparatus according to the present embodiment is provided.
- the present embodiment further includes a recording in which the program is stored.
- a medium can also be provided.
- An analysis unit for analyzing a character string indicating the utterance content obtained as a result of speech recognition A display control unit for displaying the character string indicating the utterance content and the analysis result on a display screen;
- An information processing apparatus comprising: (2) The information processing apparatus according to (1), wherein the display control unit displays a character string indicating the utterance content in a state where the character string can be edited. (3) The analysis unit corrects fluctuations in voice recognition, The information processing apparatus according to (1) or (2), wherein the display control unit displays a character string indicating an utterance content in which the fluctuation is corrected.
- the display control unit highlights and displays an application corresponding to the utterance content and / or a service corresponding to the utterance content on the display screen based on the analysis result. ).
- the analysis unit analyzes an operation on a character string indicating the utterance content, The information processing apparatus according to (4), wherein the display control unit highlights and displays the application and the service to be displayed on a display screen based on an analysis result of an operation included in the analysis result.
- the display control unit determines a candidate for the application to be displayed on the display screen and / or a candidate for the service to be displayed on the display screen based on capability information indicating the capability of the device. (4) The information processing apparatus according to (5).
- the analysis unit converts the character string indicating the utterance content into a character string indicating the utterance content based on the result of the syntax analysis on the character string indicating the utterance content, or the result of the semantic analysis on the character string indicating the utterance content.
- the processor is If there is no score greater than a reference value or the score greater than the reference value, the entire character string indicating the utterance content is used for processing, When there is the score larger than the reference value or the score equal to or higher than the reference value, one or more partial character strings having a high score among the character strings indicating the utterance contents are selected.
- the information processing apparatus according to (7), used for processing.
- the analysis unit analyzes tense with respect to a character string indicating the utterance content, When the analysis result of the tense indicates that the character string indicating the utterance content includes a word indicating tense, the processing unit uses the wording indicating the tense included in the character string indicating the utterance content for processing.
- the information processing apparatus according to any one of (1) to (9), further including a log recording control unit that records a result of the analysis as a log.
- (11) Analyzing a character string indicating utterance content obtained as a result of speech recognition; Displaying the character string indicating the utterance content and the result of the analysis on a display screen; An information processing method.
- (12) Analyzing a character string indicating the utterance content obtained as a result of speech recognition; Displaying a character string indicating the utterance content and a result of the analysis on a display screen; A program that causes a computer to execute.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Acoustics & Sound (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
- Machine Translation (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Information Transfer Between Computers (AREA)
- Debugging And Monitoring (AREA)
- Stored Programmes (AREA)
Abstract
Description
1.本実施形態に係る情報処理方法
2.本実施形態に係る情報処理装置
3.本実施形態に係る情報処理システム
4.本実施形態に係るプログラム
本実施形態に係る情報処理装置の構成について説明する前に、まず、本実施形態に係る情報処理方法について説明する。以下では、本実施形態に係る情報処理装置が、本実施形態に係る情報処理方法に係る処理を行う場合を例に挙げて、本実施形態に係る情報処理方法について説明する。
上述したように、既存の装置が用いられる場合には、発話内容のどの部分が認識されて処理が行われたのかを、ユーザは知ることができない。そのため、既存の装置が用いられる場合には、例えば“当該装置における音声発話によるその後の挙動が、ユーザに十分に伝わらない”など、ユーザの利便性を低下させうる事態が起こりうる。
次に、本実施形態に係る情報処理装置における、本実施形態に係る情報処理方法に係る処理について、より具体的に説明する。上述したように、本実施形態に係る情報処理装置は、本実施形態に係る情報処理方法に係る処理として、例えば、上記(1)の処理(解析処理)および上記(2)の処理(表示制御処理)を行う。
本実施形態に係る情報処理装置は、音声認識の結果得られる発話内容を示す文字列を解析する。
本実施形態に係る情報処理装置は、発話内容を示す文字列と、上記(1)の処理(解析処理)における解析の結果とを、表示画面に表示させる。ここで、本実施形態に係る情報処理装置は、例えば図1のCのC1に示すように、解析の結果が強調表示されるように、発話内容を示す文字列と解析の結果とを表示させる。
例えば、上記(1)の処理(解析処理)において発話内容を示す文字列に対するスコアの設定が行われた場合には、本実施形態に係る情報処理装置は、設定されている基準値と、設定されたスコアとの比較結果に基づいて、発話内容を示す文字列の一部、または全部を、発話内容に対応するアプリケーションの実行や、発話内容に対応するサービスに関する処理の実行に用いる文字列として決定する。そして、本実施形態に係る情報処理装置は、決定された文字列を用いて、発話内容に対応するアプリケーションや、発話内容に対応するサービスに関する処理を実行する。
例えば、上記(1)の処理(解析処理)において発話内容を示す文字列に対して時制が解析された場合には、本実施形態に係る情報処理装置は、時制の解析結果に基づいて、発話内容を示す文字列に含まれる時制を示す文言を処理に選択的に用いる。より具体的には、上記の場合には、本実施形態に係る情報処理装置は、時制の解析結果が、発話内容を示す文字列に時制を示す文言が含まれることを示す場合に、発話内容を示す文字列に含まれる時制を示す文言を処理に用いる。
本実施形態に係る情報処理装置は、例えば、上記第1の例に係る処理と、上記第2の例に係る処理とを組み合わせた処理を行うことも可能である。
次に、本実施形態に係る情報処理方法に係る処理の具体例を説明する。以下では、本実施形態に係る情報処理方法に係る処理を、本実施形態に係る情報処理装置が行うものとして説明する。
図2は、本実施形態に係る情報処理方法に係る処理の第1の例を示す流れ図である。ここで、例えば、図2に示すステップS104、S106の処理が、上記(1)の処理(解析処理)に該当し、図2に示すステップS108の処理が、上記(2)の処理(表示制御処理)に該当する。また、図2では、本実施形態に係る情報処理装置が、発話内容を示す音声データに対する音声認識に係る処理を行う例を示している。なお、上述したように、発話内容を示す音声データに対する音声認識に係る処理は、サーバなどの外部装置において行われてもよい。
なお、本実施形態に係る情報処理方法に係る処理は、上記第1の例に係る処理に限られない。例えば、本実施形態に係る情報処理装置は、例えば図1のCのC2に示すように、本実施形態に係る上記(2)の処理(表示制御処理)において、上記(1)の処理(解析処理)における解析の結果に基づいて、発話内容に対応するアプリケーション、および/または、発話内容に対応するサービスを、さらに表示画面に表示させることも可能である。
なお、本実施形態に係る情報処理方法に係る処理は、上記第1の例に係る処理や、上記第2の例に係る処理に限られない。例えば、本実施形態に係る情報処理装置は、さらに、上述した本実施形態に係る実行処理を行うことも可能である。
・「明日」のスコア:“1.3”
・「の」のスコア:“1”
・「XXX」のスコア:“2”
・「が」のスコア:“1”
・「見」のスコア:“0.5”
・「た」のスコア:“1”
・「い」のスコア:“1”
次に、上述した本実施形態に係る情報処理方法に係る処理を行うことが可能な、本実施形態に係る情報処理装置の構成の一例について、説明する。
図10は、本実施形態に係る情報処理装置100のハードウェア構成の一例を示す説明図である。情報処理装置100は、例えば、MPU150と、ROM152と、RAM154と、記録媒体156と、入出力インタフェース158と、操作入力デバイス160と、表示デバイス162と、通信インタフェース164とを備える。また、情報処理装置100は、例えば、データの伝送路としてのバス166で各構成要素間を接続する。
上述したように、本実施形態に係る情報処理装置は、サーバなどの外部装置において音声認識の処理が行われた結果得られる発話内容を示す文字列を、処理することも可能である。そこで、次に、本実施形態に係る情報処理装置が、外部装置において音声認識の処理が行われた結果得られる発話内容を示す文字列を処理することが可能な、本実施形態に係る情報処理システムの一例について説明する。
コンピュータを、本実施形態に係る情報処理装置として機能させるためのプログラム(例えば、“上記(1)の処理(解析処理)、および上記(2)の処理(表示制御処理)”や、“上記(1)の処理(解析処理)、上記(2)の処理(表示制御処理)、および上記実行処理”、“上記(1)の処理(解析処理)、上記(2)の処理(表示制御処理)、および上記ログ記録制御処理”、“上記(1)の処理(解析処理)、上記(2)の処理(表示制御処理)、上記実行処理、および上記ログ記録制御処理”など、本実施形態に係る情報処理方法に係る処理を実行することが可能なプログラム)が、コンピュータにおいて実行されることによって、ユーザの利便性の向上を図ることができる。また、コンピュータを、本実施形態に係る情報処理装置として機能させるためのプログラムが、コンピュータにおいて実行されることによって、上述した本実施形態に係る情報処理方法に係る処理が行われることによる効果を奏することができる。
(1)
音声認識の結果得られる発話内容を示す文字列を解析する解析部と、
前記発話内容を示す文字列と解析の結果とを、表示画面に表示させる表示制御部と、
を備える、情報処理装置。
(2)
前記表示制御部は、前記発話内容を示す文字列を、編集することが可能な状態で表示させる、(1)に記載の情報処理装置。
(3)
前記解析部は、音声認識のゆらぎを補正し、
前記表示制御部は、前記ゆらぎが補正された発話内容を示す文字列を表示させる、(1)、または(2)に記載の情報処理装置。
(4)
前記表示制御部は、前記解析の結果に基づいて、前記発話内容に対応するアプリケーション、および/または、前記発話内容に対応するサービスを、表示画面に強調して表示させる、(1)~(3)のいずれか1つに記載の情報処理装置。
(5)
前記解析部は、前記発話内容を示す文字列に対して動作を解析し、
前記表示制御部は、前記解析の結果に含まれる動作の解析結果に基づいて、表示画面に表示させる前記アプリケーション、前記サービスを強調して表示させる、(4)に記載の情報処理装置。
(6)
前記表示制御部は、装置が有する能力を示す能力情報に基づいて、表示画面に表示させる前記アプリケーションの候補、および/または、表示画面に表示させる前記サービスの候補を決定する、(4)、または(5)に記載の情報処理装置。
(7)
選択された前記発話内容に対応するアプリケーション、または、選択された前記発話内容に対応するサービスに関する処理を実行する処理部をさらに備える、(4)~(6)のいずれか1つに記載の情報処理装置。
(8)
前記解析部は、前記発話内容を示す文字列に対する構文解析の結果、または、前記構文解析の結果および前記発話内容を示す文字列に対する意味解析の結果に基づいて、前記発話内容を示す文字列にスコアを設定し、
前記処理部は、
基準値よりも大きな前記スコア、または、前記基準値以上の前記スコアがない場合には、前記発話内容を示す文字列全体を処理に用い、
前記基準値よりも大きな前記スコア、または、前記基準値以上の前記スコアがある場合には、前記発話内容を示す文字列のうちの、前記スコアが大きい1または2以上の一部の文字列を、処理に用いる、(7)に記載の情報処理装置。
(9)
前記解析部は、前記発話内容を示す文字列に対して時制を解析し、
前記処理部は、時制の解析結果が前記発話内容を示す文字列に時制を示す文言が含まれることを示す場合には、前記発話内容を示す文字列に含まれる時制を示す文言を処理に用いる、(7)、または(8)に記載の情報処理装置。
(10)
前記解析の結果をログとして記録させるログ記録制御部をさらに備える、(1)~(9)のいずれか1つに記載の情報処理装置。
(11)
音声認識の結果得られる発話内容を示す文字列を解析するステップと、
前記発話内容を示す文字列と解析の結果とを、表示画面に表示させるステップと、
を有する、情報処理方法。
(12)
音声認識の結果得られる発話内容を示す文字列を解析するステップ、
前記発話内容を示す文字列と、解析の結果とを表示画面に表示させるステップ、
をコンピュータに実行させるためのプログラム。
102 通信部
104 制御部
110 解析部
112 表示制御部
114 ログ記録制御部
116 処理部
200 音声認識処理装置
300、300A、300B 制御対象装置
400 ネットワーク
1000 情報処理システム
Claims (12)
- 音声認識の結果得られる発話内容を示す文字列を解析する解析部と、
前記発話内容を示す文字列と解析の結果とを、表示画面に表示させる表示制御部と、
を備える、情報処理装置。 - 前記表示制御部は、前記発話内容を示す文字列を、編集することが可能な状態で表示させる、請求項1に記載の情報処理装置。
- 前記解析部は、音声認識のゆらぎを補正し、
前記表示制御部は、前記ゆらぎが補正された発話内容を示す文字列を表示させる、請求項1に記載の情報処理装置。 - 前記表示制御部は、前記解析の結果に基づいて、前記発話内容に対応するアプリケーション、および/または、前記発話内容に対応するサービスを、表示画面に強調して表示させる、請求項1に記載の情報処理装置。
- 前記解析部は、前記発話内容を示す文字列に対して動作を解析し、
前記表示制御部は、前記解析の結果に含まれる動作の解析結果に基づいて、表示画面に表示させる前記アプリケーション、前記サービスを強調して表示させる、請求項4に記載の情報処理装置。 - 前記表示制御部は、装置が有する能力を示す能力情報に基づいて、表示画面に表示させる前記アプリケーションの候補、および/または、表示画面に表示させる前記サービスの候補を決定する、請求項4に記載の情報処理装置。
- 選択された前記発話内容に対応するアプリケーション、または、選択された前記発話内容に対応するサービスに関する処理を実行する処理部をさらに備える、請求項4に記載の情報処理装置。
- 前記解析部は、前記発話内容を示す文字列に対する構文解析の結果、または、前記構文解析の結果および前記発話内容を示す文字列に対する意味解析の結果に基づいて、前記発話内容を示す文字列にスコアを設定し、
前記処理部は、
基準値よりも大きな前記スコア、または、前記基準値以上の前記スコアがない場合には、前記発話内容を示す文字列全体を処理に用い、
前記基準値よりも大きな前記スコア、または、前記基準値以上の前記スコアがある場合には、前記発話内容を示す文字列のうちの、前記スコアが大きい1または2以上の一部の文字列を、処理に用いる、請求項7に記載の情報処理装置。 - 前記解析部は、前記発話内容を示す文字列に対して時制を解析し、
前記処理部は、時制の解析結果が前記発話内容を示す文字列に時制を示す文言が含まれることを示す場合には、前記発話内容を示す文字列に含まれる時制を示す文言を処理に用いる、請求項7に記載の情報処理装置。 - 前記解析の結果をログとして記録させるログ記録制御部をさらに備える、請求項1に記載の情報処理装置。
- 音声認識の結果得られる発話内容を示す文字列を解析するステップと、
前記発話内容を示す文字列と解析の結果とを、表示画面に表示させるステップと、
を有する、情報処理方法。 - 音声認識の結果得られる発話内容を示す文字列を解析するステップ、
前記発話内容を示す文字列と、解析の結果とを表示画面に表示させるステップ、
をコンピュータに実行させるためのプログラム。
Priority Applications (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
BR112015014830-1A BR112015014830B1 (pt) | 2012-12-28 | 2013-11-20 | Dispositivo e método de processamento de informação, e, meio de armazenamento de memória. |
CN201380067142.3A CN104871240A (zh) | 2012-12-28 | 2013-11-20 | 信息处理设备、信息处理方法、以及程序 |
EP21177740.4A EP3896596A1 (en) | 2012-12-28 | 2013-11-20 | Information processing device, information processing method and program |
US14/648,890 US10424291B2 (en) | 2012-12-28 | 2013-11-20 | Information processing device, information processing method, and program |
JP2014554243A JPWO2014103568A1 (ja) | 2012-12-28 | 2013-11-20 | 情報処理装置、情報処理方法、およびプログラム |
EP13867146.6A EP2940683A4 (en) | 2012-12-28 | 2013-11-20 | INFORMATION PROCESSING DEVICE, INFORMATION PROCESSING METHOD, AND PROGRAM |
CN201911128993.XA CN110889265B (zh) | 2012-12-28 | 2013-11-20 | 信息处理设备和信息处理方法 |
US16/523,675 US11100919B2 (en) | 2012-12-28 | 2019-07-26 | Information processing device, information processing method, and program |
US17/386,260 US11676578B2 (en) | 2012-12-28 | 2021-07-27 | Information processing device, information processing method, and program |
US18/310,105 US12125475B2 (en) | 2012-12-28 | 2023-05-01 | Information processing device, information processing method, and program |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012287357 | 2012-12-28 | ||
JP2012-287357 | 2012-12-28 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/648,890 A-371-Of-International US10424291B2 (en) | 2012-12-28 | 2013-11-20 | Information processing device, information processing method, and program |
US16/523,675 Continuation US11100919B2 (en) | 2012-12-28 | 2019-07-26 | Information processing device, information processing method, and program |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2014103568A1 true WO2014103568A1 (ja) | 2014-07-03 |
Family
ID=51020662
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2013/081312 WO2014103568A1 (ja) | 2012-12-28 | 2013-11-20 | 情報処理装置、情報処理方法、およびプログラム |
Country Status (6)
Country | Link |
---|---|
US (3) | US10424291B2 (ja) |
EP (2) | EP3896596A1 (ja) |
JP (2) | JPWO2014103568A1 (ja) |
CN (2) | CN110889265B (ja) |
BR (1) | BR112015014830B1 (ja) |
WO (1) | WO2014103568A1 (ja) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10424291B2 (en) | 2012-12-28 | 2019-09-24 | Saturn Licensing Llc | Information processing device, information processing method, and program |
WO2020240958A1 (ja) * | 2019-05-30 | 2020-12-03 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR102352764B1 (ko) * | 2014-11-04 | 2022-01-19 | 삼성전자주식회사 | 사용자 단말 장치, 사용자 단말 장치와 연계되는 디스플레이 장치, 연계 시스템 및 그 제어 방법 |
US10271093B1 (en) | 2016-06-27 | 2019-04-23 | Amazon Technologies, Inc. | Systems and methods for routing content to an associated output device |
CN109643548B (zh) * | 2016-06-27 | 2023-05-05 | 亚马逊技术公司 | 用于将内容路由到相关联输出设备的系统和方法 |
US10931999B1 (en) * | 2016-06-27 | 2021-02-23 | Amazon Technologies, Inc. | Systems and methods for routing content to an associated output device |
JP6957485B2 (ja) * | 2016-09-26 | 2021-11-02 | ソニーグループ株式会社 | 情報処理装置、情報処理方法、プログラム、および情報処理システム |
US10449440B2 (en) | 2017-06-30 | 2019-10-22 | Electronic Arts Inc. | Interactive voice-controlled companion application for a video game |
US10621317B1 (en) | 2017-09-14 | 2020-04-14 | Electronic Arts Inc. | Audio-based device authentication system |
US10572586B2 (en) * | 2018-02-27 | 2020-02-25 | International Business Machines Corporation | Technique for automatically splitting words |
JP7305678B2 (ja) * | 2018-12-05 | 2023-07-10 | 株式会社東芝 | 会話分析システム、方法及びプログラム |
US10926173B2 (en) * | 2019-06-10 | 2021-02-23 | Electronic Arts Inc. | Custom voice control of video game character |
KR20240044260A (ko) * | 2022-09-28 | 2024-04-04 | 엘지전자 주식회사 | 디스플레이 장치 및 그의 동작 방법 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001166914A (ja) | 1999-12-06 | 2001-06-22 | Canon Inc | 情報入力装置、情報入力方法、及び記憶媒体 |
JP2010134074A (ja) * | 2008-12-03 | 2010-06-17 | Toshiba Corp | 音声認識装置、方法、およびプログラム |
JP2012063526A (ja) * | 2010-09-15 | 2012-03-29 | Ntt Docomo Inc | 端末装置、音声認識方法および音声認識プログラム |
Family Cites Families (47)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0756933A (ja) * | 1993-06-24 | 1995-03-03 | Xerox Corp | 文書検索方法 |
US7165098B1 (en) * | 1998-11-10 | 2007-01-16 | United Video Properties, Inc. | On-line schedule system with personalization features |
US6594630B1 (en) * | 1999-11-19 | 2003-07-15 | Voice Signal Technologies, Inc. | Voice-activated control for electrical device |
US7158935B1 (en) * | 2000-11-15 | 2007-01-02 | At&T Corp. | Method and system for predicting problematic situations in a automated dialog |
US8019602B2 (en) * | 2004-01-20 | 2011-09-13 | Microsoft Corporation | Automatic speech recognition learning using user corrections |
US7929767B2 (en) * | 2004-09-22 | 2011-04-19 | Microsoft Corporation | Analyzing subordinate sub-expressions in expression recognition |
JP4542974B2 (ja) * | 2005-09-27 | 2010-09-15 | 株式会社東芝 | 音声認識装置、音声認識方法および音声認識プログラム |
US7809566B2 (en) * | 2005-10-14 | 2010-10-05 | Nuance Communications, Inc. | One-step repair of misrecognized recognition strings |
US20100153885A1 (en) * | 2005-12-29 | 2010-06-17 | Rovi Technologies Corporation | Systems and methods for interacting with advanced displays provided by an interactive media guidance application |
US9436951B1 (en) * | 2007-08-22 | 2016-09-06 | Amazon Technologies, Inc. | Facilitating presentation by mobile device of additional content for a word or phrase upon utterance thereof |
JP4967519B2 (ja) * | 2006-08-11 | 2012-07-04 | 日産自動車株式会社 | 音声認識装置 |
US9318108B2 (en) | 2010-01-18 | 2016-04-19 | Apple Inc. | Intelligent automated assistant |
US9015172B2 (en) * | 2006-09-22 | 2015-04-21 | Limelight Networks, Inc. | Method and subsystem for searching media content within a content-search service system |
US8631440B2 (en) * | 2007-04-30 | 2014-01-14 | Google Inc. | Program guide user interface |
US8140335B2 (en) * | 2007-12-11 | 2012-03-20 | Voicebox Technologies, Inc. | System and method for providing a natural language voice user interface in an integrated voice navigation services environment |
US20090210411A1 (en) * | 2008-02-15 | 2009-08-20 | Oki Electric Industry Co., Ltd. | Information Retrieving System |
JP5220451B2 (ja) | 2008-03-26 | 2013-06-26 | 日本電気株式会社 | 電話受付システム、電話受付方法、プログラム、及び記録媒体 |
KR20090107365A (ko) | 2008-04-08 | 2009-10-13 | 엘지전자 주식회사 | 이동 단말기 및 그 메뉴 제어방법 |
KR101513615B1 (ko) * | 2008-06-12 | 2015-04-20 | 엘지전자 주식회사 | 이동 단말기 및 그 음성 인식 방법 |
US8108214B2 (en) * | 2008-11-19 | 2012-01-31 | Robert Bosch Gmbh | System and method for recognizing proper names in dialog systems |
US9367608B1 (en) * | 2009-01-07 | 2016-06-14 | Guangsheng Zhang | System and methods for searching objects and providing answers to queries using association data |
US10706373B2 (en) * | 2011-06-03 | 2020-07-07 | Apple Inc. | Performing actions associated with task items that represent tasks to perform |
CN102483917B (zh) * | 2009-07-15 | 2016-03-16 | 谷歌公司 | 针对显示文本的命令 |
KR101253104B1 (ko) * | 2009-09-01 | 2013-04-10 | 한국전자통신연구원 | 패턴 데이터베이스화 장치 및 그 방법, 이를 이용한 음성 이해 장치 및 그 방법 |
JP5535238B2 (ja) * | 2009-11-30 | 2014-07-02 | 株式会社東芝 | 情報処理装置 |
WO2011068170A1 (ja) * | 2009-12-04 | 2011-06-09 | ソニー株式会社 | 検索装置、検索方法、及び、プログラム |
EP3091535B1 (en) * | 2009-12-23 | 2023-10-11 | Google LLC | Multi-modal input on an electronic device |
US8522283B2 (en) * | 2010-05-20 | 2013-08-27 | Google Inc. | Television remote control data transfer |
US8359020B2 (en) * | 2010-08-06 | 2013-01-22 | Google Inc. | Automatically monitoring for voice input based on context |
JP2015038526A (ja) * | 2011-03-31 | 2015-02-26 | 株式会社東芝 | 音声処理装置、及び音声処理方法 |
JP2012221108A (ja) * | 2011-04-06 | 2012-11-12 | Sony Corp | 情報処理装置、情報処理方法、プログラム、履歴管理サーバ、履歴管理方法および情報処理システム |
US8954329B2 (en) * | 2011-05-23 | 2015-02-10 | Nuance Communications, Inc. | Methods and apparatus for acoustic disambiguation by insertion of disambiguating textual information |
WO2013012107A1 (ko) * | 2011-07-19 | 2013-01-24 | 엘지전자 주식회사 | 전자 기기 및 그 제어 방법 |
US8488916B2 (en) * | 2011-07-22 | 2013-07-16 | David S Terman | Knowledge acquisition nexus for facilitating concept capture and promoting time on task |
WO2013022221A2 (en) * | 2011-08-05 | 2013-02-14 | Samsung Electronics Co., Ltd. | Method for controlling electronic apparatus based on voice recognition and motion recognition, and electronic apparatus applying the same |
US20130073293A1 (en) * | 2011-09-20 | 2013-03-21 | Lg Electronics Inc. | Electronic device and method for controlling the same |
US10424291B2 (en) | 2012-12-28 | 2019-09-24 | Saturn Licensing Llc | Information processing device, information processing method, and program |
DK2994908T3 (da) * | 2013-05-07 | 2019-09-23 | Veveo Inc | Grænseflade til inkrementel taleinput med realtidsfeedback |
KR102227599B1 (ko) * | 2013-11-12 | 2021-03-16 | 삼성전자 주식회사 | 음성인식 시스템, 음성인식 서버 및 디스플레이 장치의 제어방법 |
EP3089157B1 (en) * | 2013-12-26 | 2020-09-16 | Panasonic Intellectual Property Management Co., Ltd. | Voice recognition processing device, voice recognition processing method, and display device |
CN105940399B (zh) * | 2014-03-28 | 2019-04-26 | 松下知识产权经营株式会社 | 声音检索装置、声音检索方法以及显示装置 |
US10978052B2 (en) * | 2014-04-16 | 2021-04-13 | Facebook, Inc. | Email-like user interface for training natural language systems |
US9852136B2 (en) * | 2014-12-23 | 2017-12-26 | Rovi Guides, Inc. | Systems and methods for determining whether a negation statement applies to a current or past query |
US10268491B2 (en) * | 2015-09-04 | 2019-04-23 | Vishal Vadodaria | Intelli-voyage travel |
US12026214B2 (en) * | 2015-09-28 | 2024-07-02 | Developing Software LLC | Location based push notification and multi-user class social introduction |
US9978367B2 (en) * | 2016-03-16 | 2018-05-22 | Google Llc | Determining dialog states for language models |
US10304444B2 (en) * | 2016-03-23 | 2019-05-28 | Amazon Technologies, Inc. | Fine-grained natural language understanding |
-
2013
- 2013-11-20 US US14/648,890 patent/US10424291B2/en active Active
- 2013-11-20 CN CN201911128993.XA patent/CN110889265B/zh active Active
- 2013-11-20 WO PCT/JP2013/081312 patent/WO2014103568A1/ja active Application Filing
- 2013-11-20 EP EP21177740.4A patent/EP3896596A1/en active Pending
- 2013-11-20 EP EP13867146.6A patent/EP2940683A4/en not_active Ceased
- 2013-11-20 CN CN201380067142.3A patent/CN104871240A/zh active Pending
- 2013-11-20 BR BR112015014830-1A patent/BR112015014830B1/pt active IP Right Grant
- 2013-11-20 JP JP2014554243A patent/JPWO2014103568A1/ja active Pending
-
2018
- 2018-05-25 JP JP2018100317A patent/JP6603754B2/ja active Active
-
2019
- 2019-07-26 US US16/523,675 patent/US11100919B2/en active Active
-
2021
- 2021-07-27 US US17/386,260 patent/US11676578B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2001166914A (ja) | 1999-12-06 | 2001-06-22 | Canon Inc | 情報入力装置、情報入力方法、及び記憶媒体 |
JP2010134074A (ja) * | 2008-12-03 | 2010-06-17 | Toshiba Corp | 音声認識装置、方法、およびプログラム |
JP2012063526A (ja) * | 2010-09-15 | 2012-03-29 | Ntt Docomo Inc | 端末装置、音声認識方法および音声認識プログラム |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10424291B2 (en) | 2012-12-28 | 2019-09-24 | Saturn Licensing Llc | Information processing device, information processing method, and program |
US11100919B2 (en) | 2012-12-28 | 2021-08-24 | Saturn Licensing Llc | Information processing device, information processing method, and program |
WO2020240958A1 (ja) * | 2019-05-30 | 2020-12-03 | ソニー株式会社 | 情報処理装置、情報処理方法、及びプログラム |
US12033630B2 (en) | 2019-05-30 | 2024-07-09 | Sony Group Corporation | Information processing device, information processing method, and program |
Also Published As
Publication number | Publication date |
---|---|
EP2940683A1 (en) | 2015-11-04 |
US20150310854A1 (en) | 2015-10-29 |
JP2018170015A (ja) | 2018-11-01 |
EP2940683A4 (en) | 2016-08-10 |
BR112015014830A2 (pt) | 2017-07-11 |
JPWO2014103568A1 (ja) | 2017-01-12 |
BR112015014830B1 (pt) | 2021-11-16 |
US11676578B2 (en) | 2023-06-13 |
CN110889265B (zh) | 2024-01-30 |
CN104871240A (zh) | 2015-08-26 |
JP6603754B2 (ja) | 2019-11-06 |
CN110889265A (zh) | 2020-03-17 |
EP3896596A1 (en) | 2021-10-20 |
US20210358480A1 (en) | 2021-11-18 |
US11100919B2 (en) | 2021-08-24 |
US20190348024A1 (en) | 2019-11-14 |
US20230267920A1 (en) | 2023-08-24 |
US10424291B2 (en) | 2019-09-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP6603754B2 (ja) | 情報処理装置 | |
EP3190512B1 (en) | Display device and operating method therefor | |
US20200260127A1 (en) | Interactive server, display apparatus, and control method thereof | |
JP6375521B2 (ja) | 音声検索装置、音声検索方法、および表示装置 | |
JP6244560B2 (ja) | 音声認識処理装置、音声認識処理方法、および表示装置 | |
JP2014021494A (ja) | 音声認識エラー修正方法及びそれを適用した放送受信装置 | |
US11664024B2 (en) | Artificial intelligence device | |
US11412281B2 (en) | Channel recommendation device and operating method therefor | |
US20220293106A1 (en) | Artificial intelligence server and operation method thereof | |
JP2015106203A (ja) | 情報処理装置、情報処理方法、及びプログラム | |
US12087296B2 (en) | Display device and artificial intelligence server | |
KR20190117913A (ko) | 디스플레이 장치 | |
CN111344664B (zh) | 电子设备及其控制方法 | |
US11706482B2 (en) | Display device | |
US12125475B2 (en) | Information processing device, information processing method, and program | |
KR102463066B1 (ko) | 디스플레이 장치, 서버 장치 및 이들을 포함하는 디스플레이 시스템과 그 컨텐츠 제공 방법들 | |
US20240203419A1 (en) | Display device and operation method thereof | |
CN118278390A (zh) | 电子设备、文本纠错方法、文本纠错装置及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 13867146 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2014554243 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 14648890 Country of ref document: US |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2013867146 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112015014830 Country of ref document: BR |
|
ENP | Entry into the national phase |
Ref document number: 112015014830 Country of ref document: BR Kind code of ref document: A2 Effective date: 20150619 |