WO2016006088A1 - Dispositif électronique, procédé et programme - Google Patents

Dispositif électronique, procédé et programme Download PDF

Info

Publication number
WO2016006088A1
WO2016006088A1 PCT/JP2014/068478 JP2014068478W WO2016006088A1 WO 2016006088 A1 WO2016006088 A1 WO 2016006088A1 JP 2014068478 W JP2014068478 W JP 2014068478W WO 2016006088 A1 WO2016006088 A1 WO 2016006088A1
Authority
WO
WIPO (PCT)
Prior art keywords
standard
exceeds
amount
criterion
classified
Prior art date
Application number
PCT/JP2014/068478
Other languages
English (en)
Japanese (ja)
Inventor
能康 伊藤
Original Assignee
株式会社 東芝
東芝ライフスタイル株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社 東芝, 東芝ライフスタイル株式会社 filed Critical 株式会社 東芝
Priority to JP2016532376A priority Critical patent/JP6305538B2/ja
Priority to PCT/JP2014/068478 priority patent/WO2016006088A1/fr
Publication of WO2016006088A1 publication Critical patent/WO2016006088A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor

Definitions

  • This embodiment relates to an electronic device, a method, and a program for performing processing according to the voice of one or more people.
  • the speech recognition apparatus has a function of digitizing speech uttered by a human and analyzing the digitized speech data to identify words. Furthermore, the speech recognition apparatus can also determine a certain degree of command by identifying words from the speech data.
  • the speech recognition device can be used in various fields such as robots, and further applications are expected.
  • a speech recognition there may be a case where it is not enough to analyze the speech data from one or more speakers and identify the word by interpreting the command word or the command word. .
  • an object of the present embodiment is to provide an electronic device, a method, and a program that can effectively use a topic of voice or conversation content uttered by one or a plurality of speakers.
  • the amount of the portion classified as the first reference among the plurality of portions included in the audio data and the reception unit that receives the audio data uttered by one or more users A processing unit that executes processing using the first reference when the threshold is exceeded, of a plurality of parts included in the audio data, When the amount exceeds the threshold value, the processing unit performs processing using the second criterion.
  • the television receiver 100 includes a digital tuner device 102.
  • the digital tuner device 102 includes a plurality of tuners that can receive, for example, a scrambled digital terrestrial broadcast signal.
  • the digital tuner device 102 may be included in a tuner (not shown) for receiving BS / CS digital broadcasting, for example.
  • the broadcast program signals of a plurality of channels received by the digital tuner device 102 are input to the TS processing unit 103.
  • TSs Transport Streams
  • CH channels
  • the multiplexed TS includes a packet sequence of broadcast programs for each channel. Identification information for channel and packet identification is added to the packet of each channel.
  • the multiplexed TS can be stored in the recording / reproducing device 223 under the control of the microprocessor 200.
  • the packet including the control information is read by the microprocessor 200.
  • the recording / reproducing apparatus 223 includes, for example, a hard disk drive and an optical disk recording / reproducing apparatus.
  • the optical disc is a digital versatile disc (DVD) or a Blu-ray disc (BD).
  • a packet including control information sent from the TS processing unit 103 to the microprocessor 200 describes, for example, ECM (Entitlement Control Message) that is encryption information of a broadcast program, and event information such as a program name, performer, and start time.
  • ECM Entitlement Control Message
  • Event Information Table Event Information Table
  • EPG Electronic Program Guide
  • the video data included in the packet is encoded by, for example, MPEG (Moving Picture Expert Group) system, AVC (Advanced Video Coding) system, or the like.
  • the audio data in the audio packet is encoded by, for example, a PCM (Pulse Code Modulation) method, a Dolby method, an MPEG method, or the like. Therefore, the data in each packet is decoded according to each method.
  • the packet including audio data and video data among the packets separated by the TS processing unit 103 is decoded by the microprocessor 200 according to the encoding method.
  • the decoded audio data is supplied to the speaker 112 via the amplifier 111.
  • the packet including the video data is also decoded by the microprocessor 200 according to the encoding method.
  • the decoded video data is improved in image quality by the image quality improving circuit 104 and supplied to the liquid crystal display panel 106 via the liquid crystal driving circuit 105.
  • the flash memory 221 and the random access memory 222 are connected to the microprocessor 200.
  • the flash memory 221 stores parameters used by the microprocessor 200, certain image data, and the like.
  • the random access memory 222 is used for temporary data storage, program information storage, and the like.
  • the remote control receiving unit 401 receives an operation signal from the remote controller and transmits it to the microprocessor 200. If the command from the remote controller is a power on or off command, remote control receiving unit 401 can transfer the command to power control unit 402.
  • a command to turn on or off the power supply, and various operation commands for the television receiver may be input from the transmission / reception unit 403.
  • various operation commands are input from the transmission / reception unit 403.
  • the power supply circuit 411 generates a stable power supply voltage suitable for each block of the television receiver.
  • the microprocessor 200 can multiplex data, graphics, program guides, scene comments, messages, etc. to the main video signal.
  • the microprocessor 200 includes a command processing unit that processes commands such as a calculation unit and operation input. Further, the microprocessor 200 includes a communication control unit, an operation device management unit, a recording / playback control unit, a display processing unit, and a memory. Further, an EPG (Electric program guide) data processing unit, a recording control unit, and the like are included but are not shown here.
  • a command processing unit that processes commands such as a calculation unit and operation input.
  • the microprocessor 200 includes a communication control unit, an operation device management unit, a recording / playback control unit, a display processing unit, and a memory. Further, an EPG (Electric program guide) data processing unit, a recording control unit, and the like are included but are not shown here.
  • EPG Electronic program guide
  • the microprocessor 200 adjusts the entire operation sequence.
  • the command processing unit can analyze an operation command input from the outside and reflect an operation corresponding to the command to the television receiver.
  • the operation device management unit can grasp and manage device identification data such as a portable terminal or a remote controller that gives an operation signal to the microprocessor 200. When the operation signal from the portable terminal or the remote controller is input, the microprocessor 200 confirms whether or not the ID of the operation device is registered in the operation device management unit.
  • the recording / playback control unit controls the operation of the recording system that records the received program in the recording / playback apparatus 223.
  • the recording / playback control unit controls the operation of the playback system that plays back the program recorded in the recording / playback apparatus 223.
  • the playback mode includes normal playback, digest playback, and program playback based on a scene list.
  • the display processing unit can generally control the video signal displayed on the liquid crystal display panel 106.
  • the display processing unit can also perform adjustment processing such as the resolution, display size, and display area of the image signal. Further, it is possible to display a video signal of a program guide created using EPG data and a menu video signal created in the microprocessor 200 in response to an operation signal.
  • the flash memory 221 can store various data and applications stored in the microprocessor 200.
  • the communication control unit can communicate with the outside via the transmission / reception unit 403 to capture operation commands, data, contents, and the like.
  • the captured content and data can be stored in the recording / reproducing device 223 or the memories 221, 222, for example.
  • the communication control unit can transmit data, content, and the like from the television receiver to the outside.
  • the communication control unit can perform mutual communication with the server 501 via the transmission / reception unit 403 and the Internet 500.
  • the server may be a home server or an in-area server.
  • the communication control unit can perform mutual communication with the mobile terminal via the transmission / reception unit 403 and the access point. As a result, data can be transmitted from the television receiver to the portable terminal, and data can be captured from the portable terminal to the television receiver.
  • the mobile terminal can also remotely control the television receiver.
  • the portable terminal can access the server via a base station (not shown), the Internet, or the like. It is possible to download various applications and game software as well as contents provided by the server and transfer them to the microprocessor 200 via the transmission / reception unit 403.
  • the portable terminal can transfer information for acquiring content (for example, a web server address, a mail address, a network address, etc.) to the microprocessor 200 via the transmission / reception unit 403.
  • a microphone 301 is further provided.
  • Voice (audio) data collected by the microphone 301 is processed by the microphone control unit 302.
  • the microphone control unit 302 digitizes audio data.
  • the digitized audio data is sent to the microprocessor 200.
  • the microphone control unit 302 may perform voice noise removal, speaker orientation and distance specification, voice collection in a specific direction, and the like. Audio digital data can be temporarily stored in the memory 222 for a certain period of time.
  • FIG. 2 is a diagram schematically showing a block configuration of a main part of one embodiment. Most of this block configuration is configured in the microprocessor 200. However, part of processing such as speech recognition may be performed by the microphone control unit 302.
  • the audio data collected by the microphone 301 is stored in a memory device.
  • the memory device only needs to be capable of digitally recording, for example, a conversation of at least about 2 to 3 minutes of audio data. Any memory device may be used as long as it erases the old one while leaving the data of the most recent conversation.
  • the recording time is not limited, and it may be one that can record and store a longer time.
  • the sound collection method using the microphone 301 can take various forms. For example, there is a method of collecting sound via a wireless medium (radio waves, infrared rays, etc.) using a remote controller with a built-in microphone. In addition, there is a method in which ambient sound is collected by the mobile terminal and audio data is transmitted from the mobile terminal to the voice recognition unit 601. Also, the microphone can be installed in various forms. For example, various forms such as installing in a living room, installing in the vicinity of a refrigerator in a kitchen, or installing on a table for eating are selectively or collectively used.
  • the audio data read from the memory device is input to the voice recognition device 601 in the voice data processing unit 600 as the processing unit via the reception unit 303.
  • the data of the most recent conversation may be constantly input, or input may be started when a user operation is performed.
  • the voice recognition unit 601 separates voice data uttered by a person from the digitized audio data.
  • a speech sentence of a speaker is estimated.
  • the speech sentence of the speaker A and the speech sentence of the speaker B are distinguished, for example, 60 seconds in total (not limited to this length, 2 minutes, Voice data of a conversation of about 3 minutes may be recognized.
  • the speech volume (strongness) level may also be measured. Sentences that speakers A and B each utter during conversation are referred to as spoken sentences.
  • time information is added to each utterance sentence.
  • the utterance sentence estimated by the voice recognition unit 601 is input to the morphological analysis unit 602.
  • the morphological analysis unit 602 decomposes the speech sentence obtained by the speech recognition unit 601 into parts of speech, and extracts words corresponding to the parts of speech from the database of the database management unit 608 based on the parts of speech. Since the uttered sentence is a word or a set (sentence) of the words uttered by the speaker through conversation, a plurality of parts of speech are detected.
  • the intention determination unit 603 determines that the intention of the utterance is “news (symmetric)” or “want to see (playback instruction)”.
  • the topic is clear.
  • the category of the utterance is determined.
  • part-of-speech data such as nouns included in each utterance is input to a category classification unit 604 that classifies conversation categories.
  • the category classification unit 604 categorizes the extracted words (nouns) and the like by substituting synonyms and synonyms from the database.
  • the primary category of each utterance is specified.
  • a plurality of candidates may be acquired for the primary category for one utterance sentence. For example, when terms such as Mt. Fuji, mountain climbing, mountain climbing equipment, Everest, weather, mountain work, etc. are detected, a category related to mountain climbing, a category related to mountain weather, and a category related to business for mountain climbing are estimated. However, in this case, the most frequent word is “mountain”, and the categories are dispersed. If terms related to the weather such as sunny, snowstorm, snow, rain, etc. occur frequently, it is estimated that the conversation is related to “weather” as a category. Therefore, in this case, it can be narrowed down that the topic is “weather conversation about the mountain”.
  • a threshold may be set with respect to the number of times a certain word is detected, and “words” appearing exceeding the threshold may be adopted and used for category estimation.
  • the speech volume statistics unit 606 refers to the time information for each utterance sentence acquired by the clock unit 605, retroactively, for example, for 60 seconds from the last utterance (or utterance), and tabulates the number of utterances of each category within the 60 seconds. To do.
  • the topic identification unit 607 uses the number of utterances counted by the speech volume statistics unit 606, retroactively from the last utterance, and for example, the category most estimated most recently and the word most detected most recently. This is the part to be extracted. This determines the most categories and words for 60 seconds. That is, the topic of the whole conversation for 60 seconds is estimated by the most categories and words. The fact that the topic has been estimated comes to the fact that the object to be searched for is determined based on this topic. Topics are databased beforehand, for example, by combinations of categories and words. Further, the database may be increased based on the learning function.
  • the topic identified as described above (based on category and word, but the category may be dispersed) is sent to the device control unit (also referred to as topic related information extraction unit).
  • the device control unit also referred to as topic related information extraction unit.
  • the device control unit 700 is a broadcast program recommendation device stored in, for example, a television receiver or a server.
  • the device control unit 700 can take various forms, and other embodiments will be described later.
  • the device control unit 700 includes a content search unit 701. Also, a content management unit 702 that stores program information, a content presentation unit 703, and a graphic user interface (GUI) processing unit 704 are provided.
  • a content management unit 702 that stores program information
  • a content presentation unit 703 that stores program information
  • a graphic user interface (GUI) processing unit 704 are provided.
  • the content search unit 701 searches the program information of the content management unit 702 for program information related to the topic. For example, if the topic is “mountain”, a program name including “mountain” in the program information is searched. The content search unit 701 supplies the searched target program name to the content presentation unit 703. Furthermore, when the topic includes other terms, for example, “weather on the mountain”, the program can be narrowed down.
  • the content search unit 701 matches the topic with the recorded content managed by the content management unit 702 and the program information of the future program, and identifies the program that is the center of the topic.
  • This program may be one or plural.
  • the content presentation unit 703 controls the GUI processing unit 704 to display the target program name and its channel on the liquid crystal display panel. Also, the content presentation unit 703 can output the program name of the recommended program via the voice synthesis unit 611. When a plurality of programs are extracted, program names and channels of the plurality of programs are presented as a program list.
  • a database management unit 608 manages a word database and a thesaurus (synonym) database necessary for the speech recognition unit 601 and the category classification unit 604.
  • the topic specifying unit 607 described above can obtain topic data that is data of a word and / or category that is the center for specifying a topic and serves as a trigger for controlling an electronic device.
  • the topic control data 700 can be used by the device control unit 700 to present a recommended program to the user.
  • FIG. 3 is a flowchart for searching for content (program name) related to the most recent topic (for example, a topic in a period retroactive for 60 seconds from the present time) and performing a process for posting the program information.
  • content program name
  • FIG. 3 is a flowchart for searching for content (program name) related to the most recent topic (for example, a topic in a period retroactive for 60 seconds from the present time) and performing a process for posting the program information.
  • the presentation process is started, for example, by operating a specific button on the TV operation screen of the remote controller or the portable terminal (step S1).
  • step S2 the elapsed time from the previous voice input is checked (step S2). If the predetermined time has not elapsed, the dialogue is continued, so that the input voice is recognized. That is, voice input start (step S4), recording of the utterance time of the input voice (step S5), and voice recognition (step S6) are executed.
  • the voice utterance time is recorded in the memory MEMO1 as the previous (final) utterance time, and is used in step S2 for checking the passage of time.
  • the recognition character string obtained by speech recognition is subjected to morphological analysis with reference to the word database.
  • voice input start step S4
  • utterance time recording step S5
  • voice recognition step S6
  • morphological analysis step S7
  • the last utterance time is temporarily stored in the memory MEMO1.
  • the recognized character string is stored in the memory MEMO2.
  • the database DB1 is used as the word database.
  • the utterance is “I want to see the news”, it is classified as “news (noun)”, “(particle)”, “see (verb)”, “tai (auxiliary verb)”. . Further, the intention of the utterance is determined as “news (target)” or “want to see (playback instruction)” (step S8). That is, the audio data is divided into a plurality of parts and classified for each part of speech.
  • step S9 determines whether there is an instruction word such as “I want to see”. If there is no directive, the process proceeds to category classification step S11.
  • the topic including the instruction word is gathered in another holder. Then, at the time when the analysis of the conversation for 60 seconds is completed, the statistical processing of the topic including the instruction wording is performed, and a message regarding the most frequent topic is presented to the user.
  • the user can instruct, for example, by a remote controller whether or not to allow an instruction related to the topic to be executed.
  • step S9 if there is no directive, category classification is performed.
  • category classification is performed, synonyms and synonyms of the thesaurus database DB2 are used.
  • step S12 the utterance time for each category is recorded, and the processing routine returns to step S2. The utterance time is recorded in the memory MEMO3.
  • step S10 if the data processing of the conversation content for about 60 seconds has already been completed, the identified topic is sent to the device control unit. Thereby, in step S17, the target or recommended program is searched.
  • the search uses, for example, data in the content management database DB3. After the recommended program is specified, the program name is presented and the process ends (steps S17 and S18).
  • step S10 when the data processing of the conversation content for about 60 seconds has not been completed yet, the processing by the speech volume statistics unit 606 and the topic identification unit 607 is performed in steps S13, S14, and S15. That is, the speech volume statistics unit 606 refers to the time information for each utterance sentence acquired by the clock unit 605 and totals the number of utterances in each category within 60 seconds from the last utterance (or utterance). Then, the topic specifying unit 607 uses the number of utterances counted by the speech volume statistics unit 606 to determine the most frequent categories and words for 60 seconds. That is, the topic of the whole conversation for 60 seconds is estimated by the largest number of categories and words. The fact that the topic has been estimated comes to the fact that the object to be searched for is determined based on this topic.
  • step S16 the process returns from step S16 to steps S2 and S3, and the processes of steps S13 to S15 are performed again. If the topic is fixed, the process proceeds to step S17, and a recommended program is searched.
  • FIG. 4 shows an example in which the function of the television receiver 100 changes when the user is talking around the television receiver 100. Now, assume that the television receiver 100 is playing a certain program.
  • the morpheme analysis result may be acquired from the speech recognition unit.
  • the number of utterances may be statistically calculated for each utterance by the same speaker.
  • the reception unit 303 receives voice data uttered by one or more users.
  • the processing unit 600 determines the first Perform processing using criteria.
  • the processing unit 600 outputs the second reference.
  • the process can be executed using.
  • the first standard may be a keyword related to sports, for example, and the second standard may be a keyword related to drama, for example.
  • the above embodiment is an example in which the electronic device is a television receiver, and is intended to provide a recommended program to the user.
  • the first standard is a keyword related to sports
  • the second standard is a keyword related to drama.
  • various modifications can be made to the specifications of the first standard and the second standard.
  • the basic idea shown by the above embodiment can be linked with various electronic devices. Therefore, the classification of “keywords” as the first standard and the second standard can take various forms. When a designer classifies various keywords into a plurality of criteria, it is preferable to classify them according to items required by an electronic device to which the idea of the present invention is applied.
  • the processing unit when the user performs an operation for executing a first process (for example, power-on, program viewing, program search, or topic process) on the electronic device at a first time, the processing unit Sets the first period retroactively from the first time.
  • the processing unit includes the first reference among a plurality of first parts (words and / or categories) included in the first data of the speech uttered by the one or more users during the first period.
  • the second process can be executed by changing the first process using the first reference.
  • the first process is reception of a default channel or reception of a channel that was viewed last time
  • the second process is performed using the first reference. That is, as the second process, program selection adapted to the topic in the conversation for several minutes before the operation can be performed. That is, when the previous user preference and the current user preference have changed, the processing unit can follow this change and give priority to the current user preference.
  • a third process may be executed by changing the first process.
  • the above processing unit can operate without a trigger from the user.
  • the amount of the second portion classified into the first criterion among the plurality of second portions included in the second data of the speech uttered by one or a plurality of users in the second period is a threshold value. May be exceeded.
  • the processing unit can automatically execute the processing using the first reference without any explicit operation by the user.
  • the processing unit may perform the first operation without an explicit operation by the user. Processing can be performed automatically using two criteria.
  • the above execution includes, for example, a method of displaying a program name suitable for the topic in a partial area of the screen of the television receiver.
  • a method of informing the user that there is a recommended program suitable for the topic by blinking a specific lamp.
  • the speech data is divided into a plurality of parts, classified into parts of speech, statistical processing is executed, and “topic” is specified from “category” “word”.
  • this "topic identification method can be implemented in various embodiments. After the speech data is once classified into the smallest unit words, each word is combined with the preceding and following words. It may be formed into words that have meaning in words, and these words may be used as criteria (elements) for classification.
  • FIG. 5 shows a block configuration of another embodiment.
  • a microphone 901 is connected to the topic specifying unit 902 in the audio data processing unit.
  • the topic identification unit 902 can turn on and off the monitoring camera 903.
  • the surveillance camera 903 is not always necessary.
  • the topic data determined by the topic identification unit 902 is transmitted to the topic related information extraction unit 1100.
  • the topic related information extraction unit 1100 can extract information related to the topic and provide it to the user.
  • the provided information described in FIG. 2 and FIG. 3 is recommended program information.
  • the topic related information extraction unit 1100 may be installed in a certain home area 2001.
  • the topic-related information extraction unit 1100 may be provided in a television receiver that can be connected to the Internet. Further, the topic related information extraction unit 1100 may access and extract an external server 1111 in order to extract topic related information.
  • the area 2002 including the microphone 901 and the topic specifying unit 902 may be a server connected to the topic related information extracting unit 1100 via the Internet, for example, in the home.
  • the topic identification unit 902 In the case where a special topic is detected in the topic identification unit 902, for example, in the case of a scream of help, robbery, etc., and when the volume is higher than a predetermined level, for example, the monitoring camera 903 is switched to a high-definition video recording mode. It may be.
  • the topic identification unit 902 may also be installed in the external server 111.
  • FIG. 6 shows still another embodiment.
  • a microphone is installed near the refrigerator 3000 or in the kitchen.
  • a “topic” relating to cooking is determined from the food stored in the refrigerator 3000 and the conversation of the dish name.
  • the refrigerator 3000 includes a display screen 3001, a transceiver connected to the Internet, and a topic specifying unit. Note that the transceiver and the topic identification unit may be provided not in the refrigerator 3000 but in a home server, a server on the cloud, or a television receiver via a home LAN.
  • ingredients and names of dishes are transmitted to the service station 3200, for example.
  • the service station 3200 can transmit a recipe for cooking that can be cooked to the refrigerator 3000 from the words of the topic, ingredients, and the name of the dish.
  • the sample image which is cooking with the recipe transmitted simultaneously can be transmitted.
  • the user can cook with the ingredients in the refrigerator 3000 with reference to the recipe and sample images of cooking.
  • FIG. 7 shows an example implemented in a convenience store, for example.
  • This embodiment is an example in which the topic of conversation between the clerk B and the customer A is specified at the location of the accounting register C.
  • the topic identification unit switches the monitoring camera D to the high-definition video shooting mode. Can also automatically report to the police.
  • the embodiment can also be used as a support device for specifying a topic in a conference hall.
  • a topic to be identified or a topic not to be identified may be designated for a device that identifies a topic. That is, it may be possible to designate a topic area to be detected.
  • topics general news, political news, sports news, overseas, domestic, and program genres may be designated.
  • restricted word groups restricted in advance are classified in the topic specifying unit and can be specified by the user. Words based on utterances belonging to this restricted word group are used for topic determination. Excluded.
  • the topic to be detected can be restricted, and the topic detection speed can be increased.
  • the topic when a topic is specified, statistical processing of the frequency of appearance of each category and word counted within a set time (60 seconds) is performed. And the topic was identified from the category and the word that appeared most frequently within the set time. However, it is not necessary to adopt the category and word that appear most frequently, and the topic may be specified based on the category and word that appears the second most frequently. That is, the appearance frequency, which is an element for specifying the topic, may be switched between the maximum, the second, and the third.
  • the above-described topic detection may always be detected from the current conversation data, for example, three minutes ago.
  • the topic detection may be started from the time when the user starts some operation, and may be detected from the conversation data, for example, three minutes before the current time.
  • the set period for detecting the topic may be arbitrarily changed.
  • a topic detection section change button is provided in the operation menu. When this change button is pressed, a bar display indicating the date and time is obtained on the screen. When the user selects a desired setting section in the bar display with the cursor and presses the enter button, the topic detection section is set.
  • the constituent blocks of the audio data processing unit 600 described in FIG. 2 may be realized by software.
  • FIG. 8 shows another embodiment.
  • a morphological element analysis unit 602 compared to the configuration of FIG. 2, an intention determination unit 603, a category classification unit 604, a speech volume statistics unit 606, and a topic identification unit 607 are provided in an external service server 700A, for example.
  • the operation is the same as in the previous embodiment.
  • the price of the television receiver can be reduced.
  • the administrator may upgrade the new software in the server 700A.
  • the embodiment described above can include the following components.
  • the speech data processing unit morphologically analyzes speech data uttered during a set period, estimates a category from a word of part of speech obtained by morphological analysis and synonyms, synonyms and synonyms related to the word,
  • the topic that is the center of the conversation is specified based on the amount of speech of the category, and the topic data that serves as a trigger for controlling the electronic device is obtained, which is data of the word and / or category that is the center for specifying the topic. I am doing so.
  • the command word may be removed from the word used to identify the topic.
  • the topic data may be used for program search by the electronic device. Furthermore, the topic data may be acquired via a server. Further, the topic data may be used for control of the surveillance camera. Furthermore, the topic data can be used for data retrieval related to cooking.
  • a topic that is the center of conversation is specified by the amount of words and categories, it is possible to limit the words and categories to be used. It is also possible to realize most functions for executing the above-described operations by instructions of a computer program in the control unit.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

La présente invention a pour objet d'utiliser efficacement la voix ou le thème de contenu de conversation d'un ou plusieurs locuteurs. Un mode de réalisation comporte une unité de réception servant à recevoir des données de voix produites par un ou plusieurs utilisateurs, et un unité de traitement servant à exécuter un traitement en utilisant une première norme, lorsque l'ampleur d'une section catégorisée par la première norme parmi une pluralité de sections comprises dans les données de voix dépasse un seuil. De plus, l'unité de traitement exécute un traitement en utilisant une deuxième norme lorsque l'ampleur d'une section catégorisée par la deuxième norme parmi la pluralité de sections comprises dans les données de voix dépasse un seuil.
PCT/JP2014/068478 2014-07-10 2014-07-10 Dispositif électronique, procédé et programme WO2016006088A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
JP2016532376A JP6305538B2 (ja) 2014-07-10 2014-07-10 電子機器及び方法及びプログラム
PCT/JP2014/068478 WO2016006088A1 (fr) 2014-07-10 2014-07-10 Dispositif électronique, procédé et programme

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2014/068478 WO2016006088A1 (fr) 2014-07-10 2014-07-10 Dispositif électronique, procédé et programme

Publications (1)

Publication Number Publication Date
WO2016006088A1 true WO2016006088A1 (fr) 2016-01-14

Family

ID=55063761

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2014/068478 WO2016006088A1 (fr) 2014-07-10 2014-07-10 Dispositif électronique, procédé et programme

Country Status (2)

Country Link
JP (1) JP6305538B2 (fr)
WO (1) WO2016006088A1 (fr)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018001404A (ja) * 2016-07-07 2018-01-11 深▲せん▼狗尾草智能科技有限公司Shenzhen Gowild Robotics Co.,Ltd. 音声と仮想動作を同期させる方法、システムとロボット本体
WO2020144857A1 (fr) * 2019-01-11 2020-07-16 三菱電機株式会社 Dispositif de traitement d'informations, programme et procédé de traitement d'informations

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7189499B2 (ja) 2018-05-07 2022-12-14 オムロン株式会社 センサシステム

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006085565A1 (fr) * 2005-02-08 2006-08-17 Nippon Telegraph And Telephone Corporation Terminal de communication d’information, système de communication d’information, méthode de communication d’information, programme de communication d’information et support d’enregistrement sur lequel le programme est enregistré
JP2008287697A (ja) * 2007-04-16 2008-11-27 Sony Corp 音声チャットシステム、情報処理装置およびプログラム
JP2010128665A (ja) * 2008-11-26 2010-06-10 Kyocera Corp 情報端末及び会話補助プログラム
JP2014013494A (ja) * 2012-07-04 2014-01-23 Nikon Corp 表示制御装置、表示システム、表示装置、端末装置、表示制御方法及びプログラム

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2003198761A (ja) * 2001-12-26 2003-07-11 Casio Comput Co Ltd Webページ提供システム、webページ提供方法およびwebページ提供プログラム

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006085565A1 (fr) * 2005-02-08 2006-08-17 Nippon Telegraph And Telephone Corporation Terminal de communication d’information, système de communication d’information, méthode de communication d’information, programme de communication d’information et support d’enregistrement sur lequel le programme est enregistré
JP2008287697A (ja) * 2007-04-16 2008-11-27 Sony Corp 音声チャットシステム、情報処理装置およびプログラム
JP2010128665A (ja) * 2008-11-26 2010-06-10 Kyocera Corp 情報端末及び会話補助プログラム
JP2014013494A (ja) * 2012-07-04 2014-01-23 Nikon Corp 表示制御装置、表示システム、表示装置、端末装置、表示制御方法及びプログラム

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2018001404A (ja) * 2016-07-07 2018-01-11 深▲せん▼狗尾草智能科技有限公司Shenzhen Gowild Robotics Co.,Ltd. 音声と仮想動作を同期させる方法、システムとロボット本体
WO2020144857A1 (fr) * 2019-01-11 2020-07-16 三菱電機株式会社 Dispositif de traitement d'informations, programme et procédé de traitement d'informations
JPWO2020144857A1 (ja) * 2019-01-11 2021-03-11 三菱電機株式会社 情報処理装置、プログラム及び情報処理方法

Also Published As

Publication number Publication date
JPWO2016006088A1 (ja) 2017-04-27
JP6305538B2 (ja) 2018-04-04

Similar Documents

Publication Publication Date Title
US11151184B2 (en) Method and system for voice based media search
US20150082330A1 (en) Real-time channel program recommendation on a display device
US9219949B2 (en) Display apparatus, interactive server, and method for providing response information
EP2919472A1 (fr) Appareil d'affichage, procédé de commande d'un appareil d'affichage et système interactif
US9412368B2 (en) Display apparatus, interactive system, and response information providing method
US8250623B2 (en) Preference extracting apparatus, preference extracting method and preference extracting program
KR101298823B1 (ko) 구두 피드백을 프로세싱하고 디지탈 비디오 레코더(dvr) 기록 패턴을 업데이트하기 위한 설비
KR100879516B1 (ko) 디지털방송데이터를 이용한 휴대단말 사용자 관심정보수집방법 및 장치
KR20140089862A (ko) 디스플레이 장치 및 그의 제어 방법
CN103916704A (zh) 对话型接口设备及其控制方法
KR101100191B1 (ko) 멀티미디어 재생장치와 이를 이용한 멀티미디어 자료검색방법
JP2007215046A (ja) 情報処理装置、情報処理方法、情報処理プログラム、および記録媒体
US20030191629A1 (en) Interface apparatus and task control method for assisting in the operation of a device using recognition technology
JP6305538B2 (ja) 電子機器及び方法及びプログラム
EP3896985A1 (fr) Dispositif de réception et procédé de commande
JP6212719B2 (ja) 映像受信装置、情報表示方法および映像受信システム
JP2006054517A (ja) 情報提示装置、方法及びプログラム
CN112567757A (zh) 具备通知功能的电子装置及电子装置的控制方法
KR20100030474A (ko) 서비스 정보 제공방법 및 그 장치
KR102182689B1 (ko) 서버 및 그의 제어 방법
CN113228166B (zh) 指令控制装置、控制方法及非易失性存储介质
KR102118195B1 (ko) 서버 및 그의 제어 방법
KR102379049B1 (ko) 서버 및 그의 제어 방법
JP2001275052A (ja) コマーシャル情報受信装置
JP2017060059A (ja) 制御プログラム、記憶媒体、携帯通信機、番組関連情報提供装置及び番組関連情報表示方法

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 14897220

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2016532376

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 14897220

Country of ref document: EP

Kind code of ref document: A1