CN103811007B - Display device, voice acquisition device and its audio recognition method - Google Patents

Display device, voice acquisition device and its audio recognition method Download PDF

Info

Publication number
CN103811007B
CN103811007B CN201310553280.4A CN201310553280A CN103811007B CN 103811007 B CN103811007 B CN 103811007B CN 201310553280 A CN201310553280 A CN 201310553280A CN 103811007 B CN103811007 B CN 103811007B
Authority
CN
China
Prior art keywords
voice
coding line
application
display device
instruction word
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201310553280.4A
Other languages
Chinese (zh)
Other versions
CN103811007A (en
Inventor
蒋种赫
崔赞熙
柳熙涉
朴劲美
朴胜权
裵在铉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Samsung Electronics Co Ltd
Original Assignee
Samsung Electronics Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Samsung Electronics Co Ltd filed Critical Samsung Electronics Co Ltd
Priority to CN201510276510.6A priority Critical patent/CN104883587A/en
Publication of CN103811007A publication Critical patent/CN103811007A/en
Application granted granted Critical
Publication of CN103811007B publication Critical patent/CN103811007B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/06Transformation of speech into a non-audible representation, e.g. speech visualisation or speech processing for tactile aids
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/2803Home automation networks
    • H04L12/2816Controlling appliance services of a home automation network by calling their functionalities
    • H04L12/282Controlling appliance services of a home automation network by calling their functionalities based on user interaction within the home
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4126The peripheral being portable, e.g. PDAs or mobile phones
    • H04N21/41265The peripheral being portable, e.g. PDAs or mobile phones having a remote control device for bidirectional communication between the remote control device and client device
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/4104Peripherals receiving signals from specially adapted client devices
    • H04N21/4131Peripherals receiving signals from specially adapted client devices home appliance, e.g. lighting, air conditioning system, metering devices
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/41Structure of client; Structure of client peripherals
    • H04N21/422Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS]
    • H04N21/42203Input-only peripherals, i.e. input devices connected to specially adapted client devices, e.g. global positioning system [GPS] sound input device, e.g. microphone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/436Interfacing a local distribution network, e.g. communicating with another STB or one or more peripheral devices inside the home
    • H04N21/43615Interfacing a Home Network, e.g. for connecting the client to a plurality of peripherals
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/485End-user interface for client configuration
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/221Announcement of recognition results

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Human Computer Interaction (AREA)
  • Automation & Control Theory (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Acoustics & Sound (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Data Mining & Analysis (AREA)
  • Quality & Reliability (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Telephonic Communication Services (AREA)
  • Selective Calling Equipment (AREA)

Abstract

Display device, voice acquisition device and its audio recognition method are disclosed, the display device includes: the display unit for showing image;With the communication unit of multiple communication with external apparatus;And controller comprising the speech recognition engine for identifying user speech receives voice signal from voice acquisition unit, and controls the communication unit and receive candidate instruction word from least one of the multiple external device (ED) to identify the voice signal received.

Description

Display device, voice acquisition device and its audio recognition method
Technical field
It is related to display device, voice acquisition device and its speech recognition side with the consistent device and method of exemplary embodiment Method, more particularly, to display device, voice acquisition device and its audio recognition method of identification user speech.
Background technique
Speech identifying function is used in such as DTV (TV), air-conditioning, home theater, personal computer (PC) and moves In the various electronic devices of mobile phone etc..
In order to execute speech identifying function, for example the master device of TV should have the microphone for receiving user speech and identification The speech recognition engine of voice is inputted, and speech recognition engine can will input the candidate instruction word of voice and storage (instruction words) is compared, and identifies voice according to comparison result.
However, the relevant technologies electronic device with speech identifying function has the fixation device for receiving user speech, because The various input units of input voice, such as mobile phone are difficult to be utilized in this.In addition, if many candidate instruction words are provided, Discrimination will be then improved, but electronic device will compare candidate instruction word, this causes voice recognition processing speed slower.This Outside, because the memory capacity of master device is limited, the quantity of candidate instruction word cannot be constantly increasing.
Summary of the invention
Aspect accoding to exemplary embodiment provides a kind of display device comprising: display unit is shown on it Image;Communication unit, with multiple communication with external apparatus;And controller comprising identify that the speech recognition of user speech is drawn It holds up, receives voice signal from voice acquisition unit, and control the communication unit from the multiple external device (ED) at least One reception candidate instruction word is to identify the voice signal received.
Multiple voice acquisition units can be provided.If detected at least one of the multiple voice acquisition unit Voice input, then the controller receives voice signal from the voice acquisition unit for detecting voice input to it.
The voice acquisition unit may include at least one of the following: in providing in the display device Microphone, the first external microphone provided at least one of multiple external device (ED)s and with built-in microphone and The second different external microphone of first external microphone.
The external device (ED) may include at least one application that can manage candidate instruction word.
The display device can also include the native applications (native application) of management candidate instruction word.
The display device can also include storage unit, wherein store received candidate instruction word, and The speech recognition engine can identify the voice received by using the candidate instruction word of storage.
If at least one of multiple voice acquisition units detect that wake-up keyword, the controller can enable It detects the voice acquisition unit for waking up keyword, and receives voice signal from the voice acquisition unit enabled.
If having input trigger signal by manipulating the predetermined button provided in one in multiple voice acquisition units, Then the controller can enable the voice acquisition unit that trigger signal is had input by it, and obtain from the voice enabled Unit receives voice signal.
The controller can control the display unit and show speech recognition knot for the voice signal on it Fruit and candidate instruction word corresponding with speech recognition result.
The display unit can show the information of the application about management candidate instruction word on it.
The speech recognition engine can be believed by the voice for determining and receiving in the candidate instruction word received Number same or similar coding line identifies voice.
Aspect according to another exemplary embodiment provides a kind of voice acquisition device comprising: communication unit, It is communicated with the display device with speech identifying function;Voice acquisition unit receives user speech;Speech convertor, will The voice received is converted into electric voice signal;And controller, it controls the communication unit and is sent to the display device The voice signal and candidate instruction word converted are to identify the voice signal.
The voice acquisition device can also include at least one application that can manage candidate instruction word.
Aspect according to another exemplary embodiment provides a kind of audio recognition method of display device comprising: from Voice acquisition unit receives voice signal;Candidate instruction word is received from least one of multiple external device (ED)s to be received to identify The voice signal arrived;And voice signal and candidate instruction word identify the voice of user based on the received.
The audio recognition method can also include: the language that detection is input at least one of multiple voice acquisition units Sound, and receiving voice signal may include: from the voice acquisition unit reception voice signal for detecting that voice is inputted to it.
The voice acquisition unit may include at least one of the following: in providing in the display device Microphone, the first external microphone provided at least one of multiple external device (ED)s and with the display fill The second external microphone provided in the device different with the multiple external device (ED) is provided.
The external device (ED) may include at least one application for managing candidate instruction word.
The display device can also include the native applications of management candidate instruction word.
The audio recognition method can also include the received candidate instruction word of storage, and identify that voice can wrap It includes and voice is identified by using the candidate instruction word of storage.
Detection voice input may include: one wake-up keyword that detection is input in multiple voice acquisition units, And it enables and detects the voice acquisition unit for waking up keyword.
Detection voice input may include: according to the predetermined button provided in one in multiple voice acquisition units The input of manipulation detection trigger signal, and enable and pass through its voice acquisition unit for having input the trigger signal.
The audio recognition method can also include: speech recognition result and and language of the display for the voice signal The corresponding candidate instruction word of sound recognition result.
The display may include: the information of application of the display about management candidate instruction word.
Identification voice may include: the voice signal phase by determining with receiving in the candidate instruction word received With or similar coding line identify voice.
Detailed description of the invention
From referring to the drawings to the description of exemplary embodiment, above and/or other aspect be will be apparent and more It is readily appreciated that, in attached drawing:
Fig. 1 illustrates the examples of speech recognition system accoding to exemplary embodiment;
Fig. 2 is the block diagram of speech recognition system accoding to exemplary embodiment;
Fig. 3 illustrates the example of the execution of speech recognition accoding to exemplary embodiment;
Fig. 4 illustrates the example as the screen as the result is shown of speech recognition in Fig. 3;
Fig. 5 illustrates the example of the execution of speech recognition according to another exemplary embodiment;
Fig. 6 is the flow chart for showing the audio recognition method of speech recognition system accoding to exemplary embodiment;
Fig. 7 is the flow chart for showing the details for the process that voice input is detected in Fig. 6;And
Fig. 8 is the flow chart for showing the details for the process that speech recognition is executed in Fig. 6.
Specific embodiment
Exemplary embodiment is described in detail next, with reference to attached drawing.Exemplary embodiment can be come specific in a variety of manners It realizes, and is not limited to the exemplary embodiment illustrated here.For the sake of clarity, the description to well-known components is omitted, and And identical reference marker refers to identical element always.
Fig. 1 illustrates the examples of speech recognition system accoding to exemplary embodiment.
As shown in fig. 1, speech recognition system includes master device 100, multiple voice acquisition device 201 and 202, Yi Jiduo A external device (ED) 301,302 and 303.Master device 100, multiple voice acquisition device 201 and 202 and multiple external device (ED)s 301, it 302 and 303 is connected with each other to be in communication with each other.
Master device 100 includes voice acquisition unit 140 for receiving the such as microphone of user speech and for knowing Voice and communication unit 160 and multiple voice acquisition device 201 and 202 and multiple external device (ED)s 301,302 Shu Ru not be passed through With the speech recognition engine 181 of 303 communications.Master device 100 further includes being actuated to that master device 100 is made to perform various functions (clothes Business) primary (native) apply 171 and 172.Native applications 171 and 172 store corresponding with the function wherein in advance Candidate instruction word.That is, native applications 171 and 172 are included in available services scenarios (available service Scenario in).The candidate instruction word stored in native applications 171 and 172 is sent to speech recognition in speech recognition Engine 181 is so that speech recognition engine 181 is able to carry out speech recognition.
Each of multiple voice acquisition device 201 and 202 may include the such as Mike for receiving user speech The voice acquisition unit of wind, and voice signal corresponding with the voice received is sent to master device 100 for language Sound identification.
Multiple voice acquisition device 201 and 202 can receive the voice of user, and the voice is converted to electricity voice letter Number, and master device 100 is sent by the electric voice signal.Multiple voice acquisition device 201 and 202 can execute and main dress Set 100 wireless communication.Although wireless communication includes Wireless LAN, radio frequency (RF) communication, bluetooth, purple honeybee, infrared ray (IR) communication Etc., but it is not limited to this.
Multiple external device (ED)s 301,302 and 303 can according to need including for execute function (service) at least one Development and application (dev.Application).Storage is executed with by external device (ED) 301,302 and 303 wherein in advance for development and application The corresponding candidate instruction word of function.The candidate instruction word command stored in development and application is sent in speech recognition Speech recognition engine 181 is so that speech recognition engine 181 is able to carry out speech recognition.
Pre-stored candidate instruction word can be and apply in native applications 171 and 172 and in development and application The relevant coding line of function/operation.For example, if master device 100 is TV, channel change, volume adjustment with TV etc. Relevant candidate instruction word can be stored in one in native applications 171 and 172.If external device (ED) 302 is air-conditioning, It then can to the relevant candidate instruction word of the temperature of air-conditioning adjustment (increase/decrease), the intensity adjustment (strong/weak/medium) of wind etc. To be stored in application included in external device (ED) 302.
External device (ED) or voice acquisition device may include both voice acquisition unit and development and application.In such case Under, if voice is input into the voice acquisition unit in the first external device (ED) 301, it is stored in advance in the first external device (ED) Candidate instruction word in 301 development and application is sent to the speech recognition engine 181 of master device 100 to execute speech recognition.
Speech recognition system accoding to exemplary embodiment includes at least one voice acquisition unit.If detecting input To the voice of the voice acquisition unit, then speech recognition system has been detected by the voice acquisition of voice input by enabling to it Unit receives voice flow.Provided that multiple voice acquisition units, then speech recognition system can be described more by enabling The voice acquisition unit that voice inputs is had been detected by it to receive voice flow in a voice acquisition unit.Multiple voices Acquiring unit may include the built-in microphone provided in master device 100, in multiple external device (ED)s 301,302 and 303 The first external microphone for being there is provided at least one and with master device 100 and multiple external device (ED)s 301,302 and 303 not The second external microphone provided in same voice acquisition device 201 and 202.Voice acquisition device 201 and 202 and master device 100 and multiple external device (ED)s 301,302 and 303 separate.
If at least one of the multiple voice acquisition unit detects wake-up keyword, master device 100 can be with It enables and the voice acquisition unit of the wake-up keyword is detected by it, and receive voice from the voice acquisition unit enabled Signal.If passing through manipulation predetermined button (for example, event generation) at least one of the multiple voice acquisition unit Trigger signal is had input, then master device 100, which can be enabled, has input the voice acquisition unit of the input trigger signal simultaneously by it And voice signal is received from the voice acquisition unit enabled.
Master device 100 can be operated with speech recognition mode.If enabled by waking up keyword or trigger signal At least one voice acquisition unit, then master device 100 can disable other voice acquisition units to prevent speech recognition mistake Accidentally.Master device 100 can be with long-range or process speech recognition mode (distant or adjacent voice Recognition mode) operation.In order to which user is convenient, master device 100, which can be shown, is shown connected to display unit 130(slightly Afterwards will description) voice acquisition unit user interface (UI).
Master device 100 can receive candidate instruction word from least one of multiple external device (ED)s 301,302 and 303 to know The voice signal not received.The candidate instruction word received can be sent to speech recognition engine 181 to know for voice Not.
Multiple external device (ED)s 301,302 and 303 include at least one application of management candidate instruction word.Master device 100 is wrapped Include the native applications 171 and 172 of management candidate instruction word.It can be sent out by the candidate instruction word that native applications 171 and 172 manage Speech recognition engine 181 is sent to for speech recognition.
Master device 100 may be implemented as such as the display device in Fig. 2, such as TV (TV).
Fig. 2 is the block diagram of speech recognition system accoding to exemplary embodiment.
Display device 100 handles the picture signal from external image source of supply (not shown) to be based on processed image Signal shows image.
In speech recognition system accoding to exemplary embodiment, display device 100 is implemented as being based on sending out from broadcasting station Broadcast singal/broadcast message/the broadcast data penetrated handles the TV or set-top box of broadcast image.It is appreciated, however, that one In a or a number of other exemplary embodiments, other than TV or set-top box, display device 100 can be adapted for handling and show The various other equipment of diagram picture.For example, display device 100 may include personal computer (PC), portable computer etc..
Furthermore, it is to be understood that the type for the image that can be shown by display device 100 is not limited to broadcast image.For example, aobvious Showing device 100 can be shown based on the signal/data sent by various image source of supply (not shown) for example, video, static Image, application, screen display (OSD), the graphical user interface (GUI) for controlling various operations.
Accoding to exemplary embodiment, display device 100 may be implemented as intelligent TV.Intelligent TV can with real-time reception and Show broadcast singal, with real-time display broadcast singal and by the web browser function of the various contents of internet hunt, and And provide convenient user environment for doing above-mentioned item.Intelligent TV may include the opening for providing the user with interactive service Software platform, and various contents can be provided the user with by the software platform of the opening, for example, providing answering for reservation service With.The application can provide various types of services, for example, SNS, finance, news, weather, map, music, film, trip Play, e-book etc..
Display device 100 includes the speech recognition engine 181 of user speech for identification.It is corresponding with the voice identified Order (such as control command) be sent to respective application to execute operation.If application corresponding with control command is former Raw one applied in 171 and 172, then display device 100 executes operation by the application according to control command.If with control It is development and application that system, which orders corresponding application, then control command is sent to the external device (ED) 301,302 including development and application With 303.External device (ED) 301,302 and 303 can execute operation by the application according to control command.
Referring to Fig. 2, multiple voice acquisition device, such as mobile phone 200 and remote controler 300 are provided.Remote controler 300 can To serve as both voice acquisition device and external device (ED).Mobile phone 200 can be the smart phone that function is obtained with voice.
Remote controler 300 can be manipulated by user pre-set commands (control command) being sent to related device.Remote controler 300 It can be set to send a command to display device 100 or external device (ED), and may be implemented as sending a command to The comprehensive remote controller (integrated remote controller) of multiple devices.Remote controler 300 may include TV remote controler And/or air-conditioning remote control.
Voice acquisition device may be implemented as receiving the various devices of user speech, for example, cell phone, microphone are sent out Emitter etc..
As shown in the figure 2, provide multiple external device (ED)s, such as remote controler 300 and air-conditioning 400.As described above, remote controler 300 can serve as both voice acquisition device and external device (ED).
Although Fig. 2 illustrates the external device (ED) as remote controler 300 and air-conditioning 400, exemplary embodiment is not limited to In this.For example, external device (ED) may be implemented as executing various other electronic equipments of wireless communication, for example, being implemented as house Front yard movie theatre, wireless device, VCR, DVD, washing machine, refrigerator, robotic vacuum cleaner etc..If the external device (ED) Voice acquisition unit including such as microphone, then external device (ED) is also used as voice acquisition device.
External device (ED) accoding to exemplary embodiment includes the application 372 and 472 for executing function respectively.Using 372 and 472 Candidate instruction word is stored in advance, and manages the candidate instruction word.The candidate instruction word can be sent to display device 100 to be used for speech recognition.
External device (ED), that is, remote controler 300 and air-conditioning 400 can execute and by display device 100 according to the result of speech recognition The corresponding operation of the control command of transmission.
Hereinafter, referring to Fig. 2, it will be described in each element of speech recognition system.
Display device 100 may include: the picture receiver 110 for receiving picture signal;Image processor 120, processing The picture signal received from picture receiver 110;Display unit 130, based on the image letter handled by image processor 120 Number display image;Receive the first voice acquisition unit 140 of user speech;First speech convertor 150, the language that will be received Sound is converted into electric voice signal;With the first communication unit 160 of communication with external apparatus;Store the first storage unit of various data 170;And the first controller 180 of control display device 100.
Picture receiver 110 receives picture signal and sends image processor 120 for described image signal.For example, figure As receiver 110 can wirelessly receive rf signal from broadcasting station (not shown), or according to such as synthesizing Video, component video, super video, SCART(radio and television receiver Manufacturers Association), high-definition multimedia interface (HDMI) etc. standard receives picture signal in a wired manner.If described image signal includes broadcast singal, image Receiver 110 includes the tuner by channel tuner broadcast singal.
Picture signal can be received from external device (ED), the external device (ED) is for example, PC, AV equipment, smart phone, intelligence are flat Plate etc..Picture signal can be the data sent by the network of such as internet.In this case, display device 100 Network communication can be executed by the first communication unit 160, and may include additional network communication unit.Alternatively, Picture signal can be stored in the first storage unit 170(for example, flash memory, hard disk drive (HDD) etc.) in Data.First storage unit 170 can be provided in the inner/outer of display device 100.If the first storage unit 170 exists The external of display device 100 provides, then display device 100 may include that the first storage unit 170 is connected to its connector (not It shows).
Image processor 120 executes various image processing operations for described image signal, and by processed image Signal is output to display unit 130.
The image processing operations of image processor 120 may include decoding operate corresponding with various picture formats, go De-interlacing operation, frame updating rate conversion, zoom operations, the noise reduction operation for improving picture quality, details enhancing operation, row are swept Operation etc. is retouched, but it is not limited to this.Image processor 120 may be implemented as independently executing the individual of aforesaid operations Group, or be implemented as executing the system on chip (SoC) of comprehensive function.
Display unit 130 shows image based on the picture signal handled by image processor 120.Display unit 130 can To include liquid crystal display (LCD), plasma display panel (PDP), light emitting diode (LED), Organic Light Emitting Diode (OLED), surface-conduction electron emission device, carbon nanotube, nanocrystal etc., but it is not limited to this.
Display unit 130 can include additional element according to its implementation type.For example, the display list as LCD type Member 130 includes LCD panel (not shown), the back light unit (not shown) for emitting light into LCD panel and driving LCD panel Panel driving substrate (not shown).
Display unit 130 can show speech recognition result as the information about the voice identified.Speech recognition knot Fruit can be shown with the various forms of such as text, figure, icon etc..Text includes character and number.Display unit 130 Candidate instruction word can also be shown according to speech recognition result and application message.Later with reference to Fig. 4 in more detail to this into Row description.
User can be checked whether based on the speech recognition result shown on display unit 130 correctly identifies language Sound.User can manipulate the user input unit 330 in remote controler 300 to select and user's language from shown candidate instruction word The corresponding coding line of sound, or can choose and check information relevant to speech recognition result.
First voice acquisition unit 140 receives the voice of user, and may be implemented as microphone.
The voice inputted by the first voice acquisition unit 140 is converted into electric voice signal by the first speech convertor 150.Through The voice signal of conversion can be using the audio volume control of pulse code modulation (pcm) or compression.First speech convertor 150 can To be implemented as the A/D converter that the voice by user is converted into digital form.
If the first voice acquisition unit 140 is digital microphone, it does not need additional A/D conversion.In this feelings Under condition, the first voice acquisition unit 140 may include the first speech convertor 150.
First communication unit 160 and voice acquisition device and communication with external apparatus, that is, with mobile phone 200, remote controler 300 and air-conditioning 400 communicate.First communication unit 160 can execute including in infrared communication, RF, purple honeybee and bluetooth extremely Few one wireless communication.
First storage unit 170 is by the control of the first controller 180 come storing data.First storage unit 170 is by reality It is now non-volatile memory medium, such as flash memory, hard disk drive (HDD) etc..First storage unit 170 is read / write-in/is taken to change/delete/the first controller 180 of more new data access.
The data being stored in the first storage unit 170 include for example for driving the operating system of display device 100 (OS), various applications, image data and additional data for being run on OS etc..
First storage unit 170 can store the various data of user speech for identification.For example, the first storage unit 170 can store the coding line table 171(including candidate instruction word hereinafter, also referred to as candidate instruction word group), as with The corresponding voice messaging identified of the voice signal received.In coding line table 171, candidate instruction word can be by phase It should apply to manage.
First storage unit 170 can also store at least one application, for example, the first application 172 and the second application 173 with Execute the function of display device 100.First application 172 and the second application 173 will be described later by the first controller 180() Control executes the various functions of display device 100 to drive.Although Fig. 2, which is illustrated, is wherein mounted with that two are applied 172 Hes 173 display device 100, but exemplary embodiment is not limited to this.I.e., it is possible in display device 100 install three or More applications.
First application 172 and the second application 173 can manage candidate instruction word corresponding with performed function.By The candidate instruction word of 173 management of one application 172 and the second application can be registered to coding line table 171/ from coding line table 171 are deleted.
If candidate instruction word is registered to coding line table 171, speech recognition engine 181 is by using coding line table Candidate instruction word in lattice 171 executes speech recognition.
Can be registered to coding line table 171/ from coding line table 171 be deleted candidate instruction word may include by Remote controler 300(later will description) third using 372 management candidate instruction words and by air-conditioning 400 the 4th apply 472 The candidate instruction word of management.
The various elements of first controller 180 control display device 100.For example, the first controller 180 controls image procossing Device 120 handles picture signal, and in response to the order executive control operation from remote controler 300 to control display device 100 Overall operation.
For example, the first controller 180 may be implemented as the central processing unit (CPU) in conjunction with software.
First controller 180 may include the speech recognition engine 181 for identifying user speech.Speech recognition engine 181 Speech identifying function can be executed by using known speech recognition algorithm.For example, speech recognition engine 181 extract it is described The speech characteristic vector of voice signal, and by extracted speech characteristic vector and the finger for being stored in the first storage unit 170 The candidate instruction word in word table 171 is enabled to be compared to identify voice.If not identical with the speech characteristic vector The candidate instruction word being stored in coding line table 171, then speech recognition engine 181 can be by utilizing most like coding line Speech recognition result is adjusted to identify the voice.If there is multiple similar candidate instruction words, then the first controller 180 can To show multiple candidate instruction words on display unit 130, one in the multiple candidate instruction word for selection by the user.
Speech recognition engine 181 accoding to exemplary embodiment is implemented as the Embedded Speech Recognition System provided in CPU Engine 181, however it is without being limited thereto.For example, speech recognition engine 181 can be implemented as being provided separately from CPU in display device Device in 100, that is, be implemented as the additional chips of such as microcomputer.
However it is without being limited thereto, exemplary embodiment includes (hereinafter, will be by the server isolated with display device 100 Referred to as Cloud Server (not shown)) in provide speech recognition engine 181.Cloud Server passes through the network of such as internet and shows Showing device 100 communicates.The network can be cable network or wireless network.In this case, speech recognition engine 181 The Embedded Speech Recognition System engine provided in the CPU of Cloud Server is provided, or is implemented as separating with the CPU Ground provides the device in Cloud Server, that is, the additional chips of such as microcomputer.
First controller 180 can execute operation corresponding with the speech recognition result of speech recognition engine 181.Example Such as, if display device 100 is TV and user is watching film or news, speech recognition engine 181 can be identified Such as the voice of " volume rising ", " volume decline ", " more loud ", " smaller sound " etc., and the first controller 180 can root The volume of the film or news is adjusted according to the voice.
If the identification of speech recognition engine 181 is for controlling the language of the such as external device (ED) of remote controler 300 or air-conditioning 400 Sound, then the first controller 180 can control the first communication unit 160 control command is sent to it is corresponding with the voice identified External device (ED).For example, if speech recognition engine 181 identifies the voice of " increase temperature ", the first controller 180 can be with Identify the voice be for controlling air-conditioning 400, and control the first communication unit 160 send commands to air-conditioning 400 with Increase the temperature of air-conditioning 400.
Hereinafter, the detailed configuration of mobile phone 200 will be described.
As shown in Figure 2, mobile phone 200 may include the second voice acquisition unit 240 for receiving user speech, will connect The voice received be converted into electric voice signal the second speech convertor 250, with the second communication unit 260 of PERCOM peripheral communication, deposit It stores up the second storage unit 270 of data and controls the second controller 280 of mobile phone 200.
The second voice acquisition unit 240 for receiving user speech may be implemented as microphone.Second speech convertor 250 The voice received is converted into electric voice signal.Converted voice signal can using pulse code modulation (pcm) or The audio volume control of compression.Second speech convertor 250 may be implemented as the input voice of user being converted into digital form A/D converter.
If the second voice acquisition unit 240 is digital microphone, it does not need additional A/D conversion.In this feelings Under condition, the second voice acquisition unit 240 may include the second speech convertor 250.
The second communication unit 260 communicated with display device 100 can execute wire communication or wireless communication.The nothing Line communication may include at least one of RF, purple honeybee and bluetooth.
Second communication unit 260 can send display device 100 from the second speech convertor 250 for voice signal.
Second storage unit 270 can be by the control of second controller 280 come storing data.Second storage unit 270 It is implemented as the non-volatile memory medium of such as flash memory.Second storage unit 270 is read, and/write-in/is changed/is deleted Except the second controller 280 of/more new data accesses.
The data being stored in the second storage unit 270 may include for example for driving the OS of mobile phone 200, in OS Various applications, image data and additional data of upper operation etc..
Second controller 280 can control the various elements of mobile phone 200.For example, second controller 280 can respond Order is generated in user's manipulation, executes operation corresponding with the order of generation, and show on a display unit (not shown) Show result.
Second controller 280 may be implemented as the micro controller unit (MCU) in conjunction with software.
If having input user speech by the second voice acquisition unit 240, second controller 280 controls the second voice The voice of user is converted into electric voice signal and controls the second communication unit 260 for converted voice letter by converter 250 Number it is sent to display device 100.
Hereinafter, the detailed configuration of remote controler 300 will be described.
As shown in Figure 2, the remote controler 300 as voice acquisition device and external device (ED) may include: to receive user The user input unit 330 of manipulation;Receive the third voice acquisition unit 340 of user speech;Third speech convertor 350, The voice received is converted into electric voice signal;With the third communication unit 360 of PERCOM peripheral communication;The third of storing data stores Unit 370;And the third controller 380 of control remote controler 300.
User input unit 330 can send for various control commands or information by the manipulation of user and input Three controllers 380.User input unit 330 may be implemented as the Menu key provided in remote controler 300, number key etc.. If remote controler 300 is TV remote controler, user input unit 330 may include receiving the touching sensing of the touching input of user The action sensor of the movement of device (touch sensor), and/or sensing remote controler 300.
The third voice acquisition unit 340 for receiving user speech may be implemented as microphone.
The voice inputted by third voice acquisition unit 340 is converted into electric voice signal by third speech convertor 350.Through The voice signal of conversion can be using the audio volume control of pulse code modulation (pcm) or compression.Third speech convertor 350 can To be implemented as the input voice of user being converted into the A/D converter of digital form.
If third voice acquisition unit 340 is digital microphone, it does not need additional A/D conversion.In this feelings Under condition, third voice acquisition unit 340 may include third speech convertor 350.
Third communication unit 360 is communicated with display device 100.Third communication unit 360 executes wireless communication.It is described wireless Communication includes at least one of RF, purple honeybee and bluetooth.
Third communication unit 360 to display device 100 send voice signal from third speech convertor 350 and by Third storage unit 370(later will description) third apply 372 management candidate instruction words.
It may be implemented as such as fastly by the control of third controller 380 come the third storage unit 370 of storing data The non-volatile memory medium of flash memory etc..Third storage unit 370 is read, and/more new data is changed/deleted to/write-in/ Third controller 380 access.
The data being stored in third storage unit 370 include for example for driving the OS of remote controler 300, running on OS Various applications, image data and additional data etc..
Third storage unit 370 can also store at least one application, for example, the function for executing remote controler 300 Third applies 372.Third will be described using 372 by third controller 380(later) control drive, and execute remote control The various functions of device 300.Here, third will be described using 372 and the 4th using 472(later) development and application will be referred to as, with Just it is distinguished with the native applications of display device 100 172 and 173.
Although Fig. 2 illustrates the remote controler 300 for being wherein mounted with an application 372, exemplary embodiment is not limited to In this.I.e., it is possible to install two or more applications in remote controler 300.
Third can manage candidate instruction word corresponding with the function of executing using 372.By third using 372 management Candidate instruction word can be registered in the coding line table 171 of display device 100/from the coding line table of display device 100 It is deleted in 171.
Third controller 380 can control the various elements of remote controler 300.For example, third controller 380 can be in response to The user of user input unit 330, which manipulates, generates order, and control third communication unit 360 order of generation is sent to it is aobvious Showing device 100.
Third controller 380 may be implemented as the MCU in conjunction with software.
If having input user speech by third voice acquisition unit 340, third controller 380 controls third voice The voice of user is converted into electric voice signal and controls third communication unit 360 for converted voice letter by converter 350 Number it is sent to display device 100.
When communicating with display device 100, third controller 380 can send to display device 100 and store list by third The third of member 370 applies the candidate instruction word of 372 management.The candidate instruction word of transmission is registered in the instruction of display device 100 It is used to identify voice in word table 171 and by speech recognition engine 181.
If third controller 380 can be with as speech recognition as a result, sending control command by display device 100 The control command is received by third communication unit 360 and executes operation corresponding with the control command received.
Hereinafter, the detailed configuration of air-conditioning 400 will be described.
As shown in Figure 2, as the air-conditioning of external device (ED) 400 may include with the fourth communication unit 460 of PERCOM peripheral communication, 4th storage unit 470 of storing data and the 4th controller 480 for controlling air-conditioning 400.
The fourth communication unit 460 communicated with display device 100 can execute including in RF, purple honeybee and bluetooth at least One wireless communication.
Fourth communication unit 460 is sent to display device 100 later will description by the 4th storage unit 470() the 4th answer With the candidate instruction word of 472 management.
It may be implemented as such as quick flashing by the 4th storage unit 470 of the control storing data of the 4th controller 480 The non-volatile memory medium of memory.4th storage unit 470, which is read ,/write-in/changes/deletes/more new data the 4th Controller 480 accesses.
The data being stored in the 4th storage unit 470 include for example for driving the OS of air-conditioning 400, running on OS Various applications, image data and additional data etc..
4th storage unit 470 can also store at least one and apply (development and application), for example, for executing air-conditioning 400 Function the 4th apply 472.4th application 472 will be described later by the 4th controller 480() control drive, and Execute the various functions of air-conditioning 400.
Although Fig. 2 illustrates the air-conditioning 400 for being wherein mounted with an application 472, exemplary embodiment is not limited to This.I.e., it is possible to install two or more applications in air-conditioning 400.
4th application 372 can manage candidate instruction word corresponding with the function of executing.By 472 management of the 4th application Candidate instruction word can be registered in the coding line table 171 of display device 100/from the coding line table of display device 100 It is deleted in 171.
The various elements of 4th controller 480 control air-conditioning 400.For example, the 4th controller 480 can be in response to air-conditioning The user of 400 remote controler manipulates to receive control command, and according to the control command executive control operation of generation, for example, Adjust temperature.
4th controller 480 may be implemented as the MCU in conjunction with software.
When communicating with display device 100, the 4th controller 480 can send single by the 4th storage to display device 100 The candidate instruction word of 472 management of the 4th application of member 470.The candidate instruction word of transmission is registered in the instruction of display device 100 It is used to identify voice in word table 171 and by speech recognition engine 181.
If the 4th controller 480 can be with as speech recognition as a result, sending control command by display device 100 The control command is received by fourth communication unit 460 and executes operation corresponding with the control command received.
If detecting that voice inputs at least one of multiple voice acquisition units 140,240 and 340, it is used as root According to the first communication of the first controller 180 control of the display device 100 of the master device of the speech recognition system of exemplary embodiment Unit 140 receives voice signal from the voice acquisition unit that voice inputs is had been detected by it.First controller 180 passes through the One communication unit 140 is filled from least one of development and application 372 and 472 of multiple external device (ED)s 300 and 400 or from display The native applications 172 and 173 for setting 100 receive candidate instruction word to identify the voice signal received, and the candidate of transmission is referred to Word is enabled to be registered in the coding line table 171 of the first storage unit 170.Speech recognition engine 181 will be registered in coding line table Candidate instruction word in 171 is compared with the voice signal and identifies the voice.
Display device 100 can detecte the voice by the various devices input for inputting user speech.Display device 100 can Voice is identified to use the candidate instruction word provided by application, and can dynamically be registered/be deleted for speech recognition Candidate instruction word.Accordingly it is possible to prevent unnecessarily increasing the candidate instruction word of display device 100.
Display device 100 can receive speech recognition from voice acquisition unit 140, from native applications 172 and 173 to Lack one or receive candidate instruction word from development and application 372 and 472, and executes voice using speech recognition engine 181 Identification.
Hereinafter, the speech recognition of identification voice accoding to exemplary embodiment will be more fully described referring to Fig. 3 and Fig. 4 System.
Fig. 3 illustrates the example for executing speech recognition, and Fig. 4 illustrates the screen of the speech recognition result in display Fig. 3 The example of curtain.
As shown in Figure 3, display device 100 can have the candidate instruction word of registration, by least one application (including Native applications and development and application) it provides and is stored in coding line table 171.
For example, coding line A and B is by the first application 172(that is, native applications) it is sent to coding line table 171(501), and And (502) are stored in coding line table 171.Speech recognition engine 181 will be stored in the coding line in coding line table 171 A and B is registered as candidate instruction word (504).
Coding line C and D is by third application 372(that is, development and application) it is sent to coding line table 171(505), and deposited Storage is in coding line table 171 (506).Speech recognition engine 181 infuses the coding line C and D that are stored in coding line table 171 Volume is candidate instruction word (508).
Therefore, speech recognition engine 181 by by first application 172 and third using 372 send coding line A, B, C and D is registered as candidate instruction word.
For example, when coding line A, B, C and D are registered as candidate instruction word, it can detecte and be input to and display device The voice A of second voice acquisition unit 240 of 100 separation.The voice A detected converts Chinese idiom by the second speech convertor 250 Sound signal, and speech recognition engine 181(509 is sent to by the second communication unit 260 and the first communication unit 160).
The voice signal of voice A is compared by speech recognition engine 181 with candidate instruction word A, B, C and D of registration, It determines identical or similar order, and identifies voice A(510).
First controller 180 can send display unit 130(511 for recognition result), and display unit 130 can be with Speech recognition result is shown as in figure 4.
As shown in Figure 4, display unit 130 can show the UI for showing speech recognition result " A " 60 and according to voice Candidate instruction word A61, B62, C63 and D64 of recognition result.Display unit 130 can also be shown according to speech recognition result The UI of the application message (the first application) 65 of management coding line A is shown.
By the UI shown on the screen, user can check speech recognition result and candidate instruction word.If voice Recognition result does not meet his/her intention of speaking, then user can choose one in the candidate instruction word.User can be with The information of device relevant to speech recognition result is obtained by application message.
First controller 180 sends first using 172 for control command according to the speech recognition result in such as Fig. 3 (512).The voice A executive control operation that first application 172 is identified by the control of the first controller 180, basis.For example, If voice A is " volume reduction ", the volume of display device 100 is reduced.
As described in Fig. 3 and Fig. 4, display device 100 can with the coding lines of some applications registered in advance (for example, A, B, C with And D), and if detecting the voice of user, display device 100 can identify the voice, show speech recognition result, And the candidate instruction word based on registration executes corresponding control operation.
Although Fig. 3 and Fig. 4 illustrate the first application 172 for being registered as candidate instruction word and third using 372 instruction Word, and by the voice of the second voice acquisition unit 240 input user, but exemplary embodiment is not limited to this.For example, Coding line can be sent by various other native applications and development and application to register/delete candidate instruction word, and by each Kind voice acquisition unit inputs voice.
Hereinafter, the voice by the execution speech recognition referring to Fig. 5 detailed description according to another exemplary embodiment is known Other system.
Fig. 5 illustrates the example of execution speech recognition according to another exemplary embodiment.
As shown in Figure 5, it can detecte the language for being input to the third voice acquisition unit 340 isolated with display device 100 Sound E.The voice E detected is converted into voice signal by third speech convertor 350, and passes through 360 He of third communication unit First communication unit 160 is sent to speech recognition engine 181(701).
Display device 100 can have the candidate instruction word of registration.For example, coding line E and F is sent by third using 372 To coding line table 171(702), and (703) are stored in coding line table 171.Speech recognition engine 181 will be stored in Coding line E and F in coding line table 171 are registered as candidate instruction word (705).
That is, being registered to speech recognition engine 181 as candidate instruction word by the coding line E and F that third is sent using 372 In.
When coding line E and F are registered as candidate instruction word, speech recognition engine 181 is by the voice signal and note of voice E Candidate instruction the word E and F of volume are compared, and determine identical or similar coding line, and identify voice E(706).
First controller 180 can send display unit 130(707 for recognition result), and display unit 130 can be with Show speech recognition result.
First controller 180 sends third application 372(708 for control command according to speech recognition result).Third is answered The voice E executive control operation identified with 372 by the control of third controller 380, basis.If according to speech recognition knot The control command that fruit sends is the order for controlling display device 100, then the control command can be sent to first and answer 173 are applied with 172 or second.
As shown in Figure 5, if detecting the voice of user, display device 100 can will with had been enter into this The coding line (such as E and F) of the corresponding application of the device of voice is registered as candidate instruction word, the candidate instruction word based on registration It identifies voice, shows speech recognition result, and execute corresponding control operation.
Although Fig. 5, which is illustrated, inputs user speech by third voice acquisition unit 340, and third applies 372 finger Word is enabled to be registered as candidate instruction word, but exemplary embodiment is not limited to this.For example, can be obtained by various voices single Member input voice, and coding line can be sent by various native applications and development and application with registration/deletion candidate instruction word.
Hereinafter, the audio recognition method of speech recognition system accoding to exemplary embodiment is described with reference to the accompanying drawings.
Fig. 6 is the flow chart for showing the audio recognition method of speech recognition system accoding to exemplary embodiment.
As shown in Figure 6, speech recognition system, which can detecte, is input in multiple voice acquisition units 140,240 and 340 The voice of at least one (operation S810).The voice detected is converted into electric language by speech convertor 150,250 and 350 Sound signal.
First controller 180 receives the voice signal (operation S820).External voice acquisition is input to if detected The voice of unit 240 and 340 then can receive the voice signal by the first communication unit 160.
Speech recognition engine 181 registers candidate instruction word based on voice signal identification voice (operation S830).Note The candidate instruction word of volume can be the words being stored in advance in coding line table 171, either pass through native applications or exploitation It receives, and is stored in coding line table 171 using 172,173,372 and 472.
The voice (operation S840) of candidate instruction word identification user of the speech recognition engine 181 based on storage.
First controller 180 shows speech recognition result on display unit 130.Display unit 130 can show for The speech recognition result of the voice signal, candidate instruction word and application message according to speech recognition result.
First controller 180 generates control command according to speech recognition result and sends application for the control command (operation S860).Correspondingly, operation can be executed by the control command of generation.
Fig. 7 is the flow chart for showing the details for the process that voice input is detected in Fig. 6.
It as shown in Figure 7, can be can wherein input the voice of the voice of user as the display device of master device 100 Input pattern operates (operation S811).In the voice input pattern, various voice acquisition units 140,240 can be passed through With 340 input voices.
For example, the first controller 180 can from one in multiple voice acquisition units 140,240 and 340 detect about Wake up the speech (speaking) (operation S812) of keyword (wakeup keyword).The wake-up keyword makes it possible to lead to Special sound acquiring unit input voice is crossed, and can be pre-arranged.For example, the first voice of display device 100 obtains Unit 140 can set voice relevant to the TV of such as channel and volume etc. control to waking up keyword.Mobile phone 200 the second voice acquisition unit 240 can will be set as waking up keyword to the relevant voice of call, contact details etc..
Alternatively, if as manipulation predetermined button (voice input button) as a result, being obtained by multiple voices single One in member 140,240 and 340 has input trigger signal, then the first controller 180, which can detecte, obtains list by voice The voice (operation S813) of member input.For example, if user manipulates the voice input button provided in special sound acquisition device, Then detect the voice for being input to the voice-input unit of the special sound acquisition device.
According to the detection, the first controller 180 enables the voice in multiple voice acquisition units 140,240 and 340 The voice acquisition unit (operation S814) being input into.Because one of voice acquisition unit is activated, it is possible to prevent It detects unnecessary voice and prevents failure.
The voice acquisition unit that the voice signal is activated is sent to speech recognition engine 181 to execute speech recognition.
Fig. 8 is the flow chart for showing the details for the process that speech recognition is executed in Fig. 6.
As shown in Figure 8, speech recognition engine 181 can be from least one of multiple applications 172,173,372 and 472 Candidate instruction word is received, and registers the candidate instruction word (operation S830).
Speech recognition engine 181 can determine whether registered candidate instruction word is identical as the voice signal received/ Similar (operation S841).
If it is determined that then speech recognition engine 181 determines identical/phase there are identical or similar candidate instruction word As coding line and execute speech recognition, and the first controller 180 shows speech recognition result on display unit 130 (operation S850).
If it is determined that then speech recognition engine 181 may determine whether without identical or similar candidate instruction word Receive and register the candidate instruction word (operation S842) of other application.First controller 180 can according to the user's choice or It inputs to receive and register the candidate instruction word of other application, and can receive and register the time of multiple applications with preset order Select coding line.The capacity for considering the first storage unit 170 of display device 100, deletes to the property of can choose the time of earlier registration Select coding line.
That is, behaviour can be sequentially performed if identical/similar without candidate instruction word and the candidate instruction word registered Make S842 and S841 to execute speech recognition.
If deciding not to receive and registering the candidate instruction word of other application, speech recognition engine in operation S842 181 stop speech recognition, and the first controller 180 can show that speech recognition fails on display unit 130.
Because of the voice that master device detection accoding to exemplary embodiment is inputted by receiving the various devices of user speech, So various voice acquisition device can be used, and associated service can be provided by the voice acquisition device (linked services).
Candidate instruction word for speech recognition is sent by multiple applications, and is registered/is deleted.It therefore, will not need not Strategic point increases the candidate instruction word of master device, processing speed can be prevented slack-off or discrimination decline, and language can be improved The gross efficiency of sound identifying system.
User more easily identifies speech recognition result, candidate instruction word, application message and obtains dress about various voices Device of the information and offer set for the candidate instruction word of speech recognition, and improve the convenience of user.
However it is not limited to this, exemplary embodiment can be written as computer program and can using computer can It is realized in the general purpose digital computer of read record medium operation described program.The example of computer readable recording medium includes that magnetic is deposited Storage media (for example, ROM, floppy disk, hard disk etc.) and optical recording medium (for example, CD-ROM or DVD).Meanwhile exemplary reality The computer program sent by the computer-readable transmission medium of such as carrier wave can be written as by applying example, and described in the operation It receives and realizes in the general purpose digital computer of program.In addition, although not requiring in all respects, one of device Or multiple units can include the place for the computer program that operation is stored in the computer-readable medium of such as local storage Manage device or microprocessor.
While there has been shown and described that several exemplary embodiments, it will be apparent, however, to one skilled in the art, that can change Become principle and spirit of these exemplary embodiments without departing from present inventive concept, the range of present inventive concept in claim and It is limited in its equivalent.

Claims (14)

1. a kind of display device, comprising:
Display unit is configured as display image;
Storage unit is configured as storing data;
Communicator, is configured as and communication with external apparatus;And
Processor is configured as:
It is connected to the display device via the communicator based on the external device (ED), obtains and be included in the external device (ED) At least one of application the corresponding coding line of function;
The storage unit is controlled to store coding line obtained;
The coding line being stored in a storage unit is registered as into candidate instruction word in speech recognition engine,
Based on voice signal is received from voice acquisition unit, voice recognition processing is executed to know using the speech recognition engine Coding line corresponding with voice signal in the coding line that do not register,
Identify the application corresponding with voice signal of at least one application, identification coding line;And
Coding line based on identification corresponding with voice signal and external device (ED) described in the application control of identification.
2. display device according to claim 1, wherein be included in the application of at least one of external device (ED) and be stored in In the external device (ED).
3. display device according to claim 1, wherein at least one described application is configured as management described instruction Word.
4. display device according to claim 1, wherein from it is described at least one obtain using corresponding external device (ED) Coding line corresponding with the function of at least one application.
5. display device according to claim 1, wherein the processor be also configured to based on from another outside The corresponding another voice acquisition unit of device receives voice signal,
Coding line corresponding with the function of another application is obtained from another external device (ED).
6. display device according to claim 1, wherein the processor is also configured to based on voice recognition processing Unsuccessfully come based on similar with the speech recognition result of voice recognition processing one in the multiple candidate instruction words of user input selection A candidate instruction word.
7. display device according to claim 1, wherein the processor is also configured to depositing based on storage unit Reservoir state controls storage unit to delete the coding line of storage.
8. a kind of audio recognition method of display device, which comprises
It is connected to the display device based on external device (ED),
It obtains and include the corresponding coding line of function in the application of at least one of external device (ED);
Store coding line obtained;
The coding line of storage is registered as into candidate instruction word,
Based on voice signal is received from voice acquisition unit, execute in coding line of the voice recognition processing to identify registration with The corresponding coding line of voice signal,
Identify the application corresponding with voice signal of at least one application, identification coding line;And
Coding line based on identification corresponding with voice signal and external device (ED) described in the application control of identification.
9. according to the method described in claim 8, wherein, it is described using being stored in be included at least one of external device (ED) In external device (ED).
10. according to the method described in claim 8, wherein, at least one described application is configured as management described instruction word.
11. according to the method described in claim 8, wherein, from it is described at least one using corresponding external device (ED) obtain with The corresponding coding line of function of at least one application.
12. according to the method described in claim 8, further include: it is based on obtaining from another voice corresponding with another external device (ED) Unit receives voice signal, obtains coding line corresponding with the function of another application from another external device (ED).
13. according to the method described in claim 8, further include: user input selection is unsuccessfully based on based on voice recognition processing A candidate instruction word similar with the speech recognition result of voice recognition processing in multiple candidate instruction words.
14. according to the method described in claim 8, further include: the memory state of the storage device based on store instruction word is deleted Except the coding line of storage.
CN201310553280.4A 2012-11-09 2013-11-08 Display device, voice acquisition device and its audio recognition method Active CN103811007B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510276510.6A CN104883587A (en) 2012-11-09 2013-11-08 Display Apparatus, Voice Acquiring Apparatus And Voice Recognition Method Thereof

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR10-2012-0126650 2012-11-09
KR1020120126650A KR20140060040A (en) 2012-11-09 2012-11-09 Display apparatus, voice acquiring apparatus and voice recognition method thereof

Related Child Applications (1)

Application Number Title Priority Date Filing Date
CN201510276510.6A Division CN104883587A (en) 2012-11-09 2013-11-08 Display Apparatus, Voice Acquiring Apparatus And Voice Recognition Method Thereof

Publications (2)

Publication Number Publication Date
CN103811007A CN103811007A (en) 2014-05-21
CN103811007B true CN103811007B (en) 2019-12-03

Family

ID=49554021

Family Applications (2)

Application Number Title Priority Date Filing Date
CN201310553280.4A Active CN103811007B (en) 2012-11-09 2013-11-08 Display device, voice acquisition device and its audio recognition method
CN201510276510.6A Pending CN104883587A (en) 2012-11-09 2013-11-08 Display Apparatus, Voice Acquiring Apparatus And Voice Recognition Method Thereof

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN201510276510.6A Pending CN104883587A (en) 2012-11-09 2013-11-08 Display Apparatus, Voice Acquiring Apparatus And Voice Recognition Method Thereof

Country Status (7)

Country Link
US (5) US10043537B2 (en)
EP (4) EP3352471B1 (en)
JP (2) JP5868927B2 (en)
KR (1) KR20140060040A (en)
CN (2) CN103811007B (en)
RU (1) RU2677396C2 (en)
WO (1) WO2014073823A1 (en)

Families Citing this family (94)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20140060040A (en) 2012-11-09 2014-05-19 삼성전자주식회사 Display apparatus, voice acquiring apparatus and voice recognition method thereof
US20180317019A1 (en) 2013-05-23 2018-11-01 Knowles Electronics, Llc Acoustic activity detecting microphone
CN105023575B (en) * 2014-04-30 2019-09-17 中兴通讯股份有限公司 Audio recognition method, device and system
KR102147329B1 (en) 2014-06-17 2020-08-24 엘지전자 주식회사 Video display device and operating method thereof
KR102147346B1 (en) * 2014-06-23 2020-08-24 엘지전자 주식회사 Display device and operating method thereof
DE112016000287T5 (en) * 2015-01-07 2017-10-05 Knowles Electronics, Llc Use of digital microphones for low power keyword detection and noise reduction
EP3067884B1 (en) * 2015-03-13 2019-05-08 Samsung Electronics Co., Ltd. Speech recognition system and speech recognition method thereof
AU2015390534B2 (en) 2015-04-10 2019-08-22 Honor Device Co., Ltd. Speech recognition method, speech wakeup apparatus, speech recognition apparatus, and terminal
CN105185379B (en) * 2015-06-17 2017-08-18 百度在线网络技术(北京)有限公司 voiceprint authentication method and device
CN105206275A (en) * 2015-08-31 2015-12-30 小米科技有限责任公司 Device control method, apparatus and terminal
CN105222275B (en) * 2015-09-25 2018-04-13 珠海格力电器股份有限公司 Switching method, the apparatus and system of display data
CN106683668A (en) * 2015-11-05 2017-05-17 芋头科技(杭州)有限公司 Method of awakening control of intelligent device and system
KR20170082349A (en) * 2016-01-06 2017-07-14 삼성전자주식회사 Display apparatus and control methods thereof
US10095470B2 (en) 2016-02-22 2018-10-09 Sonos, Inc. Audio response playback
US9947316B2 (en) 2016-02-22 2018-04-17 Sonos, Inc. Voice control of a media playback system
US10509626B2 (en) 2016-02-22 2019-12-17 Sonos, Inc Handling of loss of pairing between networked devices
US9772817B2 (en) 2016-02-22 2017-09-26 Sonos, Inc. Room-corrected voice detection
US10264030B2 (en) 2016-02-22 2019-04-16 Sonos, Inc. Networked microphone device control
US9965247B2 (en) 2016-02-22 2018-05-08 Sonos, Inc. Voice controlled media playback system based on user profile
CN105791934A (en) * 2016-03-25 2016-07-20 福建新大陆通信科技股份有限公司 Realization method and system of intelligent STB (Set Top Box) microphone
US9978390B2 (en) 2016-06-09 2018-05-22 Sonos, Inc. Dynamic player selection for audio signal processing
US10134399B2 (en) 2016-07-15 2018-11-20 Sonos, Inc. Contextualization of voice inputs
US10152969B2 (en) 2016-07-15 2018-12-11 Sonos, Inc. Voice detection by multiple devices
US10115400B2 (en) 2016-08-05 2018-10-30 Sonos, Inc. Multiple voice services
US9942678B1 (en) 2016-09-27 2018-04-10 Sonos, Inc. Audio playback settings for voice interaction
US9743204B1 (en) 2016-09-30 2017-08-22 Sonos, Inc. Multi-orientation playback device microphones
KR102562287B1 (en) 2016-10-14 2023-08-02 삼성전자주식회사 Electronic device and audio signal processing method thereof
US10181323B2 (en) 2016-10-19 2019-01-15 Sonos, Inc. Arbitration-based voice recognition
KR102598082B1 (en) * 2016-10-28 2023-11-03 삼성전자주식회사 Image display apparatus, mobile device and operating method for the same
US10593328B1 (en) * 2016-12-27 2020-03-17 Amazon Technologies, Inc. Voice control of remote device
US11183181B2 (en) 2017-03-27 2021-11-23 Sonos, Inc. Systems and methods of multiple voice services
EP4036709B1 (en) * 2017-07-14 2024-04-24 Daikin Industries, Ltd. Environment control system
US10475449B2 (en) 2017-08-07 2019-11-12 Sonos, Inc. Wake-word detection suppression
US10048930B1 (en) 2017-09-08 2018-08-14 Sonos, Inc. Dynamic computation of system response volume
US10446165B2 (en) 2017-09-27 2019-10-15 Sonos, Inc. Robust short-time fourier transform acoustic echo cancellation during audio playback
US10621981B2 (en) 2017-09-28 2020-04-14 Sonos, Inc. Tone interference cancellation
US10482868B2 (en) 2017-09-28 2019-11-19 Sonos, Inc. Multi-channel acoustic echo cancellation
CN109584864B (en) * 2017-09-29 2023-11-24 上海寒武纪信息科技有限公司 Image processing apparatus and method
CN109584862B (en) * 2017-09-29 2024-01-12 上海寒武纪信息科技有限公司 Image processing apparatus and method
US10466962B2 (en) 2017-09-29 2019-11-05 Sonos, Inc. Media playback system with voice assistance
US11445235B2 (en) 2017-10-24 2022-09-13 Comcast Cable Communications, Llc Determining context to initiate interactivity
US20190130898A1 (en) * 2017-11-02 2019-05-02 GM Global Technology Operations LLC Wake-up-word detection
US10880650B2 (en) 2017-12-10 2020-12-29 Sonos, Inc. Network microphone devices with automatic do not disturb actuation capabilities
US10818290B2 (en) 2017-12-11 2020-10-27 Sonos, Inc. Home graph
US11343614B2 (en) 2018-01-31 2022-05-24 Sonos, Inc. Device designation of playback and network microphone device arrangements
JP7133969B2 (en) * 2018-04-27 2022-09-09 シャープ株式会社 Voice input device and remote dialogue system
US11175880B2 (en) 2018-05-10 2021-11-16 Sonos, Inc. Systems and methods for voice-assisted media content selection
US10847178B2 (en) 2018-05-18 2020-11-24 Sonos, Inc. Linear filtering for noise-suppressed speech detection
US10959029B2 (en) 2018-05-25 2021-03-23 Sonos, Inc. Determining and adapting to changes in microphone performance of playback devices
US10681460B2 (en) 2018-06-28 2020-06-09 Sonos, Inc. Systems and methods for associating playback devices with voice assistant services
US11076035B2 (en) 2018-08-28 2021-07-27 Sonos, Inc. Do not disturb feature for audio notifications
US10461710B1 (en) 2018-08-28 2019-10-29 Sonos, Inc. Media playback system with maximum volume setting
US10878811B2 (en) 2018-09-14 2020-12-29 Sonos, Inc. Networked devices, systems, and methods for intelligently deactivating wake-word engines
US10587430B1 (en) 2018-09-14 2020-03-10 Sonos, Inc. Networked devices, systems, and methods for associating playback devices based on sound codes
US11024331B2 (en) 2018-09-21 2021-06-01 Sonos, Inc. Voice detection optimization using sound metadata
US10811015B2 (en) 2018-09-25 2020-10-20 Sonos, Inc. Voice detection optimization based on selected voice assistant service
US11100923B2 (en) 2018-09-28 2021-08-24 Sonos, Inc. Systems and methods for selective wake word detection using neural network models
US10692518B2 (en) 2018-09-29 2020-06-23 Sonos, Inc. Linear filtering for noise-suppressed speech detection via multiple network microphone devices
EP3869301A4 (en) * 2018-10-15 2021-12-15 Sony Group Corporation Information processing device, information processing method, and computer program
US11899519B2 (en) 2018-10-23 2024-02-13 Sonos, Inc. Multiple stage network microphone device with reduced power consumption and processing load
EP3654249A1 (en) 2018-11-15 2020-05-20 Snips Dilated convolutions and gating for efficient keyword spotting
US11183183B2 (en) 2018-12-07 2021-11-23 Sonos, Inc. Systems and methods of operating media playback systems having multiple voice assistant services
US11132989B2 (en) 2018-12-13 2021-09-28 Sonos, Inc. Networked microphone devices, systems, and methods of localized arbitration
US10602268B1 (en) 2018-12-20 2020-03-24 Sonos, Inc. Optimization of network microphone devices using noise classification
US11315556B2 (en) 2019-02-08 2022-04-26 Sonos, Inc. Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification
US10867604B2 (en) 2019-02-08 2020-12-15 Sonos, Inc. Devices, systems, and methods for distributed voice processing
JP2020129252A (en) * 2019-02-08 2020-08-27 三菱電機株式会社 Device control system and terminal apparatus
JP7449070B2 (en) 2019-03-27 2024-03-13 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Voice input device, voice input method and its program
CN110046045B (en) * 2019-04-03 2021-07-30 百度在线网络技术(北京)有限公司 Voice wake-up data packet processing method and device
US11120794B2 (en) 2019-05-03 2021-09-14 Sonos, Inc. Voice assistant persistence across multiple network microphone devices
US11200894B2 (en) 2019-06-12 2021-12-14 Sonos, Inc. Network microphone device with command keyword eventing
US10586540B1 (en) 2019-06-12 2020-03-10 Sonos, Inc. Network microphone device with command keyword conditioning
US11361756B2 (en) 2019-06-12 2022-06-14 Sonos, Inc. Conditional wake word eventing based on environment
US10871943B1 (en) 2019-07-31 2020-12-22 Sonos, Inc. Noise classification for event detection
US11138975B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
US11138969B2 (en) 2019-07-31 2021-10-05 Sonos, Inc. Locally distributed keyword detection
US11317162B2 (en) 2019-09-26 2022-04-26 Dish Network L.L.C. Method and system for navigating at a client device selected features on a non-dynamic image page from an elastic voice cloud server in communication with a third-party search service
CN112581946A (en) * 2019-09-29 2021-03-30 百度在线网络技术(北京)有限公司 Voice control method and device, electronic equipment and readable storage medium
US11189286B2 (en) 2019-10-22 2021-11-30 Sonos, Inc. VAS toggle based on device orientation
JP2021071797A (en) * 2019-10-29 2021-05-06 富士通クライアントコンピューティング株式会社 Display device and information processing device
EP4080348A4 (en) * 2019-12-19 2023-09-13 LG Electronics Inc. Display device
US11200900B2 (en) 2019-12-20 2021-12-14 Sonos, Inc. Offline voice control
US11562740B2 (en) 2020-01-07 2023-01-24 Sonos, Inc. Voice verification for media playback
US11556307B2 (en) 2020-01-31 2023-01-17 Sonos, Inc. Local voice data processing
US11308958B2 (en) 2020-02-07 2022-04-19 Sonos, Inc. Localized wakeword verification
CN113360125A (en) * 2020-03-05 2021-09-07 西安诺瓦星云科技股份有限公司 Image display method, device and system
US11727919B2 (en) 2020-05-20 2023-08-15 Sonos, Inc. Memory allocation for keyword spotting engines
US11308962B2 (en) 2020-05-20 2022-04-19 Sonos, Inc. Input detection windowing
US11482224B2 (en) 2020-05-20 2022-10-25 Sonos, Inc. Command keywords with input detection windowing
CN111934957B (en) * 2020-07-16 2021-06-25 宁波方太厨具有限公司 Application system and method supporting WiFi and offline voice
US11698771B2 (en) 2020-08-25 2023-07-11 Sonos, Inc. Vocal guidance engines for playback devices
US11984123B2 (en) 2020-11-12 2024-05-14 Sonos, Inc. Network device interaction by range
US11551700B2 (en) 2021-01-25 2023-01-10 Sonos, Inc. Systems and methods for power-efficient keyword detection
WO2022265623A1 (en) * 2021-06-15 2022-12-22 Hewlett-Packard Development Company, L.P. Acknowledgement based audio communications

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1342308A (en) * 2000-01-05 2002-03-27 松下电器产业株式会社 Device setter, device setting system, and recorded medium where device setting program is recoded
CN202033897U (en) * 2011-01-25 2011-11-09 张国鸿 Voice remote control receiving module and electric appliance with voice remote control function

Family Cites Families (79)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5752232A (en) * 1994-11-14 1998-05-12 Lucent Technologies Inc. Voice activated device and method for providing access to remotely retrieved data
JP3697748B2 (en) 1995-08-21 2005-09-21 セイコーエプソン株式会社 Terminal, voice recognition device
US6665639B2 (en) 1996-12-06 2003-12-16 Sensory, Inc. Speech recognition in consumer electronic products
US6188985B1 (en) * 1997-01-06 2001-02-13 Texas Instruments Incorporated Wireless voice-activated device for control of a processor-based host system
KR100237385B1 (en) 1997-08-05 2000-01-15 정선종 The Implementation Method of Speech Recognizer on the Web Browser
DE69712485T2 (en) * 1997-10-23 2002-12-12 Sony Int Europe Gmbh Voice interface for a home network
CN1223508A (en) 1998-01-13 1999-07-21 黄金富 Sound-controlled infrared ray remote controller
JP2000099306A (en) * 1998-05-28 2000-04-07 Canon Inc Device and method for control and storage medium storing program for executing operation processing therefor
JP2000200395A (en) 1999-01-07 2000-07-18 Matsushita Electric Ind Co Ltd Fire prevention system
US6556970B1 (en) * 1999-01-28 2003-04-29 Denso Corporation Apparatus for determining appropriate series of words carrying information to be recognized
DE69942663D1 (en) * 1999-04-13 2010-09-23 Sony Deutschland Gmbh Merging of speech interfaces for the simultaneous use of devices and applications
JP4314680B2 (en) 1999-07-27 2009-08-19 ソニー株式会社 Speech recognition control system and speech recognition control method
US6397186B1 (en) * 1999-12-22 2002-05-28 Ambush Interactive, Inc. Hands-free, voice-operated remote control transmitter
US8271287B1 (en) 2000-01-14 2012-09-18 Alcatel Lucent Voice command remote control system
JP2001296881A (en) 2000-04-14 2001-10-26 Sony Corp Device and method for information processing and recording medium
JP2002116787A (en) 2000-07-10 2002-04-19 Matsushita Electric Ind Co Ltd Priority deciding device, priority deciding method and priority deciding program
ATE342567T1 (en) * 2000-07-28 2006-11-15 Koninkl Philips Electronics Nv SYSTEM FOR CONTROLLING A DEVICE USING VOICE COMMANDS
DE10040466C2 (en) * 2000-08-18 2003-04-10 Bosch Gmbh Robert Method for controlling voice input and output
WO2002017625A1 (en) 2000-08-21 2002-02-28 Koninklijke Philips Electronics N.V. A voice controlled remote control with downloadable set of voice commands
EP1184841A1 (en) * 2000-08-31 2002-03-06 Siemens Aktiengesellschaft Speech controlled apparatus and method for speech input and speech recognition
US20020110246A1 (en) 2001-02-14 2002-08-15 Jason Gosior Wireless audio system
US7369997B2 (en) * 2001-08-01 2008-05-06 Microsoft Corporation Controlling speech recognition functionality in a computing device
US20030061033A1 (en) * 2001-09-26 2003-03-27 Dishert Lee R. Remote control system for translating an utterance to a control parameter for use by an electronic device
CN1437377A (en) * 2002-02-04 2003-08-20 义隆电子股份有限公司 Multifunctional Information service system and its electronic equipment
US20030177012A1 (en) * 2002-03-13 2003-09-18 Brett Drennan Voice activated thermostat
KR100434545B1 (en) * 2002-03-15 2004-06-05 삼성전자주식회사 Method and apparatus for controlling devices connected with home network
JP3715584B2 (en) * 2002-03-28 2005-11-09 富士通株式会社 Device control apparatus and device control method
JP2003295893A (en) * 2002-04-01 2003-10-15 Omron Corp System, device, method, and program for speech recognition, and computer-readable recording medium where the speech recognizing program is recorded
US20040001040A1 (en) * 2002-06-28 2004-01-01 Kardach James P. Methods and apparatus for providing light to a display
US6834265B2 (en) 2002-12-13 2004-12-21 Motorola, Inc. Method and apparatus for selective speech recognition
JP2004275360A (en) * 2003-03-14 2004-10-07 Olympus Corp Endoscope system
CN2620913Y (en) * 2003-04-21 2004-06-16 叶龙 Voice controlled remote controller for domestic electric equipment
JP2005072764A (en) 2003-08-21 2005-03-17 Hitachi Ltd Equipment control system and device thereof, and equipment control method
JP2006033795A (en) 2004-06-15 2006-02-02 Sanyo Electric Co Ltd Remote control system, controller, program for imparting function of controller to computer, storage medium with the program stored thereon, and server
CN1713271A (en) * 2004-06-15 2005-12-28 三洋电机株式会社 Remote control system, controller, program product
CN1725902A (en) * 2004-07-20 2006-01-25 李廷玉 Mobile phone with remote-control function
US7085673B2 (en) 2004-08-31 2006-08-01 Hewlett-Packard Development Company, L.P. Displacement estimation system and method
US20060074658A1 (en) * 2004-10-01 2006-04-06 Siemens Information And Communication Mobile, Llc Systems and methods for hands-free voice-activated devices
KR100682897B1 (en) 2004-11-09 2007-02-15 삼성전자주식회사 Method and apparatus for updating dictionary
US8942985B2 (en) * 2004-11-16 2015-01-27 Microsoft Corporation Centralized method and system for clarifying voice commands
US20070299670A1 (en) * 2006-06-27 2007-12-27 Sbc Knowledge Ventures, Lp Biometric and speech recognition system and method
US20080037727A1 (en) * 2006-07-13 2008-02-14 Clas Sivertsen Audio appliance with speech recognition, voice command control, and speech generation
US7899673B2 (en) * 2006-08-09 2011-03-01 Microsoft Corporation Automatic pruning of grammars in a multi-application speech recognition interface
KR20080036697A (en) * 2006-10-24 2008-04-29 삼성전자주식회사 Method and apparatus for remote control in portable terminal
CN100538762C (en) 2006-12-15 2009-09-09 广东协联科贸发展有限公司 A kind of keying speech integrated remote controller
US8260618B2 (en) * 2006-12-21 2012-09-04 Nuance Communications, Inc. Method and apparatus for remote control of devices through a wireless headset using voice activation
US8140325B2 (en) * 2007-01-04 2012-03-20 International Business Machines Corporation Systems and methods for intelligent control of microphones for speech recognition applications
KR20080096239A (en) * 2007-04-27 2008-10-30 정장오 Speech recognition kitchen tv system for speech schedule control kitchen tv, home network system, household appliances
US20090018830A1 (en) * 2007-07-11 2009-01-15 Vandinburg Gmbh Speech control of computing devices
US8825468B2 (en) * 2007-07-31 2014-09-02 Kopin Corporation Mobile wireless display providing speech to speech translation and avatar simulating human attributes
US8099289B2 (en) * 2008-02-13 2012-01-17 Sensory, Inc. Voice interface and search for electronic devices including bluetooth headsets and remote systems
TWI385932B (en) * 2008-03-26 2013-02-11 Asustek Comp Inc Device and system for remote controlling
US9135809B2 (en) * 2008-06-20 2015-09-15 At&T Intellectual Property I, Lp Voice enabled remote control for a set-top box
JP5200712B2 (en) * 2008-07-10 2013-06-05 富士通株式会社 Speech recognition apparatus, speech recognition method, and computer program
US8676904B2 (en) * 2008-10-02 2014-03-18 Apple Inc. Electronic devices with voice command and contextual data processing capabilities
JP5465926B2 (en) * 2009-05-22 2014-04-09 アルパイン株式会社 Speech recognition dictionary creation device and speech recognition dictionary creation method
US20100332236A1 (en) * 2009-06-25 2010-12-30 Blueant Wireless Pty Limited Voice-triggered operation of electronic devices
WO2011027964A1 (en) 2009-09-01 2011-03-10 Lg Electronics Inc. Method for controlling external device and remote controller thereof
KR20110028103A (en) 2009-09-11 2011-03-17 주식회사 카서 Multipurpose local wireless communication apparatus and system
US20110067059A1 (en) * 2009-09-15 2011-03-17 At&T Intellectual Property I, L.P. Media control
US8510103B2 (en) * 2009-10-15 2013-08-13 Paul Angott System and method for voice recognition
KR20110052863A (en) * 2009-11-13 2011-05-19 삼성전자주식회사 Mobile device and method for generating control signal thereof
US9865263B2 (en) 2009-12-01 2018-01-09 Nuance Communications, Inc. Real-time voice recognition on a handheld device
CN101778198A (en) * 2010-01-25 2010-07-14 上海享云信息系统有限公司 Enhanced-type TV terminal system
US20110246902A1 (en) 2010-04-01 2011-10-06 Tsung-Han Tsai Method for portable phones to control computers
CN102255780A (en) 2010-05-20 2011-11-23 株式会社曙飞电子 Home network system and control method
US20120078635A1 (en) * 2010-09-24 2012-03-29 Apple Inc. Voice control system
US8914287B2 (en) * 2010-12-31 2014-12-16 Echostar Technologies L.L.C. Remote control audio link
EP2518722A3 (en) * 2011-04-28 2013-08-28 Samsung Electronics Co., Ltd. Method for providing link list and display apparatus applying the same
US20130027613A1 (en) 2011-05-03 2013-01-31 Lg Electronics Inc. Image display apparatus, portable terminal, and methods for operating the same
CN102196207B (en) 2011-05-12 2014-06-18 深圳市车音网科技有限公司 Method, device and system for controlling television by using voice
US20130073293A1 (en) * 2011-09-20 2013-03-21 Lg Electronics Inc. Electronic device and method for controlling the same
US20130144618A1 (en) 2011-12-02 2013-06-06 Liang-Che Sun Methods and electronic devices for speech recognition
US20130238326A1 (en) * 2012-03-08 2013-09-12 Lg Electronics Inc. Apparatus and method for multiple device voice control
KR101946364B1 (en) * 2012-05-01 2019-02-11 엘지전자 주식회사 Mobile device for having at least one microphone sensor and method for controlling the same
RU121608U1 (en) * 2012-06-05 2012-10-27 Закрытое акционерное общество "Титан-информационный сервис" ELECTRONIC CONTROL DEVICE
US20140074472A1 (en) * 2012-09-12 2014-03-13 Chih-Hung Lin Voice control system with portable voice control device
US9646610B2 (en) * 2012-10-30 2017-05-09 Motorola Solutions, Inc. Method and apparatus for activating a particular wireless communication device to accept speech and/or voice commands using identification data consisting of speech, voice, image recognition
KR20140060040A (en) 2012-11-09 2014-05-19 삼성전자주식회사 Display apparatus, voice acquiring apparatus and voice recognition method thereof

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1342308A (en) * 2000-01-05 2002-03-27 松下电器产业株式会社 Device setter, device setting system, and recorded medium where device setting program is recoded
CN202033897U (en) * 2011-01-25 2011-11-09 张国鸿 Voice remote control receiving module and electric appliance with voice remote control function

Also Published As

Publication number Publication date
EP3790285A1 (en) 2021-03-10
US20200184989A1 (en) 2020-06-11
US20220358949A1 (en) 2022-11-10
CN103811007A (en) 2014-05-21
RU2015121906A (en) 2016-12-27
JP5868927B2 (en) 2016-02-24
US20230121055A1 (en) 2023-04-20
JP6640502B2 (en) 2020-02-05
EP4106339A1 (en) 2022-12-21
US10043537B2 (en) 2018-08-07
JP2014096153A (en) 2014-05-22
CN104883587A (en) 2015-09-02
WO2014073823A1 (en) 2014-05-15
RU2677396C2 (en) 2019-01-16
US20170337937A1 (en) 2017-11-23
EP2731349A1 (en) 2014-05-14
EP2731349B1 (en) 2018-04-18
JP2016027484A (en) 2016-02-18
US11727951B2 (en) 2023-08-15
KR20140060040A (en) 2014-05-19
US20140136205A1 (en) 2014-05-15
EP3352471B1 (en) 2020-12-30
US10586554B2 (en) 2020-03-10
EP3352471A1 (en) 2018-07-25

Similar Documents

Publication Publication Date Title
CN103811007B (en) Display device, voice acquisition device and its audio recognition method
US20170223301A1 (en) Image processing apparatus, voice acquiring apparatus, voice recognition method thereof and voice recognition system
CN113784220B (en) Method for playing media resources, display device and mobile device
US20110068899A1 (en) Method and System for Controlling Electronic Devices
CN112055240B (en) Display device and operation prompt display method for pairing display device with remote controller
US10439838B2 (en) Control device, method of controlling the same, and integrated control system
CN113784200B (en) Communication terminal, display device and screen projection connection method
CN112543359B (en) Display device and method for automatically configuring video parameters
KR102501655B1 (en) Display apparatus, voice acquiring apparatus and voice recognition method thereof
CN111949782A (en) Information recommendation method and service equipment
CN112473121B (en) Display device and avoidance ball display method based on limb identification
CN113784186B (en) Terminal device, server, and communication control method
KR102262050B1 (en) Display apparatus, voice acquiring apparatus and voice recognition method thereof
CN114007128A (en) Display device and network distribution method
CN112040299A (en) Display device, server and live broadcast display method
CN112259096B (en) Voice data processing method and device
CN112492402B (en) Display device
CN111914114A (en) Badcase mining method and electronic equipment

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant