CN111897916A - Voice instruction recognition method and device, terminal equipment and storage medium - Google Patents
Voice instruction recognition method and device, terminal equipment and storage medium Download PDFInfo
- Publication number
- CN111897916A CN111897916A CN202010722276.6A CN202010722276A CN111897916A CN 111897916 A CN111897916 A CN 111897916A CN 202010722276 A CN202010722276 A CN 202010722276A CN 111897916 A CN111897916 A CN 111897916A
- Authority
- CN
- China
- Prior art keywords
- voice
- text
- input
- information slot
- voice text
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 30
- 230000002159 abnormal effect Effects 0.000 abstract description 22
- 230000000875 corresponding effect Effects 0.000 description 19
- 238000012905 input function Methods 0.000 description 17
- 230000008921 facial expression Effects 0.000 description 16
- 238000005516 engineering process Methods 0.000 description 15
- 230000006870 function Effects 0.000 description 12
- 238000012545 processing Methods 0.000 description 8
- 230000006854 communication Effects 0.000 description 7
- 238000004891 communication Methods 0.000 description 6
- 238000002372 labelling Methods 0.000 description 4
- 238000003058 natural language processing Methods 0.000 description 4
- 230000001276 controlling effect Effects 0.000 description 3
- 238000001514 detection method Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000001133 acceleration Effects 0.000 description 2
- 230000005540 biological transmission Effects 0.000 description 2
- 230000005484 gravity Effects 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000007599 discharging Methods 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 230000003340 mental effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002093 peripheral effect Effects 0.000 description 1
- 238000010079 rubber tapping Methods 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000010897 surface acoustic wave method Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/103—Formatting, i.e. changing of presentation of documents
- G06F40/117—Tagging; Marking up; Designating a block; Setting of attributes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3343—Query execution using phonetics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3344—Query execution using natural language analysis
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Abstract
The embodiment of the invention discloses a voice instruction identification method, a voice instruction identification device, terminal equipment and a storage medium. The voice instruction identification method provided by the embodiment of the invention comprises the steps of identifying a voice instruction input by a user and generating a voice text; acquiring an information slot value of the voice text; and judging whether to execute the voice command according to the information slot value, so that the user can conveniently and quickly stop voice input when the terminal equipment receives an abnormal voice command.
Description
Technical Field
The invention relates to the technical field of mobile communication, in particular to a voice instruction identification method, a voice instruction identification device, terminal equipment and a storage medium.
Background
ASR (Automatic Speech Recognition) is a technology for converting human Speech into text, and is applied to various terminal devices, such as smart phones, notebook computers, tablet computers, vehicle-mounted terminals, etc., with the development and progress of scientific technology, the convenience requirement of users for life is gradually increased, more and more terminal devices have the functions of collecting user Speech and performing Speech Recognition, the terminal devices collect Speech instructions input by users by using microphones and convert the Speech instructions input by users into Speech texts by using an Automatic Speech Recognition technology, so that the terminal devices can make corresponding system actions according to the Speech texts, thereby preventing users from controlling terminal systems to make corresponding actions by using text input or other operations, so that users can control terminals by Speech to achieve the same purpose, in the research and practice process of the prior art, the inventor of the invention finds that the user can have noisy background sound when inputting voice, the terminal can receive a plurality of disordered voice instructions, and the user can not conveniently and quickly close the voice input function of the terminal equipment.
Disclosure of Invention
The embodiment of the invention provides a voice instruction identification method, a voice instruction identification device, a terminal device and a storage medium, which are used for judging whether a voice text is abnormal or not according to the field of the voice text, expression intention and an information slot value, so that when the terminal device receives an abnormal voice instruction, a user can conveniently and quickly stop voice input.
The embodiment of the invention provides a voice instruction identification method, which comprises the following steps:
recognizing a voice instruction input by a user and generating a voice text;
acquiring an information slot value of the voice text;
and judging whether to execute the voice command according to the information slot value.
Optionally, in some embodiments of the present invention, the obtaining the slot value of the speech text includes:
determining the belonged field and the expression intention according to the content of the voice text;
acquiring an information slot of the voice text according to the belonged field and the expression intention;
and filling the information slot to generate an information slot value.
Optionally, in some embodiments of the present invention, the determining whether to execute the voice instruction according to the slot value includes:
and judging whether to execute the voice command according to the field of the voice text.
Optionally, in some embodiments of the present invention, the determining whether to execute the voice instruction according to the domain of the voice text includes:
determining that the voice text comprises at least two different fields, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises a field, and judging whether to execute the voice command according to the expression intention of the voice text.
Optionally, in some embodiments of the present invention, the determining whether the speech text is normal according to the expression intention includes:
determining that the voice text comprises at least two different expression intentions, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises the expression intention, and judging whether to execute the voice instruction according to the information slot value.
Optionally, in some embodiments of the present invention, the determining whether to execute the voice instruction according to the slot value includes:
determining that the voice text comprises at least two different information slot values, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises the information slot value, and executing the voice instruction.
Optionally, in some embodiments of the present invention, the stopping the voice input includes:
a stop button to increase the voice input.
Correspondingly, an embodiment of the present invention further provides a voice instruction recognition apparatus, including:
the recognition unit is used for recognizing a voice instruction input by a user and generating a voice text;
the acquisition unit is used for acquiring the information slot value of the voice text;
and the judging unit is used for judging whether to execute the voice command according to the information slot value.
Similarly, an embodiment of the present invention further provides a terminal device, including:
a memory for storing an application program;
a processor for implementing the steps of any of the speech instruction recognition methods when executing the application program.
In addition, an embodiment of the present invention further provides a storage medium, where an application program is stored on the storage medium, and the application program, when executed by a processor, implements the steps of any one of the voice instruction recognition methods.
The embodiment of the invention provides a voice instruction recognition method, wherein a user inputs a voice instruction, a terminal device utilizes a microphone to collect the voice instruction and utilizes an automatic voice recognition technology to recognize the voice instruction input by the user to generate a voice text, the terminal device utilizes an NLU technology to analyze the voice text, determines the belonged field, expression intention and information slot of the voice text, fills the information slot and generates an information slot value, the terminal firstly judges the belonged field included by the voice text, if the voice text includes at least two different belonged fields, the voice text is determined to be abnormal, the voice instruction is not executed, the terminal enables the user to stop voice input, if the voice text includes one belonged field, the voice text is determined to be normal, then the terminal judges the expression intention included by the voice text, if the voice text includes at least two different expression intentions, determining that the voice text is abnormal, not executing a voice instruction, stopping voice input by the terminal, if the voice text comprises the expression intention, determining that the voice text is normal, then the terminal judges the information slot value included by the voice text, the voice text includes at least two different information slot values, the voice text is determined to be abnormal, the voice instruction is not executed, the terminal stops the voice input, if the voice text includes one information slot value, the voice text is determined to be normal, the terminal executes the voice text, and when the voice text is abnormal, the terminal increases a stop button and/or a pause button of the voice input, or the position of the stop button and/or the pause button of the voice input is changed, and the stop button and/or the pause button of the voice input is moved from the edge position of the terminal display screen to the middle position, so that the user can conveniently and quickly stop the voice input.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
FIG. 1 is a schematic diagram of a speech command recognition scenario provided by an embodiment of the present invention;
FIG. 2 is a flow chart of a method for recognizing a voice command according to an embodiment of the present invention;
FIG. 3 is another flow chart of a method for recognizing a voice command according to an embodiment of the present invention;
FIG. 4 is a schematic structural diagram of a voice command recognition apparatus according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a terminal according to an embodiment of the present invention.
Detailed Description
The technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings in the embodiments of the present invention. It is to be understood that the described embodiments are merely exemplary of the invention, and not restrictive of the full scope of the invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
The embodiment of the invention provides a voice instruction identification method, a voice instruction identification device, terminal equipment and a storage medium. The device can be integrated in a terminal, and the terminal can be a mobile phone, a tablet computer, a notebook computer, a vehicle-mounted terminal and other equipment.
For example, as shown in fig. 1, a voice input function of the terminal device is turned on, the terminal device turns on a microphone, a user inputs a voice instruction to the terminal device, the terminal device collects the voice instruction input by the user by using the microphone, then transmits the voice instruction to a recognition unit of the terminal system, recognizes the voice instruction by using an automatic voice recognition technology, converts the voice instruction into a voice text, then transmits the voice text to an acquisition unit of the terminal system, analyzes the voice text by using an NLU technology, classifies the voice text so as to recognize an expression intention and a belonging field of the voice text, determines an information slot of the voice text, performs sequence labeling on the voice text, fills the information slot of the voice text, generates an information slot value, and finally the terminal device transmits the voice text to a judgment unit of the terminal system, the terminal judges the belonging field of the voice text firstly, if the voice text comprises at least two different belonging fields, the voice text is determined to be abnormal, a voice instruction is not executed, the terminal enables a user to stop voice input, if the voice text comprises one belonging field, the voice text is determined to be normal, then the terminal judges the expression intention of the voice text, if the voice text comprises at least two different expression intents, the voice text is determined to be abnormal, the voice instruction is not executed, the terminal stops the voice input, if the voice text comprises one expression intention, the voice text is determined to be normal, then the terminal judges the information slot value of the voice text, the voice text comprises at least two different information slot values, the voice text is determined to be abnormal, the voice instruction is not executed, and the terminal stops the voice input, if the voice text comprises the information slot value, the voice text is determined to be normal, the terminal executes the voice text, when the voice text is abnormal, the terminal increases a stop button and/or a pause button of voice input, or changes the position of the stop button and/or the pause button of the voice input, and moves the stop button and/or the pause button of the voice input to the middle position from the edge position of a display screen of the terminal, so that a user can conveniently click and stop the voice input.
The following are detailed below. It should be noted that the following description of the embodiments is not intended to limit the preferred order of the embodiments.
The embodiment will be described from the perspective of a voice instruction recognition device, which may be specifically integrated in a terminal device, where the terminal device may include a notebook computer, a tablet computer, a smart phone, a vehicle-mounted terminal, and the like.
A voice instruction recognition method, comprising: recognizing a voice instruction input by a user and generating a voice text; acquiring an information slot value of the voice text; and judging whether to execute the voice command according to the information slot value.
As shown in fig. 2, the specific flow of the voice command recognition method is as follows:
For example, referring to fig. 3 together, in the terminal device with a voice input function, the user turns on the voice input function of the terminal device, step 301 is executed, the terminal device turns on a recording device, such as a microphone, etc., the user inputs a voice command to the terminal device, step 302 is executed, the terminal device collects the voice command input by the user by using the recording device, step 303 is executed, and then the terminal device recognizes the voice command by using an automatic voice recognition technology, and converts the voice command into a voice text.
The voice instruction is the sound collected by the recording device after the recording function of the terminal device is turned on, and includes the words of the user to express the requirements of the user on the terminal device, such as weather of Shenzhen today, background noise when the user inputs the voice instruction, and the like.
For example, referring to fig. 3, after the terminal device obtains the voice text input by the user, the NLU technology is used to analyze the voice text, step 304 is executed to obtain the domain, expression intention and information slot of the voice text, step 305 is executed to fill the information slot of the voice text, and an information slot value is generated.
Among them, NLU (Natural Language Processing) is a technology for communicating with a computer using a Natural Language, and since the key point of Processing the Natural Language is to "understand" the Natural Language by the computer, Natural Language Processing is also called Natural Language Understanding (NLU), which is also called Computational linguistics (Computational linguistics). NLU is a branch subject of artificial intelligence, and researches on simulating human language communication process by an electronic computer to make the computer understand and use the natural language of human society, such as Chinese and English, to realize the natural language communication between human and computer to replace part of mental labor of human, including processing and processing of query data, answering questions, extract documents, assembly data and all related natural language information.
Optionally, referring to fig. 3, the terminal device classifies the voice text by using a natural language processing technology, so as to identify the domain and the expression intention of the voice text, then determines an information slot of the voice text according to the domain and the expression intention, and the terminal device performs sequence labeling on the voice text, and executes step 305, fills the information slot of the voice text, and generates an information slot value. For example, after the speech text is "how weather is in shenzhen today", the terminal device classifies the speech text by using natural language, recognizes that the field of the speech text belongs to the weather field, and expresses that the weather is queried, the information slots included in the speech text are dates and places, and performs the process of sequentially labeling the speech text by determining a corresponding information slot for each word of the speech text, where the information slot corresponding to "present" in the speech text is a date, the information slot corresponding to "day" in the speech text is a date, the information slot corresponding to "deep" in the speech text is a place, the information slot corresponding to "zhen" in the speech text is a place, the "present" in the speech text does not have a corresponding information slot, the "day" in the speech text does not have a corresponding information slot, the "qi" in the speech text does not have a corresponding information slot, and the "what" in the speech text does not have a corresponding information slot, no corresponding information slot exists in the 'sample' in the voice text, and no corresponding information slot exists in the 'sample' in the voice text, so that the information slot value corresponding to the information slot date is today, and the information slot value corresponding to the information slot place is Shenzhen.
And step 203, judging whether to execute the voice command according to the information slot value.
For example, referring to fig. 3 together, after the terminal device obtains the domain, the expression intention, the information slot, and the information slot value of the voice text, it determines whether the voice text is abnormal according to the expression intention, the domain, and the information slot value, step 306 is executed, the terminal device first determines the domain of the voice text, if the terminal device determines that the voice text includes at least two different domains, it indicates that the voice text is abnormal, step 309 is executed to stop the voice input, the terminal device increases the stop button and/or the pause button of the voice input, and can select to move the stop button and/or the pause button of the voice input from the edge corner of the display screen of the terminal device to the middle, so as to facilitate the user to click and stop the voice input.
Optionally, if it is determined that the voice text includes at least two different fields, it indicates that the voice text is abnormal, and if the terminal device controls the voice input function to stop by using the gesture or the facial expression, the terminal device reduces the requirement that the gesture or the facial expression controls the voice input function, for example, reduces the duration or complexity of the gesture or the facial expression, so that the terminal device is more sensitive to the gesture or the facial expression that stops the voice input function, and a user can conveniently stop the voice input of the terminal device.
Optionally, referring to fig. 3 together, when it is determined that the speech text includes one of the domains, the terminal device determines that the speech text is normal, step 307 is executed, then the terminal device determines an expression intention included in the speech text, if the terminal device determines that the speech text includes at least two different expression intentions, it indicates that the speech text is abnormal, step 309 is executed, the speech input is stopped, the terminal device increases a stop button and/or a pause button of the speech input, and the stop button and/or the pause button of the speech input can be selected to be moved from an edge corner to the middle of a display screen of the terminal device, so that the user can click to stop the speech input.
Optionally, if the terminal device determines that the voice text includes at least two different expression intentions, it indicates that the voice text is abnormal, and if the terminal device controls the voice input function to stop by using the gesture or the facial expression, the terminal device reduces the requirement that the gesture or the facial expression controls the voice input function, for example, reduces the duration or complexity of the gesture or the facial expression, so that the terminal device is more sensitive to the gesture or the facial expression that stops the voice input function, and a user can conveniently stop the voice input of the terminal device.
Optionally, referring to fig. 3 together, when it is determined that the speech text includes the expression intention, the terminal device determines that the speech text is normal, step 308 is executed, then the terminal device determines an information slot value included in the speech text, if the terminal device determines that the speech text includes at least two different information slot values, it indicates that the speech text is abnormal, step 309 is executed, the speech input is stopped, the terminal device increases a stop button and/or a pause button of the speech input, and may select to move the stop button and/or the pause button of the speech input from an edge corner of a display screen of the terminal device to the middle, so that a user can click to stop the speech input, otherwise, the speech text is normal, step 310 is executed, and the speech instruction is executed.
Optionally, if the terminal device determines that the voice text includes at least two different slot values, it indicates that the voice text is abnormal, and if the terminal device controls the voice input function to stop by using the gesture or the facial expression, the terminal device reduces the requirement for controlling the voice input function by using the gesture or the facial expression, for example, the duration or complexity of the gesture or the facial expression is reduced, so that the vehicle-mounted terminal senses the gesture or the facial expression for stopping the voice input function more sensitively, and a user can conveniently stop the voice input by using the terminal device.
In order to better implement the above method, an embodiment of the present invention may further provide a voice instruction recognition apparatus, where the voice instruction recognition apparatus may be specifically integrated in a network device, and the network device may be a terminal or another device.
For example, as shown in fig. 4, the voice instruction recognition apparatus may include a recognition unit 401, an acquisition unit 402, and a determination unit 403 as follows:
(1) recognition unit 401
The recognition unit 401 is configured to recognize a voice instruction input by a user and generate a voice text.
For example, a user turns on a voice input function of the terminal device, the recognition unit 401 of the terminal device turns on a recording device such as a microphone, the recording device collects a voice instruction input by the user, and the recognition unit 401 recognizes the voice instruction by using an automatic voice recognition technology to convert the voice instruction into a voice text.
(2) Acquisition unit 402
An obtaining unit 402, configured to obtain an information slot value of the speech text.
For example, after the terminal device acquires the voice text, the recognition unit 401 of the terminal device transmits the voice text to the acquisition unit 402, and the acquisition unit 402 analyzes the voice text by using the NLU technology to acquire the domain, the expression intention, the information slot, and the information slot value of the voice text.
Optionally, the obtaining unit 402 classifies the voice text by using a natural language processing technology, so as to identify an expression intention and a domain of the voice text, and determine an information slot of the voice text, and the obtaining unit 402 performs sequence labeling on the voice text, fills the information slot of the voice text, and generates an information slot value.
(3) Judging unit 403
A judging unit 403, configured to judge whether to execute the voice instruction according to the information slot value.
For example, the obtaining unit 402 transmits the expression intention, the belonging field, the information slot and the information slot value of the voice text to the judging unit 403, the judging unit 403 firstly determines that the voice text comprises at least two different belonging fields, which indicates that the voice text is abnormal, the terminal device increases the stop button and/or pause button of the voice input, and can select to move the stop button and/or pause button of the voice input from the edge corner to the middle of the display screen of the terminal device, so as to facilitate the user to click and stop the voice input, if the judging unit 403 determines that the voice text comprises one of the belonging fields, it determines that the voice text is normal, then the terminal device judges the expression intention comprised by the voice text, if the judging unit 403 determines that the voice text comprises at least two different expression intents, it determines that the voice text is abnormal, the terminal stops the voice input, if the voice text includes one expression intention, it is determined that the voice text is normal, then the terminal device may determine an information slot value included in the voice text, if the determining unit 403 determines that the voice text includes at least two different information slot values, it is determined that the voice text is abnormal, the terminal stops voice input, and if the voice text includes one information slot value, it is determined that the voice text is normal, the terminal executes the voice instruction.
Optionally, the determining unit 403 determines that the voice text includes at least two different fields, which indicates that the voice text is abnormal, and when the terminal device controls the voice input function to stop by using the gesture or the facial expression, the terminal device reduces the requirement for controlling the voice input function by using the gesture or the facial expression, for example, reduces the duration or complexity of the gesture or the facial expression, so that the vehicle-mounted terminal senses the gesture or the facial expression of stopping the voice input function more sensitively, and a user can conveniently stop the voice input of the terminal device.
Accordingly, an embodiment of the present invention further provides a terminal, as shown in fig. 5, the terminal may include a Radio Frequency (RF) circuit 501, a memory 502 including one or more computer-readable storage media, an input unit 503, a display unit 504, a sensor 505, an audio circuit 506, a Wireless Fidelity (WiFi) module 507, a processor 508 including one or more processing cores, and a power supply 509. Those skilled in the art will appreciate that the terminal structure shown in fig. 5 is not intended to be limiting and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components. Wherein:
the RF circuit 501 may be used for receiving and transmitting signals during information transmission and reception or during a call, and in particular, for receiving downlink information of a base station and then sending the received downlink information to the one or more processors 508 for processing; in addition, data relating to uplink is transmitted to the base station. In general, RF circuit 501 includes, but is not limited to, an antenna, at least one Amplifier, a tuner, one or more oscillators, a Subscriber Identity Module (SIM) card, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like. In addition, the RF circuitry 501 may also communicate with networks and other devices via wireless communications. The wireless communication may use any communication standard or protocol, including but not limited to Global System for Mobile communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), email, Short Message Service (SMS), and the like.
The memory 502 may be used to store software programs and modules, and the processor 508 executes various functional applications and data processing by operating the software programs and modules stored in the memory 502. The memory 502 may mainly include a program storage area and a data storage area, wherein the program storage area may store an operating system, an application program required by at least one function (such as a sound playing function, an image playing function, etc.), and the like; the storage data area may store data (such as audio data, a phonebook, etc.) created according to the use of the terminal, etc. Further, the memory 502 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device. Accordingly, the memory 502 may also include a memory controller to provide the processor 508 and the input unit 503 access to the memory 502.
The input unit 503 may be used to receive input numeric or character information and generate keyboard, mouse, joystick, optical or trackball signal inputs related to user settings and function control. In particular, in one particular embodiment, the input unit 503 may include a touch-sensitive surface as well as other input devices. The touch-sensitive surface, also referred to as a touch display screen or a touch pad, may collect touch operations by a user (e.g., operations by a user on or near the touch-sensitive surface using a finger, a stylus, or any other suitable object or attachment) thereon or nearby, and drive the corresponding connection device according to a predetermined program. Alternatively, the touch sensitive surface may comprise two parts, a touch detection means and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, sends the touch point coordinates to the processor 508, and can receive and execute commands sent by the processor 508. In addition, touch sensitive surfaces may be implemented using various types of resistive, capacitive, infrared, and surface acoustic waves. The input unit 503 may include other input devices in addition to the touch-sensitive surface. In particular, other input devices may include, but are not limited to, one or more of a physical keyboard, function keys (such as volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like.
The display unit 504 may be used to display information input by or provided to the user and various graphical user interfaces of the terminal, which may be made up of graphics, text, icons, video, and any combination thereof. The Display unit 504 may include a Display panel, and optionally, the Display panel may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-Emitting Diode (OLED), or the like. Further, the touch-sensitive surface may overlay the display panel, and when a touch operation is detected on or near the touch-sensitive surface, the touch operation is transmitted to the processor 508 to determine the type of touch event, and then the processor 508 provides a corresponding visual output on the display panel according to the type of touch event. Although in FIG. 5 the touch-sensitive surface and the display panel are two separate components to implement input and output functions, in some embodiments the touch-sensitive surface may be integrated with the display panel to implement input and output functions.
The terminal may also include at least one sensor 505, such as light sensors, motion sensors, and other sensors. Specifically, the light sensor may include an ambient light sensor that may adjust the brightness of the display panel according to the brightness of ambient light, and a proximity sensor that may turn off the display panel and/or the backlight when the terminal is moved to the ear. As one of the motion sensors, the gravity acceleration sensor can detect the magnitude of acceleration in each direction (generally, three axes), can detect the magnitude and direction of gravity when the mobile phone is stationary, and can be used for applications of recognizing the posture of the mobile phone (such as horizontal and vertical screen switching, related games, magnetometer posture calibration), vibration recognition related functions (such as pedometer and tapping), and the like; as for other sensors such as a gyroscope, a barometer, a hygrometer, a thermometer, and an infrared sensor, which can be configured in the terminal, detailed description is omitted here.
WiFi belongs to short-distance wireless transmission technology, and the terminal can help a user to receive and send e-mails, browse webpages, access streaming media and the like through the WiFi module 507, and provides wireless broadband internet access for the user. Although fig. 5 shows the WiFi module 507, it is understood that it does not belong to the essential constitution of the terminal, and may be omitted entirely as needed within the scope not changing the essence of the invention.
The processor 508 is a control center of the terminal, connects various parts of the entire mobile phone by using various interfaces and lines, and performs various functions of the terminal and processes data by operating or executing software programs and/or modules stored in the memory 502 and calling data stored in the memory 502, thereby integrally monitoring the mobile phone. Optionally, processor 508 may include one or more processing cores; preferably, the processor 508 may integrate an application processor, which primarily handles operating systems, user interfaces, application programs, etc., and a modem processor, which primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 508.
The terminal also includes a power supply 509 (e.g., a battery) for powering the various components, which may preferably be logically connected to the processor 508 via a power management system that may be used to manage charging, discharging, and power consumption. The power supply 509 may also include any component such as one or more dc or ac power sources, recharging systems, power failure detection circuitry, power converters or inverters, power status indicators, and the like.
Although not shown, the terminal may further include a camera, a bluetooth module, and the like, which will not be described herein. Specifically, in this embodiment, the processor 508 in the terminal loads the executable file corresponding to the process of one or more application programs into the memory 502 according to the following instructions, and the processor 508 runs the application programs stored in the memory 502, thereby implementing various functions: recognizing a voice instruction input by a user, generating a voice text, acquiring an information slot value of the voice text, and judging whether to execute the voice instruction according to the information slot value.
In the foregoing embodiments, the descriptions of the respective embodiments have respective emphasis, and for parts that are not described in detail in a certain embodiment, reference may be made to related descriptions of other embodiments.
The voice instruction recognition method, the voice instruction recognition device, the terminal device and the storage medium provided by the embodiments of the present invention are described in detail above, and a specific example is applied in the text to explain the principle and the implementation of the present invention, and the description of the above embodiments is only used to help understanding the technical scheme and the core idea of the present invention; those of ordinary skill in the art will understand that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.
Claims (10)
1. A voice instruction recognition method, comprising:
recognizing a voice instruction input by a user and generating a voice text;
acquiring an information slot value of the voice text;
and judging whether to execute the voice command according to the information slot value.
2. The method of claim 1, wherein the obtaining the slot value of the speech text comprises:
determining the belonged field and the expression intention according to the content of the voice text;
acquiring an information slot of the voice text according to the belonged field and the expression intention;
and filling the information slot to generate an information slot value.
3. The method of claim 1, wherein said determining whether to execute the voice command according to the slot value comprises:
and judging whether to execute the voice command according to the field of the voice text.
4. The method of claim 3, wherein the determining whether to execute the voice instruction according to the domain of the voice text comprises:
determining that the voice text comprises at least two different fields, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises a field, and judging whether to execute the voice command according to the expression intention of the voice text.
5. The method according to claim 4, wherein the determining whether the phonetic text is normal according to the expression intention comprises:
determining that the voice text comprises at least two different expression intentions, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises the expression intention, and judging whether to execute the voice instruction according to the information slot value.
6. The method of claim 5, wherein said determining whether to execute the voice command according to the slot value comprises:
determining that the voice text comprises at least two different information slot values, not executing the voice instruction, and stopping voice input;
and determining that the voice text comprises the information slot value, and executing the voice instruction.
7. The method of any of claims 4-6, wherein said ceasing speech input comprises:
a stop button to increase the voice input.
8. A voice instruction recognition apparatus, comprising:
the recognition unit is used for recognizing a voice instruction input by a user and generating a voice text;
the acquisition unit is used for acquiring the information slot value of the voice text;
and the judging unit is used for judging whether to execute the voice command according to the information slot value.
9. A terminal device, comprising:
a memory for storing an application program;
a processor for implementing the steps in the speech instruction recognition method according to any one of claims 1 to 7 when executing the application program.
10. A storage medium having an application program stored thereon, wherein the application program, when executed by a processor, implements the steps of the speech instruction recognition method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010722276.6A CN111897916B (en) | 2020-07-24 | 2020-07-24 | Voice instruction recognition method, device, terminal equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010722276.6A CN111897916B (en) | 2020-07-24 | 2020-07-24 | Voice instruction recognition method, device, terminal equipment and storage medium |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111897916A true CN111897916A (en) | 2020-11-06 |
CN111897916B CN111897916B (en) | 2024-03-19 |
Family
ID=73190897
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010722276.6A Active CN111897916B (en) | 2020-07-24 | 2020-07-24 | Voice instruction recognition method, device, terminal equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111897916B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115588432A (en) * | 2022-11-23 | 2023-01-10 | 广州小鹏汽车科技有限公司 | Voice interaction method, server and computer readable storage medium |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140207453A1 (en) * | 2013-01-22 | 2014-07-24 | Electronics And Telecommunications Research Institute | Method and apparatus for editing voice recognition results in portable device |
CN105206266A (en) * | 2015-09-01 | 2015-12-30 | 重庆长安汽车股份有限公司 | Vehicle-mounted voice control system and method based on user intention guess |
CN105320726A (en) * | 2014-05-30 | 2016-02-10 | 苹果公司 | Reducing the need for manual start/end-pointing and trigger phrases |
CN109543192A (en) * | 2018-11-30 | 2019-03-29 | 北京羽扇智信息科技有限公司 | Natural language analytic method, device, equipment and storage medium |
CN109616111A (en) * | 2018-12-24 | 2019-04-12 | 北京恒泰实达科技股份有限公司 | A kind of scene interactivity control method based on speech recognition |
CN109800407A (en) * | 2017-11-15 | 2019-05-24 | 腾讯科技(深圳)有限公司 | Intension recognizing method, device, computer equipment and storage medium |
US20190370385A1 (en) * | 2018-06-04 | 2019-12-05 | International Business Machines Corporation | Generation of domain specific type system |
CN110659970A (en) * | 2018-06-12 | 2020-01-07 | 百度在线网络技术(北京)有限公司 | Account information processing method and device based on voice recognition and electronic equipment |
CN110827816A (en) * | 2019-11-08 | 2020-02-21 | 杭州依图医疗技术有限公司 | Voice instruction recognition method and device, electronic equipment and storage medium |
CN111124121A (en) * | 2019-12-24 | 2020-05-08 | 腾讯科技(深圳)有限公司 | Voice interaction information processing method and device, storage medium and computer equipment |
CN111261157A (en) * | 2020-01-03 | 2020-06-09 | 苏州思必驰信息科技有限公司 | Control method, device and equipment for short video and storage medium |
CN111341311A (en) * | 2020-02-21 | 2020-06-26 | 深圳前海微众银行股份有限公司 | Voice conversation method and device |
CN111373473A (en) * | 2018-03-05 | 2020-07-03 | 华为技术有限公司 | Electronic equipment and method for performing voice recognition by using same |
-
2020
- 2020-07-24 CN CN202010722276.6A patent/CN111897916B/en active Active
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140207453A1 (en) * | 2013-01-22 | 2014-07-24 | Electronics And Telecommunications Research Institute | Method and apparatus for editing voice recognition results in portable device |
CN105320726A (en) * | 2014-05-30 | 2016-02-10 | 苹果公司 | Reducing the need for manual start/end-pointing and trigger phrases |
CN105206266A (en) * | 2015-09-01 | 2015-12-30 | 重庆长安汽车股份有限公司 | Vehicle-mounted voice control system and method based on user intention guess |
CN109800407A (en) * | 2017-11-15 | 2019-05-24 | 腾讯科技(深圳)有限公司 | Intension recognizing method, device, computer equipment and storage medium |
CN111373473A (en) * | 2018-03-05 | 2020-07-03 | 华为技术有限公司 | Electronic equipment and method for performing voice recognition by using same |
US20190370385A1 (en) * | 2018-06-04 | 2019-12-05 | International Business Machines Corporation | Generation of domain specific type system |
CN110659970A (en) * | 2018-06-12 | 2020-01-07 | 百度在线网络技术(北京)有限公司 | Account information processing method and device based on voice recognition and electronic equipment |
CN109543192A (en) * | 2018-11-30 | 2019-03-29 | 北京羽扇智信息科技有限公司 | Natural language analytic method, device, equipment and storage medium |
CN109616111A (en) * | 2018-12-24 | 2019-04-12 | 北京恒泰实达科技股份有限公司 | A kind of scene interactivity control method based on speech recognition |
CN110827816A (en) * | 2019-11-08 | 2020-02-21 | 杭州依图医疗技术有限公司 | Voice instruction recognition method and device, electronic equipment and storage medium |
CN111124121A (en) * | 2019-12-24 | 2020-05-08 | 腾讯科技(深圳)有限公司 | Voice interaction information processing method and device, storage medium and computer equipment |
CN111261157A (en) * | 2020-01-03 | 2020-06-09 | 苏州思必驰信息科技有限公司 | Control method, device and equipment for short video and storage medium |
CN111341311A (en) * | 2020-02-21 | 2020-06-26 | 深圳前海微众银行股份有限公司 | Voice conversation method and device |
Non-Patent Citations (2)
Title |
---|
MEHMET BERKEHAN AKÇAY 等: "Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers", 《SPEECH COMMUNICATION》, pages 56 - 76 * |
王东升 等: "面向限定领域问答系统的自然语言理解方法综述", 《计算机科学》, pages 1 - 8 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115588432A (en) * | 2022-11-23 | 2023-01-10 | 广州小鹏汽车科技有限公司 | Voice interaction method, server and computer readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN111897916B (en) | 2024-03-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN108549519B (en) | Split screen processing method and device, storage medium and electronic equipment | |
CN106293308B (en) | Screen unlocking method and device | |
CN108712566B (en) | Voice assistant awakening method and mobile terminal | |
CN106528545B (en) | Voice information processing method and device | |
US9921735B2 (en) | Apparatuses and methods for inputting a uniform resource locator | |
CN108958606B (en) | Split screen display method and device, storage medium and electronic equipment | |
CN106445596B (en) | Method and device for managing setting items | |
CN109284144B (en) | Fast application processing method and mobile terminal | |
CN108958629B (en) | Split screen quitting method and device, storage medium and electronic equipment | |
WO2015043200A1 (en) | Method and apparatus for controlling applications and operations on a terminal | |
CN112230877A (en) | Voice operation method and device, storage medium and electronic equipment | |
CN109669662A (en) | A kind of pronunciation inputting method, device, storage medium and mobile terminal | |
CN110335629B (en) | Pitch recognition method and device of audio file and storage medium | |
US20150088525A1 (en) | Method and apparatus for controlling applications and operations on a terminal | |
CN109688611B (en) | Frequency band parameter configuration method, device, terminal and storage medium | |
CN110688051A (en) | Screen recording operation method and device, computer readable storage medium and terminal | |
CN113393838A (en) | Voice processing method and device, computer readable storage medium and computer equipment | |
CN109062643A (en) | A kind of display interface method of adjustment, device and terminal | |
CN111897916B (en) | Voice instruction recognition method, device, terminal equipment and storage medium | |
CN111580911A (en) | Operation prompting method and device for terminal, storage medium and terminal | |
CN108920086B (en) | Split screen quitting method and device, storage medium and electronic equipment | |
CN115116434A (en) | Application implementation method and device, storage medium and electronic equipment | |
CN112367425B (en) | Volume adjusting method and device and terminal | |
CN111027406B (en) | Picture identification method and device, storage medium and electronic equipment | |
CN109032482B (en) | Split screen control method and device, storage medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |