CN111128168A - Voice control method, device and storage medium - Google Patents

Voice control method, device and storage medium Download PDF

Info

Publication number
CN111128168A
CN111128168A CN201911400172.7A CN201911400172A CN111128168A CN 111128168 A CN111128168 A CN 111128168A CN 201911400172 A CN201911400172 A CN 201911400172A CN 111128168 A CN111128168 A CN 111128168A
Authority
CN
China
Prior art keywords
semantic recognition
recognition result
equipment
voice
semantic
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911400172.7A
Other languages
Chinese (zh)
Inventor
靳莹雪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zebra Network Technology Co Ltd
Original Assignee
Zebra Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zebra Network Technology Co Ltd filed Critical Zebra Network Technology Co Ltd
Priority to CN201911400172.7A priority Critical patent/CN111128168A/en
Publication of CN111128168A publication Critical patent/CN111128168A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention provides a voice control method, a voice control device and a storage medium, wherein voice data is received and subjected to semantic recognition to obtain a first semantic recognition result for indicating the name and the operation type of equipment, whether the first semantic recognition result is complete or not is judged, if the first semantic recognition result is incomplete, inquiry voice is sent, response voice data of the inquiry voice is received, semantic recognition is carried out on the response voice data to obtain a second semantic recognition result, the second semantic recognition result is used for indicating position information and/or working parameters of the name of the equipment, and a control instruction for indicating the on-off state or the working parameters of the corresponding equipment is generated according to the first semantic recognition result and the second semantic recognition result. The real intention of the user is obtained through a multi-round voice interaction process, fine control over controlled equipment in a scene is achieved, and user experience is improved.

Description

Voice control method, device and storage medium
Technical Field
The invention relates to the technical field of human-computer interaction, in particular to a voice control method, a voice control device and a storage medium.
Background
With the continuous updating and upgrading of the internet of things and the development of artificial intelligence, a man-machine interaction mode develops towards the direction of user-centered personalized biological recognition and all-around perception. The voice interaction is one of the man-machine interaction modes, and is widely applied to the fields of intelligent home and intelligent vehicle-mounted control.
Semantic recognition is the key of voice interaction, and the main processing is complete semantics. For example, when analyzing natural language, most functions of intelligent voice devices such as an intelligent sound box need to extract an intention (intent) and an entity (slot) simultaneously, the combination of the intention and the entity can transmit complete semantics, and the devices execute corresponding operations according to the complete semantics.
However, when the intelligent voice device controls a multi-region device, even if a complete semantic meaning is recognized, the user intention understood by the intelligent voice device may be different from the real intention of the user, and an accurate device control effect cannot be achieved.
Disclosure of Invention
The invention provides a voice control method, a voice control device and a storage medium, which are used for realizing accurate control of intelligent equipment.
A first aspect of the present invention provides a voice control method, including:
receiving voice data, and performing semantic recognition on the voice data to obtain a first semantic recognition result, wherein the first semantic recognition result is used for indicating the name and the operation type of equipment;
judging whether the first semantic recognition result is complete or not, and if the first semantic recognition result is incomplete, sending inquiry voice;
receiving response voice data of the inquiry voice, and performing semantic recognition on the response voice data to obtain a second semantic recognition result, wherein the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name;
and generating a control instruction according to the first semantic recognition result and the second semantic recognition result, wherein the control instruction is used for indicating and controlling the switch state or the working parameter of the corresponding equipment.
In a possible implementation manner, before receiving the voice data, the method further includes:
acquiring equipment configuration information, wherein the equipment configuration information comprises equipment names, operation types, equipment numbers and position information corresponding to the equipment names;
storing the equipment configuration information into a voice-controlled equipment configuration table;
the judging whether the first semantic identification result is complete includes:
and judging whether the first semantic recognition result is complete according to the equipment configuration table.
In a possible implementation manner, the determining whether the first semantic recognition result is complete, and if the first semantic recognition result is incomplete, sending an inquiry voice includes:
determining the number of devices corresponding to the device names by inquiring the device configuration table;
and if the number of the devices is more than or equal to two, determining that the first semantic recognition result is incomplete, and sending the query voice.
In a possible implementation manner, the controlling the on-off state or the operating parameter of the corresponding device according to the first semantic recognition result and the second semantic recognition result includes:
and starting or closing the equipment corresponding to the position information or adjusting the working parameters of the equipment corresponding to the position information according to the first semantic recognition result and the second semantic recognition result.
A second aspect of the present invention provides a voice control apparatus comprising:
the receiving module is used for receiving voice data; the processing module is used for carrying out semantic recognition on the voice data to obtain a first semantic recognition result, and the first semantic recognition result is used for indicating the equipment name and the operation type;
the processing module is further configured to determine whether the first semantic identification result is complete, and if the first semantic identification result is incomplete; the sending module is used for sending inquiry voice;
the receiving module is further used for receiving response voice data of the inquiry voice; the processing module is further used for performing semantic recognition on the response voice data to obtain a second semantic recognition result, and the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name;
the processing module is further configured to generate a control instruction according to the first semantic recognition result and the second semantic recognition result, where the control instruction is used to instruct and control a switch state or a working parameter of a corresponding device.
In a possible implementation manner, before the receiving module receives the voice data, the receiving module is further configured to: acquiring equipment configuration information, wherein the equipment configuration information comprises equipment names, operation types, equipment numbers and position information corresponding to the equipment names;
the device further comprises: the storage module is used for storing the equipment configuration information into an equipment configuration table controlled by voice;
the processing module is specifically configured to determine whether the first semantic identification result is complete according to the device configuration table.
In a possible implementation manner, the processing module is specifically configured to:
determining the number of devices corresponding to the device names by inquiring the device configuration table;
if the number of the devices is larger than or equal to two, determining that the first semantic recognition result is incomplete; and the sending module is used for sending the inquiry voice.
In a possible implementation manner, the processing module is specifically configured to:
and starting or closing the equipment corresponding to the position information or adjusting the working parameters of the equipment corresponding to the position information according to the first semantic recognition result and the second semantic recognition result.
A third aspect of the present invention provides a voice control apparatus comprising:
a memory;
a processor; and
a computer program;
wherein the computer program is stored in the memory and configured to be executed by the processor to implement the speech control method according to any one of the first aspect of the invention.
A fourth aspect of the invention provides a computer readable storage medium having stored thereon a computer program for execution by a processor to implement the speech control method according to any one of the first aspect of the invention.
The invention provides a voice control method, a voice control device and a storage medium, wherein voice data is received and subjected to semantic recognition to obtain a first semantic recognition result for indicating the name and the operation type of equipment, whether the first semantic recognition result is complete or not is judged, if the first semantic recognition result is incomplete, inquiry voice is sent, response voice data of the inquiry voice is received, semantic recognition is carried out on the response voice data to obtain a second semantic recognition result, the second semantic recognition result is used for indicating position information and/or working parameters of the name of the equipment, and a control instruction for indicating the on-off state or the working parameters of the corresponding equipment is generated according to the first semantic recognition result and the second semantic recognition result. The real intention of the user is obtained through a multi-round voice interaction process, the success rate of finishing user indication by the voice control equipment is improved, fine control over controlled equipment in a scene is achieved, and user experience is improved.
Drawings
Fig. 1 is a schematic view of a scene of a voice control method according to an embodiment of the present invention;
fig. 2 is a schematic flow chart of a voice control method according to an embodiment of the present invention;
FIG. 3 is a schematic configuration diagram of a voice control method according to an embodiment of the present invention;
fig. 4 is a schematic structural diagram of a voice control apparatus according to an embodiment of the present invention;
fig. 5 is a hardware structure diagram of a voice control apparatus according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are some, but not all, embodiments of the present invention. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It will be understood that the terms "comprises" and "comprising," and any variations thereof, as used herein, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The embodiment of the invention provides a voice control method which can be applied to the field of intelligent homes and intelligent automobiles. Taking an intelligent automobile as an example, fig. 1 shows a scene schematic diagram of a voice control method provided by an embodiment of the present invention, as shown in fig. 1, the scene includes a voice control device 11 and a plurality of controlled devices 12, where the voice control device 11 is in communication connection with the plurality of controlled devices. The controlled devices may be a wiper 121, a window 122, an in-vehicle cooling device 123, an in-vehicle lighting device 124, and the like. It should be understood that the number of windows, in-vehicle cooling devices, in-vehicle lighting devices is more than one. For example, the windows include two windows in the front row, two windows in the rear row, and a top window; the in-vehicle refrigerating device comprises an in-vehicle refrigerating device in a driving area and a rear row area, and the in-vehicle illuminating device comprises a main driving illuminating device, an auxiliary driving illuminating device and a rear row illuminating device. It should be noted that the voice control device may be disposed in the vehicle as an independent device, and may also be integrated into one of the multiple controlled devices, for example, the voice control function is integrated into an in-vehicle audio device, and the audio device controls other devices in the vehicle to turn on or off according to the received voice data.
Based on the application scenario, the application provides a voice control method, which identifies and judges whether voice data sent by a user is complete or not by acquiring the voice data: in a possible case, if the semantic recognition is to extract an intention, such as "open", the voice data is considered to be incomplete, and multiple rounds of interaction are required to be started to acquire an entity corresponding to the intention, such as "wiper", so as to obtain a complete semantic recognition result. In another possible case, if the entity is extracted in the semantic recognition, the voice data is considered to be incomplete, and multiple rounds of interaction are required to be started to obtain the corresponding intention of the entity, so that a complete semantic recognition result is obtained. In another possible case, if the matching degree of the entity + intention extracted by the semantic recognition and the preset entity + intention is low, multiple rounds of interaction also need to be started to further determine the entity + intention, so as to obtain a complete and accurate semantic recognition result.
In the above scheme, as long as the voice control device extracts the complete semantics of the intention + entity, the controlled device corresponding to the control entity is started, closed or parameter-adjusted. Therefore, the control granularity of the scheme is coarse, the understood user intention is different from the real intention of the user, and accurate device control cannot be achieved.
Based on the above problems, the embodiment of the application provides a voice control method, when the semantic is determined to be complete, the device needs to be subjected to region division and region control according to the type and the number of the devices in the scene, the real intention of the user is obtained through a multi-round interaction process, the controlled device in the scene is finely controlled, and the user experience is improved.
The technical solution of the present invention will be described in detail below with specific examples. The following several specific embodiments may be combined with each other, and details of the same or similar concepts or processes may not be repeated in some embodiments.
Fig. 2 is a flowchart illustrating a voice control method according to an embodiment of the present invention. Fig. 3 is a schematic configuration diagram of a voice control method according to an embodiment of the present invention.
The method provided by the embodiment can be executed by any device for executing the voice control method, the device can be realized by software and/or hardware, and the device can be arranged on any terminal equipment or server.
As shown in fig. 2, the voice control method provided in this embodiment includes the following steps:
step 101, receiving voice data.
In this embodiment, as shown in fig. 3, before the voice control apparatus receives the voice data, the following steps are further included:
s1011, obtaining the device configuration information, wherein the device configuration information comprises the device name, the operation type, the number of the devices and the position information corresponding to the device name.
The voice control device is in communication connection with the controlled device in the scene, and after the controlled device is successfully connected with the voice control device, the device configuration information can be sent to the voice control device, so that the voice control device can acquire the device name, the operation type, the position information corresponding to the device name and the like of the controlled device.
After receiving the device configuration information sent by the multiple controlled devices in the scene, the voice control apparatus counts the number of devices of the same device type for performing accurate device control, which may be referred to in the following.
S1012, storing the equipment configuration information into the voice-controlled equipment configuration table.
In this embodiment, the device configuration table includes an operation type, location information, and a device number corresponding to each device type. Exemplarily, table 1 shows an apparatus configuration table of controlled apparatuses in a vehicle, and it can be known from table 1 that the operation types corresponding to different types of controlled apparatuses may be the same or different; different types of controlled equipment are arranged at different positions inside or outside the vehicle, and the controlled equipment executes corresponding operation according to a control instruction sent by the voice control device; the number of the controlled devices of different types may be the same or different, and may be one or more.
TABLE 1
Figure BDA0002347292220000061
As an example, before the voice control apparatus receives the voice data, the method further includes:
and receiving a wake-up instruction, wherein the wake-up instruction is used for waking up the voice control device. Before the voice control device is awakened, the voice control device is in a dormant state. Specifically, the manner in which the user wakes up the voice control device may include, but is not limited to, the following two manners:
in one possible implementation, the user wakes up the voice control device by sending a preset wake-up word. The preset awakening word can be user-defined or preset by a manufacturer. Specifically, when the user speaks a wake-up word and the voice control device detects the wake-up word, the voice control device judges whether the wake-up word is matched with a preset wake-up word, if so, the voice control device is switched from a dormant state to a working state, and whether new voice data exist is detected in real time.
In a possible implementation manner, a user manually triggers a preset physical button or a touch button on a human-computer interaction interface to wake up the voice control device, so that the voice control device is switched from a dormant state to a working state, and whether new voice data exists is detected in real time.
And 102, performing semantic recognition on the voice data to obtain a first semantic recognition result, wherein the first semantic recognition result is used for indicating the equipment name and the operation type.
In this embodiment, after receiving the voice data, the voice control apparatus may first perform noise reduction processing on the voice data to remove background noise. For example, when the sound box in the vehicle plays music, the voice control device removes the music played by the sound box in the vehicle through the noise reduction process, and extracts the voice uttered by the user. And then carrying out semantic recognition on the processed voice data to obtain a first semantic recognition result, and storing the first semantic recognition result. The first semantic recognition result includes the following cases:
in one possible case, the first semantic identification result includes a device name (corresponding entity); another possible scenario, the first semantic recognition result includes an operation type (corresponding to an intention); in yet another possible scenario, the first semantic identification result includes a device name and an operation type (corresponding entity + intent).
And 103, judging whether the first semantic identification result is complete, and if the first semantic identification result is incomplete, executing step 104.
Based on the first semantic recognition result in step 102, it is determined whether the first semantic recognition result is complete, specifically, whether the first semantic recognition result is complete from the perspective of device function implementation, for example, whether the device name and the operation type are complete, whether the functions of the devices of the same type in the application scene are independent from each other, and the like.
In a possible implementation manner, the voice control apparatus first determines whether the first semantic recognition result includes the device name and the operation type, and determines that the first semantic recognition result is incomplete if the first semantic recognition result includes only the device name or only the operation type. If the first semantic identification result comprises the device name and the operation type, whether the number of devices corresponding to the device name is more than or equal to two needs to be further judged, and if the number of devices corresponding to the device name is more than or equal to two, the first semantic identification result is determined to be incomplete.
Whether the number of the devices corresponding to the device names is greater than or equal to two is judged, and the determination can be carried out in the following mode: the voice control device inquires the equipment configuration table and determines the equipment number corresponding to the equipment name.
When the voice control device determines that the first semantic recognition result is incomplete, the voice control device can send inquiry voice according to a preset material library to obtain complete semantics, so that subsequent equipment control is executed.
Step 104, sending inquiry voice.
Wherein the query voice may be used to query the device name, the operation type, the operation parameter, or the location information of the device name. The voice control device transmits an inquiry voice based on a preset corpus.
If the first semantic recognition result includes only the device name, the transmitted inquiry voice can be used to inquire about the operation type of the device name, for example, the semantic recognition extracts the device name "wiper", the semantic recognition result is incomplete, the voice control apparatus transmits an inquiry voice "ask for whether to open the front wiper, or the rear wiper? ".
If the first semantic recognition result only includes the operation type, the transmitted query voice may be used to query the device name corresponding to the operation type, for example, the semantic recognition is extracted to the operation type "open", the semantic recognition result is incomplete, and the voice control apparatus transmits the query voice "ask what device to open? ".
If the first semantic recognition result comprises the device name and the operation type, but the number of the device names is greater than or equal to two, the sent inquiry voice can be used for inquiring the position information corresponding to the device name. For example, semantic recognition extracts an equipment name "air conditioner" and an operation type "open", and since the number of air conditioners in the vehicle is more than one, the semantic recognition result is incomplete, and the voice control device transmits a query voice "ask for which area of the air conditioner in the vehicle to open? "or" ask for a request to turn on the air conditioner in the driving area or the rear row area? ".
Optionally, besides the above situations, the query voice may be sent according to an actual situation, and the working parameter corresponding to the operation type may be obtained. For example, the semantic recognition extracts the device name "air conditioner", the operation type "on", the location information "rear area" of the device name, and the voice control apparatus may also send a query voice "how many degrees the rear air conditioner needs to be set to ask for a question? ". The scheme is more humanized, and user experience is improved.
And step 105, receiving response voice data of the inquiry voice.
And 106, performing semantic recognition on the answering voice data to obtain a second semantic recognition result, wherein the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name.
After the voice control device obtains the second semantic recognition result, if the combination of the first semantic recognition result and the second semantic recognition result can obtain a complete semantic recognition result, step 207 is executed.
It should be understood that if a complete semantic recognition result cannot be obtained through two rounds of voice interaction, a third round or more rounds of voice interaction can be performed until the complete semantic recognition result is obtained, and the user intention is accurately obtained.
And 107, generating a control instruction according to the first semantic recognition result and the second semantic recognition result, wherein the control instruction is used for indicating and controlling the switch state or the working parameter of the corresponding equipment.
Specifically, a complete semantic recognition result is obtained according to the first semantic recognition result and the second semantic recognition result. And generating a control instruction based on the complete semantic recognition result, wherein the control instruction is used for instructing to start or stop the equipment corresponding to the position information (for example, turn on or off the air conditioner of the rear row area), or adjusting the working parameters of the equipment corresponding to the position information (for example, adjusting the air conditioner temperature of the rear row area, and setting the temperature value as the temperature value fed back by the user). And after receiving the control instruction sent by the voice control device, the controlled equipment starts or closes the equipment, or adjusts the working parameters of the equipment.
The voice control method provided by the embodiment of the invention receives the voice data, performs semantic recognition on the voice data to obtain a first semantic recognition result for indicating the name and the operation type of the equipment, judges whether the first semantic recognition result is complete, if the first semantic recognition result is incomplete, sends inquiry voice, receives response voice data of the inquiry voice, performs semantic recognition on the response voice data to obtain a second semantic recognition result, wherein the second semantic recognition result is used for indicating the position information and/or the working parameters of the name of the equipment, and generates a control instruction for indicating the on-off state or the working parameters of the corresponding equipment according to the first semantic recognition result and the second semantic recognition result. According to the method, the real intention of the user is obtained through a multi-round voice interaction process, the success rate of finishing user indication by the voice control equipment is improved, the fine control of controlled equipment in a scene is realized, and the user experience is improved.
It should be noted that the above embodiments are not limited to the control scenario of the smart car, and may also be applied to the smart home control scenario. It should be understood that in a smart home scenario, different activity areas (living room, bedroom, kitchen, bathroom, etc.) may be provided with the same or different types of smart devices (table lamp, television, fan, air conditioner, sound box, etc.), and when a user sends a control voice, the user needs to obtain the real intention of the user through a multi-round voice interaction process so as to control the corresponding smart devices to be turned on, turned off or adjusted in working parameters. For example, if the user sends voice data "turn on a speaker", and if the current scene includes a plurality of speakers, the speakers are respectively installed in a bedroom, a living room, and a bathroom, the voice control device can ask for which room's speaker to turn on by sending a query voice? The real intention of the user is obtained, and accurate equipment control is achieved.
Fig. 4 is a schematic structural diagram of a voice control apparatus according to an embodiment of the present invention. As shown in fig. 4, the voice control apparatus 200 provided in this embodiment includes:
a receiving module 201, configured to receive voice data; the processing module 202 is configured to perform semantic recognition on the voice data to obtain a first semantic recognition result, where the first semantic recognition result is used to indicate a device name and an operation type;
the processing module 202 is further configured to determine whether the first semantic identification result is complete, and if the first semantic identification result is incomplete; a sending module 203, configured to send an inquiry voice;
the receiving module 201 is further configured to receive response voice data of the query voice; the processing module is further used for performing semantic recognition on the response voice data to obtain a second semantic recognition result, and the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name;
the processing module 202 is further configured to generate a control instruction according to the first semantic recognition result and the second semantic recognition result, where the control instruction is used to instruct to control a switch state or a working parameter of a corresponding device.
In a possible implementation manner, before the receiving module 201 receives the voice data, the receiving module is further configured to: acquiring equipment configuration information, wherein the equipment configuration information comprises equipment names, operation types, equipment numbers and position information corresponding to the equipment names;
the apparatus 200 further comprises: a storage module 204, configured to store the device configuration information in a device configuration table under voice control;
the processing module 202 is specifically configured to determine whether the first semantic identification result is complete according to the device configuration table.
In a possible implementation manner, the processing module 202 is specifically configured to:
determining the number of devices corresponding to the device names by inquiring the device configuration table;
if the number of the devices is larger than or equal to two, determining that the first semantic recognition result is incomplete; a sending module 203, configured to send the query voice.
In a possible implementation manner, the processing module 202 is specifically configured to:
and starting or closing the equipment corresponding to the position information or adjusting the working parameters of the equipment corresponding to the position information according to the first semantic recognition result and the second semantic recognition result.
The voice control apparatus 200 provided in this embodiment may implement the technical solutions of any of the above method embodiments, and the implementation principles and technical effects are similar, which are not described herein again.
Fig. 5 is a hardware structure diagram of a voice control apparatus according to an embodiment of the present invention. As shown in fig. 5, the voice control apparatus provided in this embodiment includes:
a memory 501;
a processor 502; and
a computer program;
the computer program is stored in the memory 501 and configured to be executed by the processor 502 to implement the technical solution of any one of the above method embodiments, and the implementation principle and the technical effect are similar, which are not described herein again.
Optionally, the memory 501 may be separate or integrated with the processor 502.
When the memory 501 is a device independent of the processor 502, the voice control apparatus further includes: a bus 503 for connecting the memory 501 and the processor 502.
Embodiments of the present invention further provide a computer-readable storage medium, on which a computer program is stored, where the computer program is executed by a processor 502 to implement the steps performed by the voice control apparatus in any of the above method embodiments.
In the embodiments provided in the present invention, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described embodiments of the apparatus are merely illustrative, and for example, the division of the modules is only one logical division, and the actual implementation may have another division, for example, a plurality of modules may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or modules, and may be in an electrical, mechanical or other form.
Modules described as separate parts may or may not be physically separate, and parts displayed as modules may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment.
In addition, functional modules in the embodiments of the present invention may be integrated into one processing unit, or each module may exist alone physically, or two or more modules are integrated into one unit. The unit formed by the modules can be realized in a hardware form, and can also be realized in a form of hardware and a software functional unit.
The integrated module implemented in the form of a software functional module may be stored in a computer-readable storage medium. The software functional module is stored in a storage medium and includes several instructions to enable a computer device (which may be a personal computer, a server, or a network device) or a processor (processor) to execute some steps of the methods in the embodiments of the present application.
It should be understood that the Processor may be a Central Processing Unit (CPU), other general purpose Processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of a method disclosed in connection with the present invention may be embodied directly in a hardware processor, or in a combination of the hardware and software modules within the processor.
The memory may comprise a high-speed RAM memory, and may further comprise a non-volatile storage NVM, such as at least one disk memory, and may also be a usb disk, a removable hard disk, a read-only memory, a magnetic or optical disk, etc.
The bus may be an Industry Standard Architecture (ISA) bus, a Peripheral Component Interconnect (PCI) bus, an Extended ISA (EISA) bus, or the like. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, the buses in the figures of the present application are not limited to only one bus or one type of bus.
The storage medium may be implemented by any type or combination of volatile or non-volatile memory devices, such as Static Random Access Memory (SRAM), electrically erasable programmable read-only memory (EEPROM), erasable programmable read-only memory (EPROM), programmable read-only memory (PROM), read-only memory (ROM), magnetic memory, flash memory, magnetic or optical disks. A storage media may be any available media that can be accessed by a general purpose or special purpose computer.
An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium. Of course, the storage medium may also be integral to the processor. The processor and the storage medium may reside in an Application Specific Integrated Circuits (ASIC). Of course, the processor and the storage medium may reside as discrete components in a voice control apparatus.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solution of the present invention, and not to limit the same; while the invention has been described in detail and with reference to the foregoing embodiments, it will be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (10)

1. A voice control method, comprising:
receiving voice data, and performing semantic recognition on the voice data to obtain a first semantic recognition result, wherein the first semantic recognition result is used for indicating the name and the operation type of equipment;
judging whether the first semantic recognition result is complete or not, and if the first semantic recognition result is incomplete, sending inquiry voice;
receiving response voice data of the inquiry voice, and performing semantic recognition on the response voice data to obtain a second semantic recognition result, wherein the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name;
and generating a control instruction according to the first semantic recognition result and the second semantic recognition result, wherein the control instruction is used for indicating and controlling the switch state or the working parameter of the corresponding equipment.
2. The method of claim 1, wherein prior to receiving the voice data, further comprising:
acquiring equipment configuration information, wherein the equipment configuration information comprises equipment names, operation types, equipment numbers and position information corresponding to the equipment names;
storing the equipment configuration information into a voice-controlled equipment configuration table;
the judging whether the first semantic identification result is complete includes:
and judging whether the first semantic recognition result is complete according to the equipment configuration table.
3. The method of claim 1, wherein the determining whether the first semantic recognition result is complete and sending a query voice if the first semantic recognition result is incomplete comprises:
determining the number of devices corresponding to the device names by inquiring the device configuration table;
and if the number of the devices is more than or equal to two, determining that the first semantic recognition result is incomplete, and sending the query voice.
4. The method according to claim 1, wherein the controlling the switch state or the operating parameter of the corresponding device according to the first semantic recognition result and the second semantic recognition result comprises:
and starting or closing the equipment corresponding to the position information or adjusting the working parameters of the equipment corresponding to the position information according to the first semantic recognition result and the second semantic recognition result.
5. A voice control apparatus, comprising:
the receiving module is used for receiving voice data; the processing module is used for carrying out semantic recognition on the voice data to obtain a first semantic recognition result, and the first semantic recognition result is used for indicating the equipment name and the operation type;
the processing module is further configured to determine whether the first semantic identification result is complete, and if the first semantic identification result is incomplete; the sending module is used for sending inquiry voice;
the receiving module is further used for receiving response voice data of the inquiry voice; the processing module is further used for performing semantic recognition on the response voice data to obtain a second semantic recognition result, and the second semantic recognition result is used for indicating the position information and/or the working parameters of the equipment name;
the processing module is further configured to generate a control instruction according to the first semantic recognition result and the second semantic recognition result, where the control instruction is used to instruct and control a switch state or a working parameter of a corresponding device.
6. The apparatus of claim 5, wherein before the receiving module receives the voice data, the receiving module is further configured to: acquiring equipment configuration information, wherein the equipment configuration information comprises equipment names, operation types, equipment numbers and position information corresponding to the equipment names;
the device further comprises: the storage module is used for storing the equipment configuration information into an equipment configuration table controlled by voice;
the processing module is specifically configured to determine whether the first semantic identification result is complete according to the device configuration table.
7. The apparatus of claim 5, wherein the processing module is specifically configured to:
determining the number of devices corresponding to the device names by inquiring the device configuration table;
if the number of the devices is larger than or equal to two, determining that the first semantic recognition result is incomplete; and the sending module is used for sending the inquiry voice.
8. The apparatus of claim 5, wherein the processing module is specifically configured to:
and starting or closing the equipment corresponding to the position information or adjusting the working parameters of the equipment corresponding to the position information according to the first semantic recognition result and the second semantic recognition result.
9. A voice control apparatus, comprising:
a memory;
a processor; and
a computer program;
wherein the computer program is stored in the memory and configured to be executed by the processor to implement the speech control method of any of claims 1 to 4.
10. A computer-readable storage medium, on which a computer program is stored, the computer program being executed by a processor to implement the speech control method according to any one of claims 1 to 4.
CN201911400172.7A 2019-12-30 2019-12-30 Voice control method, device and storage medium Pending CN111128168A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911400172.7A CN111128168A (en) 2019-12-30 2019-12-30 Voice control method, device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911400172.7A CN111128168A (en) 2019-12-30 2019-12-30 Voice control method, device and storage medium

Publications (1)

Publication Number Publication Date
CN111128168A true CN111128168A (en) 2020-05-08

Family

ID=70505600

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911400172.7A Pending CN111128168A (en) 2019-12-30 2019-12-30 Voice control method, device and storage medium

Country Status (1)

Country Link
CN (1) CN111128168A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111627442A (en) * 2020-05-27 2020-09-04 星络智能科技有限公司 Speech recognition method, processor, system, computer equipment and readable storage medium
CN111818172A (en) * 2020-07-21 2020-10-23 海信视像科技股份有限公司 Method and device for controlling intelligent equipment by management server of Internet of things
CN113472806A (en) * 2021-07-14 2021-10-01 斑马网络技术有限公司 Voice interaction method, device, system, equipment and storage medium for protecting privacy
WO2022057740A1 (en) * 2020-09-18 2022-03-24 青岛海尔空调器有限总公司 Voice control method and voice control device for air conditioner
CN114582333A (en) * 2022-02-21 2022-06-03 中国第一汽车股份有限公司 Voice recognition method and device, electronic equipment and storage medium
CN115048161A (en) * 2021-02-26 2022-09-13 华为技术有限公司 Application control method, electronic device, apparatus, and medium
CN117037788A (en) * 2023-09-11 2023-11-10 南京申瑞电力电子有限公司 Control cabinet information display device based on voice control

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105161097A (en) * 2015-07-23 2015-12-16 百度在线网络技术(北京)有限公司 Voice interaction method and apparatus
CN105516520A (en) * 2016-02-04 2016-04-20 平安科技(深圳)有限公司 Interactive voice response device
CN105590626A (en) * 2015-12-29 2016-05-18 百度在线网络技术(北京)有限公司 Continuous speech man-machine interaction method and system
CN106205615A (en) * 2016-08-26 2016-12-07 王峥嵘 A kind of control method based on interactive voice and system
CN107612968A (en) * 2017-08-15 2018-01-19 北京小蓦机器人技术有限公司 The method, equipment and system of its connected device are controlled by intelligent terminal
CN107665706A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Rapid Speech exchange method and system
CN108320747A (en) * 2018-02-08 2018-07-24 广东美的厨房电器制造有限公司 Appliances equipment control method, equipment, terminal and computer readable storage medium
CN108831468A (en) * 2018-07-20 2018-11-16 英业达科技有限公司 Intelligent sound Control management system and its method
CN108881466A (en) * 2018-07-04 2018-11-23 百度在线网络技术(北京)有限公司 Exchange method and device
GB201818906D0 (en) * 2017-09-29 2019-01-02 Google Llc Processing voice commands based on device topology
US20190013018A1 (en) * 2017-07-05 2019-01-10 Comcast Cable Communications, Llc Methods And Systems For Using Voice To Control Multiple Devices
CN109637519A (en) * 2018-11-13 2019-04-16 百度在线网络技术(北京)有限公司 Interactive voice implementation method, device, computer equipment and storage medium
CN109920413A (en) * 2018-12-28 2019-06-21 广州索答信息科技有限公司 A kind of implementation method and storage medium of kitchen scene touch screen voice dialogue
CN110060669A (en) * 2018-01-19 2019-07-26 现代自动车株式会社 Speech recognition apparatus and its control method
CN110309345A (en) * 2018-02-07 2019-10-08 上海博泰悦臻网络技术服务有限公司 Online song search method and search engine based on vehicle-mounted radio

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105161097A (en) * 2015-07-23 2015-12-16 百度在线网络技术(北京)有限公司 Voice interaction method and apparatus
CN105590626A (en) * 2015-12-29 2016-05-18 百度在线网络技术(北京)有限公司 Continuous speech man-machine interaction method and system
CN105516520A (en) * 2016-02-04 2016-04-20 平安科技(深圳)有限公司 Interactive voice response device
CN107665706A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Rapid Speech exchange method and system
CN106205615A (en) * 2016-08-26 2016-12-07 王峥嵘 A kind of control method based on interactive voice and system
US20190013018A1 (en) * 2017-07-05 2019-01-10 Comcast Cable Communications, Llc Methods And Systems For Using Voice To Control Multiple Devices
CN107612968A (en) * 2017-08-15 2018-01-19 北京小蓦机器人技术有限公司 The method, equipment and system of its connected device are controlled by intelligent terminal
GB201818906D0 (en) * 2017-09-29 2019-01-02 Google Llc Processing voice commands based on device topology
CN110060669A (en) * 2018-01-19 2019-07-26 现代自动车株式会社 Speech recognition apparatus and its control method
CN110309345A (en) * 2018-02-07 2019-10-08 上海博泰悦臻网络技术服务有限公司 Online song search method and search engine based on vehicle-mounted radio
CN108320747A (en) * 2018-02-08 2018-07-24 广东美的厨房电器制造有限公司 Appliances equipment control method, equipment, terminal and computer readable storage medium
CN108881466A (en) * 2018-07-04 2018-11-23 百度在线网络技术(北京)有限公司 Exchange method and device
CN108831468A (en) * 2018-07-20 2018-11-16 英业达科技有限公司 Intelligent sound Control management system and its method
CN109637519A (en) * 2018-11-13 2019-04-16 百度在线网络技术(北京)有限公司 Interactive voice implementation method, device, computer equipment and storage medium
CN109920413A (en) * 2018-12-28 2019-06-21 广州索答信息科技有限公司 A kind of implementation method and storage medium of kitchen scene touch screen voice dialogue

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111627442A (en) * 2020-05-27 2020-09-04 星络智能科技有限公司 Speech recognition method, processor, system, computer equipment and readable storage medium
CN111818172A (en) * 2020-07-21 2020-10-23 海信视像科技股份有限公司 Method and device for controlling intelligent equipment by management server of Internet of things
CN111818172B (en) * 2020-07-21 2022-08-19 海信视像科技股份有限公司 Method and device for controlling intelligent equipment by management server of Internet of things
WO2022057740A1 (en) * 2020-09-18 2022-03-24 青岛海尔空调器有限总公司 Voice control method and voice control device for air conditioner
CN115048161A (en) * 2021-02-26 2022-09-13 华为技术有限公司 Application control method, electronic device, apparatus, and medium
CN113472806A (en) * 2021-07-14 2021-10-01 斑马网络技术有限公司 Voice interaction method, device, system, equipment and storage medium for protecting privacy
CN113472806B (en) * 2021-07-14 2022-11-22 斑马网络技术有限公司 Voice interaction method, device, system, equipment and storage medium for protecting privacy
CN114582333A (en) * 2022-02-21 2022-06-03 中国第一汽车股份有限公司 Voice recognition method and device, electronic equipment and storage medium
CN117037788A (en) * 2023-09-11 2023-11-10 南京申瑞电力电子有限公司 Control cabinet information display device based on voice control

Similar Documents

Publication Publication Date Title
CN111128168A (en) Voice control method, device and storage medium
CN109273006B (en) Voice control method of vehicle-mounted system, vehicle and storage medium
CN105204357A (en) Contextual model regulating method and device for intelligent household equipment
CN110851221A (en) Smart home scene configuration method and device
CN113401134A (en) Contextual model self-defining method and device, electronic equipment and storage medium
CN109210703B (en) Voice control method of air conditioner and voice-controlled air conditioner
CN112201233A (en) Voice control method, system and device of intelligent household equipment and computer storage medium
CN111968611B (en) Karaoke method, vehicle-mounted terminal and computer readable storage medium
CN110570850A (en) Voice control method, device, computer equipment and storage medium
CN112311635B (en) Voice interruption awakening method and device and computer readable storage medium
CN110782891A (en) Audio processing method and device, computing equipment and storage medium
CN110553361A (en) Air conditioner control method and device, storage medium and electrical equipment
CN111261163A (en) Voice control method and system and intelligent air conditioner
CN110657561B (en) Air conditioner and voice instruction recognition method, control device and readable storage medium thereof
CN111429917A (en) Equipment awakening method and terminal equipment
CN114724558A (en) Method and device for voice control of air conditioner, air conditioner and storage medium
CN113614713A (en) Human-computer interaction method, device, equipment and vehicle
CN112133296B (en) Full duplex voice control method and device, storage medium and voice equipment
CN112152890B (en) Control system and method based on intelligent sound box
CN111023490A (en) Air conditioner control method based on face recognition and air conditioner
CN115503639A (en) Voice processing method, voice interaction method, server and storage medium
CN115479370A (en) Air conditioner control method, device and equipment and air conditioner
KR102386040B1 (en) A method, apparatus and computer readable storage medium having instructions for processing voice input, a vehicle having a voice processing function, and a user terminal
CN116105307A (en) Air conditioner control method, device, electronic equipment and storage medium
CN110853633A (en) Awakening method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200508

RJ01 Rejection of invention patent application after publication