CN112700770A - Voice control method, sound box device, computing device and storage medium - Google Patents

Voice control method, sound box device, computing device and storage medium Download PDF

Info

Publication number
CN112700770A
CN112700770A CN201911011486.8A CN201911011486A CN112700770A CN 112700770 A CN112700770 A CN 112700770A CN 201911011486 A CN201911011486 A CN 201911011486A CN 112700770 A CN112700770 A CN 112700770A
Authority
CN
China
Prior art keywords
user
personalized
equipment
control
controlled
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911011486.8A
Other languages
Chinese (zh)
Inventor
聂松洋
韩陆
李星星
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Alibaba Group Holding Ltd
Original Assignee
Alibaba Group Holding Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba Group Holding Ltd filed Critical Alibaba Group Holding Ltd
Priority to CN201911011486.8A priority Critical patent/CN112700770A/en
Publication of CN112700770A publication Critical patent/CN112700770A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/907Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/909Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9535Search customisation based on user profiles and personalisation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The embodiment of the invention discloses a voice control method, a sound box device, a computing device and a storage medium, wherein the method comprises the following steps: acquiring a user voice signal; acquiring user personalized control information for controlling a device to be controlled; and processing the voice signal and obtaining a corresponding equipment control request at least based on the user personalized control information, wherein the control request is used for controlling the target equipment to be controlled to execute equipment control operation. The embodiment of the invention can realize personalized voice control.

Description

Voice control method, sound box device, computing device and storage medium
Technical Field
The present invention relates to the field of language control technologies, and in particular, to a voice control method, a speaker device, a computing device, and a storage medium.
Background
With the development of the information-oriented society and the progress of the information technology, the speech recognition technology has undergone the development process from inexhaustible to mature, and the existing speech recognition technology can meet the basic requirements of people on speech recognition in both recognition speed and recognition accuracy.
The related speech recognition technology can only recognize words stored in a word bank, and can not normally analyze different speech habits among different users. Therefore, it is difficult for the related art to recognize personalized voices of different users, and further, personalized voice control cannot be realized based on the recognized voices.
Disclosure of Invention
The embodiment of the invention provides a voice control method, a sound box device, a computing device and a storage medium, which are used for solving the problem that personalized voice control cannot be realized.
According to a first aspect of embodiments of the present invention, there is provided a voice control method, which may include:
acquiring a user voice signal;
acquiring user personalized control information for controlling a device to be controlled;
and processing the voice signal and obtaining a corresponding equipment control request at least based on the user personalized control information, wherein the control request is used for controlling the target equipment to be controlled to execute equipment control operation.
According to a second aspect of embodiments of the present invention, there is provided a navigation method, which may include:
acquiring a user voice signal;
acquiring personalized information of a geographical position to be positioned;
and processing the voice signal and obtaining a corresponding geographic position at least based on the personalized information of the geographic position to be positioned.
According to a third aspect of embodiments of the present invention, there is provided an acoustic enclosure apparatus, which may include:
the acquisition module acquires a user voice signal;
the acquisition module is also used for acquiring user personalized control information for controlling the equipment to be controlled;
and the processing module is used for processing the voice signal and obtaining a corresponding equipment control request at least based on the user personalized control information, wherein the equipment control request is used for controlling the target equipment to be controlled to execute equipment control operation.
According to a fourth aspect of the embodiments of the present invention, there is provided an acoustic enclosure apparatus including: a memory for storing program instructions;
a processor for performing the method of the first aspect in accordance with program instructions.
According to a fifth aspect of embodiments of the present invention, there is provided a computing device, comprising: at least one processor, at least one memory, and at least one computer program, wherein the computer program is stored in the memory, the computer program comprising instructions, the processor being configured to execute the computer program to cause a computing device to implement the method as shown in the first aspect or the method as shown in the second aspect.
According to a sixth aspect of embodiments of the present invention, there is provided a computer-readable storage medium having stored therein instructions which, when run on a computer, cause the computer to perform the voice control method of the above-described aspects or the method shown in the second aspect.
According to the voice control method disclosed in the embodiment of the invention, the user voice signal is processed through the user personalized control information for controlling the device to be controlled, and the corresponding device control request is obtained, wherein the device control request is used for controlling the target device to be controlled to execute the device control operation. The embodiment of the invention can realize personalized voice control and increase the voice recognition rate according to the personalized voice.
Drawings
The present invention will be better understood from the following description of specific embodiments thereof taken in conjunction with the accompanying drawings, in which like or similar reference characters designate like or similar features.
Fig. 1 is a schematic view of an application scenario of a voice control method according to an embodiment of the present invention;
FIG. 2 is a flow chart of a voice control method according to an embodiment of the present invention;
FIG. 3 is a schematic diagram of a voice control system according to an embodiment of the present invention;
FIG. 4 is a flow chart of another speech control method according to an embodiment of the present invention;
FIG. 5 is a diagram illustrating a navigation method according to an embodiment of the present invention;
FIG. 6 is a diagram of a voice-controlled device according to an embodiment of the present invention;
fig. 7 is a schematic view of an audio device according to an embodiment of the present invention;
FIG. 8 is a diagram of a navigation device according to an embodiment of the present invention;
FIG. 9 is a schematic block diagram of a voice control apparatus of an embodiment of the present invention;
fig. 10 is a schematic block diagram of a sound box apparatus according to an embodiment of the present invention.
Detailed Description
Features and exemplary embodiments of various aspects of the present invention will be described in detail below, and in order to make objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not to be construed as limiting the invention. It will be apparent to one skilled in the art that the present invention may be practiced without some of these specific details. The following description of the embodiments is merely intended to provide a better understanding of the present invention by illustrating examples of the present invention.
It is noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
Fig. 1 is a schematic view of an application scenario of a voice control method according to an embodiment of the present invention. As shown in FIG. 1, in a scenario, a room is primarily divided by function into multiple areas (e.g., bedroom, kitchen, etc.; or, alternatively, office, conference room, etc.); in the above-described areas, a voice control device and a device to be controlled (for example, a lamp, a projector, or the like) are provided. Taking area 1 as a conference room and area 2 as an office as an example, area 1 includes voice control device 11, lamp 12, and projector 13; the area 2 includes a voice control device 21 and a lamp 22. In the area 1, a user can send a voice signal to the voice control device 11, and then can control the lamp 12 and the projector 13 in the area 1; or, the device to be controlled in the area is operated by an Application (APP) on the user terminal. Only the operation of the device to be controlled in the above-described area by voice is illustrated in fig. 1.
It should be noted that the voice control device may be a sound box device.
The embodiments provided in the present specification will be described in detail below with reference to the accompanying drawings.
Fig. 2 is a flowchart of a voice control method according to an embodiment of the present invention. As shown in fig. 2, the voice control method may include:
s210: acquiring a user voice signal;
s220: acquiring user personalized control information for controlling a device to be controlled;
s230: and processing the voice signal and obtaining a corresponding equipment control request at least based on the user individual control information, wherein the equipment control request is used for controlling the target equipment to be controlled to execute equipment control operation.
According to the voice control method disclosed in the embodiment of the invention, the user voice signal is processed through the user personalized control information for controlling the device to be controlled, and the corresponding device control request is obtained, wherein the device control request is used for controlling the target device to be controlled to execute the device control operation. The embodiment of the invention can realize personalized voice control and increase the voice recognition rate according to the personalized voice.
In the embodiment of the present invention, S210 includes:
in fig. 3, the voice control apparatus 31 receives a user voice signal from the user 10.
The voice control device 31 may be a voice acquiring device having a microphone, a speaker, and the like. In fig. 3, only a microphone is taken as an example.
In an embodiment of the present invention, the voice control device 31 may wake up the voice control device 31 when the user 10 needs to operate the device 33 (e.g., a lamp, a television, a projector, etc.); on one hand, resources can be saved, on the other hand, unnecessary operation can be avoided, and user experience is improved.
In the embodiment of the present invention, S220 includes:
and acquiring user personalized control information for controlling the equipment to be controlled from the target server.
In fig. 3, user personalization control information is obtained from the target server 32. The user personalized control information obtained from the target server 32 is only taken as an example in fig. 3.
It should be noted that the voice control device 31 and the target server 32 in the voice control system 30 are both in an online state.
In the embodiment of the present invention, S220 includes:
and acquiring the user personalized control information from the local.
It should be noted that, the user personalized control information is obtained from the local (for example, database DB); in one case, if the user personalized control information is not acquired locally, the user personalized control information may be acquired from the target server.
Since the speed of acquiring the user personalized control information from the local is generally shorter than the time of acquiring the user personalized control information from the target service end, the user personalized control information stored in the target service end can be synchronized to the local of the voice control device through the interface after being modified. The user can modify the user personalized control information through an Application (APP) on the user terminal.
It should be noted that, in consideration that the user personalized control information may relate to sensitive information, in this case, the security of obtaining the user personalized control information from the target server is better than that of obtaining the user personalized control information locally. The target server is generally a generator of the target device to be controlled, and when the user sets that the user personalized control information does not include sensitive information, the user personalized control information can be synchronized to the local after being modified, so that the user experience degree is improved conveniently.
In the embodiment of the present invention, S230 includes:
the first step is as follows: recognizing a user voice signal to obtain a first text;
the second step is that: and analyzing the first text at least based on the user personalized control information, and obtaining the equipment control request according to the analyzed first text.
In an embodiment of the present invention, the first step comprises:
and identifying the voice information of the user based on the user personalized control information to obtain a first text.
The method for recognizing the user voice information based on the user personalized control information to obtain the first text comprises the following steps:
recognizing the voice signal of the user to obtain word segmentation; and matching the word segmentation with the user personalized control information to obtain a first text.
In an embodiment of the present invention, the second step includes:
analyzing the first text, and matching the analyzed first text with the user personalized control information to obtain target equipment to be controlled matched with the analyzed first text;
and obtaining an equipment control request according to the matched target equipment to be controlled and the equipment control operation in the analyzed first text.
Specifically, analyzing the target device to be controlled based on Natural Language Understanding (Nlu) in a Dialog box (Dialog), and matching the analyzed target device to be controlled with user personalized control information to obtain a first text matching target device to be controlled; and obtaining a device control request based on the target device to be controlled and the device control operation in the parsed first text by using the Dialog Script function engine.
In an embodiment of the present invention, the user personalized control information includes at least one of the following items: the method comprises the following steps of personalized naming of the equipment to be controlled, personalized position information of the equipment to be controlled and a personalized control instruction.
The name of the device to be controlled can be defined by the user, such as "Xiaoming", or can be a name inherent to the device, such as "Lamp".
The personalized position information of the device to be controlled can be defined by the user, such as 'daughter' or 'lying down'.
The personalized control command can be defined by the user, such as 'long press', or can be turned on or off.
In one example, the user personalized control information is "long press" corresponds to "on", "cool dazzle" corresponds to "lying down", "button" corresponds to "lamp ID 11111", and the user voice signal is "long press cool dazzle button". For convenience of description, turning on the light is exemplified hereinafter. In addition, only one user personalized control information is listed here, and the user can set the control information according to the use requirement of the user, so that the setting is avoided being too single.
The first step comprises:
the voice signal of the user is recognized to obtain 'long press of a cool and dazzling button'.
The second step comprises:
according to the conditions that the long press corresponds to the on state, the cool dazzle corresponds to the sub-horizontal state, and the button corresponds to the lamp ID 11111, the long press cool dazzle button is analyzed and matched with the personalized control information of the user, the intention (intent) is obtained to be the on state, the slot (slot) is obtained to be the sub-horizontal state, the zone (domain) corresponding to the lamp ID 11111 is obtained to be the sub-horizontal state, and the control request, namely the sub-horizontal middle lamp (lamp ID 11111) is generated based on the domain + intent + slot.
It should be noted that, if the user voice signal includes the location information of the target device to be controlled, in the process of analyzing the user voice signal, an area (domain) obtained by analyzing the user voice signal is the location information of the target device to be controlled; in the case where the region (domain) is default in the user voice signal, the region (domain) may be supplemented based on the positional information where the voice control apparatus 31 is located, thereby realizing voice control.
In the embodiment of the present invention, the target device to be controlled is located in the same area as the voice control device 31, for example, the same room or the same office area.
In one example, when the number of the regions is one, as long as "intent + slot" is analyzed, the "domain" can be supplemented by the location information of the voice control device 31, and the voice control can be realized.
In addition, for a larger area, especially when there are more target devices to be controlled of the same type, it is considered to set a plurality of voice control devices 31 to realize distributed control.
In one example, when the number of the regions is at least two, if only "intent + slot" is analyzed, the voice control can be realized by supplementing the position information of the voice control device 31 receiving the voice signal of the user; if the 'domain + intent + slot' is analyzed, the voice control can be realized.
In the embodiment of the present invention, the target device to be controlled is not located in the same area as the voice control device 31.
In one example, a room includes a plurality of rooms, and a voice control device 31 is provided in the room.
In the embodiment of the invention, the user personalized control information can be corresponding to a certain user ID; or corresponding to a plurality of user IDs; the second is that the user-customized control information corresponds to the entire house and can be called based on the user ID of each room. The user ID may be a mobile phone number of the user or an account number set by another user.
In the embodiment of the present invention, after S230, the voice control method further includes:
and executing the equipment control operation by the equipment to be controlled based on the equipment control request control target.
The voice control device 31 receives a user voice signal sent by the user 10, processes the voice signal and obtains a corresponding device control request at least based on the acquired user personalized control information, and controls the device to be controlled to execute the device control operation according to the device control request. For example, the device control request is "please turn on the lamp in the bedroom", and the lamp in the bedroom is controlled to turn on by the voice control device 31.
In the embodiment of the present invention, after S230, the voice control method further includes:
and sending the equipment control request to the target server to instruct the target server to control the target equipment to be controlled to execute the equipment control operation based on the equipment control request.
In fig. 3, the voice control device 31 receives a user voice signal sent by the user 10, processes the voice signal and obtains a corresponding device control request based on at least the acquired user personalized control information, and sends the device control request to the target server 32 to instruct the target server 32 to control the target device to be controlled 33 to perform device control operation based on the device control request. For example, the device control request is "please turn on the lamp in the bedroom", the voice control device 31 sends "please turn on the lamp in the bedroom" to the target server 32, and the target server 32 controls the lamp to be turned on.
In addition, after the lamp 33 is turned on, the processing result returned by the target server 32 is received and sequentially transmitted to the voice control device 31, displayed on the control panel thereof by the voice control device 31 and/or voice-broadcasted by the voice control device 31.
In the embodiment of the present invention, the voice control device may be a sound box device, and at this time, the sound box device executes the step corresponding to the voice control method; the voice control device can also be a cloud terminal corresponding to the sound box device, at the moment, the sound box device acquires voice data, the voice data is uploaded to the cloud terminal, and the cloud terminal executes the steps corresponding to the subsequent voice control method.
Fig. 4 is a flowchart of another speech control method according to an embodiment of the present invention. As shown in fig. 4, the voice control may include:
s410: acquiring a user voice signal;
s420: acquiring user personalized control information corresponding to the user ID from a target server according to the user ID;
wherein the user-customized control information includes at least one of: the method comprises the following steps of (1) carrying out personalized naming on equipment to be controlled, personalized position information of the equipment to be controlled and a personalized control instruction;
s430: recognizing the user voice signal based on ASR, and matching with the user personalized control information to obtain a text;
s440: analyzing the Natural Language Understanding (Nlu) in a Dialog box (Dialog), and matching the analyzed Natural Language Understanding with the personalized control information of the user to obtain target equipment to be controlled with matched text; and a Dialog Script function engine (Dialog Script function engine) obtains an equipment control request based on the target equipment to be controlled and the equipment control request in the analyzed text, and sends the equipment control request to the target server, and the target server controls the target equipment to be controlled to execute equipment control operation;
s450: and sending the equipment control request to the target server to instruct the target server to control the target equipment to be controlled to execute the equipment control operation based on the equipment control request.
After the target server executes and receives the control request, returning a response message to the Dialog Script, sequentially transmitting the response message to the sound box equipment by the Dialog Script, and displaying or voice broadcasting by the sound box equipment; wherein the notification message may be "good".
In one example, the user personalized control information is "long press" corresponds to "on", "cool dazzle" corresponds to "lying down", "button" corresponds to "lamp ID 11111", and the user voice signal is "long press cool dazzle button". For convenience of description, turning on the light is exemplified hereinafter. In addition, only one user personalized control information is listed here, and the user can set the control information according to the use requirement of the user, so that the setting is avoided being too single.
The first step comprises:
the voice signal of the user is recognized to obtain 'long press of a cool and dazzling button'.
The second step comprises:
according to the conditions that the long press corresponds to the on state, the cool dazzle corresponds to the sub-horizontal state, and the button corresponds to the lamp ID 11111, the long press cool dazzle button is analyzed and matched with the personalized control information of the user, the intention (intent) is obtained to be the on state, the slot (slot) is obtained to be the sub-horizontal state, the zone (domain) corresponding to the lamp ID 11111 is obtained to be the sub-horizontal state, and the control request, namely the sub-horizontal middle lamp (lamp ID 11111) is generated based on the domain + intent + slot.
The third step comprises:
a control request "turn on a lamp in the next bed (lamp ID is 11111)" is sent to the target server to instruct the target server to control the turn-on ID 11111. Each device to be controlled has an inherent ID at the target server, so that the ID of the device to be controlled can be found, and the control of the device to be controlled can be realized.
Fig. 5 is a flowchart of a navigation method according to an embodiment of the present invention. As shown in fig. 5, the navigation method includes:
s510: acquiring a user voice signal;
s520: acquiring personalized information of a geographical position to be positioned;
the personalized information of the geographic position to be located is user-defined information related to the position, such as a small flower, which corresponds to a sub-lying position.
S530: and processing the voice signal and obtaining a corresponding geographic position at least based on the personalized information of the geographic position to be positioned.
According to the navigation method disclosed by the embodiment of the invention, the voice signal of the user is processed through the personalized information of the geographical position to be positioned, and the corresponding geographical position is obtained. The embodiment of the invention can realize personalized voice positioning, and meanwhile, the voice recognition rate can be increased according to the personalized voice.
Fig. 6 is a schematic diagram of a voice control apparatus according to an embodiment of the present invention. As shown in fig. 6, the voice control apparatus 600 includes:
an obtaining module 610, configured to obtain a user voice signal;
the obtaining module 610 is further configured to obtain user personalized control information for controlling the device to be controlled;
and the processing module 620 is configured to process the voice signal and obtain a corresponding device control request at least based on the user-customized control information, where the control request is used to control the target device to be controlled to perform a device control operation.
The voice control device disclosed in the embodiment of the invention processes the user voice signal and obtains a corresponding device control request through the user personalized control information for controlling the device to be controlled, wherein the device control request is used for controlling the target device to be controlled to execute the device control operation. The embodiment of the invention can realize personalized voice control and increase the voice recognition rate according to the personalized voice.
Optionally, the voice control apparatus further includes:
and the execution module is used for controlling the target equipment to be controlled to execute the equipment control operation according to the equipment control request.
Optionally, the voice control apparatus further includes:
and the sending module is used for sending the equipment control request to the target server so as to instruct the target server to control the target equipment to be controlled to execute the equipment control operation based on the equipment control request.
Optionally, the processing module 620 includes:
the recognition unit is used for recognizing a user voice signal to obtain a first text;
and the analysis unit is used for analyzing the first text at least based on the user personalized control information and obtaining the equipment control request according to the analyzed first text.
Optionally, the identification unit is further configured to:
and identifying the voice information of the user based on the user personalized control information to obtain a first text.
Optionally, the parsing unit is further configured to:
analyzing the first text, and matching the analyzed first text with the user personalized control information to obtain target equipment to be controlled matched with the analyzed first text;
and obtaining an equipment control request according to the matched target equipment to be controlled and the equipment control operation in the analyzed first text.
Wherein the user personalized control information comprises at least one of the following items: personalized naming of equipment to be controlled, personalized position information of equipment to be controlled and personalized control instruction
Optionally, the obtaining module 610 is further configured to:
and acquiring user personalized control information from the target server or the local storage equipment according to the user ID.
Optionally, the voice control apparatus further includes:
and the synchronization module is used for synchronizing the user personalized control information in the target server to the local storage equipment.
Fig. 7 is a schematic diagram of a sound box apparatus according to an embodiment of the present invention. As shown in fig. 7, the speaker device 700 includes:
an obtaining module 710, configured to obtain a user voice signal;
the obtaining module 710 is further configured to obtain user personalized control information for controlling the device to be controlled;
and the processing module 720 is configured to process the voice signal and obtain a corresponding device control request at least based on the user personalized control information, where the control request is used to control the target device to be controlled to perform a device control operation.
According to the loudspeaker box device disclosed by the embodiment of the invention, the user voice signal is processed through the user personalized control information for controlling the device to be controlled, and the corresponding device control request is obtained, wherein the device control request is used for controlling the target device to be controlled to execute the device control operation. The embodiment of the invention can realize personalized voice control and increase the voice recognition rate according to the personalized voice.
Optionally, the speaker device further includes:
and the execution module is used for controlling the target equipment to be controlled to execute the equipment control operation according to the equipment control request.
Optionally, the speaker device further includes:
and the sending module is used for sending the equipment control request to the target server so as to instruct the target server to control the target equipment to be controlled to execute the equipment control operation based on the equipment control request.
Optionally, the processing module 720 includes:
the recognition unit is used for recognizing a user voice signal to obtain a first text;
and the analysis unit is used for analyzing the first text at least based on the user personalized control information and obtaining the equipment control request according to the analyzed first text.
Optionally, the identification unit is further configured to:
and identifying the voice information of the user based on the user personalized control information to obtain a first text.
Optionally, the parsing unit is further configured to:
analyzing the first text, and matching the analyzed first text with the user personalized control information to obtain target equipment to be controlled matched with the analyzed first text;
and obtaining an equipment control request according to the matched target equipment to be controlled and the equipment control operation in the analyzed first text.
Wherein the user personalized control information comprises at least one of the following items: personalized naming of equipment to be controlled, personalized position information of equipment to be controlled and personalized control instruction
Optionally, the obtaining module 710 is further configured to:
and acquiring user personalized control information from the target server or the local storage equipment according to the user ID.
Optionally, the speaker device further includes:
and the synchronization module is used for synchronizing the user personalized control information in the target server to the local storage equipment.
FIG. 8 is a diagram of a navigation device according to an embodiment of the invention. As shown in fig. 8, the navigation device 800 includes:
an obtaining module 810, configured to obtain a user voice signal;
the obtaining module 810 is further configured to obtain personalized information of the geographic location to be located;
the personalized information of the geographic position to be located is user-defined information related to the position, such as a small flower, which corresponds to a sub-lying position.
The processing module 820 is further configured to process the voice signal and obtain a corresponding geographic location based on at least the personalized information of the geographic location to be located.
According to the navigation method disclosed by the embodiment of the invention, the voice signal of the user is processed through the personalized information of the geographical position to be positioned, and the corresponding geographical position is obtained. The embodiment of the invention can realize personalized voice positioning, and meanwhile, the voice recognition rate can be increased according to the personalized voice.
Fig. 9 is a schematic block diagram of a voice control device according to an embodiment of the present invention, and is a block diagram of an exemplary hardware architecture of a computing device capable of implementing the voice control method and apparatus provided by the embodiment of the present invention.
Computing device 900 includes input device 901, input interface 902, central processor 903, memory 904, output interface 905, and output device 906. The input interface 902, the central processing unit 903, the memory 904, and the output interface 905 are connected to each other via a bus 910, and the input device 901 and the output device 906 are connected to the bus 910 via the input interface 902 and the output interface 906, respectively, and further connected to other components of the computing device 900.
Specifically, the input device 901 receives input information from the outside, and transmits the input information to the central processor 903 through the input interface 902; central processor 903 processes input information based on computer-executable instructions stored in memory 904 to generate output information, stores the output information temporarily or permanently in memory 904, and then transmits the output information to output device 906 via output interface 905; output device 906 outputs the output information external to computing device 900 for use by a user.
The computing device shown in fig. 9 may also be implemented to include: a memory storing computer-executable instructions; and a processor which, when executing computer executable instructions, may implement the voice control method described in connection with fig. 2, 4, or the navigation method described in fig. 5.
Fig. 10 is a schematic block diagram of a sound box device according to an embodiment of the present invention, and is a structural diagram of an exemplary hardware architecture of a computing device capable of implementing a voice control method and apparatus provided by an example of the present invention.
The audio equipment 1000 includes an input device 1001, an input interface 1002, a central processing unit 1003, a memory 1004, an output interface 1005, and an output device 1006. The input interface 1002, the central processing unit 1003, the memory 1004, and the output interface 1005 are connected to each other through a bus 1010, and the input device 1001 and the output device 1006 are connected to the bus 1010 through the input interface 1002 and the output interface 1006, respectively, and further connected to other components of the sound box device 1000.
Specifically, the input device 1001 receives input information from the outside, and transmits the input information to the central processor 1003 via the input interface 1002; the central processor 1003 processes input information based on computer-executable instructions stored in the memory 1004 to generate output information, stores the output information temporarily or permanently in the memory 1004, and then transmits the output information to the output device 1006 through the output interface 1005; the output device 1006 outputs the output information to the outside of the speaker device 1000 for use by the user.
The computing device shown in FIG. 10 may also be implemented to include: a memory storing computer-executable instructions; and a processor which, when executing computer executable instructions, may implement the voice control method described in connection with fig. 2, 4, or the navigation method described in fig. 5.
According to an embodiment of the invention, the process described above with reference to the flow chart may be implemented as a computer software program. For example, an example of the invention includes a computer program product comprising a computer program tangibly embodied on a machine-readable medium, the computer program comprising program code for performing the method illustrated in the flow chart. In such an embodiment, the computer program may be downloaded and installed from a network, and/or installed from a removable storage medium.
In the above embodiments, the implementation may be wholly or partially realized by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product comprises one or more computer instructions which, when run on a computer, cause the computer to perform the methods described in the various embodiments above. The procedures or functions according to the embodiments of the invention are brought about in whole or in part when the computer program instructions are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center by wire (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wirelessly (e.g., infrared, wireless, microwave, etc.). The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., solid state disk), among others.
The above-described examples of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the example scheme. One of ordinary skill in the art can understand and implement it without inventive effort.
It is to be understood that the invention is not limited to the particular arrangements and instrumentality described in the above embodiments and shown in the drawings. For convenience and brevity of description, detailed description of a known method is omitted here, and for the specific working processes of the system, the module and the unit described above, reference may be made to corresponding processes in the foregoing method embodiments, which are not described herein again.
It will be apparent to those skilled in the art that the method procedures of the present invention are not limited to the specific steps described and illustrated, and that various changes, modifications and additions, or equivalent substitutions and changes in the sequence of steps within the technical scope of the present invention are possible within the technical scope of the present invention as those skilled in the art can appreciate the spirit of the present invention.

Claims (14)

1. A method of voice control, comprising:
acquiring a user voice signal;
acquiring user personalized control information for controlling a device to be controlled;
and processing the voice signal and obtaining a corresponding equipment control request at least based on the user personalized control information, wherein the control request is used for controlling target equipment to be controlled to execute equipment control operation.
2. The method of claim 1, wherein the method further comprises:
and controlling the target equipment to be controlled to execute equipment control operation according to the equipment control request.
3. The method of claim 1, wherein the method further comprises:
and sending the equipment control request to a target server to indicate the target server to control the target equipment to be controlled to execute equipment control operation based on the equipment control request.
4. The method of claim 2 or 3, wherein said processing the voice signal and deriving a corresponding control request based at least on the user-personalized control information comprises:
recognizing the user voice signal to obtain a first text;
and analyzing the first text at least based on the user personalized control information, and obtaining the equipment control request according to the analyzed first text.
5. The method of claim 4, wherein said recognizing the user speech signal resulting in a first text comprises:
and identifying the user voice information based on the user personalized control information to obtain the first text.
6. The method of claim 4, wherein parsing the first text based at least on the user-customized control information and deriving the device control request from the parsed first text comprises:
analyzing the first text, and matching the analyzed first text with the user personalized control information to obtain target equipment to be controlled matched with the analyzed first text;
and obtaining the equipment control request according to the matched target equipment to be controlled and the equipment control operation in the analyzed first text.
7. The method of claim 1, wherein the user-personalized control information comprises at least one of: the method comprises the following steps of personalized naming of the equipment to be controlled, personalized position information of the equipment to be controlled and a personalized control instruction.
8. The method of claim 1, wherein the obtaining user-personalized control information for controlling a device to be controlled comprises:
and acquiring the user personalized control information from a target server or local storage equipment according to the user ID.
9. The method of claim 8, wherein prior to retrieving the user-personalized control information from a local storage device, the method further comprises:
and synchronizing the user personalized control information in the target server to the local storage device.
10. A method of navigation, the method comprising:
acquiring a user voice signal;
acquiring personalized information of a geographical position to be positioned;
and processing the voice signal and obtaining a corresponding geographic position at least based on the personalized information of the geographic position to be positioned.
11. An acoustic enclosure apparatus, comprising:
the acquisition module acquires a user voice signal;
the acquisition module is also used for acquiring user personalized control information for controlling the equipment to be controlled;
and the processing module is used for processing the voice signal and obtaining a corresponding equipment control request at least based on the user personalized control information, wherein the control request is used for controlling target equipment to be controlled to execute equipment control operation.
12. An acoustic enclosure apparatus, comprising:
a memory for storing program instructions;
a processor for performing the method according to any one of claims 1-9 according to the program instructions.
13. A computing device, the computing device comprising: at least one processor, at least one memory, and at least one computer program, wherein the computer program is stored in the memory, the computer program comprising instructions for executing the computer program by the processor to cause the computing device to implement the method of any one of claims 1-9 or the method of claim 10.
14. A computer-readable storage medium comprising instructions that, when executed on a computer, cause the computer to perform the method of any of claims 1-9 or the method of claim 10.
CN201911011486.8A 2019-10-23 2019-10-23 Voice control method, sound box device, computing device and storage medium Pending CN112700770A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911011486.8A CN112700770A (en) 2019-10-23 2019-10-23 Voice control method, sound box device, computing device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911011486.8A CN112700770A (en) 2019-10-23 2019-10-23 Voice control method, sound box device, computing device and storage medium

Publications (1)

Publication Number Publication Date
CN112700770A true CN112700770A (en) 2021-04-23

Family

ID=75505070

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911011486.8A Pending CN112700770A (en) 2019-10-23 2019-10-23 Voice control method, sound box device, computing device and storage medium

Country Status (1)

Country Link
CN (1) CN112700770A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113380245A (en) * 2021-06-01 2021-09-10 青岛海尔空调器有限总公司 Information processing method, device and system
CN115529530A (en) * 2022-11-24 2022-12-27 深圳市百泰实业股份有限公司 Intelligent control method for internal functions of multifunctional mobile power supply sound box

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107612968A (en) * 2017-08-15 2018-01-19 北京小蓦机器人技术有限公司 The method, equipment and system of its connected device are controlled by intelligent terminal
CN108337139A (en) * 2018-01-29 2018-07-27 广州索答信息科技有限公司 Home appliance voice control method, electronic equipment, storage medium and system
CN110246499A (en) * 2019-08-06 2019-09-17 苏州思必驰信息科技有限公司 The sound control method and device of home equipment

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107612968A (en) * 2017-08-15 2018-01-19 北京小蓦机器人技术有限公司 The method, equipment and system of its connected device are controlled by intelligent terminal
CN108337139A (en) * 2018-01-29 2018-07-27 广州索答信息科技有限公司 Home appliance voice control method, electronic equipment, storage medium and system
CN110246499A (en) * 2019-08-06 2019-09-17 苏州思必驰信息科技有限公司 The sound control method and device of home equipment

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113380245A (en) * 2021-06-01 2021-09-10 青岛海尔空调器有限总公司 Information processing method, device and system
CN115529530A (en) * 2022-11-24 2022-12-27 深圳市百泰实业股份有限公司 Intelligent control method for internal functions of multifunctional mobile power supply sound box

Similar Documents

Publication Publication Date Title
US11626117B2 (en) Contingent device actions during loss of network connectivity
CN108683574B (en) Equipment control method, server and intelligent home system
US10079014B2 (en) Name recognition system
CN109658932B (en) Equipment control method, device, equipment and medium
CN108831469B (en) Voice command customizing method, device and equipment and computer storage medium
CN107112014B (en) Application focus in speech-based systems
KR102429436B1 (en) Server for seleting a target device according to a voice input, and controlling the selected target device, and method for operating the same
KR101954550B1 (en) Volume adjustment method, system and equipment, and computer storage medium
CN108133707B (en) Content sharing method and system
JP6783339B2 (en) Methods and devices for processing audio
US10388277B1 (en) Allocation of local and remote resources for speech processing
CN111880645A (en) Server for determining and controlling target device based on voice input of user and operating method thereof
KR20170032096A (en) Electronic Device, Driving Methdo of Electronic Device, Voice Recognition Apparatus, Driving Method of Voice Recognition Apparatus, and Computer Readable Recording Medium
US11373645B1 (en) Updating personalized data on a speech interface device
US11238860B2 (en) Method and terminal for implementing speech control
CN103699530A (en) Method and equipment for inputting texts in target application according to voice input information
KR20190141767A (en) Key phrase detection with audio watermarking
WO2018099000A1 (en) Voice input processing method, terminal and network server
JP2019204074A (en) Speech dialogue method, apparatus and system
CN108028044A (en) The speech recognition system of delay is reduced using multiple identifiers
EP3627498B1 (en) Method and system, for generating speech recognition training data
CN111640434A (en) Method and apparatus for controlling voice device
WO2019101099A1 (en) Video program identification method and device, terminal, system, and storage medium
CN110992937A (en) Language offline recognition method, terminal and readable storage medium
CN112700770A (en) Voice control method, sound box device, computing device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination