CN113539252A - Barrier-free intelligent voice system and control method thereof - Google Patents

Barrier-free intelligent voice system and control method thereof Download PDF

Info

Publication number
CN113539252A
CN113539252A CN202010320575.7A CN202010320575A CN113539252A CN 113539252 A CN113539252 A CN 113539252A CN 202010320575 A CN202010320575 A CN 202010320575A CN 113539252 A CN113539252 A CN 113539252A
Authority
CN
China
Prior art keywords
voice
audio
database
tag
mobile device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010320575.7A
Other languages
Chinese (zh)
Inventor
庄连豪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202010320575.7A priority Critical patent/CN113539252A/en
Publication of CN113539252A publication Critical patent/CN113539252A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/686Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, title or artist information, time, location or usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The invention provides a barrier-free intelligent voice system and a control method thereof, which can recognize a plurality of words which can be used as independent semantic units from a voice frequency, can continuously judge whether the words are one of a plurality of voice labels created by a user, can further distinguish the voice labels to represent the name, code, single or combined instruction, program, voice information, recording information and the like of a certain object or information, can further judge a target object, a program instruction and remark description pointed by the semantic label in a database according to the successfully compared voice label combination, and can start a corresponding program or trigger a remote device to actuate according to the target object, the program instruction and the remark description, and can be used as an AI intelligent voice processing engine, can effectively reduce the operation amount and increase the processing speed of the system by enabling the user to define different types of voice label combinations by himself, meanwhile, the effects of confidentiality, theft prevention, barrier-free use and no limitation of language types can be achieved.

Description

Barrier-free intelligent voice system and control method thereof
Technical Field
The present invention relates to audio recognition technology, and more particularly to a barrier-free intelligent audio system and a control method thereof, which can recognize a plurality of independent semantic units from an audio, continuously compare whether the recognized semantic units are one of a plurality of audio tags created by a user, and further determine a voice command (also called a voice code) corresponding to the audio according to the combination of the compared audio tags (which can represent the name, code, single or combined command, program, voice message, recording message, etc. of a target object or information) to start the corresponding program or trigger other controlled devices to operate.
Background
With the development of technology, mobile devices with speech recognition systems are becoming popular, and most speech recognition systems allow users to communicate with mobile devices directly in Natural Language (Natural Language) through Language understanding technology, for example, the user may issue a continuous voice command to the mobile device to "buy a airline ticket to tokyo on saturday, huahang", however, to achieve the level of spoken language understanding (speech understating), the speech recognition system needs to perform a syntax analysis (e.g. syntax analyzer) and a semantic interpretation (e.g. semantic analyzer) on the continuous speech command, after the part of speech tagging is performed on each word of the continuous voice command, the word stem is extracted, the structure tree is formed, and each vocabulary of the structure tree is endowed with a semantic role, the semantic meaning of the whole sentence voice command is analyzed, and a larger operand can be generated.
Furthermore, the grammar structure of such continuous voice commands usually follows the restriction of specific grammar rules (including syntax and lexical), and the grammar structures of different languages are different, so that if the continuous voice commands issued by the user are complicated, even many superfluous words, slightly pause without speaking, or the grammar structure expressed by the user is not correct enough, or the voice commands may be different in individual accents or the user is used according to different single languages and mixed languages, the recognition accuracy of the voice recognition system may be affected, and the training of the natural language processing model (NLP) is also difficult.
Moreover, if the voiceprint recognition technology is not adopted, the existing voice recognition system cannot distinguish whether the user has the authority to perform a specific action or not from the voice of the user, so that how to provide an audio recognition technology which can reduce the operation amount of voice recognition, reduce the influence of a grammar structure on the voice recognition system, can be used without obstacles, can verify the use authority and has the characteristics of confidentiality and theft prevention for the voice recognition system which generally adopts the language understanding technology still remains a problem to be solved.
Disclosure of Invention
In order to achieve the above object, the present invention provides a control method for an intelligent barrier-free speech system, comprising:
(1) a step of analyzing voice audio: a voice recognition unit connected to a voice database and performing a voice analysis on a voice audio received by a voice receiving unit to identify a plurality of voices therefrom, and performing a word formation analysis on the plurality of voices to identify a plurality of words therefrom that can be used as independent semantic units;
(2) a comparison and voice labeling step: the voice recognition unit is connected to a label database to judge whether a plurality of words are one of a plurality of target voice labels defined by a mobile device and whether the words are one of a plurality of instruction voice labels defined by the mobile device;
(3) a step of executing corresponding voice commands: and a processor of the mobile device enables the mobile device to execute the program instruction on the target object according to the target object pointed by the compared target voice label in the label database and the program instruction pointed by the compared instruction voice label in the label database.
To achieve the above objective, the present invention provides an intelligent barrier-free speech system suitable for a mobile device having a processor, the system comprising: a voice receiving unit, which is connected with the processor information and is used for receiving a voice audio; the communication unit is in information connection with the processor; a voice database, which stores a plurality of voice audio samples; a label database, which stores a plurality of target voice labels and a plurality of instruction voice labels; a voice recognition unit which is respectively connected with the communication unit, the voice database and the label database, and is used for receiving the voice audio sent by the voice receiving unit, performing a voice analysis on the voice audio, identifying a plurality of voices from the result based on the result of reading the voice database, and performing a word formation analysis on the plurality of voices to identify a plurality of independent words from the result; the voice recognition unit is also used for judging whether the plurality of words are one of a plurality of target voice labels defined by the mobile device or not and whether the plurality of words are one of a plurality of instruction voice labels defined by the mobile device or not based on the result of reading the label database; if the judgment result is in accordance with the target object, the voice recognition unit enables the processor to enable the mobile device to execute the program instruction on the target object according to the target object pointed by the compared target voice tag in the tag database and the program instruction pointed by the compared instruction voice tag in the tag database through the communication unit.
Further, the method also comprises a step of detecting the awakening voice: the voice recognition unit judges whether the voice receiving unit receives a predefined awakening audio, if so, the awakening audio is regarded as awakening operation, and the step of analyzing the audio is continuously executed on the voice audio.
Further, when the step of comparing the voice tags is performed, the voice recognition unit determines whether the plurality of words identified by the voice audio also include a remark voice tag defined by the mobile device, and if so, the processor adjusts the program instruction or the content of the target object according to a remark description pointed by the remark voice tag in the tag database.
Further, the method also comprises a step of verifying the use authority: and the permission verification unit judges a grade permission corresponding to the voice audio based on the compared target voice label and the compared instruction voice label so as to determine whether the mobile device can execute the program instruction based on the current grade permission when the processor executes the corresponding voice instruction step.
Further, the method also comprises a step of detecting the dormant voice: the voice recognition unit judges whether the voice receiving unit receives a predefined dormant audio or not, if so, the dormant audio is regarded as dormant operation, and the step of analyzing the voice audio is stopped.
The invention also provides a barrier-free intelligent voice system, comprising:
a voice receiving unit, which is connected with a processor information of a mobile device for receiving a voice audio;
a communication unit in information connection with the processor;
a voice database storing a plurality of voice audio samples;
a label database for storing a plurality of target voice labels and a plurality of instruction voice labels;
a voice recognition unit, which is respectively connected with the communication unit, the voice database and the label database for receiving the voice audio sent by the voice receiving unit, executing a voice analysis to the voice audio, identifying a plurality of voices based on the result of reading the voice database, and then executing a word formation analysis to the plurality of voices to identify a plurality of independent words;
the voice recognition unit is also used for judging whether the words are one of the target voice labels defined by the mobile device or not and whether the words are one of the instruction voice labels defined by the mobile device or not based on the result of reading the label database; and the voice recognition unit is also used for enabling the processor to enable the mobile device to execute the program instruction on the target object according to a target object pointed by the compared target voice label in the label database and a program instruction pointed by the compared instruction voice label in the label database through the communication unit.
Furthermore, the voice recognition unit is also used for judging whether the voice receiving unit receives a predefined awakening audio and a predefined sleeping audio, if so, the voice analysis and the word formation analysis are continuously executed on the voice audio, and if so, the voice analysis and the word formation analysis are stopped being executed on the voice audio.
Further, the voice recognition unit is also used for judging whether the words recognized by the voice audio also include a remark voice tag defined by the mobile device or not based on the result of reading the tag database, and if so, the processor adjusts the program instruction or the content of the target object according to a remark description pointed by the remark voice tag in the tag database.
Furthermore, the system also comprises an authority verification unit which is in information connection with the voice recognition unit and used for judging a grade authority corresponding to the voice audio based on the compared target voice label and the compared instruction voice label so as to enable the voice recognition unit to determine the processor and determine whether to execute the program instruction based on the current grade authority of the mobile device.
The present invention also provides a barrier-free intelligent voice system, which is suitable for a mobile device having a processor, and is characterized by comprising:
a voice receiving unit connected with the processor information for receiving a voice audio;
a voice database storing a plurality of voice audio samples;
a label database for storing a plurality of target voice labels and a plurality of instruction voice labels;
a voice recognition unit, which is respectively connected with the voice database and the label database information, and is used for receiving the voice audio sent by the voice receiving unit, executing a voice analysis on the voice audio, identifying a plurality of voices based on the result of reading the voice database, and then executing a word formation analysis on the plurality of voices so as to identify a plurality of independent words;
the voice recognition unit is also used for judging whether the words are one of a plurality of target voice labels defined by the mobile device or not and whether the words are one of a plurality of instruction voice labels defined by the mobile device or not based on the result of reading the label database; and if the judgment result is that the target object is matched with the target object, the voice recognition unit enables the processor to enable the mobile device to execute the program instruction on the target object according to a target object pointed by the compared target voice label in the label database and a program instruction pointed by the compared instruction voice label in the label database.
Drawings
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
FIG. 1 is a block diagram of an intelligent speech system according to the present invention.
Fig. 2 is a schematic information flow diagram of the intelligent speech system of the present invention.
Fig. 3 is a flowchart of a control method of the intelligent voice system of the present invention.
FIG. 4 is a schematic diagram of a step of analyzing a voice audio according to the present invention.
FIG. 5A is a schematic diagram of an implementation of the present invention.
FIG. 5B is a schematic diagram of an implementation of the present invention.
Fig. 6A is a schematic information flow diagram according to another embodiment (a) of the present invention.
Fig. 6B is a schematic diagram of an implementation scenario of another embodiment (i) of the present invention.
FIG. 6C is a schematic diagram (II) illustrating another embodiment (I)
Fig. 7 is a block diagram of another embodiment (two) of the present invention.
Fig. 8 is a flowchart of a method according to another embodiment (c).
Fig. 9 is a block diagram illustrating another embodiment (four) of the present invention.
Fig. 10 is a schematic diagram of an implementation scenario of another embodiment (five) of the present invention.
Description of the reference numerals
10 accessible intelligent voice system
101 mobile device 1011 voice receiving unit
1012 processor
1013 communication unit
102 servo 1021 voice database
1022 tag database
1023 speech recognition unit
1024 right verification unit
103 controlled device
V-voice audio
L1 target Voice tag O target object
L2 instruction Voice tag I program instruction
L3 remark Voice tag R remark description
N network
Control method of S barrier-free intelligent voice system
S5 detecting a wake-up voice
S10 analyzing voice audio
S20 comparison voice tag
S25 verifying the usage rights
S30 executing the corresponding voice command
S35 detects a dormant voice.
Detailed Description
The technical solutions in the embodiments of the present invention are clearly and completely described below. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention, but the present invention may be practiced in other ways than those specifically described and will be readily apparent to those of ordinary skill in the art without departing from the spirit of the present invention, and therefore the present invention is not limited to the specific embodiments disclosed below.
Referring to fig. 1, which is a block diagram of an intelligent speech system according to the present invention, and referring to fig. 2, an information flow diagram of the intelligent speech system 10 of the present embodiment includes:
(1) a voice receiving unit 1011 in communication with a processor 1012 of a mobile device 101 for receiving a voice audio V, wherein the voice receiving unit 1011 may be a microphone of the mobile device 101 or a wireless earphone (wireless earphone) connected to the mobile device 101 via wireless communication (e.g. bluetooth);
(2) a communication unit 1013 capable of being in communication with the processor 1012;
(3) a speech database 1021, which can store a plurality of speech audio samples, and each group of speech (phone) in each speech audio sample corresponds to a word (word) that can be an independent semantic unit, wherein the speech audio sample can be identified by a speech recognition unit 1023 according to various initials (Initial) and finals (Final) of a speech audio V to identify one or more syllables (syllables) in the speech audio sample, and the language of the speech audio sample can be chinese, english, south-minning, cantonese, japanese, korean, etc., but not limited thereto;
(4) a tag database 1022 storing a plurality of target voice tags L1 and a plurality of command voice tags L2;
(5) a voice recognition unit 1023, which is respectively in information connection with the communication unit 1013, the voice database 1021 and the tag database 1022, for receiving the voice audio V transmitted by the voice receiving unit 1011 through a network N, and performing a voice Analysis (telephonic Analysis) on the voice audio V to identify a plurality of voices (phones) based on the result of reading the voice database 1021, and then performing a morphological Analysis (morphological) on the plurality of voices to identify independent words (words) therefrom; preferably, after recognizing the above-mentioned multiple voices, the present invention can simultaneously perform a Phoneme Analysis (phonetic Analysis) on the voice audio V to recognize the Phoneme (Phoneme) of the voice in the language to which the voice belongs, so as to help the voice recognition unit 1023 to recognize the multiple voices as words because the Phoneme is a minimum voice unit with a semantic function;
(6) the voice recognition unit 1023 can also determine whether the words are one of a plurality of target voice tags L1 defined by the mobile device 101 and one of a plurality of command voice tags L2 defined by the mobile device 101 based on the results of reading the tag database 1022;
(7) the voice recognition unit 1023 can also make the processor 1012 of the mobile device 101 execute the program instruction I on the target object O according to a target object O pointed by the compared target voice tag L1 in the tag database 1022 and a program instruction I pointed by the compared instruction voice tag L2 in the tag database 1022 through the communication unit 1013;
(8) in this embodiment, the voice receiving unit 1011, the processor 1012 and the communication unit 1013 are all operable on the mobile device 101.
(9) In this embodiment, the voice database 1021, the tag database 1022 and the voice recognition unit 1023 can all run on a server 102, and the server 102 can also have a second processor, which is not described herein.
The target object O illustrated in fig. 1 may be, for example, a contact information (which may be a contact name, a contact code number, or a contact name) stored in the mobile device 101, a target object information (which may be a target object name, a target object code number, or a target object name), a trip information, a to-do item information, a to-do list information, a file address information, or a hyperlink, but not limited thereto, and may also be a controlled device (not shown) communicatively connected to the communication unit 1013 of the mobile device 101 through the network N.
The network N illustrated in fig. 1 may be, for example, a public or private network, such as a wireless network (e.g., 3G, 4G LTE, Wi-Fi), a wired network, a Local Area Network (LAN), a wide area network (WA), etc., but not limited thereto.
The Server 102 illustrated in fig. 1 may be, for example, an independent Server providing a connection service, a Virtual Machine (VM) installed and running in the Server, a Server running in the form of a Virtual Private Server (Virtual Private Server), a public cloud, a Private cloud, an edge device (edge device), or the like, but is not limited thereto.
The processor 1012 illustrated in fig. 1 may be, for example, a Central Processing Unit (CPU), a Microprocessor (MPU), a Microcontroller (MCU), an Application Processor (AP), an embedded processor, or an Application Specific Integrated Circuit (ASIC), but is not limited thereto.
For example, the voice database 1021 and the tag database 1022 illustrated in fig. 1 may be a physical database host, a cloud database, or stored in the server 102 in the form of a plurality of data tables (tables) as a relational database or a non-relational database, but not limited thereto.
Referring to fig. 3, which is a flowchart of a control method of an intelligent speech system according to the present invention, and referring to fig. 1-2, the barrier-free intelligent speech system 10 of the present embodiment is adapted to analyze a speech audio V, and includes a speech receiving unit 1011, a speech database 1021, a tag database 1022, a speech recognition unit 1023, and a processor 1012 of the mobile device 101, and the control method S of the barrier-free intelligent speech system includes the following steps:
(1) analyzing the speech audio (step S10): the voice recognition unit 1023 is connected to the voice database 1021, and performs a voice analysis on the voice audio V received by the voice receiving unit 1011 to recognize a plurality of voices therefrom, and further performs a word formation analysis on the plurality of voices to recognize a plurality of words (words) that can be used as independent semantic units, such as books, boys, etc. that can be independently formed into words, and referring to the schematic diagram of the voice audio analyzing step of fig. 4, as shown in fig. 4, the voice audio V received by the voice receiving unit 1011, whether recognized as a plurality of voices of "banker board hit", "banker board", "banker old", "banker board Call", "banker head hit" (minnan t 'au' -ke) "," banker head "or" banker head Call ", violates the grammar limit or not, can be further identified as a plurality of words consisting of "zhuang boss" and "make", zhuang cephalad "and" make ", zhuang lao zhang" and "make telephone", zhuang lao zhang and "Call", "zhuang cephalad" and "Call", or "zhuang nape lao" and "Call"; similarly, if the speech audio V is a plurality of speech sounds recognized as "banker boss photo-combination display", "party boss photo-combination display", "banker head photo-combination display", "party banker boss Show", "Show photo-combination boss", "banker head photo Show", or "party head photo Show", regardless of whether the actual expression violates the grammar rule, the speech audio V can be further recognized as a plurality of words composed of "banker boss", "combination" and "display", "banker", "year plate", "combination" and "Show", or "banker", "combination" and "Show", but this is merely an example and not a limitation;
(2) comparing the voice tags (step S20): referring to the schematic diagram of the step of analyzing the voice audio shown in fig. 4, as shown in fig. 4, the voice recognition unit 1023 can be connected to the tag database 1022, to determine whether the words are one of a plurality of target voice tags L1 defined by the mobile device 101, and whether it is one of the command voice tags L2 defined by the mobile device 101, more specifically, if the voice audio V is recognized as words consisting of "banker boss" + "make", "banker house" + "make", "banker boss" + "make Call", "banker boss" + "Call", "banker house" + "Call", then, in step S20, the voice recognition unit 1023 compares the target voice tag L1 "banker boss" and the instructed voice tag L2 "Call" or "hit", or comparing another target voice tag L1 "banker" and a command voice tag L2 "Call" or "hit"; similarly, if the speech audio V can be recognized as words consisting of "banker boss" + "matching" + "display", or "banker house" + "matching" + "display", or "banker boss" + "matching" + "Show", then the speech recognition unit 1023 can compare the target speech tag L1 "banker boss matching" with the comparison instruction speech tag L2 "display" or "Show", or compare another target speech tag L1 "banker boss matching" with the comparison instruction speech tag L2 "display" or "Show", so far, if the comparison results of the speech recognition unit 1023 for the target speech tag L1 and the instruction speech tag L2 are matching, the step S30 is continued, if not matching, the speech is not recognized, and the step S10 or step S20 can be executed again;
(3) executing the corresponding voice instruction (step S30): the processor 1012 causes the mobile device 101 to execute the program command I on the target object O according to a target object O pointed to by the compared target voice tag L1 in the tag database 1022 and a program command I pointed to by the compared command voice tag L2 in the tag database 1022; more specifically, if the voice audio V can be recognized as words consisting of "banker boss" + "make", "banker boss" + "make a Call" or "banker boss" + "Call", then the voice recognition unit 1023 in step S30 may determine whether it can correspond to the target object O "banker xiao' S contact phone" in the mobile device 101 according to the compared target voice tag L1 "banker boss", and may determine that it can correspond to the program instruction I "to execute a telephone Application (APP)" carried by the mobile device 101 on the target object according to the compared instruction voice tag L2 "Call" or "make", and execute this, as shown in the implementation scenario diagram (one) of fig. 5A; similarly, if the voice audio V can be recognized by words consisting of "zhuanggao" + "matching" + "display" or "zhuanggao" + "matching" + "Show", the voice recognition unit 1023 can determine whether it can correspond to the photo of the target object O "and the zhuanggao in the mobile device 101 according to the compared target voice tag L1" zhuanggao ", and can determine whether it can correspond to the program instruction I" to execute a photo detector program installed in the mobile device 101 "according to the compared instruction voice tag L2" Show "or" display ", and execute the program, as shown in the implementation scenario diagram (two) of fig. 5B.
Referring to fig. 6A to 6C, which are a schematic diagram of information flow and a schematic diagram of implementation situation (a) and (b) respectively, and referring to fig. 1 to 3, when the voice recognition unit 1023 of this embodiment executes step S20 (comparing the voice tags), it can determine whether the words recognized by the voice audio V also include a remark voice tag L3 defined by the mobile device 101 itself, if so, the processor 1012 of the mobile device 101 can adjust the content of the program instruction I or the target object O according to the remark voice tag L3 in a remark explanation R pointed by the tag database 1022, more specifically, if the voice audio V is exemplified by a plurality of words consisting of "banker" + "house" or "banker" + "Call" + "house", then the voice recognition unit 1023 executes step S20 (comparing the voice tags), the target voice tag L1 "banker boss" can be compared, the instruction voice tag L2 "hit" or "Call" can be compared, and the remark voice tag L3 "house" can be compared, whereby, when the voice recognition unit 1023 executes the step S30 (executes the corresponding voice instruction), the voice recognition unit 1023 can determine whether it can correspond to the target object O "shoal house contact phone" in the mobile device 101 according to the compared target voice tag L1 "shoal boss" and remark voice tag L3 "house", meanwhile, according to the compared command voice tag L2 "Call" or "Call", it can be determined that it can correspond to the program command I "execute a telephone Application (APP) carried by the mobile device 101 on the target object, and execute it, so that, that is, as shown in fig. 6B, that is, the embodiment shown in fig. 6B is an example in which the content of the target object O can be adjusted by the remark voice tag L3; similarly, if the word "banker boss" + "group" + "shows" + "notes" or "banker boss" + "group" + "Show" + "notes" is taken as an example, the voice recognition unit 1023 compares the target voice tag L1 "banker boss group" + "instructs the voice tag L2" shows "or" Show "and compares the note voice tag L3" notes "or" notes "in step S20 (compare voice tags), so that when the voice recognition unit 1023 executes step S30 (execute corresponding voice command), the voice recognition unit 1023 can determine whether the target object O" in the mobile device 101 can be corresponded to the photo "banker boss group" + "notes" or "notes" according to the compared target voice tag L1 "group" ", and can be compared to the voice tag L2" or "Show" 2 ", and according to the compared remark voice tag L3 "speak remarks" or "speak notes", determining that it can correspond to the program instruction I "execute a photo viewer program installed on the mobile device 101, and play or present remark information associated with the target object", and execute the program, so far, as shown in fig. 6C, that is, the embodiment shown in fig. 6C is an example in which the remark voice tag L3 can adjust the content of the program instruction I.
Please refer to fig. 7, which is a block diagram of another embodiment (two) of the present invention, the present embodiment is similar to the technologies illustrated in fig. 1-3, and the main difference is that the barrier-free intelligent speech system 10 of the present embodiment further includes an authority verification unit 1024, which stores a plurality of program instructions and is in information connection with the speech recognition unit 1023, the authority verification unit 1024 can determine a level authority corresponding to the speech audio V according to the result of the speech recognition unit 1023 reading the tag database 1022 based on the compared target speech tag L1 and the compared instruction speech tag L2, so that the speech recognition unit 1023 can determine whether the mobile device 101 can execute the program instruction I based on the current level authority; in other words, the control method S of the barrier-free smart speech system of the present embodiment may include the step of "verifying the usage right" (step S25): an authority verification unit 1024 determines a level authority corresponding to the voice audio V based on the compared target voice tag L1 and the compared command voice tag L2 according to the result of the tag database 1022 read by the voice recognition unit 1023, so as to determine whether the processor 1012 executes the program command I in step 30 (executing the corresponding voice command), for example, the authority verification unit 1024 determines the target voice tag L1 and the command voice tag L2 corresponding to the voice audio V, and the data table stored in the tag database 1022 belongs to the first level (highest level), so as to determine that the current level authority of the mobile device 101 is the owner, and when the level authority of the program command I is the first level, the mobile device 101 is determined to have the authority to execute the program command I, and the second level, the third level, the fourth level, and the fourth level, The level authority of the third level may be, for example, a family user who cannot execute the program instruction I belonging to the first level; if the permission verification unit 1024 determines that the target voice tag L1 and the instructed voice tag L2 corresponding to the voice audio V belong to the second hierarchy in the data table stored in the tag database 1022, and when the level permission of the program instruction I is the third hierarchy or the second hierarchy, it determines that the mobile device 101 currently has the permission to execute the program instruction I; in contrast, if the permission verification unit 1024 determines that the target voice tag L1 and the command voice tag L2 corresponding to the voice audio V belong to the third level in the data table stored in the tag database 1022, and when the level permission of the program command I is the second level, it may determine that the mobile device 101 does not currently have the permission to execute the program command I, and so on.
Please refer to fig. 8, which is a flowchart illustrating a method according to another embodiment (three), the present embodiment is similar to the techniques illustrated in fig. 1 to 3, and the main difference is that the control method S of the barrier-free intelligent speech system of the present embodiment may include a step of "detecting a wake-up speech" (step S5): the speech recognition unit 1023 first determines whether the speech receiving unit 1011 receives a predefined wake-up audio, such as "small white" or "secretary", and if so, regards the wake-up audio as a wake-up operation, and then continues to perform step S10 (analyzing the speech audio) on the speech audio V to perform speech analysis and word formation analysis on the speech audio V in a continuous manner; in addition, the control method S of the barrier-free intelligent speech system of the embodiment may further include a step of "detecting the dormant speech" (step S35): the speech recognition unit 1023 determines whether the speech receiving unit 1011 has received a predefined dormant audio, such as "closing the reminder" or "rest in secretary", and if so, the dormant audio may be regarded as a dormant operation, and stops performing the step S10 (analyzing the speech audio) on the speech audio V, i.e. suspending performing the speech analysis and the word formation analysis on the speech audio V, and only performing the step S5 continuously, but the step S35 may be performed after the step S5, the step S10 or the step S20 continuously, and the sequence is not limited to the sequence illustrated in fig. 8.
Referring to fig. 9, which is a block diagram of another embodiment (four) of the present invention, compared to the technical solutions illustrated in fig. 1-3, in the barrier-free intelligent speech system 10 of the present embodiment, the speech database 1021, the tag database 1022, the speech recognition unit 1023 and the authorization verification unit 1024 can all be mounted in the mobile device 101, and the communication unit 1013 shown in fig. 1 can be omitted, so that the information of the speech audio V, the target object O, the program instruction I and the remark R can be transmitted and received between the processor 1012 and the speech recognition unit 1023 through the network N shown in fig. 1, in other words, the barrier-free intelligent speech system 10 of the present embodiment can allow the user to directly use the mobile device 101 even in the absence of a communication network, and then the recognition and execution of the speech command can be completed at present.
Referring to fig. 10, which is an implementation scenario diagram of another embodiment (five) of the present invention, and referring to fig. 1 to fig. 3 in combination, in the barrier-free intelligent speech system 10 of the present embodiment, when the speech recognition unit 1023 enables the processor 1012 to execute the program instruction I on the target object O pointed by the tag database 1022 according to the compared target speech tag L1 and the compared instruction speech tag L2 and the program instruction I pointed by the tag database 1022 according to the compared instruction speech tag L2, the target object O may be a controlled device 103, such as a power door, a lamp, a television, an electrical appliance, etc., other than the mobile device 101, for example, the user may operate the mobile device 101, and after the speech receiving unit 1011 receives the speech audio V, the speech audio V may be recognized by the speech recognition unit 1023 as being composed of an "on" (corresponding to the instruction speech tag L2) + "TVS" (corresponding to the target speech tag L1) For example, and no matter whether the grammar structure of the voice audio V is correct, the voice recognition unit 1023 may compare the target object O "tv TVS news stand" corresponding to the target voice tag L1 and the program instruction I "corresponding to the instruction voice tag L2 with the communication unit 1013 of the mobile device 101 to wirelessly turn on the smart tv.
In another embodiment of the present invention, a computer program product for executing the method S for controlling a barrier-free intelligent audio system is provided, wherein after a plurality of program instructions of the computer program product are loaded into a computer system, the steps S5, S10, S20, S25, S30 and S35 of the method S for controlling a barrier-free intelligent audio system are at least completed.
Accordingly, the present invention can achieve at least the following advantages:
(1) no matter whether the grammar rule of the voice audio is correct or not, the invention can identify whether the voice label combination (representing the name, code number, single or combined instruction, program, voice information, recording information and the like of a certain object or information) is created by a user by comparing with the voice label combination, and can identify whether the voice label combination corresponds to a specific target object and program instruction.
(2) The invention can judge the current level authority of the voice audio by comparing the voice label combination created by the user, has the technology of verification and similar to audio encryption, and can not drive the mobile device to execute the specific function or start the controlled device except the mobile device by sending the voice audio if a third party does not know the voice label combination created by the user and the voice audio which is not from the original user.
However, the above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention; any equivalent changes and modifications within the spirit and scope of the present invention should be covered by the protection scope of the present invention.

Claims (10)

1. A control method of barrier-free intelligent voice system is disclosed, the barrier-free intelligent voice system is suitable for analyzing a voice audio, the barrier-free intelligent voice system includes a voice receiving unit, a voice database, a label database, a voice recognition unit, a permission verification unit and a processor of a mobile device, the control method includes:
a step of analyzing voice audio: the voice recognition unit is connected to the voice database, and executes voice analysis on the voice audio received by the voice receiving unit so as to recognize a plurality of voices from the voice audio, and then executes word formation analysis on the plurality of voices so as to recognize a plurality of words which can be used as independent semantic units from the voice audio;
a comparison and voice labeling step: the voice recognition unit is connected to the tag database to determine whether the words are one of a plurality of target voice tags defined by the mobile device and whether the words are one of a plurality of instruction voice tags defined by the mobile device; and
a step of executing corresponding voice commands: the processor enables the mobile device to execute the program instruction on a target object according to a target object pointed by the compared target voice tag in the tag database and according to a program instruction pointed by the compared instruction voice tag in the tag database.
2. The method of claim 1, further comprising a step of detecting the wake-up voice: the voice recognition unit judges whether the voice receiving unit receives a predefined awakening audio, if so, the awakening audio is regarded as awakening operation, and the step of analyzing the audio is continuously executed on the voice audio.
3. The method as claimed in claim 1, wherein when the comparing step is performed, the voice recognition unit determines whether the words identified by the voice audio also include a remark voice tag defined by the mobile device, and if so, the processor adjusts the program command or the content of the target object according to a remark description pointed by the remark voice tag in the tag database.
4. The method for controlling a barrier-free intelligent voice system of claim 1, further comprising a step of verifying the usage right: and the permission verification unit judges a grade permission corresponding to the voice audio based on the compared target voice label and the compared instruction voice label so as to determine whether the mobile device can execute the program instruction based on the current grade permission when the processor executes the corresponding voice instruction step.
5. The method for controlling an intelligent voice system without obstruction as claimed in claim 1 or 2, further comprising a step of detecting a dormant voice: the voice recognition unit judges whether the voice receiving unit receives a predefined dormant audio or not, if so, the dormant audio is regarded as dormant operation, and the step of analyzing the voice audio is stopped.
6. An intelligent barrier-free speech system, comprising:
a voice receiving unit, which is connected with a processor information of a mobile device for receiving a voice audio;
a communication unit in information connection with the processor;
a voice database storing a plurality of voice audio samples;
a label database for storing a plurality of target voice labels and a plurality of instruction voice labels;
a voice recognition unit, which is respectively connected with the communication unit, the voice database and the label database for receiving the voice audio sent by the voice receiving unit, executing a voice analysis to the voice audio, identifying a plurality of voices based on the result of reading the voice database, and then executing a word formation analysis to the plurality of voices to identify a plurality of independent words;
the voice recognition unit is also used for judging whether the words are one of the target voice labels defined by the mobile device or not and whether the words are one of the instruction voice labels defined by the mobile device or not based on the result of reading the label database; and
the voice recognition unit is also used for enabling the processor to enable the mobile device to execute the program instruction on the target object according to a target object pointed by the compared target voice label in the label database and a program instruction pointed by the compared instruction voice label in the label database through the communication unit.
7. The barrier-free intelligent voice system of claim 6, wherein the voice recognition unit is also configured to determine whether the voice receiving unit receives a predefined wake-up audio and a predefined sleep audio, if so, then continue to perform the voice analysis and the word formation analysis on the voice audio, and if not, then stop performing the voice analysis and the word formation analysis on the voice audio.
8. The barrier-free intelligent speech system of claim 6, wherein the speech recognition unit is further configured to determine whether the words identified by the speech audio also include a remark speech tag defined by the mobile device based on the result of reading the tag database, and if so, enable the processor to adjust the program command or the content of the target object according to a remark description pointed to by the remark speech tag in the tag database.
9. The barrier-free intelligent voice system of claim 6, further comprising an authority verification unit in communication with the voice recognition unit for determining a level authority corresponding to the voice audio based on the compared target voice tag and the compared instruction voice tag, so that the voice recognition unit can determine whether to execute the program instruction based on the level authority of the mobile device.
10. An intelligent barrier-free speech system adapted for use with a mobile device having a processor, comprising:
a voice receiving unit connected with the processor information for receiving a voice audio;
a voice database storing a plurality of voice audio samples;
a label database for storing a plurality of target voice labels and a plurality of instruction voice labels;
a voice recognition unit, which is respectively connected with the voice database and the label database information, and is used for receiving the voice audio sent by the voice receiving unit, executing a voice analysis on the voice audio, identifying a plurality of voices based on the result of reading the voice database, and then executing a word formation analysis on the plurality of voices so as to identify a plurality of independent words;
the voice recognition unit is also used for judging whether the words are one of a plurality of target voice labels defined by the mobile device or not and whether the words are one of a plurality of instruction voice labels defined by the mobile device or not based on the result of reading the label database; and
if the judgment result is that the target object is matched with the target object, the voice recognition unit enables the processor to enable the mobile device to execute the program instruction on the target object according to a target object pointed by the compared target voice label in the label database and a program instruction pointed by the compared instruction voice label in the label database.
CN202010320575.7A 2020-04-22 2020-04-22 Barrier-free intelligent voice system and control method thereof Pending CN113539252A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010320575.7A CN113539252A (en) 2020-04-22 2020-04-22 Barrier-free intelligent voice system and control method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010320575.7A CN113539252A (en) 2020-04-22 2020-04-22 Barrier-free intelligent voice system and control method thereof

Publications (1)

Publication Number Publication Date
CN113539252A true CN113539252A (en) 2021-10-22

Family

ID=78094052

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010320575.7A Pending CN113539252A (en) 2020-04-22 2020-04-22 Barrier-free intelligent voice system and control method thereof

Country Status (1)

Country Link
CN (1) CN113539252A (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103077714A (en) * 2013-01-29 2013-05-01 华为终端有限公司 Information identification method and apparatus
CN103839547A (en) * 2012-11-27 2014-06-04 英业达科技有限公司 System for loading corresponding instruction elements by comparing voice operation signals and method thereof
CN104462262A (en) * 2014-11-21 2015-03-25 北京奇虎科技有限公司 Method and device for achieving voice search and browser client side
CN107180631A (en) * 2017-05-24 2017-09-19 刘平舟 A kind of voice interactive method and device
CN108198550A (en) * 2017-12-29 2018-06-22 江苏惠通集团有限责任公司 A kind of voice collecting terminal and system
CN108665895A (en) * 2018-05-03 2018-10-16 百度在线网络技术(北京)有限公司 Methods, devices and systems for handling information
CN110211583A (en) * 2019-05-31 2019-09-06 北京声赫科技有限公司 A kind of voice interactive method and interactive voice equipment based on intelligent line traffic control
CN110349578A (en) * 2019-06-21 2019-10-18 北京小米移动软件有限公司 Equipment wakes up processing method and processing device
CN110718225A (en) * 2019-11-25 2020-01-21 深圳康佳电子科技有限公司 Voice control method, terminal and storage medium
CN110767232A (en) * 2019-09-29 2020-02-07 深圳和而泰家居在线网络科技有限公司 Speech recognition control method and device, computer equipment and computer storage medium

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103839547A (en) * 2012-11-27 2014-06-04 英业达科技有限公司 System for loading corresponding instruction elements by comparing voice operation signals and method thereof
CN103077714A (en) * 2013-01-29 2013-05-01 华为终端有限公司 Information identification method and apparatus
CN104462262A (en) * 2014-11-21 2015-03-25 北京奇虎科技有限公司 Method and device for achieving voice search and browser client side
CN107180631A (en) * 2017-05-24 2017-09-19 刘平舟 A kind of voice interactive method and device
CN108198550A (en) * 2017-12-29 2018-06-22 江苏惠通集团有限责任公司 A kind of voice collecting terminal and system
CN108665895A (en) * 2018-05-03 2018-10-16 百度在线网络技术(北京)有限公司 Methods, devices and systems for handling information
CN110211583A (en) * 2019-05-31 2019-09-06 北京声赫科技有限公司 A kind of voice interactive method and interactive voice equipment based on intelligent line traffic control
CN110349578A (en) * 2019-06-21 2019-10-18 北京小米移动软件有限公司 Equipment wakes up processing method and processing device
CN110767232A (en) * 2019-09-29 2020-02-07 深圳和而泰家居在线网络科技有限公司 Speech recognition control method and device, computer equipment and computer storage medium
CN110718225A (en) * 2019-11-25 2020-01-21 深圳康佳电子科技有限公司 Voice control method, terminal and storage medium

Similar Documents

Publication Publication Date Title
US20200234706A1 (en) Promoting voice actions to hotwords
CN108520743B (en) Voice control method of intelligent device, intelligent device and computer readable medium
US20210335351A1 (en) Voice Characterization-Based Natural Language Filtering
US10013985B2 (en) Systems and methods for audio command recognition with speaker authentication
KR102371697B1 (en) Operating Method for Voice function and electronic device supporting the same
KR100586767B1 (en) System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input
US20240071382A1 (en) Temporary account association with voice-enabled devices
CN106210239A (en) The maliciously automatic identifying method of caller's vocal print, device and mobile terminal
CN103903621A (en) Method for voice recognition and electronic equipment
US9251808B2 (en) Apparatus and method for clustering speakers, and a non-transitory computer readable medium thereof
CN110047481A (en) Method for voice recognition and device
CN109462482A (en) Method for recognizing sound-groove, device, electronic equipment and computer readable storage medium
CN112669842A (en) Man-machine conversation control method, device, computer equipment and storage medium
WO2022265896A1 (en) Natural language processing routing
CN111768789B (en) Electronic equipment, and method, device and medium for determining identity of voice generator of electronic equipment
CN109273004B (en) Predictive speech recognition method and device based on big data
KR101775532B1 (en) Multimedia device for providing voice recognition service by using at least two of database and the method for controlling the same
CN113539252A (en) Barrier-free intelligent voice system and control method thereof
TWM601446U (en) Barrier-free intelligent voice system
US11705126B2 (en) Barrier-free intelligent voice system and control method thereof
CN112037772B (en) Response obligation detection method, system and device based on multiple modes
US11605380B1 (en) Coordinating content-item output across multiple electronic devices
US11481188B1 (en) Application launch delay and notification
CN112185374A (en) Method and device for determining voice intention
CN113870857A (en) Voice control scene method and voice control scene system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination