NL2035518A - Intelligent voice ai pacifying method - Google Patents
Intelligent voice ai pacifying method Download PDFInfo
- Publication number
- NL2035518A NL2035518A NL2035518A NL2035518A NL2035518A NL 2035518 A NL2035518 A NL 2035518A NL 2035518 A NL2035518 A NL 2035518A NL 2035518 A NL2035518 A NL 2035518A NL 2035518 A NL2035518 A NL 2035518A
- Authority
- NL
- Netherlands
- Prior art keywords
- speech
- voice
- user
- timbre
- intelligent
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 16
- 238000010276 construction Methods 0.000 claims abstract description 19
- 238000004891 communication Methods 0.000 claims description 10
- 230000006854 communication Effects 0.000 claims description 10
- 230000006870 function Effects 0.000 claims description 7
- 230000008451 emotion Effects 0.000 claims description 4
- 230000010354 integration Effects 0.000 claims description 3
- 230000003213 activating effect Effects 0.000 claims 1
- 238000005516 engineering process Methods 0.000 abstract description 9
- 238000013473 artificial intelligence Methods 0.000 abstract description 6
- 230000004927 fusion Effects 0.000 abstract 1
- 238000007726 management method Methods 0.000 abstract 1
- 230000000875 corresponding effect Effects 0.000 description 43
- 230000001960 triggered effect Effects 0.000 description 8
- 230000002996 emotional effect Effects 0.000 description 5
- 238000010586 diagram Methods 0.000 description 4
- 230000000694 effects Effects 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/004—Artificial life, i.e. computing arrangements simulating life
- G06N3/006—Artificial life, i.e. computing arrangements simulating life based on simulated virtual individual or collective life forms, e.g. social simulations or particle swarm optimisation [PSO]
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L13/00—Speech synthesis; Text to speech systems
- G10L13/02—Methods for producing synthetic speech; Speech synthesisers
- G10L13/033—Voice editing, e.g. manipulating the voice of the synthesiser
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L51/00—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail
- H04L51/02—User-to-user messaging in packet-switching networks, transmitted according to store-and-forward or real-time protocols, e.g. e-mail using automatic reactions or user delegation, e.g. automatic replies or chatbot-generated messages
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Theoretical Computer Science (AREA)
- Computing Systems (AREA)
- General Physics & Mathematics (AREA)
- Biomedical Technology (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Life Sciences & Earth Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Computer Networks & Wireless Communication (AREA)
- Signal Processing (AREA)
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The present invention provides an intelligent voice AI pacifying method, and relates to the technical field of artificial intelligence. Based on a policy management method of power industry of classification tree fusion technology, the pacifying 5 Inethod, includes the following steps: SI. performing preliminary construction of the intelligent voice AI; SZ. collecting and inputting the required, voice; S3. performing compilation of a voice code stream based on the collected voice; and S4. emitting the voice fran a user and triggering corresponding instructions IO according to the voice. According to the present invention, AI is started after the following steps of: performing preliminary construction of the intelligent voice AI; collecting and inputting the required voice; performing compilation of voice code stream based on the collected voice; and emitting voice from the user and 15 triggering corresponding instructions according to the voice. (+ Fig. l)
Description
P1852 /NLpd
INTELLIGENT VOICE AI PACIFYING METHOD
The present invention relates to the technical field of arti- ficial intelligence, and in particular, relates to an intelligent voice AI pacifying method.
Background technology
Intelligent voice, or intelligent voice technology, is the realization of human-computer language communication, including automatic speech recognition (ASR) and text to speech (TTS). The research of intelligent voice technology starts with the ASR, which can be traced back to the 1950s. As information technology develops, the intelligent voice technology has become the most convenient and effective means of people's information acquisition and communication. The artificial intelligence is a branch of com- puter science, which attempts to understand the essence of intel- ligence and produce a new intelligent machine that can respond in a way similar to human intelligence. The research in this field includes robotics, language recognition, image recognition, natu- ral language processing and expert syspraaks, etc. Since the birth of artificial intelligence, it has become increasingly mature in terms of theory and technology, and its fields of application are also expanding. It can be envisioned that in the future, scien- tific and technological products brought by the artificial intel- ligence will be “containers” of human wisdom.
AI voice function is a technology that a machine automatical- ly converts human voice into texts. AI voice, using ASR, TTS, se- mantic understanding and other artificial intelligence technolo- gies, can interact with customers in a natural and smooth way through anthropomorphic voice, text and other ways, so as to pro- vide independent on-line Q&A, consulting, business services, etc.
The existing intelligent voice has indifferent tone, and poorer intelligent effect, can not be effectively communicated with the user, and can not to pacify the users according to their emotional changes.
Solved technical problems
Aiming at the shortcomings of the prior art, the present in- vention provides an intelligent voice AI pacifying method, solving the problems that the existing intelligent voice tone is indiffer- ent, the intelligent effect is poor, the communication with the user can not be effectively carried out, and the corresponding emotional pacifying work can not be carried out according to the emotional transformation of the user.
Technical solutions
In order to realize the above object, the present invention is realized through the following technical solutions: an intelli- gent voice AI pacifying method, including the following steps:
Sl. preliminary construction of intelligent voice AI is per- formed
The intelligence of AI is constructed; although difficult and complicated operations are not need to carry out by AI, correct and smooth communication to basic answers are required by AI, and corresponding information should be ensured to be triggered rea- sonably by AI according to voice data in a database.
S2. the required voice is collected and input
Voice is collected through a corresponding device, so as to obtain corresponding voice data; further, the collected voice data is sent to a cloud database, and the voice data in the cloud data- base is performed automatic voice recognition via a server; the voice data is not limited to external voice, and the voice may comprise communication recording, audio in video, and voice sent in chat softwares; the corresponding voice data is obtained through analysis and integration; results are recognized in the cloud database; and after successful recognition, intelligent voice services can be performed through AI according to the corre- sponding voice.
S3. compilation of a voice code stream is performed based on the collected voice
The voice code stream to be sent is obtained according to need; the audio data in the cloud database is read and answered by
AI, and the voice control information for controlling an audio mixing strategy is obtained according to the voice code stream; the voice code stream is compiled according to the collected voice; imitated timbre and tone are compiled and imitated by audio mixing and compiling, such that the voice emitted by AI is the same as that of the user; and answer and pacifying behaviors can be more effective through the consistent voice.
S4. voice is emitted from a user and corresponding instruc- tions are triggered based on the voice
When the user phonates the voice, instructions are triggered; according to different instructions sent by the user, AI is start- ed according to the corresponding instruction is triggered by the voice; further, along with the words said by the user, AI performs extraction from the cloud database, and the server performs recog- nition, thus entering a corresponding working mode according to the user’s options, which is used for pacifying work in different situations, such as missing, sad, painful, fearful, angry, etc; according to the recognition of the user’s tone, timbre, and spo- ken voice, AI selects the corresponding working mode to extract the corresponding voice data for answering, so as to carry out the work of answering and pacifying.
Preferably, 81: preliminary construction of intelligent voice
AT is performed
The most important thing in the intelligent voice AI pacify- ing is the construction of intelligent voice AI; if the construc- tion of intelligent voice AI cannot be completed, the reasonable construction work cannot be completed in the follow-up; the intel- ligent voice AI not only answers by extracting audio data files, but also determines emotion of the user according to the words, keywords, timbre and tone of the user, such that the corresponding audio data files can be reasonably mobilized to play and pacify the user, corresponding to determining the state of designated agents according to the trigger information of user state, or ac- cording to the voice instructions sent by the user, the corre- sponding text and audio are mobilized to play coherently to an- swer.
Preferably, S3: compilation of a voice code stream is per- formed based on the collected voice
According to the obtained voice code stream to be sent, the obtained voice code stream and voice control information are sent to a voice server, and the voice server receives at least one voice stream; according to functions corresponding to each posi- tion, a corresponding factor is searched in a timbre database as an analog timbre to be output; the timbre database is used for storing the position range of line segments corresponding to each timbre and the function corresponding to the position range, so as to compile the corresponding voice; after the compilation, the voice is stored in the server and the cloud database, waiting for
AI to perform a retrieval work; the audio mixing and compiling are used to compile and imitate the imitated timbre and tone, such that the voice emitted by AI is the same as that of the user, and it is more effective to perform answer and pacify by a consistent voice.
Beneficial effects
The present invention provides an intelligent voice AI paci- fying method. The present invention has the following beneficial effects:
According to the present invention, the corresponding in- structions are started based on the voice through the following steps of: performing preliminary construction of the intelligent voice AI; collecting and inputting the required voice; performing compilation of voice code stream based on the collected voice; and emitting voice from the user. According to the present invention, the intelligent voice AI not only answers by extracting audio data files, but also determines emotion of the user according to the words, keywords, timbre and tone of the user, such that the corre- sponding audio data files can be reasonably mobilized to play and pacify the user, corresponding to determining the state of desig- nated agents according to the trigger information of user state, or according to the voice instructions sent by the user, the cor- responding text and audio are mobilized to play coherently to an- swer, solving the problems that the existing intelligent voice tone is indifferent, the intelligent effect is poor, the communi- cation with the user can not be effectively carried out, and the corresponding emotional pacifying work can not be carried out ac-
cording to the emotional transformation of the user.
Description of attached figures
FIG. 1 is a flow chart of an intelligent voice AI pacifying method according to the present invention; 5 FIG. 2 is a preliminary construction diagram of an intelli- gent voice AI pacifying method according to the present invention;
FIG. 3 is an operating diagram of collecting and inputting the required voice of an intelligent voice AI pacifying method ac- cording to the present invention;
FIG. 4 is a diagram of compiling a voice code stream based on the collected voice of an intelligent voice AI pacifying method according to the present invention; and
FIG. 5 is a diagram of emitting voice form a user and start- ing corresponding instructions based on the voice of an intelli- gent voice AI pacifying method according to the present invention.
Specific embodiments
The technical solutions of the examples in the present inven- tion will be described clearly and completely with reference to the accompanying drawings of the examples in the present invention below. Obviously, the examples described are only some, rather than all examples of the present invention. Based on the examples of the present invention, all the other examples obtained by the ordinary skilled in the art without creative efforts fall within the scope of protection of the present invention.
Example 1:
As shown in FIGs. 1-5, the example of the present invention provides an intelligent voice AI pacifying method, including the following steps:
S1. preliminary construction of intelligent voice AI is per- formed
The intelligence of AI is constructed; although difficult and complicated operations are not required to carry out by AI, cor- rect and smooth communication to basic answers are required by Al, and corresponding information should be ensured to be triggered reasonably by AI according to voice data in a database.
S2. the required voice is collected and input
Voice is collected through a corresponding device, so as to obtain corresponding voice data; further, the collected voice data is sent to a cloud database, and the voice data in the cloud data- base is performed automatic voice recognition via a server; the voice data is not limited to external voice, and the voice may comprise communication recording, audio in video, and voice sent in chat softwares; the corresponding voice data is obtained through analysis and integration; results are recognized in the cloud database; and after successful recognition, intelligent voice services can be performed through AI according to the corre- sponding voice.
S3. compilation of a voice code stream is performed based on the collected voice
The voice code stream to be sent is obtained according to need; the audio data in the cloud database is read and answered by
AI, and the voice control information for controlling an audio mixing strategy is obtained according to the voice code stream; the voice code stream is compiled according to the collected voice; imitated timbre and tone are compiled and imitated by audio mixing and compiling, such that the voice emitted by AI is the same as that of the user; and answer and pacifying behaviors can be more effective through the consistent voice.
S4. voice is emitted from a user and corresponding instruc- tions are triggered based on the voice
When the user phonates the voice, instructions are triggered; according to different instructions sent by the user, AI is start- ed according to the corresponding instruction is triggered by the voice; further, along with the words said by the user, AI performs extraction from the cloud database, and the server performs recog- nition, thus entering a corresponding working mode according to the user's options, which is used for pacifying work in different situations, such as missing, sad, painful, fearful, angry, etc; according to the recognition of the user’s tone, timbre, and spo- ken voice, AI selects the corresponding working mode to extract the corresponding voice data for answering, so as to carry out the work of answering and pacifying.
Example 2:
Sl. specific construction of intelligent voice AI is per-
formed
The most important thing in the intelligent voice AI pacify- ing is the construction of intelligent voice AI; if the construc- tion of intelligent voice AI cannot be completed, the reasonable construction work cannot be completed in the follow-up; the intel- ligent voice AI not only answers by extracting audio data files, but also determines emotion of the user according to the words, keywords, timbre and tone of the user, such that the corresponding audio data files can be reasonably mobilized to play and pacify the user, corresponding to determining the state of designated agents according to the trigger information of user state, or ac- cording to the voice instructions sent by the user, the corre- sponding text and audio are mobilized to play coherently to an- swer.
S3. compilation of a voice code stream is performed based on the collected voice
According to the obtained voice code stream to be sent, the obtained voice code stream and voice control information are sent to a voice server, and the voice server receives at least one voice stream; according to functions corresponding to each posi- tion, a corresponding factor is searched in a timbre database as an analog timbre to be output; the timbre database is used for storing the position range of line segments corresponding to each timbre and the function corresponding to the position range, so as to compile the corresponding voice; after the compilation, the voice is stored in the server and the cloud database, waiting for
AI to perform a retrieval work; the audio mixing and compiling are used to compile and imitate the imitated timbre and tone, such that the voice emitted by AI is the same as that of the user, and it is more effective to perform answer and pacify by a consistent voice.
Although the examples of the present invention have been shown and described, for those ordinary skilled in the art, it can be understood as various changes, modifications, replacements and variations can be made on these examples within the principle and spirit of the present invention. The scope of the present inven- tion is defined by the attached claims and the equivalent thereof.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
NL2035518A NL2035518B1 (en) | 2023-07-31 | 2023-07-31 | Intelligent voice ai pacifying method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
NL2035518A NL2035518B1 (en) | 2023-07-31 | 2023-07-31 | Intelligent voice ai pacifying method |
Publications (2)
Publication Number | Publication Date |
---|---|
NL2035518A true NL2035518A (en) | 2023-09-11 |
NL2035518B1 NL2035518B1 (en) | 2024-04-16 |
Family
ID=87972071
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
NL2035518A NL2035518B1 (en) | 2023-07-31 | 2023-07-31 | Intelligent voice ai pacifying method |
Country Status (1)
Country | Link |
---|---|
NL (1) | NL2035518B1 (en) |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200265829A1 (en) * | 2019-02-15 | 2020-08-20 | International Business Machines Corporation | Personalized custom synthetic speech |
WO2021034786A1 (en) * | 2019-08-21 | 2021-02-25 | Dolby Laboratories Licensing Corporation | Systems and methods for adapting human speaker embeddings in speech synthesis |
-
2023
- 2023-07-31 NL NL2035518A patent/NL2035518B1/en active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200265829A1 (en) * | 2019-02-15 | 2020-08-20 | International Business Machines Corporation | Personalized custom synthetic speech |
WO2021034786A1 (en) * | 2019-08-21 | 2021-02-25 | Dolby Laboratories Licensing Corporation | Systems and methods for adapting human speaker embeddings in speech synthesis |
Also Published As
Publication number | Publication date |
---|---|
NL2035518B1 (en) | 2024-04-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN105843381B (en) | Data processing method for realizing multi-modal interaction and multi-modal interaction system | |
US6526395B1 (en) | Application of personality models and interaction with synthetic characters in a computing system | |
JP7313455B2 (en) | speaking agent | |
CN109522835A (en) | Children's book based on intelligent robot is read and exchange method and system | |
CN106055662A (en) | Emotion-based intelligent conversation method and system | |
CN106294726A (en) | Based on the processing method and processing device that robot role is mutual | |
KR102293743B1 (en) | AI Chatbot based Care System | |
CN109002515A (en) | A kind of method and apparatus of intelligent response | |
CN109065052A (en) | A kind of speech robot people | |
CN109101663A (en) | A kind of robot conversational system Internet-based | |
CN116049360A (en) | Intelligent voice dialogue scene conversation intervention method and system based on client image | |
KR20200059112A (en) | System for Providing User-Robot Interaction and Computer Program Therefore | |
Alam et al. | Comparative study of speaker personality traits recognition in conversational and broadcast news speech. | |
Ritschel et al. | Shaping a social robot's humor with natural language generation and socially-aware reinforcement learning | |
CN110347811A (en) | A kind of professional knowledge question and answer robot system based on artificial intelligence | |
CN110335014A (en) | Interview method, apparatus and computer readable storage medium | |
Origlia et al. | FANTASIA: a framework for advanced natural tools and applications in social, interactive approaches | |
NL2035518B1 (en) | Intelligent voice ai pacifying method | |
CN115222857A (en) | Method, apparatus, electronic device and computer readable medium for generating avatar | |
CN116895087A (en) | Face five sense organs screening method and device and face five sense organs screening system | |
CN111291539B (en) | File editing control method, device, computer device and storage medium | |
CN111209376A (en) | AI digital robot operation method | |
López et al. | Lifeline dialogues with roberta | |
CN109948153A (en) | It is a kind of to be related to man-machine communication's system of video and audio multimedia information processing | |
Pelé et al. | Let’s find a restaurant with nestor a 3d embodied conversational agent on the web |