CN117854492A - Intelligent interaction method, system, equipment and storage medium based on large model - Google Patents
Intelligent interaction method, system, equipment and storage medium based on large model Download PDFInfo
- Publication number
- CN117854492A CN117854492A CN202410104210.9A CN202410104210A CN117854492A CN 117854492 A CN117854492 A CN 117854492A CN 202410104210 A CN202410104210 A CN 202410104210A CN 117854492 A CN117854492 A CN 117854492A
- Authority
- CN
- China
- Prior art keywords
- feature sequence
- text
- audio
- voice
- module
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 230000003993 interaction Effects 0.000 title claims abstract description 56
- 238000000034 method Methods 0.000 title claims abstract description 53
- 238000012545 processing Methods 0.000 claims abstract description 24
- 238000012549 training Methods 0.000 claims description 11
- 238000013136 deep learning model Methods 0.000 claims description 10
- 238000013507 mapping Methods 0.000 claims description 9
- 230000008451 emotion Effects 0.000 claims description 7
- 230000002452 interceptive effect Effects 0.000 claims 1
- 238000013473 artificial intelligence Methods 0.000 abstract description 2
- 230000015572 biosynthetic process Effects 0.000 description 6
- 238000004590 computer program Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 6
- 238000003786 synthesis reaction Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 5
- 238000003058 natural language processing Methods 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000008909 emotion recognition Effects 0.000 description 3
- 238000000605 extraction Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012423 maintenance Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 230000033764 rhythmic process Effects 0.000 description 2
- 238000006467 substitution reaction Methods 0.000 description 2
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 230000008439 repair process Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000013518 transcription Methods 0.000 description 1
- 230000035897 transcription Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/18—Speech classification or search using natural language modelling
- G10L15/1822—Parsing for meaning understanding
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/16—Speech classification or search using artificial neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/04—Time compression or expansion
- G10L21/055—Time compression or expansion for synchronising with other signals, e.g. video signals
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
- G10L2015/0631—Creating reference templates; Clustering
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Evolutionary Computation (AREA)
- Machine Translation (AREA)
Abstract
The invention relates to the technical field of artificial intelligence, in particular to an intelligent interaction method, an intelligent interaction system, intelligent interaction equipment and an intelligent interaction storage medium based on a large model, and aims to solve the technical problem of poor man-machine interaction experience caused by lack of higher-layer semantics and finer voice characteristic information of a lower layer. The voice information is extracted through the voice big model, the semantics of a higher layer are reserved, and meanwhile, the voice-text alignment module is used for obtaining the first text token feature sequence aligned with the audio feature sequence, so that the finer voice characteristics of a lower layer can be obtained during the processing of the language big model, the intention of a user is better identified, and the experience of human-computer interaction of the user is improved.
Description
Technical Field
The invention relates to the technical field of artificial intelligence, and particularly provides an intelligent interaction method, system, equipment and storage medium based on a large model.
Background
The main man-machine interaction system in the current stage is commonly completed by a plurality of algorithm modules such as ASR speech recognition, LLM large language model, TTS speech synthesis system and the like in combination with complex algorithm flow and exception handling logic development, and is widely applied to various intelligent man-machine interaction scenes. The ASR speech recognition algorithm module is used for recognizing the speech input of the person into text content; the LLM large language model receives the text output of the ASR module, and outputs a corresponding text response after user intention recognition and semantic understanding are carried out; and finally, the text content generated by the LLM large language model is sent to a TTS language synthesis module, and high-quality, natural and anthropomorphic voices are generated and played.
Although the current algorithm models formed by human-computer interaction are optimized and performance improved independently under respective modes, the method still faces several problems which are difficult to solve:
1. the ASR has poor self-adaptive capacity and robustness to scenes (far-near field and environmental noise), speakers (accent and speech speed), dialects and the like, so that the recognition effect is poor, and interaction failure is caused in the first step;
2. because the text semantic information contained in the training data of the ASR is less, the ASR model usually adopts an external NGRAM or RNN language model to improve the accuracy of transcription, but the external language model is still constructed based on the text content of a single language, and is only suitable for the language and the scene with characteristics, and has low universality;
3. the LLM large language model only has text single-mode information, and during interaction, auxiliary language information contained in voice is lacking, so that user intention cannot be understood correctly, generated content south-beam North rut is caused, and actual interaction experience is influenced;
4. the text front-end module for TTS voice synthesis converts the text into a correct reading method, pronunciation and rhythm pause, and a large amount of manpower is needed for maintenance and abnormal repair, but new word layers of the text are endless, the text content is rich and various, the prediction accuracy is greatly influenced, and the synthesis intelligibility and the naturalness are reduced;
5. the algorithm models are mutually independent, and the generated recognition errors are amplified in cascade in the interaction flow, so that the overall interaction effect is reduced.
Therefore, in the man-machine interaction scheme, due to the lack of higher-layer semantics and finer-layer voice characteristic information, the man-machine interaction experience is often poor, and the technical problem to be solved is needed.
Disclosure of Invention
The invention is proposed to overcome the above-mentioned drawbacks, and to solve or at least partially solve the technical problem of poor human-computer interaction experience caused by lack of higher-level semantics and lower-level fine voice characteristic information in the traditional human-computer interaction scheme.
In a first aspect, the present invention provides a large model-based intelligent interaction method, including:
extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module;
acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module;
a second text token feature sequence corresponding to the input prompt word text is obtained through a tokenization module;
acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence;
the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type.
In one embodiment, the tokenization module refers to encoding any piece of text into a sequence of text token features.
In one embodiment, the speech-text alignment module is configured to complete an alignment mapping of the audio feature sequence to the text token feature sequence.
In one embodiment, the first text token feature sequence and the second text token feature sequence are spliced to obtain a third text token feature sequence.
In one embodiment, the hint word type includes at least one of: the identification type and the audio output type.
In one embodiment, if the prompt word type is the recognition type, the third text token feature sequence is subjected to deserialization operation through the language big model, and text reply content is output.
In one embodiment, if the prompt word type is an audio output type, the fourth text token feature sequence is obtained after the third text token feature sequence is processed by the language big model.
In one embodiment, the audio feature sequence aligned with the fourth text token feature sequence is acquired by a text-to-speech alignment module; the text-voice alignment module is used for realizing alignment mapping from the text token feature sequence to the audio feature sequence.
In one embodiment, the target speech information is output based on the aligned audio feature sequence.
In one embodiment, the large speech model refers to a deep learning model trained from a large amount of speech data, wherein the training method comprises at least one of: the method is supervised, self-supervised and semi-supervised;
the language big model is a deep learning model trained by a large amount of text data; wherein the training method comprises at least one of the following: has supervision, self-supervision and semi-supervision.
In one embodiment, a frame-level high-dimensional feature sequence of input speech is extracted as an audio feature sequence by a speech large model.
In one embodiment, the audio feature sequence carries audio content information, audio emotion information, audio prosody information, audio voiceprint information, audio scene information, audio event information.
In a second aspect, the present invention provides a large model-based intelligent interaction system comprising: the device comprises a first acquisition module, an alignment module, a second acquisition module, a third acquisition module and a processing module;
the first acquisition module is used for extracting an audio feature sequence of the voice information through the voice big model and transmitting the audio feature sequence to the voice-text alignment module;
the alignment module is used for acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module;
the second acquisition module is used for acquiring a second text token feature sequence corresponding to the input prompt word text through the tokenization module;
the third acquisition module is used for acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence;
and the processing module is used for determining a processing strategy for the third text token feature sequence based on the prompt word type by the language big model.
In a third aspect, a computer device is provided, including a processor and a storage device, where the memory stores a program, and the processor implements the large model-based intelligent interaction method according to any one of the above solutions when executing the program.
In a fourth aspect, a computer readable storage medium is provided, in which a program is stored, the program when executed implementing the intelligent interaction method based on a large model according to any one of the above-mentioned methods.
The technical scheme provided by the invention has at least one or more of the following beneficial effects:
the technical scheme for implementing the invention comprises the following steps: extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module; acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module; a second text token feature sequence corresponding to the input prompt word text is obtained through a tokenization module; acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence; the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type. The voice information is extracted through the voice big model, the semantics of a higher layer are reserved, and meanwhile, the voice-text alignment module is used for obtaining the first text token feature sequence aligned with the audio feature sequence, so that the finer voice characteristics of a lower layer can be obtained during the processing of the language big model, the intention of a user is better identified, and the experience of human-computer interaction of the user is improved.
Further, the scheme can skip the intermediate results such as the recognition result of ASR and the front-end conversion of TTS text, is not limited by the accuracy of the two modules any more, and can directly perform the end-to-end voice-to-voice interaction, thereby integrally improving the interaction effect.
Further, through different prompt texts (prompt words), the end-to-end multi-mode human-computer interaction system can simultaneously perform tasks such as single-mode text interaction, ASR Speech recognition, voiceprint recognition, emotion recognition and the like, so that the tasks of the NLP text and the Speech are unified under the framework of the system.
Drawings
The present disclosure will become more readily understood with reference to the accompanying drawings. As will be readily appreciated by those skilled in the art: the drawings are for illustrative purposes only and are not intended to limit the scope of the present invention. Moreover, like numerals in the figures are used to designate like parts, wherein:
FIG. 1 is a flow diagram of the main steps of a large model-based intelligent interaction method according to one embodiment of the present invention;
FIG. 2 is a schematic block diagram of the main structure of a large model-based intelligent interaction system according to one embodiment of the present invention.
Detailed Description
Some embodiments of the invention are described below with reference to the accompanying drawings. It should be understood by those skilled in the art that these embodiments are merely for explaining the technical principles of the present invention, and are not intended to limit the scope of the present invention.
In the description of the present invention, a "module," "processor" may include hardware, software, or a combination of both. A module may comprise hardware circuitry, various suitable sensors, communication ports, memory, or software components, such as program code, or a combination of software and hardware. The processor may be a central processor, a microprocessor, an image processor, a digital signal processor, or any other suitable processor. The processor has data and/or signal processing functions. The processor may be implemented in software, hardware, or a combination of both. Non-transitory computer readable storage media include any suitable medium that can store program code, such as magnetic disks, hard disks, optical disks, flash memory, read-only memory, random access memory, and the like. The term "a and/or B" means all possible combinations of a and B, such as a alone, B alone or a and B. The term "at least one A or B" or "at least one of A and B" has a meaning similar to "A and/or B" and may include A alone, B alone or A and B. The singular forms "a", "an" and "the" include plural referents.
Noun interpretation:
large speech model: refers to a deep learning model trained with a large amount of speech data, wherein the training method comprises at least one of: the method is supervised, self-supervised and semi-supervised; generally speaking, a large speech model is a deep learning model with a large parameter scale, and the purpose of understanding rich speech bottom information such as speech content, emotion, scene/event, rhythm, voiceprint and the like is achieved by training on a large amount of speech data. The parameter scale is large, for example: billions of parameters or more, etc.
Language big model: a deep learning model trained by a large amount of text data; wherein the training method comprises at least one of the following: has supervision, self-supervision and semi-supervision. Generally speaking, a language large model refers to a deep learning model with a larger parameter scale, such as: billions of parameters or more, etc.
ASR speech recognition: refers to automatic continuous speech recognition, recognizing text information contained in a piece of audio.
TTS speech synthesis: refers to a text-to-speech synthesis system, which can synthesize any piece of text content into complete playable audio.
Token tokenization: refers to encoding text into a token sequence id that is employed in training a large language model.
A speech-text alignment module: for completing an alignment mapping of an audio feature sequence to a text token feature sequence, for example: and finishing the alignment mapping from the high-dimensional audio feature sequence to the text token feature sequence at the voice frame level.
And a tokenization module: refers to encoding any piece of text into a sequence of text token features.
Text-to-speech alignment module: for implementing an alignment mapping of text token feature sequences to audio feature sequences.
Vocoder module: and outputting complete voice audio based on the audio feature sequence output by the text-to-voice alignment module.
FIG. 1 is a flow diagram of the main steps of a large model-based intelligent interaction method according to one embodiment of the present invention; as shown in fig. 1, the method mainly comprises the following steps S101-S105:
step S101, extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module;
in this embodiment, the audio feature sequence of the voice information is extracted through the large voice model. Specifically: the input of the voice mode is original audio, and a high-dimensional characteristic sequence at a frame level is extracted through a voice big model to serve as an audio characteristic sequence, for example: T-D audio feature sequence, where "T" represents length and "D" represents dimension.
In this embodiment, the audio feature sequence carries audio content information, audio emotion information, audio prosody information, audio voiceprint information, audio scene information, and audio event information.
In this embodiment, the large speech model is a deep learning model trained by a large amount of speech data, where the training method includes at least one of the following: has supervision, self-supervision and semi-supervision.
Because the large voice model is trained by a large amount of voice data, the voice information is subjected to audio feature extraction by the large voice model to obtain higher-layer semantics (such as audio content information, audio emotion information, audio prosody information, audio voiceprint information, audio scene information and audio event information).
Step S102, obtaining a text token feature sequence aligned with the audio feature sequence through the voice-text alignment module;
in this embodiment, the voice-text alignment module is configured to complete alignment mapping from the audio feature sequence to the text token feature sequence. The voice-text alignment module is required to be independently set and trained and is used for receiving the audio feature sequence extracted from the voice large model and outputting the audio feature sequence as a corresponding text token feature sequence.
Specifically, based on the obtained t×d audio feature sequence, the text token feature sequence is converted into a t×1-dimensional text token feature sequence through alignment of a speech-text alignment module.
Because the voice big model performs audio feature extraction on the voice information to obtain higher-layer semantics, the aligned text token feature sequence also contains finer voice characteristics of a lower layer, so that the subsequent language big model can better recognize user intention, and the user man-machine interaction experience is improved.
Step S103, obtaining a second text token feature sequence corresponding to the input prompt word text through a tokenization module;
in this embodiment, the tokenizing module is configured to encode an arbitrary section of text into a token feature sequence.
Specifically, the text mode is input as a Prompt word, for example: { please identify text content in audio }, { please identify speaker in audio }, { please identify emotion in audio }, { please generate reply audio based on audio content }, { please directly reply text }, etc., after tokenization, a Length 1-dimensional text token feature sequence is obtained.
Step S104, a third text token feature sequence is obtained based on the first text token feature sequence and the second text token feature sequence;
in this embodiment, the first text token feature sequence and the second text token feature sequence are spliced to obtain a third text token feature sequence.
Specifically, a text token feature sequence in Length 1 dimension and a text token feature sequence in T1 dimension are spliced to obtain a text token feature sequence in length+t 1 dimension.
Step S105, the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type.
In this embodiment, the language big model refers to a deep learning model trained by a large amount of text data; wherein the training method comprises at least one of the following: has supervision, self-supervision and semi-supervision.
In this embodiment, the language big model determines a prompt word type based on the prompt word text association information, where the prompt word type includes at least one of the following: the identification type and the audio output type.
Specifically, the language big model determines a processing strategy for a text token feature sequence of "(length+t) x 1 dimension based on a Prompt type (Prompt word type);
if the prompt word type is an identification type (for example { please identify emotion in audio }, { please directly reply text }), performing anti-serialization operation on the text token feature sequence of (length+T) ×1 dimension through the language big model, and outputting text reply content.
If the prompt word type is an audio output type (for example { please generate reply audio based on audio content }), processing a text token feature sequence of (length+T) ×1 dimension through the language big model, and then obtaining a fourth text token feature sequence;
acquiring an audio feature sequence aligned with the fourth text token feature sequence through a text-to-speech alignment module; the text-voice alignment module is used for realizing alignment mapping from the text token feature sequence to the audio feature sequence. For example, by a text-to-speech alignment module, a T1-dimensional audio feature sequence aligned with the fourth text token feature sequence is obtained.
The text-voice alignment module is required to be independently set and trained and is used for receiving a text token feature sequence output by the language big model and outputting the text token feature sequence as an audio feature sequence.
Through the alignment of the text and the voice, correct pronunciation and prosody sequence prediction can be completed directly based on priori knowledge of a language big model, and a large amount of manual maintenance of a front-end text regularization, pronunciation and prosody prediction model is not needed.
In this embodiment, the target voice information is output based on the aligned audio feature sequence.
Specifically, based on the aligned T1-dimensional audio feature sequence, target voice information is output through a vocoder.
In this embodiment, the first text token feature sequence, the second text token feature sequence, and the third text token feature sequence are all presented in the form of a token sequence.
In the field of natural language processing, machine learning models typically have Token as their input unit. Token may be understood as the smallest unit in text where Token may be a word, number, punctuation, single letter, or any single element that can be a text analysis.
By fusion alignment of the end-to-end voice big model and the language big model, information sharing among modes is realized, so that end-to-end output of voice-to-voice is completed, an intermediate result of ASR recognition is not needed in the interaction process, and cascade errors are reduced.
The technical scheme provided by the invention has at least one or more of the following beneficial effects:
the technical scheme for implementing the invention comprises the following steps: extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module; acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module; a second text token feature sequence corresponding to the input prompt word text is obtained through a tokenization module; acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence; the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type. The voice information is extracted through the voice big model, the semantics of a higher layer are reserved, and meanwhile, the voice-text alignment module is used for obtaining the first text token feature sequence aligned with the audio feature sequence, so that the finer voice characteristics of a lower layer can be obtained during the processing of the language big model, the intention of a user is better identified, and the experience of human-computer interaction of the user is improved.
Further, the scheme can skip the intermediate results such as the recognition result of ASR and the front-end conversion of TTS text, is not limited by the accuracy of the two modules any more, and can directly perform the end-to-end voice-to-voice interaction, thereby integrally improving the interaction effect.
Further, through different prompt texts (prompt words), the end-to-end multi-mode human-computer interaction system can simultaneously perform tasks such as single-mode text interaction, ASR Speech recognition, voiceprint recognition, emotion recognition and the like, so that the tasks of the NLP text and the Speech are unified under the framework of the system.
The invention also provides a main structure block diagram schematic diagram of the intelligent interaction system based on the large model of one embodiment. As shown in fig. 2, the intelligent interaction system based on the large model according to one embodiment of the present invention mainly includes: a first acquisition module 201, an alignment module 202, a second acquisition module 203, a third acquisition module 204, a processing module 205;
a first obtaining module 201, configured to extract an audio feature sequence of the voice information through the large voice model and transmit the audio feature sequence to a voice-text alignment module;
an alignment module 202, configured to obtain, by using the speech-text alignment module, a first text token feature sequence aligned with the audio feature sequence;
a second obtaining module 203, configured to obtain, by using a tokenizing module, a second text token feature sequence corresponding to the input prompt word text;
a third obtaining module 204, configured to obtain a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence;
a processing module 205, configured to determine, by using the language big model, a processing policy for the third text token feature sequence based on the prompt word type.
In some embodiments, one or more of the first acquisition module 201, the alignment module 202, the second acquisition module 203, the third acquisition module 204, and the processing module 205 may be combined together into one module. In one embodiment, the specific implementation functions may be described with reference to the steps of the above examples.
The technical scheme provided by the invention has at least one or more of the following beneficial effects:
the technical scheme for implementing the invention comprises the following steps: extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module; acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module; a second text token feature sequence corresponding to the input prompt word text is obtained through a tokenization module; acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence; the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type. The voice information is extracted through the voice big model, the semantics of a higher layer are reserved, and meanwhile, the voice-text alignment module is used for obtaining the first text token feature sequence aligned with the audio feature sequence, so that the finer voice characteristics of a lower layer can be obtained during the processing of the language big model, the intention of a user is better identified, and the experience of human-computer interaction of the user is improved.
Further, the scheme can skip the intermediate results such as the recognition result of ASR and the front-end conversion of TTS text, is not limited by the accuracy of the two modules any more, and can directly perform the end-to-end voice-to-voice interaction, thereby integrally improving the interaction effect.
Further, through different prompt texts (prompt words), the end-to-end multi-mode human-computer interaction system can simultaneously perform tasks such as single-mode text interaction, ASR Speech recognition, voiceprint recognition, emotion recognition and the like, so that the tasks of the NLP text and the Speech are unified under the framework of the system.
The foregoing big model-based intelligent interaction system is used for executing the big model-based intelligent interaction method embodiment shown in fig. 2, and the technical principles of the two are similar, the technical problems to be solved and the technical effects to be produced are similar, and those skilled in the art can clearly understand that, for convenience and brevity of description, the specific working process and related description of the big model-based intelligent interaction system may refer to what is described in the big model-based intelligent interaction method embodiment, and will not be repeated herein.
It will be appreciated by those skilled in the art that the present invention may implement all or part of the above-described methods according to the above-described embodiments, or may be implemented by means of a computer program for instructing relevant hardware, where the computer program may be stored in a computer readable storage medium, and where the computer program may implement the steps of the above-described embodiments of the method when executed by a processor. Wherein the computer program comprises computer program code which may be in source code form, object code form, executable file or some intermediate form etc. The computer readable storage medium may include: any entity or device, medium, usb disk, removable hard disk, magnetic disk, optical disk, computer memory, read-only memory, random access memory, electrical carrier wave signals, telecommunications signals, software distribution media, and the like capable of carrying the computer program code. It should be noted that the computer readable storage medium may include content that is subject to appropriate increases and decreases as required by jurisdictions and by jurisdictions in which such computer readable storage medium does not include electrical carrier signals and telecommunications signals.
Further, the present invention also provides a computer apparatus, which includes a processor and a storage device, where the storage device may be configured to store a program for executing the large model-based intelligent interaction method of the above method embodiment, and the processor may be configured to execute the program in the storage device, where the program includes, but is not limited to, a program for executing the large model-based intelligent interaction method of the above method embodiment. For convenience of explanation, only those portions of the embodiments of the present invention that are relevant to the embodiments of the present invention are shown, and specific technical details are not disclosed, please refer to the method portions of the embodiments of the present invention. The computer device may be a computer device formed including various electronic devices.
Further, the invention also provides a computer readable storage medium. In one computer-readable storage medium embodiment according to the present invention, the computer-readable storage medium may be configured to store a program for performing the large model-based intelligent interaction method of the above-described method embodiment, which may be loaded and executed by a processor to implement the large model-based intelligent interaction method described above. For convenience of explanation, only those portions of the embodiments of the present invention that are relevant to the embodiments of the present invention are shown, and specific technical details are not disclosed, please refer to the method portions of the embodiments of the present invention. The computer readable storage medium may be a storage device including various electronic devices, and optionally, the computer readable storage medium in the embodiments of the present invention is a non-transitory computer readable storage medium.
Further, it should be understood that, since the respective modules are merely set for illustrating the functional units of the system of the present invention, the physical devices corresponding to the modules may be the processor itself, or a part of software in the processor, a part of hardware, or a part of a combination of software and hardware. Accordingly, the number of individual modules in the figures is merely illustrative.
Those skilled in the art will appreciate that the various modules in the system may be adaptively split or combined. Such splitting or combining of specific modules does not cause the technical solution to deviate from the principle of the present invention, and therefore, the technical solution after splitting or combining falls within the protection scope of the present invention.
Thus far, the technical solution of the present invention has been described in connection with the preferred embodiments shown in the drawings, but it is easily understood by those skilled in the art that the scope of protection of the present invention is not limited to these specific embodiments. Equivalent modifications and substitutions for related technical features may be made by those skilled in the art without departing from the principles of the present invention, and such modifications and substitutions will be within the scope of the present invention.
Claims (15)
1. An intelligent interaction method based on a large model is characterized by comprising the following steps:
extracting an audio feature sequence of voice information through a voice big model and transmitting the audio feature sequence to a voice-text alignment module;
acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module;
a second text token feature sequence corresponding to the input prompt word text is obtained through a tokenization module;
acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence;
the language big model determines a processing strategy for the third text token feature sequence based on the prompt word type.
2. The method of claim 1, wherein the tokenizing module encodes any piece of text into a sequence of token characteristics.
3. The method of claim 1, wherein the speech-to-text alignment module is configured to perform an alignment mapping of the audio feature sequence to the text token feature sequence.
4. The method of claim 1, wherein the first text token feature sequence and the second text token feature sequence are spliced to obtain a third text token feature sequence.
5. The method of claim 1, wherein the hint word type includes at least one of: the identification type and the audio output type.
6. The method of claim 5, wherein if the prompt word type is an identification type, performing a deserialization operation on the third text token feature sequence through the language big model, and outputting text reply content.
7. The method of claim 5, wherein if the prompt word type is an audio output type, then processing the third text token feature sequence through the language big model to obtain a fourth text token feature sequence.
8. The method of claim 7, wherein the audio feature sequence aligned with the fourth text token feature sequence is obtained by a text-to-speech alignment module; the text-voice alignment module is used for realizing alignment mapping from the text token feature sequence to the audio feature sequence.
9. The method of claim 8, wherein target speech information is output based on the aligned audio feature sequence.
10. The method of claim 1, wherein the large speech model is a deep learning model trained from a large amount of speech data, wherein the training method comprises at least one of: the method is supervised, self-supervised and semi-supervised;
the language big model is a deep learning model trained by a large amount of text data; wherein the training method comprises at least one of the following: has supervision, self-supervision and semi-supervision.
11. The method according to claim 1 or 10, characterized in that the frame-level high-dimensional feature sequence of the input speech is extracted as an audio feature sequence by a speech large model.
12. The method of claim 11, wherein the sequence of audio features carries audio content information, audio emotion information, audio prosody information, audio voiceprint information, audio scene information, audio event information.
13. An intelligent interactive system based on a large model, comprising: the device comprises a first acquisition module, an alignment module, a second acquisition module, a third acquisition module and a processing module;
the first acquisition module is used for extracting an audio feature sequence of the voice information through the voice big model and transmitting the audio feature sequence to the voice-text alignment module;
the alignment module is used for acquiring a first text token feature sequence aligned with the audio feature sequence through the voice-text alignment module;
the second acquisition module is used for acquiring a second text token feature sequence corresponding to the input prompt word text through the tokenization module;
the third acquisition module is used for acquiring a third text token feature sequence based on the first text token feature sequence and the second text token feature sequence;
and the processing module is used for determining a processing strategy for the third text token feature sequence based on the prompt word type by the language big model.
14. A computer device comprising a processor and a memory, wherein the memory has a program stored therein, characterized in that the processor implements the method of any of claims 1 to 12 when executing the program.
15. A computer readable storage medium storing a program, characterized in that the program when executed implements the method of any one of claims 1 to 12.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410104210.9A CN117854492A (en) | 2024-01-24 | 2024-01-24 | Intelligent interaction method, system, equipment and storage medium based on large model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202410104210.9A CN117854492A (en) | 2024-01-24 | 2024-01-24 | Intelligent interaction method, system, equipment and storage medium based on large model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117854492A true CN117854492A (en) | 2024-04-09 |
Family
ID=90534630
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202410104210.9A Pending CN117854492A (en) | 2024-01-24 | 2024-01-24 | Intelligent interaction method, system, equipment and storage medium based on large model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117854492A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118098274A (en) * | 2024-04-19 | 2024-05-28 | 腾讯科技(深圳)有限公司 | Model training method and device, electronic equipment and storage medium |
-
2024
- 2024-01-24 CN CN202410104210.9A patent/CN117854492A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN118098274A (en) * | 2024-04-19 | 2024-05-28 | 腾讯科技(深圳)有限公司 | Model training method and device, electronic equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11295721B2 (en) | Generating expressive speech audio from text data | |
WO2022083083A1 (en) | Sound conversion system and training method for same | |
Bell et al. | Prosodic adaptation in human-computer interaction | |
WO2017076222A1 (en) | Speech recognition method and apparatus | |
CN116364055B (en) | Speech generation method, device, equipment and medium based on pre-training language model | |
CN111667812A (en) | Voice synthesis method, device, equipment and storage medium | |
CN111477216A (en) | Training method and system for pronunciation understanding model of conversation robot | |
CN117854492A (en) | Intelligent interaction method, system, equipment and storage medium based on large model | |
CN112581963A (en) | Voice intention recognition method and system | |
CN112735404A (en) | Ironic detection method, system, terminal device and storage medium | |
CN113744727A (en) | Model training method, system, terminal device and storage medium | |
CN114783424A (en) | Text corpus screening method, device, equipment and storage medium | |
CN114974218A (en) | Voice conversion model training method and device and voice conversion method and device | |
CN116978381A (en) | Audio data processing method, device, computer equipment and storage medium | |
CN113628608A (en) | Voice generation method and device, electronic equipment and readable storage medium | |
CN115966197A (en) | Speech synthesis method, speech synthesis device, electronic equipment and storage medium | |
EP4068279B1 (en) | Method and system for performing domain adaptation of end-to-end automatic speech recognition model | |
CN113299270B (en) | Method, device, equipment and storage medium for generating voice synthesis system | |
CN113257225B (en) | Emotional voice synthesis method and system fusing vocabulary and phoneme pronunciation characteristics | |
CN113160801B (en) | Speech recognition method, device and computer readable storage medium | |
CN115547345A (en) | Voiceprint recognition model training and related recognition method, electronic device and storage medium | |
CN114170997A (en) | Pronunciation skill detection method, pronunciation skill detection device, storage medium and electronic equipment | |
CN114299989A (en) | Voice filtering method and device, electronic equipment and storage medium | |
CN113920997A (en) | Voice awakening method and device, electronic equipment and operating machine | |
CN115424616A (en) | Audio data screening method, device, equipment and computer readable medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |