CN111862980A - Incremental semantic processing method - Google Patents

Incremental semantic processing method Download PDF

Info

Publication number
CN111862980A
CN111862980A CN202010787458.1A CN202010787458A CN111862980A CN 111862980 A CN111862980 A CN 111862980A CN 202010787458 A CN202010787458 A CN 202010787458A CN 111862980 A CN111862980 A CN 111862980A
Authority
CN
China
Prior art keywords
semantic feature
semantic
voice
module
text information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010787458.1A
Other languages
Chinese (zh)
Inventor
蔡勇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Zebra Network Technology Co Ltd
Original Assignee
Zebra Network Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Zebra Network Technology Co Ltd filed Critical Zebra Network Technology Co Ltd
Priority to CN202010787458.1A priority Critical patent/CN111862980A/en
Publication of CN111862980A publication Critical patent/CN111862980A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The invention provides an increment semantic processing method, which is applied to an electronic device, wherein the electronic device comprises a voice detection module, a voice recognition module and a semantic processing module; the voice recognition module accumulates the received voice signals according to the time sequence in a segmentation mode and converts each segment of voice signals into text information in real time; the voice recognition module sequentially sends the text information of each voice signal segment to the semantic processing module according to the time sequence, and the semantic processing module receives the text information of each voice signal segment in real time, wherein when the semantic processing module receives the text information of each voice signal segment, the semantic feature value of the text information of the voice signal segment is extracted, so that the voice signal is processed in parallel with the voice recognition module. According to the embodiment of the invention, the delay time in man-machine language interaction can be reduced, and the user experience is improved.

Description

Incremental semantic processing method
Technical Field
The application relates to the technical field of natural language processing, in particular to an incremental semantic processing method.
Background
At present, all man-machine voice interaction is divided into three parts of voice recognition, semantic processing and application program. Firstly, the speech recognition part judges the end of the user speaking according to the speech breakpoint detection, then all the characters are sent to the semantic processing engine for processing, and then the application program module executes the subsequent action according to the semantic processing result. The three links are all executed in series, each link causes certain time delay, and bad experience is brought to the use of a user. With several types of automobile product experience in the current mainstream, as shown in fig. 4, the voice detection loop is saved by 500ms, the semantic processing loop is saved by 200ms, and the application program module processing loop is saved by 500 and 1000 ms. The three links add up with a delay of about 1-3 s.
Disclosure of Invention
In view of this, the present application provides an incremental semantic processing method, which can reduce the delay time during human-computer language interaction, quickly understand the voice signal content of a user, give feedback, and improve the voice interaction experience of the user.
In order to solve the technical problem, the following technical scheme is adopted in the application:
in a first aspect, the present application provides an incremental semantic processing method applied to an electronic device, where the electronic device includes a voice detection module, a voice recognition module, and a semantic processing module, and the method includes:
when the voice detection module detects a voice signal, informing the voice recognition module to process the voice signal;
the voice recognition module converts the received voice signals into segmented text information in real time according to a time sequence;
the voice recognition module sequentially sends the text information of each voice signal segment to the semantic processing module according to the time sequence, the semantic processing module receives the text information of each voice signal segment in real time, and when the semantic processing module receives the text information of each voice signal segment, the semantic feature value of the text information of the voice signal segment is extracted, so that the voice signal is processed in parallel with the voice recognition module.
As an embodiment of the first aspect of the present application, the method further comprises:
the semantic processing module sequentially accumulates the semantic feature values corresponding to each section of text information according to a time sequence, scores the semantic feature values obtained for the first time or scores the semantic feature values accumulated based on the semantic feature values obtained for the first time, and sends the semantic feature values or the accumulated semantic feature values to the application program module when the scores of the semantic feature values or the accumulated semantic feature values are larger than a threshold value;
and the application program module receives the semantic feature value corresponding to each section of voice signal in real time and adjusts the output content according to the received semantic feature value each time.
As an embodiment of the first aspect of the present application, the receiving, by the application program module, a semantic feature value corresponding to each segment of speech signal in real time, and adjusting output content according to the received semantic feature value each time includes:
and after receiving the semantic feature values, the application program module obtains a command corresponding to the semantic feature values according to semantic feature value matching, and switches to an application interface for executing the command so as to output interface contents corresponding to the semantic feature values through the application interface.
As an embodiment of the first aspect of the present application, when the voice detection module detects that the voice signal is ended, the application program module executes a command corresponding to a semantic feature value of a complete segment of the voice signal, where the complete segment of the voice signal is the voice signal within a period from when the detection module starts timing the voice signal to when the voice signal is ended.
As an embodiment of the first aspect of the present application, the semantic feature value includes a unique primary key value, and the primary key value is used to identify the first intention of the voice signal.
As an embodiment of the first aspect of the present application, the semantic feature value may further include N secondary key values (N ≧ 0), which are used to further define the range of the first intent of the primary key value.
As an embodiment of the first aspect of the present application, the primary key value and the secondary key value of the voice information value are updated in real time according to the voice signal.
As an embodiment of the first aspect of the present application, the finest granularity information of the text information is a single chinese character or a single word.
As an embodiment of the first aspect of the present application, the storage type of the semantic feature value is a json data type.
As an embodiment of the first aspect of the present application, the semantic processing module 103 employs any one of LSTM, CNN, or transform model.
The technical scheme of the application has at least one of the following beneficial effects:
according to the incremental semantic processing method, the processing time of the voice signals can be shortened, the voice signal content of the user can be understood quickly, feedback is given, delay in the voice interaction process is reduced, and the voice interaction experience of the user is improved.
Drawings
FIG. 1 is a block diagram of an incremental semantic processing method according to an embodiment of the present disclosure;
FIG. 2 is a flowchart of an incremental semantic processing method according to an embodiment of the present application;
FIG. 3 is a schematic diagram of the runtime of each module according to an embodiment of the present application;
FIG. 4 is a schematic diagram of the runtime of various modules of the prior art;
fig. 5 is a system flowchart of an incremental semantic processing method according to an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The following description will discuss embodiments of the present application with reference to specific examples.
The embodiment of the application is mainly applied to all human-computer voice interaction scenes such as automobiles and home furnishing, and as shown in fig. 1, the electronic device comprises a voice detection module 101, a voice recognition module 102, a semantic processing module 103 and an application program module 200, and the modules run in parallel, so that the processing time of a human-computer interaction process can be effectively reduced. For example, a user wants to listen to a song, say to an electronic device: "i want to listen to liu de hua forgetting water", when the voice detection module 101 recognizes the voice signal of the user, the voice recognition module 102 processes the voice signal in real time, for example, the user says "i want to listen to liu de hua forgetting water", the real-time processing result may be that "i want to listen to" liu de hua "," forgetting water ", and such segmented text information, further, the voice recognition module 102 sequentially sends each piece of text information to the semantic processing module 103 according to a time sequence, the semantic processing module 103 sequentially receives the contents in the above mentioned double quotations, and extracts semantic feature values in the segmented text information, in the practical application of the embodiment of the present application, an intent may be named by creating a json object for storing semantic feature values, for example, when the semantic processing module 103 receives" i want to listen to "text information, it may be determined that the semantic feature value intent is music playing, is marked as { "intent": "palette _ music" }, further, the semantic processing module 103 will score the feature value, and when the score is greater than the threshold, send the semantic feature value to the application module 200, for example, { "intent": "palette _ music" } is sent to the application module 200, and the application module 200 displays the music playing interface. That is, as shown in fig. 3, from the beginning of human speech, the speech recognition module 102, the semantic processing module 103 and the application module 200 process speech signals in parallel, generally, the time of human speech is at least 1 second, while the time of semantic processing module 103 and the application module 200 are mostly between several tens milliseconds and several hundreds milliseconds, when the speech detection module 101 detects the end of speech signal, it takes time t1, the speech recognition module recognizes speech signal and sends text information converted from speech signal to the parallel semantic processing module 103 in real time, the semantic processing module 102 finishes working, delay t2, the semantic processing module 103 transmits the result to the application module 200 in real time, the application module 200 finishes working, delay t3, in practical application, the user performs command from the end of speech to the application module 200, the total delay Td is t1+ t2+ t3, while 1 second of the speaking time of the user is enough for the semantic processing module 103 and the application program module 200 to finish the operation, t2 and t3 are only the communication time among the speech recognition module, the semantic processing module 103 and the application program module 200, and are about tens of milliseconds, so that the overall delay time Td is far shorter than the delay time of t1+ t2+ t3 in the prior art shown in fig. 4, and thus, in the human-computer speech interaction scene, the method can shorten the processing time of the speech signal, quickly understand the speech signal content of the user, give feedback, reduce the delay time in the speech interaction process, and improve the speech interaction experience of the user.
An incremental semantic processing method according to the present application is described below with reference to the accompanying drawings, and fig. 2 shows a flowchart of an incremental semantic processing method, where the method is applied to an electronic device, the electronic device includes a voice detection module, a voice recognition module, and a semantic processing module, and as shown in fig. 2, the method includes:
step S210, when the voice detection module detects the voice signal, the voice recognition module is notified to process the voice signal. That is, the voice detection module detects the voice signal first, and when the user is judged to speak, the voice recognition module is immediately informed to process the voice signal.
In step S220, the speech recognition module converts the received speech signal into segmented text information in real time according to the time sequence. That is, the speech processing module first processes the speech signal in chronological order to convert the speech signal into text information in real time, wherein the speech processing module typically converts the speech signal into segmented text information, for example, a user says: in some embodiments of the present application, the result of the finest granularity of the text information converted by the speech processing module is "i", "want", "listen", "Liu", "De", "Hua", "forget", "Emotion" and "Water".
Step S230, the voice recognition module sequentially sends the text information of each segment of voice signal to the semantic processing module according to the time sequence, and the semantic processing module receives the text information of each segment of voice signal in real time, wherein when the semantic processing module receives each segment of text information of a voice signal, the semantic feature value of the text information of the segment of voice signal is extracted to implement parallel processing of the voice signal with the voice recognition module. That is to say, through the previous step S220, segmented text information is obtained, the speech recognition module sends each piece of text information to the semantic processing module according to the time sequence, the semantic processing module extracts a semantic feature value of the text information, for example, when the semantic processing module receives "i want to listen" in real time, a semantic feature value play _ music may be extracted, in the practical application of the embodiment of the present application, a json object is created, named intent, and used for storing the semantic feature value, and the semantic purpose { "intent": "palette _ music" } identification.
In some embodiments of the present application, the method further comprises:
step S240, the semantic processing module sequentially accumulates the semantic feature values corresponding to each section of text information according to a time sequence, and scores the semantic feature values obtained for the first time or the semantic feature values accumulated based on the semantic feature values obtained for the first time, and when the score of the semantic feature value or the semantic feature value accumulated is greater than a threshold value, the semantic processing module sends the semantic feature value or the semantic feature value accumulated to the application program module. That is, further, when the semantic processing module continues to receive the "liu de hua" text message, the semantic feature value "singer" may be extracted as "liu de hua", and added to the above semantics, and the semantic feature value may be expressed as { "intent": "play _ music", "singer": the method includes the steps of marking by Liu Dehua, further, when a semantic processing module continues to receive text information of 'forgetting to do water', extracting a semantic feature value 'song' as 'forgetting to do water', accumulating the semantic feature value 'song' into the semantic meaning, and using { 'intent': "play _ music", "singer": "Liu De Hua", "song": "forget to water" }.
As shown in fig. 4, each time the semantic processing module extracts a semantic feature value, the semantic feature value may be scored, for example, when the text information received by the semantic processing module is "i want to hear", the semantic feature value extracted by the semantic processing module is { "intent": "palette _ music", but this semantic is less complete, so the semantic feature value score does not exceed the threshold, and therefore this semantic is not passed to the application module. When the semantic processing module receives the "liu de hua" text information, the semantic processing module extracts "singer" ═ liu de hua ", and then the semantic processing module pairs {" intent ": "play _ music", "singer": the semantic score of "Liu Dehua" will be greater than the threshold, at which time the semantic processing module will: "play _ music", "singer": "Liu De Hua" } to the application module.
And the application program module receives the semantic feature value corresponding to each section of voice signal in real time and adjusts the output content according to the received semantic feature value each time. That is, when the application module receives the semantic feature value { "intent": "play _ music", "singer": "Liu De Hua", will show the interface of preparing to play Liu De Hua's song at the application program interface.
In some embodiments of the present application, the receiving, by the application program module, a semantic feature value corresponding to each segment of the speech signal in real time, and adjusting the output content according to the received semantic feature value each time includes:
after receiving the semantic feature values, the application program module obtains commands corresponding to the semantic feature values according to semantic feature value matching, and switches to an application interface of the commands to be executed so as to output interface contents corresponding to the semantic feature values through the application interface. Wherein, when the semantic feature value received by the application program module is { "intent": "play _ music", "singer": when the song is played, the application module displays the interface of the song to be played in the application interface according to the semantic, but the song name is not available, so that the song to be played is the other song of the song to be played in the Liu De Hua. When the semantic processing module continues to extract the semantic feature value "song" - "forgetting water", obtaining { "intent": "play _ music", "singer": "Liu De Hua", "song": and the score of the semantics exceeds a threshold value, and when the score is sent to the application program module, the application program module changes the interface for playing the music from any song for playing Liudebua to the forgetting water for playing Liudebua.
Step S250, when the voice detection module detects that the voice signal is ended, the application program module executes a command corresponding to the semantic feature value of the complete segment of voice signal, where the complete segment of voice signal is the voice signal within a period from when the detection module starts timing from the detection of the voice signal to when the voice signal is ended. That is, when the voice detection module detects the end of the voice signal, the application module starts to execute the command, for example, in step S250, the interface of the application module is already at the playing interface of the liudeluxe forgetting water, and when the voice detection module detects the end of the voice signal when the user stops speaking, the application module starts to play the liudeluxe forgetting water.
In some embodiments of the present application, the semantic feature value includes a unique primary key value, the primary key value identifying a first intent of the speech signal. For example, in the above-described embodiment, "intent" is a primary key value of the semantic feature value, and thus, it can be known that the user wants to listen to music.
In some embodiments of the present application, the semantic feature values may also include N secondary key values (N ≧ 0) used to further define the range of first intent of the primary key value. For example, in the above-described embodiment, "singer" is "liu de hua" and "song" is "forgetful water" is an auxiliary key value of the semantic feature value, and thus, it is possible to know that the user specifically wants to listen to the song forgetful water of liu de hua.
In some embodiments of the present application, the semantic processing module employs any one of an LSTM, CNN, or Transformer model, where the LSTM, CNN, or Transformer model is a mature neural model and is easy to train, and the semantic processing module can quickly and accurately extract semantic feature values through the LSTM, CNN, or Transformer model.
Therefore, the incremental semantic processing method can be used for paralleling the voice recognition module, the semantic processing module and the application program module, shortening the processing time of voice signals, quickly understanding the voice signal content of a user, giving feedback, reducing the time delay in the voice interaction process and improving the voice interaction experience of the user.
It is noted that, in the examples and descriptions of this patent, relational terms such as first and second, and the like are used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or method that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or method. Without further limitation, the use of the verb "comprise a" to define an element does not exclude the presence of another, same element in a process, method, article, or method that comprises the element.
The foregoing is a preferred embodiment of the present application, and it should be noted that, for those skilled in the art, several modifications and refinements can be made without departing from the principle described in the present application, and these modifications and refinements should be regarded as the protection scope of the present application.

Claims (10)

1. An incremental semantic processing method is applied to an electronic device, and is characterized in that the electronic device comprises a voice detection module, a voice recognition module and a semantic processing module, and the method comprises the following steps:
when the voice detection module detects a voice signal, informing the voice recognition module to process the voice signal;
the voice recognition module converts the voice signals received according to the time sequence into segmented text information in real time;
the voice recognition module sequentially sends the text information of each section of the voice signals to the semantic processing module according to the time sequence, the semantic processing module receives the text information of each section of the voice signals in real time, and when the semantic processing module receives the text information of each section of the voice signals, the semantic feature value of the text information of each section of the voice signals is extracted, so that the voice signals are processed in parallel with the voice recognition module.
2. The method of claim 1, further comprising:
the semantic processing module sequentially accumulates the semantic feature values corresponding to each section of text information according to a time sequence, scores the semantic feature values obtained for the first time or scores the semantic feature values accumulated based on the semantic feature values obtained for the first time, and sends the semantic feature values or the accumulated semantic feature values to the application program module when the scores of the semantic feature values or the accumulated semantic feature values are larger than a threshold value;
and the application program module receives the semantic feature value corresponding to each section of voice signal in real time and adjusts the output content according to the received semantic feature value each time.
3. The method of claim 2, wherein the receiving, by the application module, the semantic feature value corresponding to each segment of the speech signal in real time, and adjusting the output content according to the semantic feature value received each time comprises:
and after receiving the semantic feature values, the application program module obtains a command corresponding to the semantic feature values according to semantic feature value matching, and switches to an application interface for executing the command so as to output interface contents corresponding to the semantic feature values through the application interface.
4. The method according to any of claims 1-3, wherein when the speech detection module detects the end of the speech signal, the application module executes a command corresponding to a semantic feature value of a complete segment of the speech signal, the complete segment of the speech signal being the speech signal within a period of time from when the detection module started timing the detection of the speech signal to when the speech signal ended.
5. The method of claim 1, wherein the semantic feature value comprises a unique primary key value, and wherein the primary key value is used to identify the first intent of the speech signal.
6. The method of claim 5, wherein the semantic feature value further comprises N secondary key values (N ≧ 0) used to further define the range of the first intent of the primary key value.
7. The method of claim 6, wherein the primary key value and the secondary key value of the voice information value are updated in real time according to the voice signal.
8. The method of claim 1, wherein the finest granularity information of the text information is a single chinese character or a single word.
9. The method of claim 1, wherein the storage type of the semantic feature value is a json data type.
10. The method of claim 1, wherein the semantic processing module 103 employs any one of LSTM, CNN or Transformer models.
CN202010787458.1A 2020-08-07 2020-08-07 Incremental semantic processing method Pending CN111862980A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010787458.1A CN111862980A (en) 2020-08-07 2020-08-07 Incremental semantic processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010787458.1A CN111862980A (en) 2020-08-07 2020-08-07 Incremental semantic processing method

Publications (1)

Publication Number Publication Date
CN111862980A true CN111862980A (en) 2020-10-30

Family

ID=72971028

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010787458.1A Pending CN111862980A (en) 2020-08-07 2020-08-07 Incremental semantic processing method

Country Status (1)

Country Link
CN (1) CN111862980A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112700769A (en) * 2020-12-26 2021-04-23 科大讯飞股份有限公司 Semantic understanding method, device, equipment and computer readable storage medium

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0543329A2 (en) * 1991-11-18 1993-05-26 Kabushiki Kaisha Toshiba Speech dialogue system for facilitating human-computer interaction
JP2001013992A (en) * 1999-07-02 2001-01-19 Nec Corp Voice understanding device
US20050114132A1 (en) * 2003-11-21 2005-05-26 Acer Inc. Voice interactive method and system
WO2017084360A1 (en) * 2015-11-17 2017-05-26 乐视控股(北京)有限公司 Method and system for speech recognition
CN107305541A (en) * 2016-04-20 2017-10-31 科大讯飞股份有限公司 Speech recognition text segmentation method and device
CN107665706A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Rapid Speech exchange method and system
US10102851B1 (en) * 2013-08-28 2018-10-16 Amazon Technologies, Inc. Incremental utterance processing and semantic stability determination
US20180308487A1 (en) * 2017-04-21 2018-10-25 Go-Vivace Inc. Dialogue System Incorporating Unique Speech to Text Conversion Method for Meaningful Dialogue Response

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0543329A2 (en) * 1991-11-18 1993-05-26 Kabushiki Kaisha Toshiba Speech dialogue system for facilitating human-computer interaction
JP2001013992A (en) * 1999-07-02 2001-01-19 Nec Corp Voice understanding device
US20050114132A1 (en) * 2003-11-21 2005-05-26 Acer Inc. Voice interactive method and system
US10102851B1 (en) * 2013-08-28 2018-10-16 Amazon Technologies, Inc. Incremental utterance processing and semantic stability determination
WO2017084360A1 (en) * 2015-11-17 2017-05-26 乐视控股(北京)有限公司 Method and system for speech recognition
CN107305541A (en) * 2016-04-20 2017-10-31 科大讯飞股份有限公司 Speech recognition text segmentation method and device
CN107665706A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Rapid Speech exchange method and system
US20180308487A1 (en) * 2017-04-21 2018-10-25 Go-Vivace Inc. Dialogue System Incorporating Unique Speech to Text Conversion Method for Meaningful Dialogue Response

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112700769A (en) * 2020-12-26 2021-04-23 科大讯飞股份有限公司 Semantic understanding method, device, equipment and computer readable storage medium

Similar Documents

Publication Publication Date Title
CN112100349A (en) Multi-turn dialogue method and device, electronic equipment and storage medium
CN111326154B (en) Voice interaction method and device, storage medium and electronic equipment
CN101447185A (en) Audio frequency rapid classification method based on content
CN111161726B (en) Intelligent voice interaction method, device, medium and system
CN110673821B (en) Intelligent device awakening feedback method and intelligent device
CN108766431B (en) Automatic awakening method based on voice recognition and electronic equipment
CN111583912A (en) Voice endpoint detection method and device and electronic equipment
CN111261151A (en) Voice processing method and device, electronic equipment and storage medium
CN108595406B (en) User state reminding method and device, electronic equipment and storage medium
CN111081254B (en) Voice recognition method and device
CN111862980A (en) Incremental semantic processing method
CN110956958A (en) Searching method, searching device, terminal equipment and storage medium
CN111048068B (en) Voice wake-up method, device and system and electronic equipment
CN113593565B (en) Intelligent home device management and control method and system
CN114372476B (en) Semantic truncation detection method, device, equipment and computer readable storage medium
CN110222161B (en) Intelligent response method and device for conversation robot
CN113012687B (en) Information interaction method and device and electronic equipment
CN112466287B (en) Voice segmentation method, device and computer readable storage medium
CN113113013A (en) Intelligent voice interaction interruption processing method, device and system
CN111985248A (en) Information interaction method and device
CN111508481A (en) Training method and device of voice awakening model, electronic equipment and storage medium
CN113420115B (en) Information processing method and device based on man-machine voice dialogue
CN110704585A (en) Question answering method, device and computer readable medium
CN117894321A (en) Voice interaction method, voice interaction prompting system and device
CN112201250B (en) Semantic analysis method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination