CN108563633A - A kind of method of speech processing and server - Google Patents

A kind of method of speech processing and server Download PDF

Info

Publication number
CN108563633A
CN108563633A CN201810272758.9A CN201810272758A CN108563633A CN 108563633 A CN108563633 A CN 108563633A CN 201810272758 A CN201810272758 A CN 201810272758A CN 108563633 A CN108563633 A CN 108563633A
Authority
CN
China
Prior art keywords
semantics recognition
unified standard
terminal
server
scene
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810272758.9A
Other languages
Chinese (zh)
Other versions
CN108563633B (en
Inventor
黄珊珊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN201810272758.9A priority Critical patent/CN108563633B/en
Publication of CN108563633A publication Critical patent/CN108563633A/en
Application granted granted Critical
Publication of CN108563633B publication Critical patent/CN108563633B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The embodiment of the invention discloses a kind of method of speech processing and servers, execute the pretreatment process under special scene for realizing according to unified standard information, extend the application scenarios of semantics recognition.The embodiment of the present invention provides a kind of method of speech processing, including:Semantics recognition is carried out to the collected voice signal of terminal by the semantics recognition model trained in advance, obtains semantics recognition result;The critical field for being matched with special scene is extracted from the semantics recognition result;Semantic analysis is carried out to the semantics recognition result, obtains the semantic level under the special scene;The unified standard information under the special scene is generated according to the critical field and the semantic level;The pretreatment process under the special scene is executed according to the unified standard information.

Description

A kind of method of speech processing and server
Technical field
The present invention relates to field of computer technology more particularly to a kind of method of speech processing and server.
Background technology
Artificial intelligence (Artificial Intelligence, AI) is research, develops for simulating, extending and extending people Intelligence theory, a new technological sciences of method, technology and application system.Artificial intelligence is one of computer science Branch, it attempts to understand the essence of intelligence, and produces a kind of new intelligence that can be made a response in such a way that human intelligence is similar Energy machine, the research in the field includes robot, language identification, image recognition, natural language processing and expert system etc..
With the development of Internet technology, semantics recognition is more and more extensive to be applied in each application scenarios, at present Semantics recognition technology used in various softwares on the market typically just is used to do the language identification translation on basis, voice conversion Word or word convert voice, for example, in navigation application program, can by carrying out semantics recognition to voice input by user, To provide navigation Service etc. to the user.
Wherein, after carrying out semantics recognition to information such as texts to be resolved, lack and semantics recognition result is done further Processing, therefore the further scheme of semantics recognition result can not be deeply excavated, limit to the applied field for semantics recognition Scape.
Invention content
An embodiment of the present invention provides a kind of method of speech processing and servers, are held for realizing according to unified standard information Pretreatment process under the special scene of row, extends the application scenarios of semantics recognition.
In order to solve the above technical problems, the embodiment of the present invention provides following technical scheme:
In a first aspect, the embodiment of the present invention provides a kind of method of speech processing, including:
Semantics recognition is carried out to the collected voice signal of terminal by the semantics recognition model trained in advance, obtains language Adopted recognition result;
The critical field for being matched with special scene is extracted from the semantics recognition result;
Semantic analysis is carried out to the semantics recognition result, obtains the semantic level under the special scene;
The unified standard information under the special scene is generated according to the critical field and the semantic level;
The pretreatment process under the special scene is executed according to the unified standard information.
Second aspect, the embodiment of the present invention also provide a kind of server, including:
Semantics recognition module, for the semantics recognition model by training in advance to the collected voice signal of terminal into Row semantics recognition obtains semantics recognition result;
Field extraction module, for extracting the critical field for being matched with special scene from the semantics recognition result;
Grade analysis module is obtained for carrying out semantic analysis to the semantics recognition result under the special scene Semantic level;
Information generating module, for being generated under the special scene according to the critical field and the semantic level Unified standard information;
Message processing module, for executing the pretreatment process under the special scene according to the unified standard information.
In second aspect, aforementioned first aspect and various possible realizations can also be performed in the comprising modules of server The step of described in mode, refers to aforementioned explanation to first aspect and in various possible realization methods.
The third aspect, the embodiment of the present invention provide a kind of server, which includes:Processor, memory;Memory For storing instruction;Processor is used to execute the instruction in memory so that server is executed as any in aforementioned first aspect The method of item.
Fourth aspect, an embodiment of the present invention provides a kind of computer readable storage medium, the computer-readable storage Instruction is stored in medium, when run on a computer so that computer executes the method described in above-mentioned various aspects.
5th aspect, an embodiment of the present invention provides a kind of computer program products including instruction, when it is in computer When upper operation so that computer executes the method described in above-mentioned various aspects.
As can be seen from the above technical solutions, the embodiment of the present invention has the following advantages:
In embodiments of the present invention, by the semantics recognition model that trains in advance to the collected voice signal of terminal into Row semantics recognition, obtains semantics recognition as a result, extracting the critical field for being matched with special scene from semantics recognition result, right Semantics recognition result carries out semantic analysis, the semantic level under special scene is obtained, according to critical field and semantic level The unified standard information under special scene is generated, the pretreatment process under special scene is executed according to unified standard information.This Apply for that the voice input in embodiment to user carries out semantics recognition and rank judges, according to the default system to speech production standard One standard information carries out different pretreatments further according to unified standard information, the voice body under different special scenes may be implemented It tests and optimizes.Judgement and pretreatment to the voice signal of user can greatly reduce the letter during user's progress vocational work Breath processing energy, realizes and executes the pretreatment process under special scene according to unified standard information, extend the application of semantics recognition Scene.
Description of the drawings
To describe the technical solutions in the embodiments of the present invention more clearly, make required in being described below to embodiment Attached drawing is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for For those skilled in the art, other drawings may also be obtained based on these drawings.
Fig. 1 is system architecture schematic diagram provided in an embodiment of the present invention;
Fig. 2 is a kind of process blocks schematic diagram of method of speech processing provided in an embodiment of the present invention;
Fig. 3 is the execution flow diagram of method of speech processing provided in an embodiment of the present invention;
Fig. 4 is the schematic diagram of the abnormal prompt scene of special attendance checking provided in an embodiment of the present invention;
Fig. 5 is the product process schematic diagram provided in an embodiment of the present invention for asking for leave single;
Fig. 6 is the schematic diagram of the feedback scene of traffic special circumstances provided in an embodiment of the present invention;
Fig. 7 is the product process schematic diagram of traffic feedback provided by the embodiments of the present application;
Fig. 8-a are a kind of composed structure schematic diagram of server provided in an embodiment of the present invention;
Fig. 8-b are the composed structure schematic diagram of another server provided in an embodiment of the present invention;
Fig. 8-c are a kind of composed structure schematic diagram of information generating module provided in an embodiment of the present invention;
Fig. 8-d are the composed structure schematic diagram of another server provided in an embodiment of the present invention;
Fig. 9 is the composed structure schematic diagram that method of speech processing provided in an embodiment of the present invention is applied to server.
Specific implementation mode
An embodiment of the present invention provides a kind of method of speech processing and servers, are held for realizing according to unified standard information Pretreatment process under the special scene of row, extends the application scenarios of semantics recognition.
In order to make the invention's purpose, features and advantages of the invention more obvious and easy to understand, below in conjunction with the present invention Attached drawing in embodiment, technical scheme in the embodiment of the invention is clearly and completely described, it is clear that disclosed below Embodiment be only a part of the embodiment of the present invention, and not all embodiments.Based on the embodiments of the present invention, this field The every other embodiment that technical staff is obtained, shall fall within the protection scope of the present invention.
Term " comprising " and " having " in description and claims of this specification and above-mentioned attached drawing and they Any deformation, it is intended that it includes so as to a series of process comprising units, method, system, product or to set to cover non-exclusive It is standby to be not necessarily limited to those units, but may include not listing clearly or solid for these processes, method, product or equipment The other units having.
Fig. 1 is the application scenario diagram of method of speech processing in one embodiment.Referring to Fig.1, which includes logical Cross the terminal 11 and server 12 of network connection.The terminal 11 can have the equipment for determining voice collecting function, i.e. terminal 11 The voice that send out of user can be acquired by microphone, be also equipped in the terminal application program (Application, APP), user can operate application program.In one embodiment, terminal 11 can be mobile terminal, and mobile terminal can wrap Include at least one of mobile phone, tablet computer, laptop, personal digital assistant and Wearable etc..Server 12 can To be realized with the independent server server cluster that either multiple physical servers form.
Server 12 can get the voice signal of user, such as server 12 and terminal 11 communicate, and are obtained from terminal 11 Get voice signal.The server may be implemented to carry out semantics recognition to voice, and keyword is extracted by semantics recognition result Section, which is to be matched with preset special scene, the special scene can there are many, such as enterprise office automation (Office Automation, OA) manages scene or alert process scene etc..Server is carried out according to semantics recognition result Semanteme classification, so that it is determined that going out the corresponding semantic level of current speech signal, semantic level refers to according to voice recognition result institute Including the semantic content grade or severity that mark off, server can also generate unified standard information, thereby using Unified standard information completes the pretreatment process under special scene, and voice experience under different special scenes and excellent may be implemented Change.
Next a kind of method of speech processing provided by the embodiments of the present application is described from the angle of server, please refers to figure Shown in 2, method of speech processing includes the following steps:
101, semantics recognition is carried out to the collected voice signal of terminal by the semantics recognition model trained in advance, obtained To semantics recognition result.
In the embodiment of the present application, user can acquire the voice signal of the user with using terminal, to which terminal can be with Voice signal is collected, then terminal is interacted with server, and the collected voice signal of institute is sent to server.Server can be with Speech samples are trained using machine learning method, to extract phonetic feature, and build semantics recognition model, server is adopted Semantics recognition model can identify semanteme from voice data.Server is using semantic model trained in advance to user Voice signal carry out semantics recognition, to obtain semantics recognition as a result, semantics recognition result refers to being identified from voice signal Character information.
In some embodiments of the invention, step 101 acquires terminal by the semantics recognition model trained in advance Before the voice signal arrived carries out semantics recognition, method of speech processing provided by the embodiments of the present application further includes following steps:
The APP operated in terminal is determined according to the acquisition port of voice signal;
Corresponding semantics recognition model is determined according to the APP operated in terminal.
Wherein, in the embodiment of the present invention, the entrance of terminal transmission voice can be directed to special scene (i.e. proprietary scene).Than Such as business management software, attendance APP is installed in the terminal of user, then the voice signal that user is sent out by attendance APP can be with It being collected by terminal, server can determine that semantics recognition model is to be directed to scene of asking for leave according to the APP operated in terminal, For another example user uses traffic APP, then can be collected by terminal by the traffic APP voice signals sent out, takes Business device determines that semantics recognition model is to be directed to traffic scene according to the APP operated in terminal, and different scenes is different Model library.
In some embodiments of the invention, method of speech processing provided by the embodiments of the present application further includes following steps:
If critical field can not be extracted from semantics recognition result, triggering terminal resurveys voice signal.
Wherein, server can use semantics recognition Model Identification voice signal, if the voice signal can not extract pass Key field can not then execute the subsequent step in the embodiment of the present invention, and server can notify terminal to resurvey voice signal, Then terminal can prompt user to re-emit voice, be lost to avoid that can not identify that critical field causes to identify from voice signal The case where losing.
102, the critical field for being matched with special scene is extracted from semantics recognition result.
In the embodiment of the present application, after server gets semantics recognition result by semantics recognition model, for special With scene, critical field is extracted from the semantics recognition result.Wherein, special scene refers to the voice signal that user sends out Processing scene, in practical applications have a variety of realization methods, the special scene need combine terminal used by a user with And server determines, such as the special scene can there are many, such as OA attendance management scene or audio alert Treatment stations Scape etc..
It should be noted that the critical field extracted required under different special scenes is different, illustrate It is bright as follows, under OA attendance management scenes, need the critical field extracted to may include:It reason of asking for leave and asks for leave the time limit, Audio alert is handled under scene, needs the critical field extracted to may include:Generation event and event location.For another example, it is swimming Under scene of playing, the critical field extracted is needed to may include:Game role and game progress etc..
103, semantic analysis is carried out to semantics recognition result, obtains the semantic level under special scene.
In the embodiment of the present application, after server gets semantics recognition result by semantics recognition model, in addition to holding Except row step 102, step 103 can also be performed, do not have sequencing between step 102 and step 103, can also hold side by side Row, does not limit herein.Server carries out semantic analysis to semantics recognition result, to true according to preset partition of the level standard Semantic level of the current speech signal under special scene is made, i.e. server can be gone out by semantics recognition interpretation of result special With the specific semantic level under scene.For example, being preset under special scene, there are three ranks:Level-one, two level, three-level, according to working as Before the semantics recognition result that gets which rank be determined as.In the embodiment of the present invention, server can be to semantics recognition result It carries out going deep into excavation, to analyze the corresponding semantic level of semantics recognition result, for different semantic levels, then it represents that Significance level under special scene representated by the voice signal of user differs, therefore carries out semantic analysis by server, The semantic level indicated by the voice signal of user can be automatically identified, can be touched based on the semantic level that server analysis goes out Send out pretreatment process specific, to save artificial hearing voice signal the trouble that processes.
It should be noted that semantics recognition is the solution model of a proprietary scene of reply, such as in OA attendance pipes It manages under scene, by judging party's voice voice emergency and semantics recognition, presets three semantic levels, respectively:Special feelings Condition, emergency episode, general problem, specific classification can recognize that at present the voice signal of user by proprietary model lab setting Situation of asking for leave is some in above three rank, so as to complete semantics recognition result for the semantic level identified Pretreatment.
It is illustrated below, according to cause field analysis of asking for leave, if there is following two:1, contain " I ", " in person " etc. Field judges that user describes situation object as user itself.2, contain the fields such as " body ", " discomfort ", " safety " and judge description object It is free, safe and healthy, psychological to be hindered, come to harm or situations such as potential injury, while meeting above-mentioned two situations judgement User's special circumstances rank is level-one.If there is following two:1, contain kinship address field or name class name field Judge that user describes situation object as user relatives or important relationship people.2, contain the fields such as " body ", " discomfort ", " safety " to judge Description object is free, safe and healthy, psychological to be hindered, is come to harm or situations such as potential injury, while meeting above two Situation judges user's special circumstances rank for two level.If there is following two:1, contain " I ", the fields such as " in person " judge user Description situation object is user itself.2, it obtains:It asks for leave cause field, while meeting above-mentioned two situations and judging the special feelings of user Condition rank is three-level.
104, the unified standard information under special scene is generated according to critical field and semantic level.
In embodiments of the present invention, server gets critical field by abovementioned steps 102 and is obtained by step 103 After getting semantic level, critical field and semantic level can be generated unified standard information, this unifies standard information can To be the document with set form, this unifies to include critical field and semantic level in standard information, to which this is unified Standard information can as semantics recognition result is done further excavate after obtained information, based on this unify standard information can be with Execute the pretreatment process to semantics recognition result.For example, the list of asking for leave of unified standard is automatically generated according to critical field, Semantic level is marked, to which this unifies in the list of asking for leave of standard just to carry the information such as type and significance level of asking for leave.
It should be noted that different special scenes is directed in the embodiment of the present invention, used unified standard information tool There is different reference formats, it is to be understood that under OA attendance management scenes and in the case where audio alert handles scene, need Using the standard information of different systems, because the key content of required record differs under different special scenes.
In embodiments of the present invention, step 104 generates the system under special scene according to critical field and semantic level One standard information, including:
The unified standard template with special scene matching is obtained from unified standard template library;
Preset position in unified standard template fills the field contents of critical field, obtains filled with field contents Unified standard template;
The unified standard template filled with field contents is labeled using semantic level, obtains unified standard information.
Wherein, server-side, which can preserve, unifies standard template library, and there are many unified standard moulds for storage in the template library Plate can store unified standard template for different special scenes and be obtained first when server needs to generate unified standard information Get the unified standard template of special scene matching, then this unifies in standard template that the empty content that some need is filled is arranged, The field contents that critical field can be then filled in the preset position in unified standard template, obtain the system filled with field contents Then one standard template is labeled the unified standard template filled with field contents using semantic level, is uniformly made Formula information, after being labeled using semantic level, can by unified standard acquisition of information to critical field and semantic level, The information obtained after further excavation can be done to which this unifies standard information as to semantics recognition result.
It is illustrated below, by taking special scene is OA attendance management scenes as an example, can may include with critical field:It asks for leave Cause field and deadline field of asking for leave, then asking for leave for the unified standard generated may include singly:" reason of asking for leave word occurs because interim It the reason of section ", can not turn out for work on time at present, therefore " deadline field of asking for leave " day of asking for leave, the list of asking for leave for unifying standard can also wrap It includes:It the address name asked for leave and asks for leave the date, and marks asked for leave semantic level.
105, the pretreatment process under special scene is executed according to unified standard information.
In embodiments of the present invention, server generates unified standard letter by the deep excavation to semantics recognition result Breath, this is unified to include critical field and semantic level in standard information, can be used as to language to which this unifies standard information Adopted recognition result does the information obtained after further excavation, and unifying standard information based on this can execute to semantics recognition result Pretreatment process, can solve the user experience optimization problem of a variety of special scenes, such as can ask for leave, medical treatment, report, face When report etc. users' scene experience and optimization.In the embodiment of the present invention, judgement and pre- place of the server to user speech information Reason, can greatly reduce information processing energy of user during carrying out vocational work, unified with standard, efficient mode pair Different information are distributed processing, improve the treatment effeciency to voice signal.
It should be noted that it is different that the pretreatment process executed is needed under different special scenes, and institute Need the pretreatment process executed that can also be determined according to the semantic level marked in unified standard information, such as unified standard Semantic level in information is severity level, then needs emergency notice intended recipient, if the semantic class in unified standard information Not Wei insignificant rank, then can take plain mode notify or need not notify, herein for illustrative purposes only.
In some embodiments of the invention, the pretreatment process under special scene is executed according to unified standard information, wrapped It includes:
Will unified standard information notify under special scene with the matched intended recipient of semantic level.
Wherein, server execute pretreated mode can there are many, such as can be with according to the severity of semantic level When determining whether to notify that intended recipient, such as semantic level are high-level, first object recipient is notified using short message mode, When semantic level is low level, using lettergram mode by the second intended recipient, so as to identify under different special scenes The pretreatment mode to match with semantic level.By taking special scene is OA attendance management scenes as an example, server passes through special feelings Condition attendance proprietary model can get critical field, and according to semantics recognition result include " with the office special feelings of attendance The fields such as the illness, safety, traffic accident, the contingency that occur in condition " analyze semantic level, according to the OA attendance management field Pretreatment mode under scape notifies attendance management person.
In some embodiments of the invention, method of speech processing provided by the embodiments of the present application further includes following steps:
The handling result to unified standard information is obtained from intended recipient;
Handling result is sent to terminal.
Wherein, unified standard information is sent to intended recipient by server, then the intended recipient can parse the system One standard information is unified in standard information to determine that critical field and semantic level, intended recipient can also be directed to from this This is unified standard information and is handled, and handling result is generated, by taking special scene is OA attendance management scenes as an example, intended recipient It can examine to ask for leave and whether pass through, and generate handling result.Being established between intended recipient and server has communication, and server is also The handling result of intended recipient is received, it is terminal that server forwards the handling result again, to which terminal can be with feedback processing As a result user is given so that user can get the feedback of the voice signal sent to oneself.
By above to the description of the embodiment of the present invention it is found that being adopted to terminal by the semantics recognition model trained in advance The voice signal that collects carries out semantics recognition, obtain semantics recognition as a result, extracted from semantics recognition result be matched with it is special The critical field of scene carries out semantic analysis to semantics recognition result, the semantic level under special scene is obtained, according to key Field and semantic level generate the unified standard information under special scene, are executed under special scene according to unified standard information Pretreatment process.The voice of user is inputted in the embodiment of the present application and carries out semantics recognition and rank judgement, according to default pair The unified standard information of speech production standard carries out different pretreatments further according to unified standard information, may be implemented different special With under scene voice experience and optimization.Judgement and pretreatment to the voice signal of user can greatly reduce user into industry The information processing energy worked during making is realized and executes the pretreatment process under special scene according to unified standard information, expands Open up the application scenarios of semantics recognition.
For ease of being better understood from and implementing the said program of the embodiment of the present invention, corresponding application scenarios of illustrating below come It is specifically described.
A kind of method of speech processing is provided in the embodiment of the present invention, dual factors may be implemented and combine output, dual factors are Refer to the output of critical field output and semantic level, critical field illustrates the corresponding semantic content of voice signal, semantic level Illustrate the corresponding severity of voice signal (i.e. rank).By being pre-designed extraction to critical field and to semantic analysis As a result classification, the document and corresponding appropriate level for generating certain standard carry out the system division of labor.To reach to semantics recognition knot Fruit automatically processes purpose.Can optimize significantly it is various in the prior art notify, ask for leave, applying, the voice input of mail, automatically The operation flow of information pre-processing is completed in grading.Wherein, be pre-designed refers to training semantics recognition by machine learning algorithm Model, such as:I+sick+hospital combines and is free of third person pronoun, judge that user needs hospital of asking for leave.System is divided the work Refer to that " degree " rank is gone out by semantics recognition, automated system operation different stage needs the next-step operation carried out, such as:Judge For sick leave rank, system automatically generated is asked for leave list, examination & approval automatically by.And responsible person is sent an SMS to, allow company to understand employee Physical condition, if it is determined that leave of absence rank, that needs automatically generates list of asking for leave.
As shown in figure 3, semantics recognition technology provided in an embodiment of the present invention need to be suitable in exclusive user's scene.User Server can be sent speech to using terminal, server carries out the semantics recognition for having weight according to scene proprietary model library, this Place's weight refers to identifying semantic level, while obtaining the rank of critical field and Scene Semantics degree, and critical field is given birth to At the document of unified standard, different respective handlings is carried out to document or system according to different stage.It can in the embodiment of the present invention , according to pre-set rank, to do the pretreatment of semantics recognition result to semantics recognition.
Next by taking enterprise's OA manages scene as an example, in specifically office scene, employee is not temporarily because special circumstances can The situation turned out for work is not within minority.As shown in figure 4, the signal of the abnormal prompt scene of special attendance checking provided in an embodiment of the present invention Figure, mainly comprises the following processes:
S01, special circumstances attendance proprietary model library is established.Weight exports when carrying out semantics recognition to current voice signal With the relevant fields such as illness, safety, traffic accident, contingency for occurring in office attendance special circumstances scene.
S02, user's generation special circumstances can not normally turn out for work, and initiate speech explanation, ask for leave.
The exclusive speech model library of S03, special circumstances attendance carries out speech recognition, extraction critical field " a situation arises " and " asking for leave the time limit ".
Wherein, field that a situation arises obtains result:Because the reason of XXXXXX, (X field content inputted voice according to user Identification extraction).
Deadline field of asking for leave obtains result:It please vacation in X days (X can be equal to 0.5).
If critical field can not be obtained:Voice is re-entered in prompt;It is transferred to step S02.
S04, user semantic judge that setting application is graded:Level-one, two level, three-level.
Level-one situation:It is the problems such as according to the personal safety containing user of user speech input content or health situation, all kinds of Bursty state leads to user itself freely, safety, health, and psychology hindered, is come to harm or situations such as potential injury, system It is judged as level-one situation;
The independent editing short message of system prompts attendance management person, prompts attendance management person to understand problem severity and pays attention to making Go out respective handling.
Two level situation:According to accidents such as user speech input content relatives containing user safety or health, user indicates The case where respective handling need to be made and help to lead to not turn out for work, system is judged as two level situation.
Three-level situation:Other special events there is a situation where causing user that can not turn out for work, system is judged as three-level situation.
S05, the list of asking for leave that unified standard is automatically generated according to critical field, mark application rank, just know type of asking for leave And significance level, and mail is sent to attendance management person.
S06, attendance management person are to single audit of asking for leave.
Audit passes through:System replys user and asks for leave success.
Postpone:Attendance responsible person judges whether user is special circumstances according to verification under line, and system replys user and postpones examining Core, time limit of asking for leave terminate after image attendance responsible person and are described in detail again, it may include the following two kinds returns the result:
It is:System replys user and asks for leave success.
It is no:User asked for leave after the time limit, can execute, and no special situation attendance records spacious class.
S07, all attendance vacation lists and result are recorded into attendance data library.
As described in Figure 5, it is the product process schematic diagram provided in an embodiment of the present invention for asking for leave single, includes mainly following mistake Journey:
S11, terminal send voice signal.
S12, server carry out noise reduction process to voice signal.
S13, server carry out end-point detection and speech enhan-cement.
S14, server carry out feature extraction and feature compensation.
S15, server carry out matching search by language model and acoustic model.
Wherein, special circumstances attendance proprietary model, weight output and the disease occurred in office attendance special circumstances when identification The relevant fields such as disease, safety, traffic accident, contingency.Language model and acoustic model are the one of current speech recognition technology Kind Floor layer Technology.
S16, server export recognition result, may include:It obtains " asking for leave the time limit ", obtains " reason of asking for leave ".
S17, according to asking for leave, reason carries out field analysis, determines rank.
Wherein, field analysis is carried out according to reason of asking for leave, may include following process:
10, contain " I ", the fields such as " in person " judge that user describes situation object as user itself.
20, contain " body ", " discomfort ", the fields such as " safety " judge that description object is free, safe and healthy, psychological and hindered Hinder, come to harm or situations such as potential injury.
Meeting above-mentioned two situations simultaneously judges user's special circumstances rank for level-one.
11, containing kinship call field or name class name field judge user describe situation object for user relatives or Important relationship people.
21, contain " body ", " discomfort ", the fields such as " safety " judge that description object is free, safe and healthy, psychological and hindered Hinder, come to harm or situations such as potential injury.
Meeting above-mentioned two situations simultaneously judges user's special circumstances rank for two level.
12, contain " I ", the fields such as " in person " judge that user describes situation object as user itself.
22, it obtains:It asks for leave cause field.
Meeting above-mentioned two situations simultaneously judges user's special circumstances rank for three-level.
S18, it generates unified standard and asks for leave list.
Wherein, it may include singly following content that unified standard, which is asked for leave,:
Occur because interim【Transfer cause field of asking for leave】The reason of, it can not turn out for work, therefore ask for leave on time at present【Jump goes for vacation Limit field】It.
It please dummy:【Obtain address name】.
Date:【Acquisition system current date】.
In the embodiment of the present invention, application of asking for leave is initiated by one key of voice, is judged by speech analysis processing system, is turned Metaplasia becomes the list of asking for leave of standard system, is committed to OA and asks for leave and apply for module, and sends mail in attendance responsible person.At system Most of links in completing to ask for leave flow are managed, unnecessary manual operation is avoided, user is made to submit application of asking for leave, side at the first time Just company's relevant person in charge is known and makes corresponding measure in time.
Next by taking road traffic condition scene as an example, at present road traffic condition can only the one-side inquiry of citizen, or The one-side feedback road conditions of person, make citizen lack sense of participation to dredging for traffic conditions.As shown in fig. 6, being the embodiment of the present invention The schematic diagram of the feedback scene of the traffic special circumstances of offer, mainly comprises the following processes:
S21, traffic special circumstances proprietary model library is established, weight exports when carrying out semantics recognition to current voice signal With the accident occurred in vehicular traffic special circumstances scene, wipe touch, injures and deaths, congestion, the relevant fields such as transit equipment failure.
S22, user have found traffic special circumstances, initiate speech explanation, provide information.
The exclusive speech model library of S23, traffic special circumstances carries out speech recognition, extraction critical field " event occurs " and " event location ".
Wherein, event field occurs and obtains result:(X field content inputs language to the case where XXXXXX has occurred according to user Sound identification extraction).
Event location field obtains result:(X field content inputs speech recognition according to user and carries for the position of XXXXXX at present It takes).
If critical field can not be obtained:Voice is re-entered in prompt;It is transferred to step 2.
S24, user semantic judge that setting application is graded:Level-one, two level, three-level.
Level-one situation:There is a situation where a wide range of collision or traffic accidents, and casualties, system to be caused to be judged as level-one feelings Condition;
The independent editing short message prompting region traffic police of system and internal system top set prompt message processing person, region traffic police on duty Live request for information is arrived at the first time.
Two level situation:The traffic accidents such as small range scraping occur or disagreeing there are car owner causes road surface crowded, system judges For two level situation;
Three-level situation:Traffic congestion causes road trip slow;System is judged as three-level situation.
S25, the feedback table that unified standard is automatically generated according to critical field, are audited by background information administrator and are believed Breath.
It verifies and confirms:According to feedback position distribution region traffic police (jump procedure S26) on duty.
Verification is not required to in-situ processing:Operation mark feedback form " being not required to in-situ processing ".
S26, region traffic police scene confirm:
Confirm no problem or has settled a dispute by the parties concerned themselves (jump procedure S29).
There are problems to need in-situ processing:The user terminal special circumstances database displaying data, user can be looked by system Ask the traffic problems (redirecting 7) occurred near having confirmed that.
S27, in-situ processing situation.
Background information processing person is fed back after the completion of S28, processing.
S29, background information processing person confirm that the Message Processing finishes Flow ends;If being traffic police's in-situ processing with the situation Situation, while deleting the data of user terminal special circumstances database displaying.
Include mainly as follows as shown in fig. 7, for the product process schematic diagram of traffic feedback provided by the embodiments of the present application Process:
S31, terminal send voice signal.
S32, server carry out noise reduction process to voice signal.
S33, server carry out end-point detection and speech enhan-cement.
S34, server carry out feature extraction and feature compensation.
S35, server carry out matching search by language model and acoustic model.
Wherein, traffic special circumstances proprietary model, the thing occurred in weight output and vehicular traffic special circumstances when identification Therefore the relevant fields such as grazing, injures and deaths, congestion, transit equipment failure.Language model and acoustic model are current speech recognition technologies A kind of Floor layer Technology.
S36, server export recognition result, may include:" event occurs " is obtained, is obtained " event location ".
S37, field analysis is carried out according to the event of generation, determines rank.
Wherein, field analysis is carried out according to the event of generation, may include following process:
13, contain the fields such as " I ", " in person ", " other people ", " someone " and judge that user describes the behaviour of situation object.
23, contain the fields such as " body ", " bleeding ", " injury " and judge the safe and healthy of description object, come to harm or potential Situations such as injury.
According with platform above-mentioned two situations simultaneously judges user's special circumstances rank for level-one.
14, containing the fields such as " vehicle ", " collision ", " friction " judges user's description object for vehicle etc..
24, contain the fields such as " removing ", " responsibility ", " road is crowded " and judge that user object is residing because a variety of causes blocks the traffic The case where being smoothed out,
Meeting above-mentioned two situations simultaneously judges user's special circumstances rank for two level.
15, containing the fields such as " road " " traffic ", " crowded " judges user's description object for traffic conditions.
25, it obtains and event field occurs.
Meeting above-mentioned two situations simultaneously judges user's special circumstances rank for three-level.
S38, unified standard information feedback is generated.
Wherein, unified standard information feedback may include following content:
At present【Transfer event location field】Position;
It has occurred【Transfer generation event field】The case where.
Please the near roads car owner near roads vehicle that pays attention to detouring (two, three-level)/ask independently open up Emergency Vehicle Lane relief wound Member's (level-one).
In the embodiment of the present invention, the description of road traffic special circumstances is initiated by one key of voice, by speech analysis processing system System carries out judgement classification, is converted into the feedback form of standard system, and being sent in background information according to different stage sequence handles people Member;Information processing personnel successively carry out Notification Validation by rank to region traffic police;Information feedback after confirmation can to user terminal It looks into, periphery car owner is reminded to make the corresponding behave such as detour, evacuation.
In the present embodiment, it is the voice input to being used in different user scene, carries out semantics recognition and degree judges, According to the default word or file to speech production standard pattern, different pre- places is carried out to document or system further according to different stage Reason, can solve it is a variety of in the prior art ask for leave, medical treatment, report, the interim experience and optimization for users' scene such as reporting.To user The judgement and pretreatment of voice messaging can greatly reduce information processing essence of user during carrying out vocational work using product Power.Unified with standard, efficient mode is distributed processing to different information.
It can be provided according in various degree to difference by pre-processing to semantics recognition result in the embodiment of the present invention The distribution that information is sought unity of standard.It can be provided for different user scene by the Essential Elements Of Analysis with setting degree rank Different information processing manners.With general and universality.
It should be noted that for each method embodiment above-mentioned, for simple description, therefore it is all expressed as a series of Combination of actions, but those skilled in the art should understand that, the present invention is not limited by the described action sequence because According to the present invention, certain steps can be performed in other orders or simultaneously.Secondly, those skilled in the art should also know It knows, embodiment described in this description belongs to preferred embodiment, and involved action and module are not necessarily of the invention It is necessary.
For ease of preferably implementing the said program of the embodiment of the present invention, the phase for implementing said program is also provided below Close device.
It please refers to shown in Fig. 8-a, a kind of server 800 provided in an embodiment of the present invention may include:Semantics recognition module 801, field extraction module 802, grade analysis module 803, information generating module 804 and message processing module 805, wherein
Semantics recognition module 801 believes the collected voice of terminal for the semantics recognition model by training in advance Number carry out semantics recognition, obtain semantics recognition result;
Field extraction module 802, for extracting the keyword for being matched with special scene from the semantics recognition result Section;
Grade analysis module 803 is obtained for carrying out semantic analysis to the semantics recognition result in the special scene Under semantic level;
Information generating module 804, for being generated in the special field according to the critical field and the semantic level Unified standard information under scape;
Message processing module 805, for executing the pretreated stream under the special scene according to the unified standard information Journey.
In some embodiments of the present application, as shown in Fig. 8-b, the server 800 further includes:Model determining module 806, for the semantics recognition module 801 by the semantics recognition model that trains in advance to the collected voice signal of terminal Before carrying out semantics recognition, the application program operated in the terminal is determined according to the acquisition port of the voice signal APP;Corresponding semantics recognition model is determined according to the APP operated in the terminal.
In some embodiments of the present application, as shown in Fig. 8-c, described information generation module 804, including:
Template acquiring unit 8041 is made for being obtained from unified standard template library with the unified of the special scene matching Formula template;
Template fills unit 8042 fills the critical field for the preset position in the unified standard template Field contents obtain the unified standard template filled with field contents;
Template marks unit 8043, for using the semantic level to the unified standard mould filled with field contents Plate is labeled, and obtains the unified standard information.
In some embodiments of the present application, described information processing module 805 is specifically used for the unified standard information Notify under the special scene with the matched intended recipient of the semantic level.
In some embodiments of the present application, as shown in Fig. 8-d, the server 800 further includes:
Transceiver module 807, for obtaining the handling result to the unified standard information from the intended recipient;By institute It states handling result and is sent to the terminal.
In some embodiments of the present application, the semantics recognition module 801 can not be from the semantics recognition if being additionally operable to As a result the critical field is extracted in, is triggered the terminal and is resurveyed the voice signal.
By above to the description of the embodiment of the present invention it is found that being adopted to terminal by the semantics recognition model trained in advance The voice signal that collects carries out semantics recognition, obtain semantics recognition as a result, extracted from semantics recognition result be matched with it is special The critical field of scene carries out semantic analysis to semantics recognition result, the semantic level under special scene is obtained, according to key Field and semantic level generate the unified standard information under special scene, are executed under special scene according to unified standard information Pretreatment process.The voice of user is inputted in the embodiment of the present application and carries out semantics recognition and rank judgement, according to default pair The unified standard information of speech production standard carries out different pretreatments further according to unified standard information, may be implemented different special With under scene voice experience and optimization.Judgement and pretreatment to the voice signal of user can greatly reduce user into industry The information processing energy worked during making is realized and executes the pretreatment process under special scene according to unified standard information, expands Open up the application scenarios of semantics recognition.
Fig. 9 is a kind of server architecture schematic diagram provided in an embodiment of the present invention, which can be because of configuration or property Energy is different and generates bigger difference, may include one or more central processing units (central processing Units, CPU) 1122 (for example, one or more processors) and memory 1132, one or more storage applications The storage medium 1130 (such as one or more mass memory units) of program 1142 or data 1144.Wherein, memory 1132 and storage medium 1130 can be of short duration storage or persistent storage.The program for being stored in storage medium 1130 may include one A or more than one module (diagram does not mark), each module may include to the series of instructions operation in server.More into One step, central processing unit 1122 could be provided as communicating with storage medium 1130, and storage medium is executed on server 1100 Series of instructions operation in 1130.
Server 1100 can also include one or more power supplys 1126, one or more wired or wireless nets Network interface 1150, one or more input/output interfaces 1158, and/or, one or more operating systems 1141, example Such as Windows ServerTM, Mac OS XTM, UnixTM, LinuxTM, FreeBSDTM etc..
The server architecture shown in Fig. 9 can be based on by the step performed by server in above-described embodiment.
In addition it should be noted that, the apparatus embodiments described above are merely exemplary, wherein described as separation The unit of part description may or may not be physically separated, the component shown as unit can be or It can not be physical unit, you can be located at a place, or may be distributed over multiple network units.It can be according to reality Border needs to select some or all of module therein to achieve the purpose of the solution of this embodiment.In addition, provided by the invention In device embodiment attached drawing, the connection relation between module indicates there is communication connection between them, specifically can be implemented as one Item or a plurality of communication bus or signal wire.Those of ordinary skill in the art are without creative efforts, you can with Understand and implements.
Through the above description of the embodiments, it is apparent to those skilled in the art that the present invention can borrow Help software that the mode of required common hardware is added to realize, naturally it is also possible to by specialized hardware include application-specific integrated circuit, specially It is realized with CPU, private memory, special components and parts etc..Under normal circumstances, all functions of being completed by computer program can It is easily realized with corresponding hardware, moreover, for realizing that the particular hardware structure of same function can also be a variety of more Sample, such as analog circuit, digital circuit or special circuit etc..But it is more for the purpose of the present invention in the case of software program it is real It is now more preferably embodiment.Based on this understanding, technical scheme of the present invention substantially in other words makes the prior art The part of contribution can be expressed in the form of software products, which is stored in the storage medium that can be read In, such as the floppy disk of computer, USB flash disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory Device (RAM, Random Access Memory), magnetic disc or CD etc., including some instructions are with so that a computer is set Standby (can be personal computer, server or the network equipment etc.) executes the method described in each embodiment of the present invention.
In conclusion the above embodiments are merely illustrative of the technical solutions of the present invention, rather than its limitations;Although with reference to upper Stating embodiment, invention is explained in detail, it will be understood by those of ordinary skill in the art that:It still can be to upper The technical solution recorded in each embodiment is stated to modify or equivalent replacement of some of the technical features;And these Modification or replacement, the spirit and scope for various embodiments of the present invention technical solution that it does not separate the essence of the corresponding technical solution.

Claims (14)

1. a kind of method of speech processing, which is characterized in that including:
Semantics recognition is carried out to the collected voice signal of terminal by the semantics recognition model trained in advance, obtains semantic knowledge Other result;
The critical field for being matched with special scene is extracted from the semantics recognition result;
Semantic analysis is carried out to the semantics recognition result, obtains the semantic level under the special scene;
The unified standard information under the special scene is generated according to the critical field and the semantic level;
The pretreatment process under the special scene is executed according to the unified standard information.
2. according to the method described in claim 1, it is characterized in that, it is described by the semantics recognition model that trains in advance to end Before holding collected voice signal to carry out semantics recognition, the method further includes:
The application APP operated in the terminal is determined according to the acquisition port of the voice signal;
Corresponding semantics recognition model is determined according to the APP operated in the terminal.
3. according to the method described in claim 1, it is characterized in that, described according to the critical field and the semantic level The unified standard information under the special scene is generated, including:
The unified standard template with the special scene matching is obtained from unified standard template library;
Preset position in the unified standard template fills the field contents of the critical field, obtains filled in field The unified standard template of appearance;
The unified standard template filled with field contents is labeled using the semantic level, obtains the unified system Formula information.
4. according to the method described in claim 1, it is characterized in that, described described special according to the unified standard information execution Pretreatment process under scene, including:
By the unified standard information notify under the special scene with the matched intended recipient of the semantic level.
5. according to the method described in claim 4, it is characterized in that, the method further includes:
The handling result to the unified standard information is obtained from the intended recipient;
The handling result is sent to the terminal.
6. method according to claim 1 to 4, which is characterized in that the method further includes:
If the critical field can not be extracted from the semantics recognition result, triggers the terminal and resurvey the voice Signal.
7. a kind of server, which is characterized in that including:
Semantics recognition module carries out language for the semantics recognition model by training in advance to the collected voice signal of terminal Justice identification, obtains semantics recognition result;
Field extraction module, for extracting the critical field for being matched with special scene from the semantics recognition result;
Grade analysis module obtains the language under the special scene for carrying out semantic analysis to the semantics recognition result Adopted rank;
Information generating module, for generating the system under the special scene according to the critical field and the semantic level One standard information;
Message processing module, for executing the pretreatment process under the special scene according to the unified standard information.
8. server according to claim 7, which is characterized in that the server further includes:Model determining module, is used for The semantics recognition module carries out semantic knowledge by the semantics recognition model trained in advance to the collected voice signal of terminal Before not, the application APP operated in the terminal is determined according to the acquisition port of the voice signal;According to described The APP operated in terminal determines corresponding semantics recognition model.
9. server according to claim 7, which is characterized in that described information generation module, including:
Template acquiring unit, for obtaining the unified standard template with the special scene matching from unified standard template library;
Template fills unit, in the field that the preset position unified in standard template fills the critical field Hold, obtains the unified standard template filled with field contents;
Template marks unit, for using the semantic level to the unified standard template filled with field contents into rower Note obtains the unified standard information.
10. server according to claim 7, which is characterized in that described information processing module is specifically used for the system One standard information notify under the special scene with the matched intended recipient of the semantic level.
11. server according to claim 10, which is characterized in that the server further includes:
Transceiver module, for obtaining the handling result to the unified standard information from the intended recipient;By the processing As a result it is sent to the terminal.
12. the server according to any one of claim 7 to 11, which is characterized in that the semantics recognition module is also used In if the critical field can not be extracted from the semantics recognition result, triggers the terminal and resurvey the voice letter Number.
13. a kind of computer readable storage medium, including instruction, when run on a computer so that computer executes such as Method described in claim 1 to 6 any one.
14. a kind of server, which is characterized in that the terminal includes:Processor and memory;
The memory, for storing instruction;
The processor is executed for executing the described instruction in the memory as described in any one of claim 1 to 6 Method.
CN201810272758.9A 2018-03-29 2018-03-29 Voice processing method and server Active CN108563633B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810272758.9A CN108563633B (en) 2018-03-29 2018-03-29 Voice processing method and server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810272758.9A CN108563633B (en) 2018-03-29 2018-03-29 Voice processing method and server

Publications (2)

Publication Number Publication Date
CN108563633A true CN108563633A (en) 2018-09-21
CN108563633B CN108563633B (en) 2021-05-14

Family

ID=63533366

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810272758.9A Active CN108563633B (en) 2018-03-29 2018-03-29 Voice processing method and server

Country Status (1)

Country Link
CN (1) CN108563633B (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109634554A (en) * 2018-12-18 2019-04-16 三星电子(中国)研发中心 Method and apparatus for output information
CN109918479A (en) * 2019-02-28 2019-06-21 百度在线网络技术(北京)有限公司 For handling the method and device of information
CN110489458A (en) * 2019-07-31 2019-11-22 广州竞德信息技术有限公司 Analysis method based on semantics recognition
CN110534093A (en) * 2019-08-26 2019-12-03 河北微幼趣教育科技有限公司 Rest method, server, client to speaking for children identification
CN111179925A (en) * 2019-12-04 2020-05-19 北京永洪商智科技有限公司 Report layout system and method based on voice recognition
CN111797636A (en) * 2020-07-21 2020-10-20 苏州思必驰信息科技有限公司 Offline semantic parsing method and system
CN111951782A (en) * 2019-04-30 2020-11-17 京东方科技集团股份有限公司 Voice question and answer method and device, computer readable storage medium and electronic equipment

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1529861A (en) * 2000-11-07 2004-09-15 System for creation of database and structured information from revbal input
CN201503630U (en) * 2009-05-31 2010-06-09 北京电子科技职业学院 Voice security and protection device
CN105513593A (en) * 2015-11-24 2016-04-20 南京师范大学 Intelligent human-computer interaction method drove by voice
CN105786797A (en) * 2016-02-23 2016-07-20 北京云知声信息技术有限公司 Information processing method and device based on voice input
CN106057200A (en) * 2016-06-23 2016-10-26 广州亿程交通信息有限公司 Semantic-based interaction system and interaction method
CN106339602A (en) * 2016-08-26 2017-01-18 丁腊春 Intelligent consulting robot
CN106557971A (en) * 2016-11-18 2017-04-05 畅捷通信息技术股份有限公司 Based on the financial data processing method of speech recognition technology, system and terminal
CN107004410A (en) * 2014-10-01 2017-08-01 西布雷恩公司 Voice and connecting platform
US9772816B1 (en) * 2014-12-22 2017-09-26 Google Inc. Transcription and tagging system
CN107545029A (en) * 2017-07-17 2018-01-05 百度在线网络技术(北京)有限公司 Voice feedback method, equipment and the computer-readable recording medium of smart machine
CN107657017A (en) * 2017-09-26 2018-02-02 百度在线网络技术(北京)有限公司 Method and apparatus for providing voice service
CN107798032A (en) * 2017-02-17 2018-03-13 平安科技(深圳)有限公司 Response message treating method and apparatus in self-assisted voice session

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1529861A (en) * 2000-11-07 2004-09-15 System for creation of database and structured information from revbal input
CN201503630U (en) * 2009-05-31 2010-06-09 北京电子科技职业学院 Voice security and protection device
CN107004410A (en) * 2014-10-01 2017-08-01 西布雷恩公司 Voice and connecting platform
US9772816B1 (en) * 2014-12-22 2017-09-26 Google Inc. Transcription and tagging system
CN105513593A (en) * 2015-11-24 2016-04-20 南京师范大学 Intelligent human-computer interaction method drove by voice
CN105786797A (en) * 2016-02-23 2016-07-20 北京云知声信息技术有限公司 Information processing method and device based on voice input
CN106057200A (en) * 2016-06-23 2016-10-26 广州亿程交通信息有限公司 Semantic-based interaction system and interaction method
CN106339602A (en) * 2016-08-26 2017-01-18 丁腊春 Intelligent consulting robot
CN106557971A (en) * 2016-11-18 2017-04-05 畅捷通信息技术股份有限公司 Based on the financial data processing method of speech recognition technology, system and terminal
CN107798032A (en) * 2017-02-17 2018-03-13 平安科技(深圳)有限公司 Response message treating method and apparatus in self-assisted voice session
CN107545029A (en) * 2017-07-17 2018-01-05 百度在线网络技术(北京)有限公司 Voice feedback method, equipment and the computer-readable recording medium of smart machine
CN107657017A (en) * 2017-09-26 2018-02-02 百度在线网络技术(北京)有限公司 Method and apparatus for providing voice service

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
许晶: "基于智能手机平台的语音识别后文本处理的应用", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109634554A (en) * 2018-12-18 2019-04-16 三星电子(中国)研发中心 Method and apparatus for output information
CN109918479A (en) * 2019-02-28 2019-06-21 百度在线网络技术(北京)有限公司 For handling the method and device of information
CN109918479B (en) * 2019-02-28 2021-07-20 百度在线网络技术(北京)有限公司 Method and device for processing information
CN111951782A (en) * 2019-04-30 2020-11-17 京东方科技集团股份有限公司 Voice question and answer method and device, computer readable storage medium and electronic equipment
CN110489458A (en) * 2019-07-31 2019-11-22 广州竞德信息技术有限公司 Analysis method based on semantics recognition
CN110534093A (en) * 2019-08-26 2019-12-03 河北微幼趣教育科技有限公司 Rest method, server, client to speaking for children identification
CN111179925A (en) * 2019-12-04 2020-05-19 北京永洪商智科技有限公司 Report layout system and method based on voice recognition
CN111797636A (en) * 2020-07-21 2020-10-20 苏州思必驰信息科技有限公司 Offline semantic parsing method and system
CN111797636B (en) * 2020-07-21 2023-06-16 思必驰科技股份有限公司 Offline semantic analysis method and system

Also Published As

Publication number Publication date
CN108563633B (en) 2021-05-14

Similar Documents

Publication Publication Date Title
CN108563633A (en) A kind of method of speech processing and server
US20170162202A1 (en) Automated voice-to-reporting/management systems and method for voice call-ins of events/crimes
RU2700498C2 (en) Personal emergency response system with prognostic risk assessment of emergency call
CN101656799B (en) Automatic conversation system and conversation scenario editing device
CN106133825A (en) Broad sense phrase in automatic speech recognition system
JP5394118B2 (en) Behavior analysis device and call center system
CN105874530A (en) Predicting recognition quality of a phrase in automatic speech recognition systems
Silva et al. Emergency medical systems analysis by simulation and optimization
US10708424B1 (en) AI assistant for interacting with customers across multiple communication modes
Revetria et al. Improving healthcare using cognitive computing based software: an application in emergency situation
CN115665325B (en) Intelligent outbound method, device, electronic equipment and storage medium
KR20210075511A (en) Method and Apparatus for Recommending Disaster Response
WO2021136457A1 (en) Multi-source alarm handling method and apparatus based on unified alarm platform, and related device
CN116611471A (en) Multi-party sharing system and method for deep learning large model
Beynon‐Davies Declarations of significance: Exploring the pragmatic nature of information models
CN113643798A (en) Method and device for matching caregivers for disabled persons and computer equipment
Steen-Tveit et al. Using Audio-Logs for Analyzing the Development of a Common Operational Picture in Multi-agency Emergency Response
Liu et al. An agent-based architecture of the Digital Twin for an Emergency Department
Hepenstal et al. How analysts think: A preliminary study of human needs and demands for AI-based conversational agents
US20220366146A1 (en) Virtual assistants for emergency dispatchers
Bublyk et al. Decision Support System Design For Low-Voice Emergency Medical Calls At Smart City Based On Chatbot Management In Social Networks.
US8051026B2 (en) Rules collector system and method with user interaction
Kurbesov et al. Automated voice recognition of emotions through the use of neural networks
CN112686518A (en) Monitoring system and method for inbound personnel
US11895270B2 (en) Phone tree traversal system and method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant