CN104535071B - A kind of phonetic navigation method and device - Google Patents

A kind of phonetic navigation method and device Download PDF

Info

Publication number
CN104535071B
CN104535071B CN201410742287.5A CN201410742287A CN104535071B CN 104535071 B CN104535071 B CN 104535071B CN 201410742287 A CN201410742287 A CN 201410742287A CN 104535071 B CN104535071 B CN 104535071B
Authority
CN
China
Prior art keywords
speech recognition
navigation
point
interest
recognition result
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201410742287.5A
Other languages
Chinese (zh)
Other versions
CN104535071A (en
Inventor
谢波
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BAIDU INTERNATIONAL TECHNOLOGY (SHENZHEN) Co.,Ltd.
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201410742287.5A priority Critical patent/CN104535071B/en
Publication of CN104535071A publication Critical patent/CN104535071A/en
Application granted granted Critical
Publication of CN104535071B publication Critical patent/CN104535071B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C21/00Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
    • G01C21/26Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
    • G01C21/34Route searching; Route guidance
    • G01C21/36Input/output arrangements for on-board computers
    • G01C21/3605Destination input or retrieval
    • G01C21/3608Destination input or retrieval using speech input, e.g. using speech recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/95Retrieval from the web
    • G06F16/953Querying, e.g. by the use of web search engines
    • G06F16/9537Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Artificial Intelligence (AREA)
  • Computational Linguistics (AREA)
  • Multimedia (AREA)
  • Automation & Control Theory (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Navigation (AREA)

Abstract

The embodiment of the invention provides a kind of phonetic navigation method and devices.On the one hand, the first phonetic order that the embodiment of the present invention passes through acquisition user;To determine speech recognition modeling according to current navigation state;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain the first speech recognition result, and according to first speech recognition result, executes navigation operation.Therefore, technical solution provided in an embodiment of the present invention is able to solve in navigation procedure in the prior art the problem that operating cost is higher, Voice Navigation process efficiency is lower.

Description

A kind of phonetic navigation method and device
[technical field]
The present invention relates to human-computer interaction technology more particularly to a kind of phonetic navigation methods and device.
[background technique]
With deep and Internet communication technology the development of global position system GPS civil nature, people are more and more Ground uses navigation client.Navigation client can provide a user the navigation function such as path planning and guidance based on electronic map Energy.Since its trip to people brings very big convenience, it is increasingly becoming essential a part in life.With Attention rate and use degree to navigation client is higher and higher, and people are not only satisfied with base provided by navigation client This navigation feature, and be desirable to it and be capable of providing more accurate property, more humanized navigation Service.
Sound prompt function is the important component of the provided navigation feature of navigation client, due to navigation client Particularity will be absorbed in driving in driver's startup procedure and pay attention to surface conditions, therefore cannot frequently observe navigation client Interface is held to obtain route relevant information, therefore, sound prompt function is just particularly important.However, navigation client only has There is sound prompt function, there is still a need for users when providing navigation Service before road, is manually entered destination letter on driving vehicle Breath is needed to be operated manually completion after stopping and be checked if user needs to obtain other navigation informations in driving procedure, Therefore the operating cost at present in navigation procedure is higher, and the treatment effeciency of Voice Navigation is relatively low.
[summary of the invention]
In view of this, the embodiment of the invention provides a kind of phonetic navigation method and devices, to solve in the prior art The problem that operating cost is higher in navigation procedure, Voice Navigation process efficiency is lower.
The one side of the embodiment of the present invention provides a kind of phonetic navigation method, comprising:
Acquire the first phonetic order of user;
According to current navigation state, speech recognition modeling is determined;
Using the speech recognition modeling, speech recognition is carried out to the first phonetic order of the user, to obtain first Speech recognition result;
According to first speech recognition result, navigation operation is executed.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the acquisition are used First phonetic order at family, comprising:
Second phonetic order of monitoring users;
Model is waken up using voice, speech recognition is carried out to second phonetic order, to obtain the second speech recognition knot Fruit;
If second speech recognition result meets preset wake-up condition, the first phonetic order of user is acquired.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the foundation are worked as Preceding navigational state, determines speech recognition modeling, comprising:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the foundation are worked as Preceding navigational state, determines speech recognition modeling, comprising:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one It is a: client control instruction, navigation hint instruction and point of interest search instruction.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, it is described according to institute The first speech recognition result is stated, navigation operation is executed, comprising:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export The search result to match with the interest point name;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtain simultaneously The search result that output matches with the point of interest typonym.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, it is described according to institute The first speech recognition result is stated, navigation operation is executed, comprising:
It is right according to the client control instruction if in first speech recognition result including client control instruction The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search It takes and exports and the search result that matches of interest point information in point of interest search instruction.
The one side of the embodiment of the present invention provides a kind of voice guiding device, comprising:
Voice collecting unit, for acquiring the first phonetic order of user;
Model treatment unit, for determining speech recognition modeling according to current navigation state;
Voice recognition unit carries out language to the first phonetic order of the user for utilizing the speech recognition modeling Sound identification, to obtain the first speech recognition result;
Navigate execution unit, for executing navigation operation according to first speech recognition result.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation,
The voice collecting unit is also used to the second phonetic order of monitoring users;
Described device further include:
Voice wakeup unit carries out speech recognition to second phonetic order for waking up model using voice, to obtain Obtain the second speech recognition result;If second speech recognition result meets preset wake-up condition, the voice collecting is triggered The first phonetic order of unit acquisition user.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, at the model Unit is managed, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, at the model Unit is managed, is specifically used for:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one It is a: client control instruction, navigation hint instruction and point of interest search instruction.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the navigation are held Row unit, is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export The search result to match with the interest point name;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtain simultaneously The search result that output matches with the point of interest typonym.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the navigation are held Row unit, is specifically used for:
It is right according to the client control instruction if in first speech recognition result including client control instruction The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search It takes and exports and the search result that matches of interest point information in point of interest search instruction.
As can be seen from the above technical solutions, the embodiment of the present invention has the advantages that
The first phonetic order that the embodiment of the present invention passes through acquisition user;To determine voice according to current navigation state Identification model;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain First speech recognition result, and according to first speech recognition result, execute navigation operation.Compared with prior art, originally The technical solution that inventive embodiments provide can be automatic to execute navigation operation according to the phonetic order of user, thus realize to Family provides navigation Service, does not need user and be operated manually to realize navigation feature, therefore be able to solve the prior art The problem that operating cost is higher in middle navigation procedure and Voice Navigation process efficiency is lower, can reduce the behaviour in navigation procedure Make cost, improve the treatment effeciency of Voice Navigation, guarantees driving safety.
[Detailed description of the invention]
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below will be to needed in the embodiment attached Figure is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for this field For those of ordinary skill, without any creative labor, it can also be obtained according to these attached drawings other attached Figure.
Fig. 1 is the exemplary system figure that technical solution provided by the embodiment of the present invention uses;
Fig. 2 is the flow diagram of phonetic navigation method provided by the embodiment of the present invention;
Fig. 3 is the functional block diagram of voice guiding device provided by the embodiment of the present invention.
[specific embodiment]
For a better understanding of the technical solution of the present invention, being retouched in detail to the embodiment of the present invention with reference to the accompanying drawing It states.
It will be appreciated that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.Base Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts it is all its Its embodiment, shall fall within the protection scope of the present invention.
The term used in embodiments of the present invention is only to be not intended to be limiting merely for for the purpose of describing particular embodiments The present invention.In the embodiment of the present invention and the "an" of singular used in the attached claims, " described " and "the" It is also intended to including most forms, unless the context clearly indicates other meaning.
It should be appreciated that term "and/or" used herein is only a kind of incidence relation for describing affiliated partner, indicate There may be three kinds of relationships, for example, A and/or B, can indicate: individualism A, exist simultaneously A and B, individualism B these three Situation.In addition, character "/" herein, typicallys represent the relationship that forward-backward correlation object is a kind of "or".
It will be appreciated that though speech recognition knot may be described using term first, second etc. in embodiments of the present invention Fruit, but these keywords should not necessarily be limited by these terms.These terms are only used to for keyword being distinguished from each other out.For example, not taking off In the case where from range of embodiment of the invention, the first speech recognition result can also be referred to as the second speech recognition result, similar Ground, the second speech recognition result can also be referred to as the first speech recognition result.
Depending on context, word as used in this " if " can be construed to " ... when " or " when ... When " or " in response to determination " or " in response to detection ".Similarly, depend on context, phrase " if it is determined that " or " if detection (condition or event of statement) " can be construed to " when determining " or " in response to determination " or " when the detection (condition of statement Or event) when " or " in response to detection (condition or event of statement) ".
The system that technical solution provided by the embodiment of the present invention uses is as shown in Figure 1, mainly by client and server It forming, method and apparatus provided by the embodiment of the present invention are realized in client-side, it is mainly used for the phonetic order according to user, Voice Navigation service is provided a user, realizes the speech navigation function of client.
The embodiment of the present invention provides a kind of phonetic navigation method, referring to FIG. 2, it is language provided by the embodiment of the present invention The flow diagram of sound air navigation aid, as shown, method includes the following steps:
S201 acquires the first phonetic order of user.
S202 determines speech recognition modeling according to current navigation state.
S203 carries out speech recognition to the first phonetic order of the user, to obtain using the speech recognition modeling First speech recognition result.
S204 executes navigation operation according to first speech recognition result.
Based on above-mentioned phonetic navigation method, the method for S201 is specifically described in the embodiment of the present invention.The step is specific May include:
Firstly, the second phonetic order of client monitoring users.Then, client wakes up model to described the using voice Two phonetic orders carry out speech recognition, to obtain the second speech recognition result.Finally, client judges the second speech recognition knot Whether fruit meets preset wake-up condition, if second speech recognition result meets preset wake-up condition, client acquisition The first phonetic order of user.
It should be noted that, in order to avoid the misrecognition of the phonetic order to user, needing to realize during Voice Navigation Real-time voice arousal function, client is under voice wake-up states, if the phonetic order of the user listened to meets wake-up bars Part, into speech recognition state, to just start the phonetic order of reception user.
For example, client can open a listening thread, the second phonetic order for lasting monitoring users.Client Model is waken up using voice, to the progress speech recognition of the second phonetic order is listened to obtain, to obtain the second speech recognition result.It should Voice wakes up model and goes out included in the second phonetic order to wake up word for identification.If the client determine that the second voice is known Include preset wake-up word in other result, then illustrate that the second speech recognition result meets preset wake-up condition, client can be with Speech recognition state is converted to by voice wake-up states, so that client can acquire the first phonetic order of user.
Preferably, client can use audio collecting device to acquire above-mentioned first phonetic order.
For example, client can use microphone and acquire the first voice when client is located on mobile phone or tablet computer Instruction.
Based on above-mentioned phonetic navigation method, the method for S202 is specifically described in the embodiment of the present invention.The step is specific May include:
Client judges itself current navigation state, if the current navigation state is before starting navigation, to determine institute's predicate Sound identification model is the first model;Wherein, first model goes out emerging included in first phonetic order for identification Interesting point search instruction.If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;Its In, second model goes out included in first language instruction for identification at least one in giving an order: client Control instruction, navigation hint instruction and point of interest search instruction.
It should be noted that before starting navigation and navigating under two kinds of navigational states in the embodiment of the present invention, institute The speech recognition modeling used is different.Before starting navigation, need to specify according to user by the first phonetic order Interest point information, carry out point of interest search, therefore, it is necessary to be identified included in the first phonetic order using the first model Point of interest search instruction.Due to navigating, need according to user through the first phonetic order come specified dependent instruction, Client control, navigation hint or point of interest search are carried out, therefore, it is necessary to identify that the first voice refers to using the second model Dependent instruction included in order.
Based on above-mentioned phonetic navigation method, the method for S203 is specifically described in the embodiment of the present invention.The step is specific May include:
If the current navigation state of client is before starting navigation, using the first model to the first voice of the user of acquisition Instruction carries out speech recognition, alternatively, if the current navigation state of client is to navigate, using the second model to acquisition The first phonetic order of user carries out speech recognition.
For example, client carries out the side of speech recognition using the first model or the second model to the first phonetic order Method can include but is not limited to:
Firstly, client pre-processes the first phonetic order, may include filtering processing, sampling and quantification treatment, Windowing process, endpoint detection processing and preemphasis processing etc..Then, client is extracted to by pretreated first phonetic order Characteristic information.Finally, client is by the characteristic information of extraction and the first model if carrying out speech recognition using the first model In characteristic information matched, character information corresponding to the highest characteristic information of score will be matched as the first speech recognition As a result.If carrying out speech recognition using the second model, client is by the feature in the characteristic information of extraction and the second model Information is matched, and will match the highest characteristic information of score as the first speech recognition result.
It should be noted that enabling to the first language when carrying out speech recognition to the first phonetic order using the first model The characteristic information of sound instruction and the matching score of the characteristic information of the point of interest search instruction in the first model are higher, thus may be used Using the characteristic information that instructs point of interest search as the first speech recognition result.Similarly, using the second model to the first voice When instruction carries out speech recognition, the characteristic information of the first phonetic order and the feature of the dependent instruction in the second model are enabled to The matching score of information is higher, in this manner it is possible to using the characteristic information of these dependent instructions as the first speech recognition result.This It, can be targetedly from user using different speech recognition modelings for the difference of current navigation state in inventive embodiments Phonetic order in identify the navigation needs of user, so as to improve the accuracy and reliability of Voice Navigation.
Based on above-mentioned phonetic navigation method, the method for S204 is specifically described in the embodiment of the present invention.The step is specific May include:
When the instruction of the point of interest search included in the first speech recognition result identified above with the first model, If the instruction of point of interest search included in the first speech recognition result is interest point name, client is obtained and is exported emerging with this The search result that interest point title matches.Alternatively, if the instruction of point of interest search included in the first speech recognition result is emerging Interesting vertex type title, client obtain and export the search result to match with the point of interest typonym.
For example, client obtains the side of the search result to match with interest point name or point of interest typonym Method can include but is not limited to following two:
The first: client can be searched in the local database according to interest point name or point of interest typonym Rope, to obtain the search result to match.
Second: as shown in Figure 1, if client there is no search result in the local database, client can be with Initiate the searching request for the interest point name or point of interest typonym to server, to from server obtain with it is emerging The search result that interest point title or point of interest typonym match.
In addition, client can use the voice broadcast search result after obtaining the search result, to realize search knot The output of fruit.Alternatively, client can also show the search result, after obtaining the search result to realize search result Output.
For example, the first speech recognition result is " I will go to Beihai park ", client is according to first speech recognition result Included in interest point name " Beihai park ", obtain in the local database or in server and " Beihai park " phase The search result matched, as the search result can include but is not limited at least one in following information: the address of " Beihai park ", Phone, arrival " Beihai park " routing information, with duration required for the range information of current location, arrival and mark out The electronic map of " Beihai park ".Client can broadcast the routing information of arrival " Beihai park ", and display marks out simultaneously The electronic map of " Beihai park ", to realize the search result to match to user's output with " Beihai park ".
For example, the first speech recognition result is " I will go neighbouring KFC ", client is according to first speech recognition As a result point of interest typonym included in " neighbouring KFC ", obtain in the local database or in server with The search result that " neighbouring KFC " matches, as the search result can include but is not limited at least one in following information It is a: the address of " neighbouring KFC ", phone, KFC nearest apart from current location in arrival " neighbouring KFC " road Duration required for the range information of diameter information and current location, arrival and the electronic map for marking out nearest KFC.Visitor Family end, which can be broadcasted, reaches the routing information, and display marks out the electronic map of nearest KFC simultaneously, with realize to The search result that family output matches with " neighbouring KFC ".
When in the first speech recognition result identified above with the second model including client control instruction, foundation The client control instruction carries out at least one in following control: amplification/diminution map, increase/reduction to the client Volume and unlatching/closing projecting function.
For example, the projecting function can be and electronic map is projected on the front windshield of automobile, user is facilitated to look into See electronic map.
Alternatively, when in the first speech recognition result identified above with the second model including navigation hint instruction, Instructed according to the navigation hint, obtain and export at least one in following navigation hint information: arrive at the destination it is remaining away from From information, the required time information arrived at the destination, traffic information and place road information.
For example, place road information can include but is not limited to the camera information of the title of place road, place road Or speed-limiting messages of place road etc..
Wherein, client can use the above-mentioned navigation hint information of voice broadcast, alternatively, can also show that above-mentioned navigation mentions Show information.
Alternatively, when being instructed in the first speech recognition result identified above with the second model comprising point of interest search When, it is instructed according to the point of interest search, obtains and export and match with interest point information in point of interest search instruction Search result.
For example, can be instructed according to the point of interest search, the point of interest near search client in navigation procedure, such as add Petrol station, service area etc., or also can use point of interest search instruction, the destination of Present navigation is updated, such as by destination It is changed to family or company etc..
In addition, client executes after navigating operation, according to the first speech recognition result if it find that not having in a period of time The first phonetic order is collected again, then client, which can choose, returns to voice wake-up states by speech recognition state, if user Also want to continue to use speech navigation function, then needs to enter speech recognition state using wake-up word triggering client again.
In the embodiment of the present invention, the client can also be and be handed over using voice other than it can be navigation client Mutual mode provides a user the client of the information of audio form.The client can be located at navigation terminal, smart television or On person's user equipment;The user equipment can include but is not limited to personal computer (Personal Computer, PC), a Personal digital assistant (Personal Digital Assistant, PDA), radio hand-held equipment, tablet computer (Tablet Computer), mobile phone, MP3 player, MP4 player etc..
It should be noted that the executing subject of S201~S204 can be voice guiding device, which can be located at this The application of ground terminal, or can also be the plug-in unit or Software Development Kit (Software being located locally in the application of terminal Development Kit, SDK) etc. functional units, the embodiment of the present invention is to this without being particularly limited to.
It is understood that the application can be mounted in the application program (nativeApp) in terminal, or may be used also To be a web page program (webApp) of browser in terminal, the embodiment of the present invention is to this without limiting.
The embodiment of the present invention, which further provides, realizes the Installation practice of each step and method in above method embodiment.
Referring to FIG. 3, its functional block diagram for voice guiding device provided by the embodiment of the present invention.As shown, The device includes:
Voice collecting unit 301, for acquiring the first phonetic order of user;
Model treatment unit 302, for determining speech recognition modeling according to current navigation state;
Voice recognition unit 303 carries out the first phonetic order of the user for utilizing the speech recognition modeling Speech recognition, to obtain the first speech recognition result;
Navigate execution unit 304, for executing navigation operation according to first speech recognition result.
Preferably, the voice collecting unit 301 is also used to the second phonetic order of monitoring users;
Described device further include: voice wakeup unit 305, for waking up model to second phonetic order using voice Speech recognition is carried out, to obtain the second speech recognition result;If second speech recognition result meets preset wake-up condition, Trigger the first phonetic order of the voice collecting unit acquisition user.
Preferably, the model treatment unit 302, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
Preferably, the model treatment unit 302, is specifically used for:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one It is a: client control instruction, navigation hint instruction and point of interest search instruction.
Preferably, the navigation execution unit 304, is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export The search result to match with the interest point name;Alternatively, if point of interest included in first speech recognition result Search instruction is point of interest typonym, obtains and exports the search result to match with the point of interest typonym.
Preferably, the navigation execution unit 304, is specifically used for:
It is right according to the client control instruction if in first speech recognition result including client control instruction The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search It takes and exports and the search result that matches of interest point information in point of interest search instruction.
Method shown in Fig. 2 is able to carry out by each unit in this present embodiment, the part that the present embodiment is not described in detail, It can refer to the related description to Fig. 2.
The technical solution of the embodiment of the present invention has the advantages that
The first phonetic order that the embodiment of the present invention passes through acquisition user;To determine voice according to current navigation state Identification model;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain First speech recognition result, and according to first speech recognition result, execute navigation operation.Compared with prior art, originally The technical solution that inventive embodiments provide can be automatic to execute navigation operation according to the phonetic order of user, thus realize to Family provides navigation Service, does not need user and be operated manually to realize navigation feature, therefore be able to solve the prior art The problem that operating cost is higher in middle navigation procedure and Voice Navigation process efficiency is lower, can reduce the behaviour in navigation procedure Make cost, improve the treatment effeciency of Voice Navigation, guarantees driving safety.
In addition, can realize that voice wakes up function in technical solution provided in an embodiment of the present invention during Voice Navigation Can, only when the phonetic order of user meets wake-up condition, it just will do it Voice Navigation, therefore can be to avoid the language to user The misrecognition of sound instruction, improves the accuracy of Voice Navigation.
It is apparent to those skilled in the art that for convenience and simplicity of description, the system of foregoing description, The specific work process of device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided by the present invention, it should be understood that disclosed system, device and method can be with It realizes by another way.For example, the apparatus embodiments described above are merely exemplary, for example, the unit It divides, only a kind of logical function partition, there may be another division manner in actual implementation, for example, multiple units or group Part can be combined or can be integrated into another system, or some features can be ignored or not executed.Another point, it is shown Or the mutual coupling, direct-coupling or communication connection discussed can be through some interfaces, device or unit it is indirect Coupling or communication connection can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme 's.
It, can also be in addition, the functional units in various embodiments of the present invention may be integrated into one processing unit It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list Member both can take the form of hardware realization, can also realize in the form of hardware adds SFU software functional unit.
The above-mentioned integrated unit being realized in the form of SFU software functional unit can store and computer-readable deposit at one In storage media.Above-mentioned SFU software functional unit is stored in a storage medium, including some instructions are used so that a computer It is each that device (can be personal computer, server or network equipment etc.) or processor (Processor) execute the present invention The part steps of embodiment the method.And storage medium above-mentioned includes: USB flash disk, mobile hard disk, read-only memory (Read- Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic or disk etc. it is various It can store the medium of program code.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention Within mind and principle, any modification, equivalent substitution, improvement and etc. done be should be included within the scope of the present invention.

Claims (8)

1. a kind of phonetic navigation method, which is characterized in that the described method includes:
Acquire the first phonetic order of user;
According to current navigation state, the corresponding speech recognition modeling of the current navigation state, the current navigation state are determined It include: before starting navigation and to navigate;Wherein, if the current navigation state is before starting navigation, to determine the voice Identification model is the first model, and first model goes out point of interest search included in first phonetic order for identification Instruction;If the current navigation state is to navigate, determine that the speech recognition modeling is the second model, second mould Type goes out included in first phonetic order for identification at least one in giving an order: client control instruction, navigation Hint instructions and point of interest search instruction;
Using the corresponding speech recognition modeling of the current navigation state, voice knowledge is carried out to the first phonetic order of the user Not, to obtain the first speech recognition result;
According to first speech recognition result, navigation operation is executed.
2. the method according to claim 1, wherein the first phonetic order of the acquisition user, comprising:
Second phonetic order of monitoring users;
Model is waken up using voice, speech recognition is carried out to second phonetic order, to obtain the second speech recognition result;
If second speech recognition result meets preset wake-up condition, the first phonetic order of user is acquired.
3. execution is led the method according to claim 1, wherein described according to first speech recognition result Boat operation, comprising:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export and institute State the search result that interest point name matches;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtains and export The search result to match with the point of interest typonym.
4. execution is led the method according to claim 1, wherein described according to first speech recognition result Boat operation, comprising:
If including client control instruction in first speech recognition result, according to the client control instruction, to described Client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing projecting function;
If being instructed in first speech recognition result comprising navigation hint, is instructed according to the navigation hint, obtain and export At least one in following navigation hint information: the Distance Remaining information that arrives at the destination, the required time information arrived at the destination, Traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instruct, is obtained simultaneously according to the point of interest search The search result that output matches with interest point information in point of interest search instruction.
5. a kind of voice guiding device, which is characterized in that described device includes:
Voice collecting unit, for acquiring the first phonetic order of user;
Model treatment unit, for determining the corresponding speech recognition modeling of the current navigation state according to current navigation state, The current navigation state includes: before starting navigation and to navigate;
Voice recognition unit, for utilizing the corresponding speech recognition modeling of the current navigation state, to the first of the user Phonetic order carries out speech recognition, to obtain the first speech recognition result;
Navigate execution unit, for executing navigation operation according to first speech recognition result;
Wherein, the model treatment unit, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model, first mould before starting navigation Type goes out the instruction of point of interest search included in first phonetic order for identification;
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model, second mould Type goes out included in first phonetic order for identification at least one in giving an order: client control instruction, navigation Hint instructions and point of interest search instruction.
6. device according to claim 5, which is characterized in that
The voice collecting unit is also used to the second phonetic order of monitoring users;
Described device further include:
Voice wakeup unit, for waking up model to second phonetic order progress speech recognition, to obtain the using voice Two speech recognition results;If second speech recognition result meets preset wake-up condition, the voice collecting unit is triggered Acquire the first phonetic order of user.
7. device according to claim 5, which is characterized in that the navigation execution unit is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export and institute State the search result that interest point name matches;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtains and export The search result to match with the point of interest typonym.
8. device according to claim 5, which is characterized in that the navigation execution unit is specifically used for:
If including client control instruction in first speech recognition result, according to the client control instruction, to described Client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing projecting function;
If being instructed in first speech recognition result comprising navigation hint, is instructed according to the navigation hint, obtain and export At least one in following navigation hint information: the Distance Remaining information that arrives at the destination, the required time information arrived at the destination, Traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instruct, is obtained simultaneously according to the point of interest search The search result that output matches with interest point information in point of interest search instruction.
CN201410742287.5A 2014-12-05 2014-12-05 A kind of phonetic navigation method and device Active CN104535071B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410742287.5A CN104535071B (en) 2014-12-05 2014-12-05 A kind of phonetic navigation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410742287.5A CN104535071B (en) 2014-12-05 2014-12-05 A kind of phonetic navigation method and device

Publications (2)

Publication Number Publication Date
CN104535071A CN104535071A (en) 2015-04-22
CN104535071B true CN104535071B (en) 2018-12-14

Family

ID=52850646

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410742287.5A Active CN104535071B (en) 2014-12-05 2014-12-05 A kind of phonetic navigation method and device

Country Status (1)

Country Link
CN (1) CN104535071B (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105139851A (en) * 2015-09-17 2015-12-09 努比亚技术有限公司 Desktop application icon organization mobile terminal and method
CN105681579B (en) * 2016-03-11 2020-01-10 Oppo广东移动通信有限公司 Terminal and screen control method and device thereof in navigation state
CN105890615A (en) * 2016-04-11 2016-08-24 深圳市轱辘软件开发有限公司 Navigation method and device
CN107305483A (en) * 2016-04-25 2017-10-31 北京搜狗科技发展有限公司 A kind of voice interactive method and device based on semantics recognition
CN106289296B (en) * 2016-09-05 2020-03-24 广州极飞科技有限公司 Road navigation method and device
CN107014390A (en) * 2017-03-16 2017-08-04 北京云知声信息技术有限公司 A kind of route reminding method and device
CN107329730B (en) * 2017-07-03 2021-03-16 科大讯飞股份有限公司 Voice prompt message generation method and device
CN107600075A (en) * 2017-08-23 2018-01-19 深圳市沃特沃德股份有限公司 The control method and device of onboard system
CN107678316A (en) * 2017-10-27 2018-02-09 姜俊 A kind of environment inside car regulating system and method
CN108168540A (en) * 2017-12-22 2018-06-15 福建中金在线信息科技有限公司 A kind of intelligent glasses air navigation aid, device and intelligent glasses
CN108392269B (en) * 2017-12-29 2021-08-03 广州布莱医疗科技有限公司 Operation assisting method and operation assisting robot
CN108307069B (en) * 2018-01-29 2020-08-07 Oppo广东移动通信有限公司 Navigation operation method, navigation operation device and mobile terminal
CN108827328A (en) * 2018-04-24 2018-11-16 联想(北京)有限公司 Voice interactive method, device, system and computer-readable medium
CN108806682B (en) * 2018-06-12 2020-12-01 奇瑞汽车股份有限公司 Method and device for acquiring weather information
CN109065045A (en) * 2018-08-30 2018-12-21 出门问问信息科技有限公司 Audio recognition method, device, electronic equipment and computer readable storage medium
CN110770820A (en) * 2018-08-30 2020-02-07 深圳市大疆创新科技有限公司 Speech recognition method, apparatus, photographing system, and computer-readable storage medium
CN108986811B (en) * 2018-08-31 2021-05-28 北京新能源汽车股份有限公司 Voice recognition detection method, device and equipment
CN109000679B (en) * 2018-09-21 2021-03-05 斑马网络技术有限公司 Path prediction method, device, system and storage medium
CN110136705B (en) * 2019-04-10 2022-06-14 华为技术有限公司 Man-machine interaction method and electronic equipment
CN110057379B (en) * 2019-05-29 2021-09-03 广州小鹏汽车科技有限公司 Secondary navigation method and device for vehicle navigation and vehicle
CN111949780B (en) * 2020-07-31 2021-12-31 八维通科技有限公司 Automatic interaction method along trip route
CN114485718A (en) * 2022-01-05 2022-05-13 腾讯科技(深圳)有限公司 Voice navigation method, device, electronic equipment, storage medium and program product
CN114964300B (en) * 2022-06-22 2023-03-28 深圳市智远联科技有限公司 Voice recognition method and navigation device
CN114913855A (en) * 2022-07-11 2022-08-16 广州小鹏汽车科技有限公司 Voice interaction method, server and storage medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102063901A (en) * 2010-12-02 2011-05-18 深圳市凯立德欣软件技术有限公司 Voice identification method for position service equipment and position service equipment
CN103674012A (en) * 2012-09-21 2014-03-26 高德软件有限公司 Voice customizing method and device and voice identification method and device
CN103971681A (en) * 2014-04-24 2014-08-06 百度在线网络技术(北京)有限公司 Voice recognition method and system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102063901A (en) * 2010-12-02 2011-05-18 深圳市凯立德欣软件技术有限公司 Voice identification method for position service equipment and position service equipment
CN103674012A (en) * 2012-09-21 2014-03-26 高德软件有限公司 Voice customizing method and device and voice identification method and device
CN103971681A (en) * 2014-04-24 2014-08-06 百度在线网络技术(北京)有限公司 Voice recognition method and system

Also Published As

Publication number Publication date
CN104535071A (en) 2015-04-22

Similar Documents

Publication Publication Date Title
CN104535071B (en) A kind of phonetic navigation method and device
CN107240398B (en) Intelligent voice interaction method and device
CN108305633B (en) Speech verification method, apparatus, computer equipment and computer readable storage medium
CN104794122B (en) Position information recommendation method, device and system
CN107665710B (en) Mobile terminal voice data processing method and device
EP3754651B1 (en) Dynamic language model adaptation for generating voice-based search queries
JP5158174B2 (en) Voice recognition device
US9443202B2 (en) Adaptation of context models
US9188456B2 (en) System and method of fixing mistakes by going back in an electronic device
CN108989541A (en) Session initiation device, system, vehicle and method based on situation
US10388280B2 (en) Method and apparatus for managing multiple voice operation trigger phrases
CN105004348A (en) Voice navigation method and system
CN106959690B (en) Method, device and equipment for searching unmanned vehicle and storage medium
WO2011092639A1 (en) Systems, methods, and apparatuses for providing context-based navigation services
JP2011179917A (en) Information recording device, information recording method, information recording program, and recording medium
CN110972112B (en) Subway running direction determining method, device, terminal and storage medium
EP2660562A1 (en) Route Guidance Apparatus and Method with Voice Recognition
JP2012088370A (en) Voice recognition system, voice recognition terminal and center
US9791925B2 (en) Information acquisition method, information acquisition system, and non-transitory recording medium for user of motor vehicle
US10593323B2 (en) Keyword generation apparatus and keyword generation method
CN113706915A (en) Parking prompting method, device, equipment and storage medium
CN109900284A (en) Navigation route planning method and apparatus
JP2022103675A (en) Information processing device, information processing method, and program
CN114118582A (en) Destination prediction method, destination prediction device, electronic terminal and storage medium
CN113032681A (en) Method, apparatus, electronic device, and medium for map search

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20200409

Address after: 518000 301, floor 3, unit D, productivity building, No. 5, Gaoxin Middle Road, Science Park, Nanshan District, Shenzhen City, Guangdong Province

Co-patentee after: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.

Patentee after: BAIDU INTERNATIONAL TECHNOLOGY (SHENZHEN) Co.,Ltd.

Address before: 100085 Baidu building, No. 10, ten Street, Haidian District, Beijing

Patentee before: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd.