CN104535071B - A kind of phonetic navigation method and device - Google Patents
A kind of phonetic navigation method and device Download PDFInfo
- Publication number
- CN104535071B CN104535071B CN201410742287.5A CN201410742287A CN104535071B CN 104535071 B CN104535071 B CN 104535071B CN 201410742287 A CN201410742287 A CN 201410742287A CN 104535071 B CN104535071 B CN 104535071B
- Authority
- CN
- China
- Prior art keywords
- speech recognition
- navigation
- point
- interest
- recognition result
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 50
- 238000012544 monitoring process Methods 0.000 claims description 7
- 230000003321 amplification Effects 0.000 claims description 6
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 6
- 230000009467 reduction Effects 0.000 claims description 6
- 230000002618 waking effect Effects 0.000 claims description 3
- 230000001960 triggered effect Effects 0.000 claims description 2
- 230000008569 process Effects 0.000 abstract description 8
- 230000006870 function Effects 0.000 description 14
- 238000001514 detection method Methods 0.000 description 5
- 230000001419 dependent effect Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 235000013399 edible fruits Nutrition 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 238000004891 communication Methods 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000000605 extraction Methods 0.000 description 2
- 230000037007 arousal Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002045 lasting effect Effects 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 238000005192 partition Methods 0.000 description 1
- 238000011002 quantification Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/34—Route searching; Route guidance
- G01C21/36—Input/output arrangements for on-board computers
- G01C21/3605—Destination input or retrieval
- G01C21/3608—Destination input or retrieval using speech input, e.g. using speech recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/90—Details of database functions independent of the retrieved data types
- G06F16/95—Retrieval from the web
- G06F16/953—Querying, e.g. by the use of web search engines
- G06F16/9537—Spatial or temporal dependent retrieval, e.g. spatiotemporal queries
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Databases & Information Systems (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Multimedia (AREA)
- Automation & Control Theory (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Navigation (AREA)
Abstract
The embodiment of the invention provides a kind of phonetic navigation method and devices.On the one hand, the first phonetic order that the embodiment of the present invention passes through acquisition user;To determine speech recognition modeling according to current navigation state;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain the first speech recognition result, and according to first speech recognition result, executes navigation operation.Therefore, technical solution provided in an embodiment of the present invention is able to solve in navigation procedure in the prior art the problem that operating cost is higher, Voice Navigation process efficiency is lower.
Description
[technical field]
The present invention relates to human-computer interaction technology more particularly to a kind of phonetic navigation methods and device.
[background technique]
With deep and Internet communication technology the development of global position system GPS civil nature, people are more and more
Ground uses navigation client.Navigation client can provide a user the navigation function such as path planning and guidance based on electronic map
Energy.Since its trip to people brings very big convenience, it is increasingly becoming essential a part in life.With
Attention rate and use degree to navigation client is higher and higher, and people are not only satisfied with base provided by navigation client
This navigation feature, and be desirable to it and be capable of providing more accurate property, more humanized navigation Service.
Sound prompt function is the important component of the provided navigation feature of navigation client, due to navigation client
Particularity will be absorbed in driving in driver's startup procedure and pay attention to surface conditions, therefore cannot frequently observe navigation client
Interface is held to obtain route relevant information, therefore, sound prompt function is just particularly important.However, navigation client only has
There is sound prompt function, there is still a need for users when providing navigation Service before road, is manually entered destination letter on driving vehicle
Breath is needed to be operated manually completion after stopping and be checked if user needs to obtain other navigation informations in driving procedure,
Therefore the operating cost at present in navigation procedure is higher, and the treatment effeciency of Voice Navigation is relatively low.
[summary of the invention]
In view of this, the embodiment of the invention provides a kind of phonetic navigation method and devices, to solve in the prior art
The problem that operating cost is higher in navigation procedure, Voice Navigation process efficiency is lower.
The one side of the embodiment of the present invention provides a kind of phonetic navigation method, comprising:
Acquire the first phonetic order of user;
According to current navigation state, speech recognition modeling is determined;
Using the speech recognition modeling, speech recognition is carried out to the first phonetic order of the user, to obtain first
Speech recognition result;
According to first speech recognition result, navigation operation is executed.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the acquisition are used
First phonetic order at family, comprising:
Second phonetic order of monitoring users;
Model is waken up using voice, speech recognition is carried out to second phonetic order, to obtain the second speech recognition knot
Fruit;
If second speech recognition result meets preset wake-up condition, the first phonetic order of user is acquired.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the foundation are worked as
Preceding navigational state, determines speech recognition modeling, comprising:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the foundation are worked as
Preceding navigational state, determines speech recognition modeling, comprising:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one
It is a: client control instruction, navigation hint instruction and point of interest search instruction.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, it is described according to institute
The first speech recognition result is stated, navigation operation is executed, comprising:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export
The search result to match with the interest point name;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtain simultaneously
The search result that output matches with the point of interest typonym.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, it is described according to institute
The first speech recognition result is stated, navigation operation is executed, comprising:
It is right according to the client control instruction if in first speech recognition result including client control instruction
The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function
Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint
Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination
Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search
It takes and exports and the search result that matches of interest point information in point of interest search instruction.
The one side of the embodiment of the present invention provides a kind of voice guiding device, comprising:
Voice collecting unit, for acquiring the first phonetic order of user;
Model treatment unit, for determining speech recognition modeling according to current navigation state;
Voice recognition unit carries out language to the first phonetic order of the user for utilizing the speech recognition modeling
Sound identification, to obtain the first speech recognition result;
Navigate execution unit, for executing navigation operation according to first speech recognition result.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation,
The voice collecting unit is also used to the second phonetic order of monitoring users;
Described device further include:
Voice wakeup unit carries out speech recognition to second phonetic order for waking up model using voice, to obtain
Obtain the second speech recognition result;If second speech recognition result meets preset wake-up condition, the voice collecting is triggered
The first phonetic order of unit acquisition user.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, at the model
Unit is managed, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, at the model
Unit is managed, is specifically used for:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one
It is a: client control instruction, navigation hint instruction and point of interest search instruction.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the navigation are held
Row unit, is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export
The search result to match with the interest point name;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtain simultaneously
The search result that output matches with the point of interest typonym.
The aspect and any possible implementation manners as described above, it is further provided a kind of implementation, the navigation are held
Row unit, is specifically used for:
It is right according to the client control instruction if in first speech recognition result including client control instruction
The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function
Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint
Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination
Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search
It takes and exports and the search result that matches of interest point information in point of interest search instruction.
As can be seen from the above technical solutions, the embodiment of the present invention has the advantages that
The first phonetic order that the embodiment of the present invention passes through acquisition user;To determine voice according to current navigation state
Identification model;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain
First speech recognition result, and according to first speech recognition result, execute navigation operation.Compared with prior art, originally
The technical solution that inventive embodiments provide can be automatic to execute navigation operation according to the phonetic order of user, thus realize to
Family provides navigation Service, does not need user and be operated manually to realize navigation feature, therefore be able to solve the prior art
The problem that operating cost is higher in middle navigation procedure and Voice Navigation process efficiency is lower, can reduce the behaviour in navigation procedure
Make cost, improve the treatment effeciency of Voice Navigation, guarantees driving safety.
[Detailed description of the invention]
In order to illustrate the technical solution of the embodiments of the present invention more clearly, below will be to needed in the embodiment attached
Figure is briefly described, it should be apparent that, drawings in the following description are only some embodiments of the invention, for this field
For those of ordinary skill, without any creative labor, it can also be obtained according to these attached drawings other attached
Figure.
Fig. 1 is the exemplary system figure that technical solution provided by the embodiment of the present invention uses;
Fig. 2 is the flow diagram of phonetic navigation method provided by the embodiment of the present invention;
Fig. 3 is the functional block diagram of voice guiding device provided by the embodiment of the present invention.
[specific embodiment]
For a better understanding of the technical solution of the present invention, being retouched in detail to the embodiment of the present invention with reference to the accompanying drawing
It states.
It will be appreciated that described embodiments are only a part of the embodiments of the present invention, instead of all the embodiments.Base
Embodiment in the present invention, it is obtained by those of ordinary skill in the art without making creative efforts it is all its
Its embodiment, shall fall within the protection scope of the present invention.
The term used in embodiments of the present invention is only to be not intended to be limiting merely for for the purpose of describing particular embodiments
The present invention.In the embodiment of the present invention and the "an" of singular used in the attached claims, " described " and "the"
It is also intended to including most forms, unless the context clearly indicates other meaning.
It should be appreciated that term "and/or" used herein is only a kind of incidence relation for describing affiliated partner, indicate
There may be three kinds of relationships, for example, A and/or B, can indicate: individualism A, exist simultaneously A and B, individualism B these three
Situation.In addition, character "/" herein, typicallys represent the relationship that forward-backward correlation object is a kind of "or".
It will be appreciated that though speech recognition knot may be described using term first, second etc. in embodiments of the present invention
Fruit, but these keywords should not necessarily be limited by these terms.These terms are only used to for keyword being distinguished from each other out.For example, not taking off
In the case where from range of embodiment of the invention, the first speech recognition result can also be referred to as the second speech recognition result, similar
Ground, the second speech recognition result can also be referred to as the first speech recognition result.
Depending on context, word as used in this " if " can be construed to " ... when " or " when ...
When " or " in response to determination " or " in response to detection ".Similarly, depend on context, phrase " if it is determined that " or " if detection
(condition or event of statement) " can be construed to " when determining " or " in response to determination " or " when the detection (condition of statement
Or event) when " or " in response to detection (condition or event of statement) ".
The system that technical solution provided by the embodiment of the present invention uses is as shown in Figure 1, mainly by client and server
It forming, method and apparatus provided by the embodiment of the present invention are realized in client-side, it is mainly used for the phonetic order according to user,
Voice Navigation service is provided a user, realizes the speech navigation function of client.
The embodiment of the present invention provides a kind of phonetic navigation method, referring to FIG. 2, it is language provided by the embodiment of the present invention
The flow diagram of sound air navigation aid, as shown, method includes the following steps:
S201 acquires the first phonetic order of user.
S202 determines speech recognition modeling according to current navigation state.
S203 carries out speech recognition to the first phonetic order of the user, to obtain using the speech recognition modeling
First speech recognition result.
S204 executes navigation operation according to first speech recognition result.
Based on above-mentioned phonetic navigation method, the method for S201 is specifically described in the embodiment of the present invention.The step is specific
May include:
Firstly, the second phonetic order of client monitoring users.Then, client wakes up model to described the using voice
Two phonetic orders carry out speech recognition, to obtain the second speech recognition result.Finally, client judges the second speech recognition knot
Whether fruit meets preset wake-up condition, if second speech recognition result meets preset wake-up condition, client acquisition
The first phonetic order of user.
It should be noted that, in order to avoid the misrecognition of the phonetic order to user, needing to realize during Voice Navigation
Real-time voice arousal function, client is under voice wake-up states, if the phonetic order of the user listened to meets wake-up bars
Part, into speech recognition state, to just start the phonetic order of reception user.
For example, client can open a listening thread, the second phonetic order for lasting monitoring users.Client
Model is waken up using voice, to the progress speech recognition of the second phonetic order is listened to obtain, to obtain the second speech recognition result.It should
Voice wakes up model and goes out included in the second phonetic order to wake up word for identification.If the client determine that the second voice is known
Include preset wake-up word in other result, then illustrate that the second speech recognition result meets preset wake-up condition, client can be with
Speech recognition state is converted to by voice wake-up states, so that client can acquire the first phonetic order of user.
Preferably, client can use audio collecting device to acquire above-mentioned first phonetic order.
For example, client can use microphone and acquire the first voice when client is located on mobile phone or tablet computer
Instruction.
Based on above-mentioned phonetic navigation method, the method for S202 is specifically described in the embodiment of the present invention.The step is specific
May include:
Client judges itself current navigation state, if the current navigation state is before starting navigation, to determine institute's predicate
Sound identification model is the first model;Wherein, first model goes out emerging included in first phonetic order for identification
Interesting point search instruction.If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;Its
In, second model goes out included in first language instruction for identification at least one in giving an order: client
Control instruction, navigation hint instruction and point of interest search instruction.
It should be noted that before starting navigation and navigating under two kinds of navigational states in the embodiment of the present invention, institute
The speech recognition modeling used is different.Before starting navigation, need to specify according to user by the first phonetic order
Interest point information, carry out point of interest search, therefore, it is necessary to be identified included in the first phonetic order using the first model
Point of interest search instruction.Due to navigating, need according to user through the first phonetic order come specified dependent instruction,
Client control, navigation hint or point of interest search are carried out, therefore, it is necessary to identify that the first voice refers to using the second model
Dependent instruction included in order.
Based on above-mentioned phonetic navigation method, the method for S203 is specifically described in the embodiment of the present invention.The step is specific
May include:
If the current navigation state of client is before starting navigation, using the first model to the first voice of the user of acquisition
Instruction carries out speech recognition, alternatively, if the current navigation state of client is to navigate, using the second model to acquisition
The first phonetic order of user carries out speech recognition.
For example, client carries out the side of speech recognition using the first model or the second model to the first phonetic order
Method can include but is not limited to:
Firstly, client pre-processes the first phonetic order, may include filtering processing, sampling and quantification treatment,
Windowing process, endpoint detection processing and preemphasis processing etc..Then, client is extracted to by pretreated first phonetic order
Characteristic information.Finally, client is by the characteristic information of extraction and the first model if carrying out speech recognition using the first model
In characteristic information matched, character information corresponding to the highest characteristic information of score will be matched as the first speech recognition
As a result.If carrying out speech recognition using the second model, client is by the feature in the characteristic information of extraction and the second model
Information is matched, and will match the highest characteristic information of score as the first speech recognition result.
It should be noted that enabling to the first language when carrying out speech recognition to the first phonetic order using the first model
The characteristic information of sound instruction and the matching score of the characteristic information of the point of interest search instruction in the first model are higher, thus may be used
Using the characteristic information that instructs point of interest search as the first speech recognition result.Similarly, using the second model to the first voice
When instruction carries out speech recognition, the characteristic information of the first phonetic order and the feature of the dependent instruction in the second model are enabled to
The matching score of information is higher, in this manner it is possible to using the characteristic information of these dependent instructions as the first speech recognition result.This
It, can be targetedly from user using different speech recognition modelings for the difference of current navigation state in inventive embodiments
Phonetic order in identify the navigation needs of user, so as to improve the accuracy and reliability of Voice Navigation.
Based on above-mentioned phonetic navigation method, the method for S204 is specifically described in the embodiment of the present invention.The step is specific
May include:
When the instruction of the point of interest search included in the first speech recognition result identified above with the first model,
If the instruction of point of interest search included in the first speech recognition result is interest point name, client is obtained and is exported emerging with this
The search result that interest point title matches.Alternatively, if the instruction of point of interest search included in the first speech recognition result is emerging
Interesting vertex type title, client obtain and export the search result to match with the point of interest typonym.
For example, client obtains the side of the search result to match with interest point name or point of interest typonym
Method can include but is not limited to following two:
The first: client can be searched in the local database according to interest point name or point of interest typonym
Rope, to obtain the search result to match.
Second: as shown in Figure 1, if client there is no search result in the local database, client can be with
Initiate the searching request for the interest point name or point of interest typonym to server, to from server obtain with it is emerging
The search result that interest point title or point of interest typonym match.
In addition, client can use the voice broadcast search result after obtaining the search result, to realize search knot
The output of fruit.Alternatively, client can also show the search result, after obtaining the search result to realize search result
Output.
For example, the first speech recognition result is " I will go to Beihai park ", client is according to first speech recognition result
Included in interest point name " Beihai park ", obtain in the local database or in server and " Beihai park " phase
The search result matched, as the search result can include but is not limited at least one in following information: the address of " Beihai park ",
Phone, arrival " Beihai park " routing information, with duration required for the range information of current location, arrival and mark out
The electronic map of " Beihai park ".Client can broadcast the routing information of arrival " Beihai park ", and display marks out simultaneously
The electronic map of " Beihai park ", to realize the search result to match to user's output with " Beihai park ".
For example, the first speech recognition result is " I will go neighbouring KFC ", client is according to first speech recognition
As a result point of interest typonym included in " neighbouring KFC ", obtain in the local database or in server with
The search result that " neighbouring KFC " matches, as the search result can include but is not limited at least one in following information
It is a: the address of " neighbouring KFC ", phone, KFC nearest apart from current location in arrival " neighbouring KFC " road
Duration required for the range information of diameter information and current location, arrival and the electronic map for marking out nearest KFC.Visitor
Family end, which can be broadcasted, reaches the routing information, and display marks out the electronic map of nearest KFC simultaneously, with realize to
The search result that family output matches with " neighbouring KFC ".
When in the first speech recognition result identified above with the second model including client control instruction, foundation
The client control instruction carries out at least one in following control: amplification/diminution map, increase/reduction to the client
Volume and unlatching/closing projecting function.
For example, the projecting function can be and electronic map is projected on the front windshield of automobile, user is facilitated to look into
See electronic map.
Alternatively, when in the first speech recognition result identified above with the second model including navigation hint instruction,
Instructed according to the navigation hint, obtain and export at least one in following navigation hint information: arrive at the destination it is remaining away from
From information, the required time information arrived at the destination, traffic information and place road information.
For example, place road information can include but is not limited to the camera information of the title of place road, place road
Or speed-limiting messages of place road etc..
Wherein, client can use the above-mentioned navigation hint information of voice broadcast, alternatively, can also show that above-mentioned navigation mentions
Show information.
Alternatively, when being instructed in the first speech recognition result identified above with the second model comprising point of interest search
When, it is instructed according to the point of interest search, obtains and export and match with interest point information in point of interest search instruction
Search result.
For example, can be instructed according to the point of interest search, the point of interest near search client in navigation procedure, such as add
Petrol station, service area etc., or also can use point of interest search instruction, the destination of Present navigation is updated, such as by destination
It is changed to family or company etc..
In addition, client executes after navigating operation, according to the first speech recognition result if it find that not having in a period of time
The first phonetic order is collected again, then client, which can choose, returns to voice wake-up states by speech recognition state, if user
Also want to continue to use speech navigation function, then needs to enter speech recognition state using wake-up word triggering client again.
In the embodiment of the present invention, the client can also be and be handed over using voice other than it can be navigation client
Mutual mode provides a user the client of the information of audio form.The client can be located at navigation terminal, smart television or
On person's user equipment;The user equipment can include but is not limited to personal computer (Personal Computer, PC), a
Personal digital assistant (Personal Digital Assistant, PDA), radio hand-held equipment, tablet computer (Tablet
Computer), mobile phone, MP3 player, MP4 player etc..
It should be noted that the executing subject of S201~S204 can be voice guiding device, which can be located at this
The application of ground terminal, or can also be the plug-in unit or Software Development Kit (Software being located locally in the application of terminal
Development Kit, SDK) etc. functional units, the embodiment of the present invention is to this without being particularly limited to.
It is understood that the application can be mounted in the application program (nativeApp) in terminal, or may be used also
To be a web page program (webApp) of browser in terminal, the embodiment of the present invention is to this without limiting.
The embodiment of the present invention, which further provides, realizes the Installation practice of each step and method in above method embodiment.
Referring to FIG. 3, its functional block diagram for voice guiding device provided by the embodiment of the present invention.As shown,
The device includes:
Voice collecting unit 301, for acquiring the first phonetic order of user;
Model treatment unit 302, for determining speech recognition modeling according to current navigation state;
Voice recognition unit 303 carries out the first phonetic order of the user for utilizing the speech recognition modeling
Speech recognition, to obtain the first speech recognition result;
Navigate execution unit 304, for executing navigation operation according to first speech recognition result.
Preferably, the voice collecting unit 301 is also used to the second phonetic order of monitoring users;
Described device further include: voice wakeup unit 305, for waking up model to second phonetic order using voice
Speech recognition is carried out, to obtain the second speech recognition result;If second speech recognition result meets preset wake-up condition,
Trigger the first phonetic order of the voice collecting unit acquisition user.
Preferably, the model treatment unit 302, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model before starting navigation;
Wherein, first model goes out the instruction of point of interest search included in first phonetic order for identification.
Preferably, the model treatment unit 302, is specifically used for:
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model;
Wherein, second model goes out included in first language instruction for identification in giving an order at least one
It is a: client control instruction, navigation hint instruction and point of interest search instruction.
Preferably, the navigation execution unit 304, is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export
The search result to match with the interest point name;Alternatively, if point of interest included in first speech recognition result
Search instruction is point of interest typonym, obtains and exports the search result to match with the point of interest typonym.
Preferably, the navigation execution unit 304, is specifically used for:
It is right according to the client control instruction if in first speech recognition result including client control instruction
The client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing project function
Energy;
If being instructed in first speech recognition result comprising navigation hint, instruct, is obtained simultaneously according to the navigation hint
Export at least one in following navigation hint information: the Distance Remaining information arrived at the destination, the required time arrived at the destination
Information, traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instructs, obtain according to the point of interest search
It takes and exports and the search result that matches of interest point information in point of interest search instruction.
Method shown in Fig. 2 is able to carry out by each unit in this present embodiment, the part that the present embodiment is not described in detail,
It can refer to the related description to Fig. 2.
The technical solution of the embodiment of the present invention has the advantages that
The first phonetic order that the embodiment of the present invention passes through acquisition user;To determine voice according to current navigation state
Identification model;And then the speech recognition modeling is utilized, speech recognition is carried out to the first phonetic order of the user, to obtain
First speech recognition result, and according to first speech recognition result, execute navigation operation.Compared with prior art, originally
The technical solution that inventive embodiments provide can be automatic to execute navigation operation according to the phonetic order of user, thus realize to
Family provides navigation Service, does not need user and be operated manually to realize navigation feature, therefore be able to solve the prior art
The problem that operating cost is higher in middle navigation procedure and Voice Navigation process efficiency is lower, can reduce the behaviour in navigation procedure
Make cost, improve the treatment effeciency of Voice Navigation, guarantees driving safety.
In addition, can realize that voice wakes up function in technical solution provided in an embodiment of the present invention during Voice Navigation
Can, only when the phonetic order of user meets wake-up condition, it just will do it Voice Navigation, therefore can be to avoid the language to user
The misrecognition of sound instruction, improves the accuracy of Voice Navigation.
It is apparent to those skilled in the art that for convenience and simplicity of description, the system of foregoing description,
The specific work process of device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein.
In several embodiments provided by the present invention, it should be understood that disclosed system, device and method can be with
It realizes by another way.For example, the apparatus embodiments described above are merely exemplary, for example, the unit
It divides, only a kind of logical function partition, there may be another division manner in actual implementation, for example, multiple units or group
Part can be combined or can be integrated into another system, or some features can be ignored or not executed.Another point, it is shown
Or the mutual coupling, direct-coupling or communication connection discussed can be through some interfaces, device or unit it is indirect
Coupling or communication connection can be electrical property, mechanical or other forms.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit
The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple
In network unit.It can select some or all of unit therein according to the actual needs to realize the mesh of this embodiment scheme
's.
It, can also be in addition, the functional units in various embodiments of the present invention may be integrated into one processing unit
It is that each unit physically exists alone, can also be integrated in one unit with two or more units.Above-mentioned integrated list
Member both can take the form of hardware realization, can also realize in the form of hardware adds SFU software functional unit.
The above-mentioned integrated unit being realized in the form of SFU software functional unit can store and computer-readable deposit at one
In storage media.Above-mentioned SFU software functional unit is stored in a storage medium, including some instructions are used so that a computer
It is each that device (can be personal computer, server or network equipment etc.) or processor (Processor) execute the present invention
The part steps of embodiment the method.And storage medium above-mentioned includes: USB flash disk, mobile hard disk, read-only memory (Read-
Only Memory, ROM), random access memory (Random Access Memory, RAM), magnetic or disk etc. it is various
It can store the medium of program code.
The foregoing is merely illustrative of the preferred embodiments of the present invention, is not intended to limit the invention, all in essence of the invention
Within mind and principle, any modification, equivalent substitution, improvement and etc. done be should be included within the scope of the present invention.
Claims (8)
1. a kind of phonetic navigation method, which is characterized in that the described method includes:
Acquire the first phonetic order of user;
According to current navigation state, the corresponding speech recognition modeling of the current navigation state, the current navigation state are determined
It include: before starting navigation and to navigate;Wherein, if the current navigation state is before starting navigation, to determine the voice
Identification model is the first model, and first model goes out point of interest search included in first phonetic order for identification
Instruction;If the current navigation state is to navigate, determine that the speech recognition modeling is the second model, second mould
Type goes out included in first phonetic order for identification at least one in giving an order: client control instruction, navigation
Hint instructions and point of interest search instruction;
Using the corresponding speech recognition modeling of the current navigation state, voice knowledge is carried out to the first phonetic order of the user
Not, to obtain the first speech recognition result;
According to first speech recognition result, navigation operation is executed.
2. the method according to claim 1, wherein the first phonetic order of the acquisition user, comprising:
Second phonetic order of monitoring users;
Model is waken up using voice, speech recognition is carried out to second phonetic order, to obtain the second speech recognition result;
If second speech recognition result meets preset wake-up condition, the first phonetic order of user is acquired.
3. execution is led the method according to claim 1, wherein described according to first speech recognition result
Boat operation, comprising:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export and institute
State the search result that interest point name matches;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtains and export
The search result to match with the point of interest typonym.
4. execution is led the method according to claim 1, wherein described according to first speech recognition result
Boat operation, comprising:
If including client control instruction in first speech recognition result, according to the client control instruction, to described
Client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing projecting function;
If being instructed in first speech recognition result comprising navigation hint, is instructed according to the navigation hint, obtain and export
At least one in following navigation hint information: the Distance Remaining information that arrives at the destination, the required time information arrived at the destination,
Traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instruct, is obtained simultaneously according to the point of interest search
The search result that output matches with interest point information in point of interest search instruction.
5. a kind of voice guiding device, which is characterized in that described device includes:
Voice collecting unit, for acquiring the first phonetic order of user;
Model treatment unit, for determining the corresponding speech recognition modeling of the current navigation state according to current navigation state,
The current navigation state includes: before starting navigation and to navigate;
Voice recognition unit, for utilizing the corresponding speech recognition modeling of the current navigation state, to the first of the user
Phonetic order carries out speech recognition, to obtain the first speech recognition result;
Navigate execution unit, for executing navigation operation according to first speech recognition result;
Wherein, the model treatment unit, is specifically used for:
If the current navigation state is to determine that the speech recognition modeling is the first model, first mould before starting navigation
Type goes out the instruction of point of interest search included in first phonetic order for identification;
If the current navigation state is to navigate, determine that the speech recognition modeling is the second model, second mould
Type goes out included in first phonetic order for identification at least one in giving an order: client control instruction, navigation
Hint instructions and point of interest search instruction.
6. device according to claim 5, which is characterized in that
The voice collecting unit is also used to the second phonetic order of monitoring users;
Described device further include:
Voice wakeup unit, for waking up model to second phonetic order progress speech recognition, to obtain the using voice
Two speech recognition results;If second speech recognition result meets preset wake-up condition, the voice collecting unit is triggered
Acquire the first phonetic order of user.
7. device according to claim 5, which is characterized in that the navigation execution unit is specifically used for:
If the instruction of point of interest search included in first speech recognition result is interest point name, obtains and export and institute
State the search result that interest point name matches;Alternatively,
If the instruction of point of interest search included in first speech recognition result is point of interest typonym, obtains and export
The search result to match with the point of interest typonym.
8. device according to claim 5, which is characterized in that the navigation execution unit is specifically used for:
If including client control instruction in first speech recognition result, according to the client control instruction, to described
Client carries out at least one in following control: amplification/diminution map, increase/reduction volume and unlatching/closing projecting function;
If being instructed in first speech recognition result comprising navigation hint, is instructed according to the navigation hint, obtain and export
At least one in following navigation hint information: the Distance Remaining information that arrives at the destination, the required time information arrived at the destination,
Traffic information and place road information;
If being instructed in first speech recognition result comprising point of interest search, instruct, is obtained simultaneously according to the point of interest search
The search result that output matches with interest point information in point of interest search instruction.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410742287.5A CN104535071B (en) | 2014-12-05 | 2014-12-05 | A kind of phonetic navigation method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410742287.5A CN104535071B (en) | 2014-12-05 | 2014-12-05 | A kind of phonetic navigation method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104535071A CN104535071A (en) | 2015-04-22 |
CN104535071B true CN104535071B (en) | 2018-12-14 |
Family
ID=52850646
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410742287.5A Active CN104535071B (en) | 2014-12-05 | 2014-12-05 | A kind of phonetic navigation method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104535071B (en) |
Families Citing this family (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105139851A (en) * | 2015-09-17 | 2015-12-09 | 努比亚技术有限公司 | Desktop application icon organization mobile terminal and method |
CN105681579B (en) * | 2016-03-11 | 2020-01-10 | Oppo广东移动通信有限公司 | Terminal and screen control method and device thereof in navigation state |
CN105890615A (en) * | 2016-04-11 | 2016-08-24 | 深圳市轱辘软件开发有限公司 | Navigation method and device |
CN107305483A (en) * | 2016-04-25 | 2017-10-31 | 北京搜狗科技发展有限公司 | A kind of voice interactive method and device based on semantics recognition |
CN106289296B (en) * | 2016-09-05 | 2020-03-24 | 广州极飞科技有限公司 | Road navigation method and device |
CN107014390A (en) * | 2017-03-16 | 2017-08-04 | 北京云知声信息技术有限公司 | A kind of route reminding method and device |
CN107329730B (en) * | 2017-07-03 | 2021-03-16 | 科大讯飞股份有限公司 | Voice prompt message generation method and device |
CN107600075A (en) * | 2017-08-23 | 2018-01-19 | 深圳市沃特沃德股份有限公司 | The control method and device of onboard system |
CN107678316A (en) * | 2017-10-27 | 2018-02-09 | 姜俊 | A kind of environment inside car regulating system and method |
CN108168540A (en) * | 2017-12-22 | 2018-06-15 | 福建中金在线信息科技有限公司 | A kind of intelligent glasses air navigation aid, device and intelligent glasses |
CN108392269B (en) * | 2017-12-29 | 2021-08-03 | 广州布莱医疗科技有限公司 | Operation assisting method and operation assisting robot |
CN108307069B (en) * | 2018-01-29 | 2020-08-07 | Oppo广东移动通信有限公司 | Navigation operation method, navigation operation device and mobile terminal |
CN108827328A (en) * | 2018-04-24 | 2018-11-16 | 联想(北京)有限公司 | Voice interactive method, device, system and computer-readable medium |
CN108806682B (en) * | 2018-06-12 | 2020-12-01 | 奇瑞汽车股份有限公司 | Method and device for acquiring weather information |
CN109065045A (en) * | 2018-08-30 | 2018-12-21 | 出门问问信息科技有限公司 | Audio recognition method, device, electronic equipment and computer readable storage medium |
CN110770820A (en) * | 2018-08-30 | 2020-02-07 | 深圳市大疆创新科技有限公司 | Speech recognition method, apparatus, photographing system, and computer-readable storage medium |
CN108986811B (en) * | 2018-08-31 | 2021-05-28 | 北京新能源汽车股份有限公司 | Voice recognition detection method, device and equipment |
CN109000679B (en) * | 2018-09-21 | 2021-03-05 | 斑马网络技术有限公司 | Path prediction method, device, system and storage medium |
CN110136705B (en) * | 2019-04-10 | 2022-06-14 | 华为技术有限公司 | Man-machine interaction method and electronic equipment |
CN110057379B (en) * | 2019-05-29 | 2021-09-03 | 广州小鹏汽车科技有限公司 | Secondary navigation method and device for vehicle navigation and vehicle |
CN111949780B (en) * | 2020-07-31 | 2021-12-31 | 八维通科技有限公司 | Automatic interaction method along trip route |
CN114485718A (en) * | 2022-01-05 | 2022-05-13 | 腾讯科技(深圳)有限公司 | Voice navigation method, device, electronic equipment, storage medium and program product |
CN114964300B (en) * | 2022-06-22 | 2023-03-28 | 深圳市智远联科技有限公司 | Voice recognition method and navigation device |
CN114913855A (en) * | 2022-07-11 | 2022-08-16 | 广州小鹏汽车科技有限公司 | Voice interaction method, server and storage medium |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102063901A (en) * | 2010-12-02 | 2011-05-18 | 深圳市凯立德欣软件技术有限公司 | Voice identification method for position service equipment and position service equipment |
CN103674012A (en) * | 2012-09-21 | 2014-03-26 | 高德软件有限公司 | Voice customizing method and device and voice identification method and device |
CN103971681A (en) * | 2014-04-24 | 2014-08-06 | 百度在线网络技术(北京)有限公司 | Voice recognition method and system |
-
2014
- 2014-12-05 CN CN201410742287.5A patent/CN104535071B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102063901A (en) * | 2010-12-02 | 2011-05-18 | 深圳市凯立德欣软件技术有限公司 | Voice identification method for position service equipment and position service equipment |
CN103674012A (en) * | 2012-09-21 | 2014-03-26 | 高德软件有限公司 | Voice customizing method and device and voice identification method and device |
CN103971681A (en) * | 2014-04-24 | 2014-08-06 | 百度在线网络技术(北京)有限公司 | Voice recognition method and system |
Also Published As
Publication number | Publication date |
---|---|
CN104535071A (en) | 2015-04-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104535071B (en) | A kind of phonetic navigation method and device | |
CN107240398B (en) | Intelligent voice interaction method and device | |
CN108305633B (en) | Speech verification method, apparatus, computer equipment and computer readable storage medium | |
CN104794122B (en) | Position information recommendation method, device and system | |
CN107665710B (en) | Mobile terminal voice data processing method and device | |
EP3754651B1 (en) | Dynamic language model adaptation for generating voice-based search queries | |
JP5158174B2 (en) | Voice recognition device | |
US9443202B2 (en) | Adaptation of context models | |
US9188456B2 (en) | System and method of fixing mistakes by going back in an electronic device | |
CN108989541A (en) | Session initiation device, system, vehicle and method based on situation | |
US10388280B2 (en) | Method and apparatus for managing multiple voice operation trigger phrases | |
CN105004348A (en) | Voice navigation method and system | |
CN106959690B (en) | Method, device and equipment for searching unmanned vehicle and storage medium | |
WO2011092639A1 (en) | Systems, methods, and apparatuses for providing context-based navigation services | |
JP2011179917A (en) | Information recording device, information recording method, information recording program, and recording medium | |
CN110972112B (en) | Subway running direction determining method, device, terminal and storage medium | |
EP2660562A1 (en) | Route Guidance Apparatus and Method with Voice Recognition | |
JP2012088370A (en) | Voice recognition system, voice recognition terminal and center | |
US9791925B2 (en) | Information acquisition method, information acquisition system, and non-transitory recording medium for user of motor vehicle | |
US10593323B2 (en) | Keyword generation apparatus and keyword generation method | |
CN113706915A (en) | Parking prompting method, device, equipment and storage medium | |
CN109900284A (en) | Navigation route planning method and apparatus | |
JP2022103675A (en) | Information processing device, information processing method, and program | |
CN114118582A (en) | Destination prediction method, destination prediction device, electronic terminal and storage medium | |
CN113032681A (en) | Method, apparatus, electronic device, and medium for map search |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right | ||
TR01 | Transfer of patent right |
Effective date of registration: 20200409 Address after: 518000 301, floor 3, unit D, productivity building, No. 5, Gaoxin Middle Road, Science Park, Nanshan District, Shenzhen City, Guangdong Province Co-patentee after: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd. Patentee after: BAIDU INTERNATIONAL TECHNOLOGY (SHENZHEN) Co.,Ltd. Address before: 100085 Baidu building, No. 10, ten Street, Haidian District, Beijing Patentee before: BEIJING BAIDU NETCOM SCIENCE AND TECHNOLOGY Co.,Ltd. |