CN111768779A - Interaction method, information processing method, vehicle and server - Google Patents

Interaction method, information processing method, vehicle and server Download PDF

Info

Publication number
CN111768779A
CN111768779A CN202010599619.4A CN202010599619A CN111768779A CN 111768779 A CN111768779 A CN 111768779A CN 202010599619 A CN202010599619 A CN 202010599619A CN 111768779 A CN111768779 A CN 111768779A
Authority
CN
China
Prior art keywords
information
navigation route
vehicle
server
map application
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010599619.4A
Other languages
Chinese (zh)
Other versions
CN111768779B (en
Inventor
赵永亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Zhipeng Internet Of Vehicles Technology Co ltd
Original Assignee
Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd filed Critical Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd
Priority to CN202010599619.4A priority Critical patent/CN111768779B/en
Publication of CN111768779A publication Critical patent/CN111768779A/en
Application granted granted Critical
Publication of CN111768779B publication Critical patent/CN111768779B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • G10L15/30Distributed recognition, e.g. in client-server systems, for mobile phones or network applications
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1095Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Navigation (AREA)

Abstract

The application discloses an interaction method for a navigation route of a vehicle-mounted map application program. The interaction method comprises the following steps: acquiring voice interaction information of a user aiming at a navigation route; a server for sending voice interaction information; receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and the information template corresponding to the navigation route information; and executing the operation corresponding to the operation instruction. According to the interaction method, the navigation route information of the vehicle-mounted map application program graphical user interface is synchronized to the server, synchronization and consistency of local and cloud information are achieved, the server grasps more vehicle-mounted map application program graphical user interface information, the possibility of interaction through voice and the navigation route is provided, and voice interaction is enabled to be more intelligent. The application also discloses an information processing method, a vehicle, a server and a computer readable storage medium.

Description

Interaction method, information processing method, vehicle and server
Technical Field
The application relates to the technical field of voice recognition, in particular to an interaction method, an information processing method, a vehicle, a server and a computer-readable storage medium for a navigation route of a vehicle-mounted map application program.
Background
With the development of artificial intelligence technology, the voice intelligent platform or the voice assistant can recognize the voice input of the user and generate corresponding operation instructions under certain conditions, so that great convenience is provided for the user to operate the terminal device, the intelligence of the terminal device is improved, and the voice intelligent platform or the voice assistant is widely applied to human-computer interaction of automobiles. However, in the related art, voice interaction still stays at a relatively early stage, and only simple interaction can be realized, but for relatively complex functions, the intelligence is poor because the voice interaction cannot be realized. For example, in-vehicle navigation maps typically do not support voice interaction in a pre-navigation computation navigation route scenario, but can only operate through a graphical interactive interface.
Disclosure of Invention
In view of the above, embodiments of the present application provide an interaction method, an information processing method, a vehicle, a server, and a computer-readable storage medium.
The application provides an interaction method for a navigation route of a vehicle-mounted map application program, wherein the vehicle-mounted map application program comprises navigation route information, and the interaction method comprises the following steps:
acquiring voice interaction information of a user aiming at a navigation route;
sending the voice interaction information and the navigation route information to a server;
receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and an information template corresponding to the navigation route information;
and executing the operation corresponding to the operation instruction.
In some implementations, the navigation route information includes control information of a graphical user interface of the navigation route.
In some embodiments, the control information includes one or more of a list of navigation routes, a control indicating that navigation is initiated, a control indicating that navigation is exited, a control indicating that a route exploration mode is entered, and a control indicating that a route point of the navigation route is set.
In some embodiments, the matching, by the server, the voice interaction information and the navigation route information with the information template, and generating the operation instruction according to a result of the matching, and the receiving the operation instruction generated by the server according to the voice interaction information, the information, and the information template corresponding to the navigation route information includes:
receiving an execution instruction generated by the server according to successful matching;
the executing the operation corresponding to the operation instruction comprises:
and performing operation corresponding to the execution instruction on the navigation route.
In some embodiments, the receiving the operation instruction generated by the server according to the voice interaction information, the navigation route information and the information template corresponding to the navigation route information includes:
receiving a feedback instruction generated by the server according to the matching failure;
the executing the operation corresponding to the operation instruction comprises:
and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
In some embodiments, the performing the operation corresponding to the execution instruction on the navigation route includes:
judging whether the vehicle-mounted map application program intercepts the execution instruction;
and if the execution instruction is not intercepted by the vehicle-mounted map application program, performing operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program.
In some embodiments, the performing the operation corresponding to the execution instruction on the navigation route further includes:
if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
and performing operation corresponding to the execution instruction on the navigation route through the vehicle-mounted map application program.
The application provides an information processing method, which comprises the following steps:
receiving navigation route information uploaded by a vehicle-mounted map application program; and
and processing the navigation route information to obtain a corresponding information template.
In some embodiments, the processing the navigation routing information to obtain an information template includes:
and generalizing an expression mode interacted with the navigation route information to obtain the information template.
In some embodiments, the information processing method further includes:
receiving voice interaction information which is sent by the vehicle and aims at a navigation route;
matching the information template with the voice interaction information and the navigation route information;
and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
The application provides a vehicle, the operating system of vehicle installs on-vehicle map application, on-vehicle map application includes navigation route information, the vehicle includes:
the voice acquisition module is used for acquiring voice interaction information of a user aiming at the navigation route;
the communication module is used for sending the voice interaction information and the navigation route information to a server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and an information template corresponding to the navigation route information;
and the control module is used for executing the operation corresponding to the operation instruction.
The application provides a server, including:
the communication module is used for receiving navigation route information uploaded by a vehicle-mounted map application program; and
and the processing module is used for processing the navigation route information to obtain a corresponding information template.
A non-transitory computer-readable storage medium containing computer-executable instructions that, when executed by one or more processors, cause the processors to perform the method of interacting with or processing information for navigating routes by an in-vehicle map application is provided.
In the interaction method, the information processing method, the vehicle, the server and the computer-readable storage medium for the vehicle-mounted map application program navigation route, navigation route information of a vehicle-mounted map application program graphical user interface is synchronized to the server side, synchronization and consistency of local and cloud information are achieved, the server grasps more vehicle-mounted map application program graphical user interface information, the possibility of interaction with the navigation route through voice is provided, and voice interaction is enabled to be more intelligent.
Drawings
The foregoing and/or additional aspects and advantages of the present application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
FIG. 1 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
FIG. 2 is a block schematic diagram of a vehicle according to certain embodiments of the present application.
FIG. 3 is a schematic diagram of a scenario of an interaction method according to some embodiments of the present application.
FIG. 4 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
FIG. 5 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
Fig. 6 is a schematic flow chart of an information processing method according to some embodiments of the present application.
FIG. 7 is a block diagram of a server in accordance with certain embodiments of the present application.
FIG. 8 is a schematic illustration of a vehicle and server interaction in accordance with certain embodiments of the present application.
Fig. 9 is a schematic flow chart of an information processing method according to some embodiments of the present application.
Fig. 10 is a schematic flow chart diagram of an information processing method according to some embodiments of the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present application, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are exemplary and intended to be used for explaining the present application and should not be construed as limiting the present application.
Referring to fig. 1, the present application provides an interactive method for a navigation route of a vehicle map application. The method comprises the following steps:
s10: acquiring voice interaction information of a user aiming at a navigation route;
s20: sending voice interaction information and navigation route information to a server;
s30: receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and the information template corresponding to the navigation route information;
s40: and executing the operation corresponding to the operation instruction.
The embodiment of the application provides a vehicle. The vehicle includes a display area, an electro-acoustic element, a communication element, and a processor. The display area of the vehicle may include a dashboard screen, an on-board display area screen, and a heads-up display that may be implemented on a vehicle windshield, among others. An on-board system operating on a vehicle presents the presented content to a User using a Graphical User Interface (GUI). The display area includes a number of UI elements, and different display areas may present the same or different UI elements. The UI elements may include card objects, application icons or interfaces, folder icons, multimedia file icons, and controls for making interactive operations, among others. The electroacoustic element is used for acquiring voice interaction information of a user for the navigation route. The communication element is used for sending the voice interaction information and the navigation route information to the server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and the information template corresponding to the navigation route information. The processor is used for executing the operation corresponding to the operation instruction.
Referring to fig. 2, an embodiment of the present application further provides a vehicle 100, and the interaction method according to the embodiment of the present application may be implemented by the vehicle 100 according to the embodiment of the present application.
Specifically, the operating system of the vehicle 100 is installed with an in-vehicle map application, and the vehicle 100 includes a voice acquisition module 102, a communication module 104, and a control module 106. The S10 may be implemented by the voice acquisition module 102, the S20, S30 may be implemented by the communication module 104, and the S40 may be implemented by the control module 106. In other words, the voice obtaining module 102 is configured to obtain the voice interaction information of the user for the navigation route. The communication module 104 is configured to send the voice interaction information and the navigation route information to the server, and receive an operation instruction generated by the server according to the voice interaction information, the navigation route information, and an information template corresponding to the navigation route information. The control module 106 is configured to execute an operation corresponding to the operation instruction.
In the interaction method of the vehicle-mounted map application program and the vehicle 100, the navigation route information of the graphical user interface of the vehicle-mounted map application program is synchronized to the server, so that the synchronization and consistency of the local and cloud information are realized, the server grasps more information of the graphical user interface of the vehicle-mounted map application program, the possibility of interaction with the navigation route through voice is provided, and the voice interaction is more intelligent.
Specifically, the intelligent display area of the vehicle can provide a convenient entrance for a user to control the vehicle and interact with the vehicle, a voice assistant function is added in the vehicle-mounted operating system, voice information input by the user can be analyzed through voice recognition and semantic recognition under a certain condition, a corresponding control instruction is generated conveniently, and convenience is further provided for interaction between the user and the vehicle. However, for navigating the in-vehicle map application, voice interaction still stays at a relatively early stage, and only simple interaction can be realized, for example, zooming in and zooming out on the display scale of the graphical user interface of the in-vehicle map application is realized through voice. For a complex function, for example, for a scene where a certain information point is selected, route calculation is initiated and a plurality of navigation routes are obtained, a user can interact with the calculated navigation route only through input in a graphical user interface, such as clicking and the like, for example, operations such as switching different routes for viewing, selecting a certain route for navigation and the like are performed, and interaction cannot be achieved through voice. For the situation that the vehicle is in the driving mode at present, a user interacts through a graphical user interface of a vehicle-mounted map application program while driving, and certain safety risks exist.
In the embodiment, after waking up the voice assistant, the user inputs voice information, and obtains the information of the graphical user interface of the navigation route to the destination, which is obtained by calculation and displayed on the graphical user interface of the current vehicle-mounted map application program of the vehicle, while obtaining the voice information. The navigation route information comprises information of two aspects of display form and display structure. The display form is also a presentation form of the navigation route, for example, the navigation route may be presented in a form of a single window list, multiple windows, and the like, and the display structure is also a concrete structure of the display form of the windows and the like, for example, the number of rows and columns included in the list, the positions of the included controls, the distribution positions of the controls in the windows, the display hierarchy, and the like.
And after the user wakes up the voice assistant locally, inputting voice interaction information of the navigation route. The vehicle sends the voice interaction information and the navigation route information to a server of a cloud service provider, the server analyzes the voice interaction information by using the navigation route information as auxiliary information, so that an operation instruction is generated and is transmitted back to the local vehicle, and the vehicle executes corresponding operation according to the operation instruction.
The navigation route information is synchronized to the server through a voice software development kit, and the voice interaction software development kit is a hub for voice interaction between the vehicle-mounted map application program and the server. In one aspect, a software development kit defines a specification for generating voice interaction information. On the other hand, the voice software development kit can realize the synchronization of the navigation route information in the vehicle-mounted map application program to the server and the transmission of the operation instruction generated by the server for the voice interaction information to the vehicle-mounted map application program.
In one example, the in-vehicle map application may invoke an information synchronization method provided by the software development kit to synchronize the navigation routing information to the software development kit.
And the software development toolkit performs information fault tolerance and normalization check on the received navigation route information. Specifically, error information possibly existing in the navigation route information is corrected according to the voice interaction generation specification, so that the navigation route information data are guaranteed to meet the generation specification and can be identified and analyzed by the server. In addition, the software development kit checks the data of the navigation route information in the in-vehicle map application according to the generation specification of the voice interaction. For example, it is checked whether the attributes of the data are correct, whether the encoding of elements in the data is unique, and the like. And if the attribute configuration is correct, namely the attribute configuration meets the generation specification, releasing the navigation route information. Otherwise, feedback is given to the vehicle-mounted map application program, for example, an error log is fed back, or a prompt is given on a vehicle-mounted map application program interface.
The parsing of the voice interaction information generally includes two parts of voice recognition and semantic parsing, and the voice recognition may be performed locally, for example, the voice interaction information may be recognized by a text-to-speech module of the vehicle to convert the voice into text. Of course, the voice recognition can also be carried out in the server, thereby reducing the processing burden of the operating system of the central control display screen of the vehicle end. Semantic parsing can be completed in a server, and generally, understanding of voice interaction information is achieved through steps of word segmentation, analysis and the like of a text.
The navigation route information can make the server more clearly determine the current interactive scene of the vehicle when performing semantic analysis, and effectively limit the scope of the semantic analysis. For example, in a state where the navigation route list is displayed, the user wishes to select the first route for interactive navigation, and the user issues a voice command of "driving". If the server does not synchronously acquire the navigation route information, the driving route cannot be used during semantic analysis, and only a driving instruction is generated. And under the condition of synchronously acquiring the navigation route information, the server can judge that the user hopes to control the vehicle to navigate to the destination by taking the calculated first route as the navigation route, so that the vehicle is controlled to navigate to the destination by adopting the first route.
For another example, in the state where the user is shown in the navigation route list, the user wants to view one of the driving routes and issues a voice command "view nth route". If the server does not synchronously acquire the navigation route information, the actual semantics of the user cannot be clarified during the semantic analysis, and only an unrecognizable error prompt is generated. And under the condition of synchronously acquiring the navigation route information, the server can judge that the user wants to view the Nth route in the navigation route list, so that the corresponding route is displayed on the graphical user interface of the vehicle-mounted map application program.
Therefore, the intelligence of voice control and the success rate of hitting the real intention can be improved, and the user experience is better.
The information template of the navigation route information is formed after processing according to functions and contents in a navigation route graphic user interface uploaded by a vehicle. The information template is stored in the server, so that after the navigation route information uploaded by the user is received, the server can confirm the information template corresponding to the current navigation route information through matching with the information template, and accordingly the current interaction scene of the user is obtained, the intention of the user can be judged according to the voice interaction information, and the real intention of the user can be analyzed by the voice interaction information in an auxiliary mode according to the navigation route interface scene of the user interaction.
In addition, in the application, a driver can perform voice interaction with the vehicle-mounted map application program at any time in the driving process, such as a driving or parking state, so that the adjustment of the vehicle-mounted map scale is realized. Particularly, in the driving state, the voice input is adopted to replace the manual input of a user to interact with the vehicle-mounted map application program, and the driving safety can be considered.
In this embodiment, the navigation route information includes control information of a graphical user interface of the navigation route.
Specifically, in the process of actually using the vehicle-mounted map application program, after a user selects a certain information point in the map, for example, a destination of the current driving, a request for obtaining a route to the information point is initiated, and after calculation, the vehicle-mounted map application program returns to the navigation route result list. The user may perform further operations on the navigation route in the navigation route list, such as navigating to, switching routes, and the like.
The contents are distributed and displayed by corresponding controls, the navigation route information is the control information of the graphical user interface of the navigation route obtained by current calculation, and the vehicle-mounted map application program distributes the navigation route information through the voice interaction control library control, so that a layout data structure capable of being controlled by voice is constructed. In the data structure design process, a control supporting graphic interaction operation needs to be replaced by a control supporting voice interaction, namely a control in a voice interaction control library. For example, the linear layout control LinearLayout in the original structure is replaced by a linear layout control XLinearLayout supporting voice interaction operation packaged by a voice interaction control library. For another example, the text control TextView in the original structure is replaced by the text control XTextView supporting voice interaction operation packaged by the voice interaction control library.
Controls generally include, but are not limited to, the following information: an element identification, an element type, an action type of the element, a phonetic utterance of the element, and the like. Wherein the element identification is unique for each element by which the element can be found. The element types may include groups, text, images, and the like. The action type of an element may include clicking, sliding, and the like. The phonetic interpretation of an element includes waking up a certain operation keyword, etc.
Referring to fig. 3, the control information includes one or more of a navigation route list, a control indicating that navigation is started, a control indicating that navigation is exited, a control indicating that a route search mode is entered, and a control indicating that a route point of the navigation route is set.
Specifically, the in-vehicle map application may perform layout for controls corresponding to following navigation route interactions, such as start of navigation, route exploration mode, exit, route information, route point settings, and the like.
The navigation route list is used for displaying navigation routes calculated by adopting different standards, and the navigation route list can support voice interaction switching or select a certain route in the route information list and initiate navigation according to the route. During the interaction, a voice feedback such as "switched to a route" may be provided. And in the graphical user interface, the selected route is highlighted in both the route information list and the map.
The control representing initiation of navigation may be a "start navigation" control that may support initiation of navigation with the currently selected route through voice interaction. And in the navigation process, the voice assistant can provide voice feedback for the user, for example, a voice prompt for prompting the user when the network is not smooth is given, such as 'the network is not smooth, and off-line navigation is started for you'.
The control for indicating entering the route exploring mode can be a control for 'route exploring mode', and the control for 'route exploring mode' can support voice interaction to enter the route exploring mode from a common navigation mode.
The control representing exit navigation may be an "exit" control that may support a voice interaction exit route calculation scenario.
The control for setting the route points of the navigation route can be a 'route point setting' control, and the 'route point setting' control can support voice interaction to add the route points, search the route points and calculate the route again according to the added route points.
Referring to fig. 4, in some embodiments, the server matches the voice interaction information and the navigation route information with the information template, and generates an operation instruction according to a matching result. S30 includes:
s31: receiving an execution instruction generated by the server according to successful matching;
s40 includes:
s41: and performing operation corresponding to the execution instruction on the navigation route.
In some embodiments, S31 may be implemented by the communication module 104 and S41 may be implemented by the control module 106. That is, the communication module 104 is configured to receive the execution instruction generated by the server according to the matching success. The control module 106 is configured to perform an operation corresponding to the execution instruction on the navigation route.
In some embodiments, the communication element is configured to receive an execution instruction generated by the server upon a successful match. The processor is used for carrying out operation corresponding to the execution instruction on the navigation route.
Specifically, after the voice assistant is awakened each time, different vehicles upload the voice interaction information and the navigation route information to the server together. The server can obtain a large amount of historical data of the navigation route information along with the use of the user, and the collected large amount of navigation route information is supplemented, expanded and sorted through machine learning or manual labeling and other modes, so that the understanding of the server on the navigation route information is enriched, and the sorted content can form a corresponding information template and is stored in the server, so that the accuracy and the recognition efficiency of semantic recognition are improved in the subsequent use process of the user.
In an actual process, if the user uses the voice assistant for the first time, there may not be a pre-stored information template at the server side, in this case, the server directly performs semantic recognition according to the navigation route information to assist the voice interaction information, in this case, the requirement for the voice interaction information may be severe, but the semantic meaning of the user can also be understood to a certain extent due to the navigation route information as an assist. If the voice assistant is not used for the first time, after the server receives the navigation route information, the current graphical user interface can be identified according to the control information of the navigation route information, and then an information template corresponding to the control information is called, so that the voice interaction information and the navigation route information can be matched with the information template to analyze the real intention of the user.
It can be understood that the same user may express the same voice interaction instruction differently in the previous and subsequent implementation processes, and different users may also express the same instruction differently. And the set information template is generalized aiming at each possible expression mode of voice interaction. The richer the content of the information template, the higher the probability and success rate of recognizing the voice interaction instruction.
For example, for "route information" in the navigation route information, the expression that the user views a certain route may be expanded to view the nth item, view the nth route, help me switch to the nth route, switch to the nth item, navigate to the nth route, and the like. These representations are stored in the information template.
The speech-to-text conversion module of the vehicle performs speech recognition on the speech interaction information, and of course, the speech recognition may also be performed by the speech-to-text conversion module of the server. And comparing the uploaded information with the information template to realize the analysis of the voice interaction information semantics. And under the condition that the matching is successful, generating an execution instruction corresponding to the interactive information, returning to the vehicle, and executing the execution instruction on the navigation route list by the vehicle.
For example, when a user wants to switch a route, voice interaction information such as 'view the nth route' is sent out, the voice interaction information and navigation route information are sent to a server together, the server can obtain the display form, the structural frame layout and a control which can carry out interaction of the current navigation route according to the navigation route information, the voice interaction information and the navigation route information are matched with an information template, after the matching, the semantic meaning of the interaction information is confirmed to be that the route displayed in a map is switched to the nth route in a route list, an execution instruction for switching the route to the nth route is generated, after the vehicle-mounted map application receives the execution instruction, the vehicle-mounted map application highlights the nth route in the display list, and the nth route is highlighted in the route list.
Referring again to fig. 4, in some embodiments, S30 includes:
s32: receiving a feedback instruction generated by the server according to the matching failure;
s40 includes:
s42: and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
In some embodiments, S32 may be implemented by the communication module 104 and S42 may be implemented by the control module 106. That is, the communication module 104 is configured to receive a feedback instruction generated by the server according to the matching failure. The control module 106 is configured to broadcast the information of the matching failure according to the feedback instruction to prompt the user.
In some embodiments, the communication element is to receive a feedback instruction generated by the server based on the failure to match. And the processor is used for broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
Specifically, for interaction which is not supported on the navigation route or voice interaction information which cannot be subjected to semantic analysis, the server also gives feedback which cannot be identified, and the application program can broadcast the feedback information in modes of voice, text popup display and the like, so that the user is prompted that input information is invalid.
For the voice interaction information which cannot be identified, the vehicle-mounted map application program can monitor the interaction operation of the user through the graphical interaction interface within the preset time period of the broadcast feedback prompt, and reports the interaction operation to the server, relevant personnel manually detect the interaction operation of the voice interaction information and the graphical user interface, judge whether the voice interaction information and the graphical user interface are related, and if the voice interaction information and the graphical user interface are related, expand the expression of the voice interaction information into an information template corresponding to the execution instruction. And if no association exists, ignoring the reported information.
For example, the user wishes to navigate to a destination with a default route. And sending voice interaction information of 'driving', matching the voice interaction information and the navigation route information with the information template, confirming that the voice interaction information and the navigation route information cannot be matched with the current information template after matching, generating a feedback instruction, and broadcasting the information which cannot be identified after the vehicle-mounted map application program receives the feedback instruction. The user then manually clicks on the route and switches to the drive tab item. The vehicle-mounted map application program reports the operation of the user to the server, and related workers judge that the expression of the driving is related to the operation of viewing the driving route, so that the driving can be added to the control related to the route and the corresponding voice interaction instruction information template.
Referring to fig. 5, in some embodiments, S41 includes:
s411: judging whether the vehicle-mounted map application program intercepts an execution instruction;
s412: and if the vehicle-mounted map application program does not intercept the execution instruction, performing operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program.
In some embodiments, S411, S412 may be implemented by the control module 106. That is, the control module 106 is configured to determine whether the execution instruction is intercepted by the vehicle-mounted map application program, and perform an operation corresponding to the execution instruction on the navigation route through the software development kit of the vehicle-mounted map application program when the execution instruction is not intercepted by the vehicle-mounted map application program.
In some embodiments, the processor determines whether the execution instruction is intercepted by the vehicle-mounted map application program, and is used for performing an operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program under the condition that the execution instruction is not intercepted by the vehicle-mounted map application program.
Specifically, an execution instruction is generated after the server is successfully matched, and the execution instruction is returned. According to the business requirement, different objects are usually selected to process the execution instruction. For example, if a relatively simple, single operation is performed, the execution instructions may be processed directly by the software development kit. And if more personalized subsequent operations are needed on the basis of the basic operations, the execution instructions are processed by the vehicle-mounted map application program.
In the specific implementation process, the processing mechanism is preset, and after the vehicle-mounted map application program receives the execution instruction, the vehicle-mounted map application program selects whether to intercept the execution instruction according to different execution instruction processing mechanisms. And if the vehicle-mounted map application program is not intercepted, the execution instruction is processed and executed by a software development kit.
Referring again to fig. 5, in some embodiments, S41 further includes:
s413: if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
s414: and performing operation corresponding to the execution instruction on the navigation route through the vehicle-mounted map application program.
In some embodiments, S413, S414 may be implemented by the control module 106. That is, the control module 106 is configured to pass through the execution instruction to the vehicle-mounted map application program through the software development kit when the vehicle-mounted map application program intercepts the execution instruction, and is configured to perform an operation corresponding to the execution instruction on the navigation route through the vehicle-mounted map application program.
In some embodiments, the processor is used for transmitting the execution instruction to the vehicle-mounted map application program through the software development kit in the case that the execution instruction is intercepted by the vehicle-mounted map application program, and is used for carrying out an operation corresponding to the execution instruction on the navigation route through the vehicle-mounted map application program.
In the specific implementation process, the processing mechanism is preset, and after the execution instruction is received, the vehicle-mounted map application program selects whether to intercept the execution instruction according to different execution instruction processing mechanisms. If the in-vehicle map application intercepts the first execution instruction, the software development kit will not process the execution instruction, but instead pass the execution instruction through to the in-vehicle map application, which processes the execution instruction.
In one example, for an "exit" interaction, since the operation is relatively simple and there are generally no subsequent operations, then the setup may be performed by a software development kit. The vehicle-mounted map application program does not intercept an execution instruction related to exit interaction, the software development kit processes the execution instruction, and triggers click processing on an exit tag, so that an exit route calculation scene is realized.
For the "via point setting" interaction, since the user is usually further recalculating the route including the via point when adding the via point, the setting may be performed by the in-vehicle map application. The vehicle-mounted map application program intercepts an execution instruction related to 'passing point setting', the software development kit does not process, and the vehicle-mounted map application program triggers and adds the setting of the passing point and automatically triggers and calculates the setting including the destination and the passing point.
In another example, taking "route information" interaction, i.e. interaction of switching navigation routes as an example, if the application program does not intercept, the software development kit performs processing to trigger the click processing of the route and switch to other routes, but does not perform an operation of navigating with the switched route, i.e. if the user wishes to further initiate navigation, a manual operation is also required.
And if the application program is intercepted and the software development kit does not process, the vehicle-mounted map application program triggers the processing of switching the route and automatically triggers the operation of initiating navigation by the switched route. With better intelligence and operational efficiency.
Referring to fig. 6, the present application further provides an information processing method for processing the voice interaction information sent from the vehicle 100 to the information server 200 in the above embodiment. The information processing method comprises the following steps:
s50: receiving navigation route information uploaded by a vehicle-mounted map application program; and
s60: the navigation route information is processed to obtain a corresponding information template.
The embodiment of the application provides a server. The server includes a communication element and a processor. The communication element is used for receiving navigation route information synchronized by the vehicle-mounted map application program through the software development tool kit. The processor is used for processing the navigation route information to obtain an information template.
Referring to fig. 7, an embodiment of the present application further provides a server 200, and an information processing method according to the embodiment of the present application may be implemented by the server 200 according to the embodiment of the present application.
Specifically, the server 200 includes a communication module 202 and a processing module 204. S50 may be implemented by the communication module 202, and S60 may be implemented by the processing module 204. In other words, the communication module 202 is configured to receive navigation route information uploaded by the in-vehicle map application. The processing module 204 is configured to process the navigation route information to obtain a corresponding information template.
Referring to fig. 8, the server 200 of the present embodiment communicates with the vehicle 100, and in the process of implementing voice control on the vehicle 100, the navigation route information on the vehicle-mounted map application is synchronized to the server, so that synchronization and consistency between local information and cloud information are achieved, the server grasps more information of the vehicle-mounted map application interface, and provides a possibility of interaction with the navigation route through voice, and voice interaction is more intelligent.
The server receives navigation route information sent by different vehicles, and an information template corresponding to the navigation route is constructed according to control information contained in the navigation route information.
The information template may include the same element and different elements, or common elements and personalized elements, for the same navigation roadmap user interface. Based on the same or common elements in the graphical user interface, the server 200 may build a basic framework of the current navigation route as a basis for the information template. According to different elements in the graphical user interface, the server 200 can know the specific information of the current navigation route, so that the content of the information template is enriched. The information template has the significance of mastering more user interaction information and providing more accurate assistance for voice recognition.
Referring to fig. 9, in some embodiments, S60 includes:
s61: and generalizing an expression mode of interaction with the navigation route information to obtain an information template.
In some embodiments, S61 may be implemented by the processing module 204, that is, the processing module 204 is configured to generalize the expression for interacting with the navigation route information to obtain the information template.
In some embodiments, the processor is configured to generalize the expression of interaction with the navigation routing information to obtain the information template.
In particular, voice interaction refers to generally comprising two parts, an instruction object and a manner of operation. Correspondingly, the instruction object, namely the control in the graphical user interface included in the navigation route information, corresponds to the information template, and the expression mode of the control is generalized. That is, the same instruction object is generalized, so that different expression modes correspond to the instruction object.
For example, for the "route information" control, the generalized processing may include expressions of a route, a line, an nth route, and the like.
The operation mode is interaction with the control and generalization processing is carried out on the expression mode of the interaction with the control, namely generalization processing is carried out on the same operation mode, so that different expression modes correspond to the interaction operation.
For example, for the "route information", the generalization processing may include expressions of looking up the nth item, looking up the nth route, helping me to switch to the nth route, switching to the nth item, navigating to the nth route, going to the route with the shortest time, going to the route with the least traffic lights, going to the route with the shortest route, and the like.
For "start navigation", the generalization process may include expressions of navigation, driving, walking, driving, start navigation, and the like.
After a certain amount of voice interaction information is collected, the information template can be expanded manually, the information template has richer contents, and the same instruction has more expression modes, so that the analysis of the voice interaction information can be better assisted.
Referring to fig. 10, in some embodiments, the information processing method further includes:
s70: receiving voice interaction information aiming at a navigation route sent by a vehicle;
s80: matching the information template with the voice interaction information and the navigation route information;
s90: and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
In some embodiments, S70 may be implemented by communication module 202. S80 may be implemented by the processing module 204, and S90 may be implemented by the communication module 202 and the processing module 204. In other words, the communication module 202 is used to receive the voice interaction information for the navigation route sent by the vehicle 100. The processing module 204 matches the information template according to the voice interaction information and the navigation route information, and is configured to generate an execution instruction or a feedback instruction according to a matching result. The communication module 202 is also used to send execution instructions or feedback instructions to the vehicle 100.
In some embodiments, the communication element is configured to receive voice interaction information for a navigation route sent by a vehicle. The processor is used for matching the information template with the voice interaction information and the navigation route information and generating an execution instruction or a feedback instruction according to a matching result. The communication element is also used for sending the execution instruction or the feedback instruction to the vehicle.
Specifically, the vehicle sends the voice interaction information to a server at the cloud end, the server matches the voice interaction information and the navigation route information with the information template, a feedback instruction is generated after the matching is successful and is transmitted back to the vehicle, and then the vehicle executes corresponding operation on the navigation route according to the execution instruction.
For example, when the user wants to switch the route, voice interactive information such as "view nth route" is issued, and the vehicle uploads the voice interactive information together with the navigation route information to the server 200. After receiving the interactive information, the server 200 matches the interactive information and the navigation route information with the information template, confirms that the semantics of the interactive information is to switch the navigation route in the vehicle-mounted map application program to the Nth route in the route list after matching, generates an execution instruction for switching the route to the Nth route, and sends the execution instruction back to the vehicle.
The embodiment of the application also provides a computer readable storage medium. One or more non-transitory computer-readable storage media containing computer-executable instructions that, when executed by one or more processors, cause the processors to perform the method of interacting or processing information for navigating routes of an in-vehicle map application of any of the embodiments described above.
It will be understood by those skilled in the art that all or part of the processes of the methods of the above embodiments may be implemented by hardware instructions of a computer program, which may be stored in a non-volatile computer-readable storage medium, and when executed, may include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), or the like.
The above examples only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the present application. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (13)

1. An interaction method for a navigation route of a vehicle-mounted map application program, wherein the vehicle-mounted map application program comprises navigation route information, and the interaction method comprises the following steps:
acquiring voice interaction information of a user aiming at a navigation route;
sending the voice interaction information and the navigation route information to a server;
receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and an information template corresponding to the navigation route information;
and executing the operation corresponding to the operation instruction.
2. The interaction method according to claim 1, wherein the navigation route information includes control information of a graphical user interface of the navigation route.
3. The interaction method according to claim 2, wherein the control information includes one or more of a navigation route list, a control for starting navigation, a control for exiting navigation, a control for entering a route search mode, and a control for setting a route point of the navigation route.
4. The interaction method according to claim 3, wherein the server matches the voice interaction information and the navigation route information with the information template and generates the operation instruction according to a result of the matching, and the receiving the operation instruction generated by the server according to the voice interaction information, the information and the information template corresponding to the navigation route information comprises:
receiving an execution instruction generated by the server according to successful matching;
the executing the operation corresponding to the operation instruction comprises:
and performing operation corresponding to the execution instruction on the navigation route.
5. The interaction method according to claim 4, wherein the receiving the operation instruction generated by the server according to the voice interaction information, the navigation route information and the information template corresponding to the navigation route information comprises:
receiving a feedback instruction generated by the server according to the matching failure;
the executing the operation corresponding to the operation instruction comprises:
and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
6. The interaction method according to claim 4, wherein the performing the operation corresponding to the execution instruction on the navigation route comprises:
judging whether the vehicle-mounted map application program intercepts the execution instruction;
and if the execution instruction is not intercepted by the vehicle-mounted map application program, performing operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program.
7. The interaction method according to claim 6, wherein the performing the operation corresponding to the execution instruction on the navigation route further comprises:
if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
and performing operation corresponding to the execution instruction on the navigation route through the vehicle-mounted map application program.
8. An information processing method characterized by comprising:
receiving navigation route information uploaded by a vehicle-mounted map application program; and
and processing the navigation route information to obtain a corresponding information template.
9. The information processing method according to claim 8, wherein the processing the navigation route information to obtain an information template includes:
and generalizing an expression mode interacted with the navigation route information to obtain the information template.
10. The information processing method according to claim 8, characterized by further comprising:
receiving voice interaction information which is sent by the vehicle and aims at a navigation route;
matching the information template with the voice interaction information and the navigation route information;
and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
11. A vehicle, characterized in that an operating system of the vehicle is installed with an in-vehicle map application including navigation route information, the vehicle comprising:
the voice acquisition module is used for acquiring voice interaction information of a user aiming at the navigation route;
the communication module is used for sending the voice interaction information and the navigation route information to a server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation route information and an information template corresponding to the navigation route information;
and the control module is used for executing the operation corresponding to the operation instruction.
12. A server, comprising:
the communication module is used for receiving navigation route information uploaded by a vehicle-mounted map application program; and
and the processing module is used for processing the navigation route information to obtain a corresponding information template.
13. A non-transitory computer-readable storage medium of computer-executable instructions, that when executed by one or more processors, cause the processors to perform the method of interacting with in-vehicle map application navigation routes of any of claims 1-7 or the method of information processing of claims 8-10.
CN202010599619.4A 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server Active CN111768779B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010599619.4A CN111768779B (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010599619.4A CN111768779B (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Publications (2)

Publication Number Publication Date
CN111768779A true CN111768779A (en) 2020-10-13
CN111768779B CN111768779B (en) 2024-02-06

Family

ID=72722399

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010599619.4A Active CN111768779B (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Country Status (1)

Country Link
CN (1) CN111768779B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112525214A (en) * 2020-11-24 2021-03-19 广州橙行智动汽车科技有限公司 Interaction method and device for map card, vehicle and readable medium
CN113113015A (en) * 2020-11-17 2021-07-13 广州小鹏汽车科技有限公司 Interaction method, information processing method, vehicle and server
CN113370897A (en) * 2021-06-28 2021-09-10 北京小鹏汽车有限公司 Vehicle map display control method and device
CN113470412A (en) * 2021-06-28 2021-10-01 北京小鹏汽车有限公司 Vehicle map real-time road condition presentation control method and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150088412A1 (en) * 2013-09-26 2015-03-26 Google Inc. Systems and Methods for Providing Navigation Data to a Vehicle
CN109579866A (en) * 2019-01-18 2019-04-05 深圳壹账通智能科技有限公司 Intelligent navigation method, device, computer equipment and storage medium
CN111179930A (en) * 2019-12-31 2020-05-19 苏州思必驰信息科技有限公司 Method and system for realizing intelligent voice interaction in driving process

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150088412A1 (en) * 2013-09-26 2015-03-26 Google Inc. Systems and Methods for Providing Navigation Data to a Vehicle
CN109579866A (en) * 2019-01-18 2019-04-05 深圳壹账通智能科技有限公司 Intelligent navigation method, device, computer equipment and storage medium
CN111179930A (en) * 2019-12-31 2020-05-19 苏州思必驰信息科技有限公司 Method and system for realizing intelligent voice interaction in driving process

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113113015A (en) * 2020-11-17 2021-07-13 广州小鹏汽车科技有限公司 Interaction method, information processing method, vehicle and server
CN112525214A (en) * 2020-11-24 2021-03-19 广州橙行智动汽车科技有限公司 Interaction method and device for map card, vehicle and readable medium
CN112525214B (en) * 2020-11-24 2024-05-28 广州橙行智动汽车科技有限公司 Interaction method and device of map card, vehicle and readable medium
CN113370897A (en) * 2021-06-28 2021-09-10 北京小鹏汽车有限公司 Vehicle map display control method and device
CN113470412A (en) * 2021-06-28 2021-10-01 北京小鹏汽车有限公司 Vehicle map real-time road condition presentation control method and device
CN113470412B (en) * 2021-06-28 2022-10-14 北京小鹏汽车有限公司 Vehicle map real-time road condition presentation control method and device

Also Published As

Publication number Publication date
CN111768779B (en) 2024-02-06

Similar Documents

Publication Publication Date Title
CN111722905A (en) Interaction method, information processing method, vehicle and server
CN111768779B (en) Interaction method, information processing method, vehicle and server
CN105989841B (en) Vehicle-mounted voice control method and device
CN111722825A (en) Interaction method, information processing method, vehicle and server
US8700408B2 (en) In-vehicle apparatus and information display system
US9922651B1 (en) Avionics text entry, cursor control, and display format selection via voice recognition
JP7042240B2 (en) Navigation methods, navigation devices, equipment and media
US9618352B2 (en) Apparatus and method for controlling navigator via speech dialogue
US20190120649A1 (en) Dialogue system, vehicle including the dialogue system, and accident information processing method
US10008204B2 (en) Information processing system, and vehicle-mounted device
CN109903770A (en) The devices and methods therefor of language mistake for correcting user
EP3044781B1 (en) Vehicle interface system
CN111768777A (en) Voice control method, information processing method, vehicle and server
CN111753039A (en) Adjustment method, information processing method, vehicle and server
CN110767219B (en) Semantic updating method, device, server and storage medium
CN117809641A (en) Terminal equipment and voice interaction method based on query text rewriting
CN113113015A (en) Interaction method, information processing method, vehicle and server
CN113436627A (en) Voice interaction method, device, system, vehicle and medium
CN107885583B (en) Operation triggering method and device
EP4086580B1 (en) Voice interaction method, apparatus and system, vehicle, and storage medium
KR102371513B1 (en) Dialogue processing apparatus and dialogue processing method
US11620994B2 (en) Method for operating and/or controlling a dialog system
US9858918B2 (en) Root cause analysis and recovery systems and methods
CN118173086A (en) Voice interaction method, vehicle, device and storage medium
US20240210197A1 (en) requesting and receiving reminder instructions in a navigation session

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information

Address after: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant after: Guangzhou Xiaopeng Automatic Driving Technology Co.,Ltd.

Address before: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant before: Guangzhou Xiaopeng Internet of vehicles Technology Co.,Ltd.

CB02 Change of applicant information
TA01 Transfer of patent application right

Effective date of registration: 20201209

Address after: Room 209, No.8, Songgang street, Cencun, Changxing street, Tianhe District, Guangzhou City, Guangdong Province, 510000

Applicant after: Guangzhou Zhipeng Internet of vehicles Technology Co.,Ltd.

Address before: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant before: Guangzhou Xiaopeng Automatic Driving Technology Co.,Ltd.

TA01 Transfer of patent application right
GR01 Patent grant
GR01 Patent grant