CN111722905A - Interaction method, information processing method, vehicle and server - Google Patents

Interaction method, information processing method, vehicle and server Download PDF

Info

Publication number
CN111722905A
CN111722905A CN202010599635.3A CN202010599635A CN111722905A CN 111722905 A CN111722905 A CN 111722905A CN 202010599635 A CN202010599635 A CN 202010599635A CN 111722905 A CN111722905 A CN 111722905A
Authority
CN
China
Prior art keywords
information
navigation
vehicle
navigation scene
server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010599635.3A
Other languages
Chinese (zh)
Inventor
赵永亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangzhou Zhipeng Internet Of Vehicles Technology Co ltd
Original Assignee
Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd filed Critical Guangzhou Xiaopeng Internet of Vehicle Technology Co Ltd
Priority to CN202010599635.3A priority Critical patent/CN111722905A/en
Publication of CN111722905A publication Critical patent/CN111722905A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/44Arrangements for executing specific programs
    • G06F9/451Execution arrangements for user interfaces
    • G06F9/452Remote windowing, e.g. X-Window System, desktop virtualisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Navigation (AREA)

Abstract

The application discloses an interaction method for a navigation scene of a vehicle-mounted map application program. The interaction method comprises the following steps: acquiring voice interaction information of a user aiming at a navigation scene; sending voice interaction information and navigation scene information to a server; receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and the information template corresponding to the navigation scene information; and executing the operation corresponding to the operation instruction. According to the interaction method, the graphical user interface information of the navigation scene of the vehicle-mounted map application program is synchronized to the server, so that synchronization and consistency of local and cloud information are achieved, the server grasps more information of the graphical user interface of the vehicle-mounted map application program, the possibility of interaction in the navigation scene through voice is provided, and voice interaction is more intelligent. The application also discloses an information processing method, a vehicle, a server and a computer readable storage medium.

Description

Interaction method, information processing method, vehicle and server
Technical Field
The present application relates to the field of speech recognition technologies, and in particular, to an interaction method for a navigation scene, an information processing method, a vehicle, a server, and a computer-readable storage medium.
Background
With the development of artificial intelligence technology, the voice intelligent platform or the voice assistant can recognize the voice input of the user and generate corresponding operation instructions under certain conditions, so that great convenience is provided for the user to operate the terminal device, the intelligence of the terminal device is improved, and the voice intelligent platform or the voice assistant is widely applied to human-computer interaction of automobiles. However, in the related art, voice interaction still stays at a relatively early stage, and only simple interaction can be realized, but for relatively complex functions, the intelligence is poor because the voice interaction cannot be realized. For example, in-vehicle navigation maps typically do not support voice interaction in the navigation scene, but can only be operated through a graphical interactive interface.
Disclosure of Invention
In view of the above, embodiments of the present application provide an information processing method, an interaction method, a server, a terminal, and a computer-readable storage medium.
The application provides an interaction method for a navigation scene of a vehicle-mounted map application program, wherein the vehicle-mounted map application program comprises navigation scene information, and the interaction method comprises the following steps:
acquiring voice interaction information of a user aiming at a navigation scene;
sending the voice interaction information and the navigation scene information to a server;
receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and an information template corresponding to the navigation scene information;
and executing the operation corresponding to the operation instruction.
In some implementations, the navigation scenario information includes control information of a graphical user interface of the navigation scenario.
In some embodiments, the control information includes one or more of a traffic information window, a control indicating that navigation is started, a control indicating that a navigation scene exits, a control indicating a special location graph, a control indicating that a navigation route is switched, a control indicating that a via point of the navigation route is set, a control indicating that a navigation view angle is switched, and a control indicating that a navigation route is previewed.
In some embodiments, the matching, by the server, the voice interaction information and the navigation scenario information with the information template, and generating the operation instruction according to a result of the matching, where the receiving the operation instruction generated by the server according to the voice interaction information, the information, and the information template corresponding to the navigation scenario information includes:
receiving an execution instruction generated by the server according to successful matching;
the executing the operation corresponding to the operation instruction comprises:
and performing operation corresponding to the execution instruction on the navigation scene.
In some embodiments, the receiving the operation instruction generated by the server according to the voice interaction information, the navigation scenario information, and the information template corresponding to the navigation scenario information includes:
receiving a feedback instruction generated by the server according to the matching failure;
the executing the operation corresponding to the operation instruction comprises:
and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
In some embodiments, the performing, by the navigation scene, an operation corresponding to the execution instruction includes:
judging whether the vehicle-mounted map application program intercepts the execution instruction;
and if the vehicle-mounted map application program does not intercept the execution instruction, performing operation corresponding to the execution instruction on the navigation scene through a software development kit of the vehicle-mounted map application program.
In some embodiments, the performing the operation corresponding to the execution instruction on the navigation scene further includes:
if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
and performing operation corresponding to the execution instruction on the navigation scene through the vehicle-mounted map application program.
The application provides an information processing method, which comprises the following steps:
receiving navigation scene information uploaded by a vehicle-mounted map application program; and
and processing the navigation scene information to obtain a corresponding information template.
In some embodiments, the processing the navigation scenario information to obtain an information template comprises:
and generalizing an expression mode interacted with the navigation scene information to obtain the information template.
In some embodiments, the information processing method further includes:
receiving voice interaction information aiming at a navigation scene sent by the vehicle;
matching the information template with the voice interaction information and the navigation scene information;
and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
The application provides a vehicle, the operating system of vehicle installs on-vehicle map application, on-vehicle map application includes navigation scene information, the vehicle includes:
the voice acquisition module is used for acquiring voice interaction information of a user aiming at a navigation scene;
the communication module is used for sending the voice interaction information and the navigation scene information to a server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and an information template corresponding to the navigation scene information;
and the control module is used for executing the operation corresponding to the operation instruction.
The application provides a server, including:
the communication module is used for receiving navigation scene information uploaded by a vehicle-mounted map application program; and
and the processing module is used for processing the navigation scene information to obtain a corresponding information template.
A non-transitory computer-readable storage medium containing computer-executable instructions that, when executed by one or more processors, cause the processors to perform the method of interacting with or processing information for navigating a scene by an in-vehicle map application is provided.
According to the navigation scene interaction method, the information processing method, the vehicle, the server and the computer readable storage medium, the graphical user interface information of the navigation scene of the vehicle-mounted map application program is synchronized to the server, synchronization and consistency of local and cloud information are achieved, the server grasps more graphical user interface information of the vehicle-mounted map application program, the possibility of interaction in the navigation scene through voice is provided, and voice interaction is enabled to be more intelligent.
Drawings
The foregoing and/or additional aspects and advantages of the present application will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
FIG. 1 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
FIG. 2 is a block schematic diagram of a vehicle according to certain embodiments of the present application.
FIG. 3 is a schematic diagram of a scenario of an interaction method according to some embodiments of the present application.
FIG. 4 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
FIG. 5 is a flow chart diagram illustrating an interaction method according to some embodiments of the present application.
Fig. 6 is a schematic flow chart of an information processing method according to some embodiments of the present application.
FIG. 7 is a block diagram of a server in accordance with certain embodiments of the present application.
FIG. 8 is a schematic illustration of a vehicle and server interaction in accordance with certain embodiments of the present application.
Fig. 9 is a schematic flow chart of an information processing method according to some embodiments of the present application.
Fig. 10 is a schematic flow chart diagram of an information processing method according to some embodiments of the present application.
Detailed Description
Reference will now be made in detail to embodiments of the present application, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the drawings are exemplary and intended to be used for explaining the present application and should not be construed as limiting the present application.
Referring to fig. 1, the present application provides an interactive method for a navigation scene of a vehicle map application. The method comprises the following steps:
s10: acquiring voice interaction information of a user aiming at a navigation scene;
s20: sending voice interaction information and navigation scene information to a server;
s30: receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and the information template corresponding to the navigation scene information;
s40: and executing the operation corresponding to the operation instruction.
The embodiment of the application provides a vehicle. The vehicle includes a display area, an electro-acoustic element, a communication element, and a processor. The display area of the vehicle may include a dashboard screen, an on-board display area screen, and a heads-up display that may be implemented on a vehicle windshield, among others. An on-board system operating on a vehicle presents the presented content to a User using a Graphical User Interface (GUI). The display area includes a number of UI elements, and different display areas may present the same or different UI elements. The UI elements may include card objects, application icons or interfaces, folder icons, multimedia file icons, and controls for making interactive operations, among others. The electroacoustic element is used for acquiring voice interaction information of a user aiming at the navigation scene. The communication element is used for sending the voice interaction information and the navigation scene information to the server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and the information template corresponding to the navigation scene information. The processor is used for executing the operation corresponding to the operation instruction.
Referring to fig. 2, an embodiment of the present application further provides a vehicle 100, and the interaction method according to the embodiment of the present application may be implemented by the vehicle 100 according to the embodiment of the present application.
Specifically, the operating system of the vehicle 100 is installed with an in-vehicle map application, and the vehicle 100 includes a voice acquisition module 102, a communication module 104, and a control module 106. The S10 may be implemented by the voice acquisition module 102, the S20, S30 may be implemented by the communication module 104, and the S40 may be implemented by the control module 106. In other words, the voice obtaining module 102 is configured to obtain the voice interaction information of the user for the navigation scene. The communication module 104 is configured to send the voice interaction information and the navigation scene information to the server, and receive an operation instruction generated by the server according to the voice interaction information, the navigation scene information, and an information template corresponding to the navigation scene information. The control module 106 is configured to execute an operation corresponding to the operation instruction.
In the interaction method of the vehicle-mounted map application program and the vehicle 100, navigation scene information of the graphical user interface of the vehicle-mounted map application program is synchronized to the server, so that synchronization and consistency of local and cloud information are realized, the server grasps more information of the graphical user interface of the vehicle-mounted map application program, the possibility of interaction in the navigation scene through voice is provided, and voice interaction is more intelligent.
Specifically, the intelligent display area of the vehicle can provide a convenient entrance for a user to control the vehicle and interact with the vehicle, a voice assistant function is added in the vehicle-mounted operating system, voice information input by the user can be analyzed through voice recognition and semantic recognition under a certain condition, a corresponding control instruction is generated conveniently, and convenience is further provided for interaction between the user and the vehicle. However, for the vehicle map application, voice interaction still stays at a relatively early stage, and only simple interaction can be realized, for example, zooming in and zooming out on the display scale of the graphical user interface of the vehicle map application is realized through voice. For more complex functions, for example, for a navigation scene, a user can only switch a currently driving route to other routes for navigation through input in a graphical user interface of the navigation scene, such as clicking and the like, and for example, the current driving road and a route full view angle are switched, and interaction cannot be realized through voice. For the situation that the vehicle is in the driving mode at present, a user interacts through a graphical user interface of a vehicle-mounted map application program while driving, and certain safety risks exist.
In the embodiment, after waking up the voice assistant, the user inputs voice information, and obtains the information of the graphical user interface including the navigation line, the road condition, and the like displayed on the navigation scene interface of the current vehicle-mounted map application program of the vehicle while obtaining the voice information. The navigation scene information comprises information of two aspects of display form and display structure. The display form is also a presentation form of the navigation scene, for example, the navigation scene may be presented in the form of multiple windows, a combination of a window and a card, and the display structure is also a specific structure of the display form of the window and the like, for example, the number of rows and columns of the sub-content included in the window, the included control, the position of the control distributed in the scene, the display hierarchy, and the like.
And after the user wakes up the voice assistant locally, inputting voice interaction information interacted with the navigation scene. The vehicle sends the voice interaction information and the navigation scene information to a server of a cloud service provider, the server analyzes the voice interaction information by using the navigation scene information as auxiliary information, so that an operation instruction is generated and is transmitted back to the local vehicle, and the vehicle executes corresponding operation according to the operation instruction.
The navigation scene information is synchronized to the server through a voice software development kit, and the voice software development kit is a hub for voice interaction between the vehicle-mounted map application program and the server. In one aspect, a software development kit defines a specification for generating voice interaction information. On the other hand, the voice software development kit can realize the synchronization of the navigation scene information in the vehicle-mounted map application program to the server and the transmission of the operation instruction generated by the server for the voice interaction information to the vehicle-mounted map application program.
In one example, the in-vehicle map application may invoke an information synchronization method provided by the software development kit to synchronize the navigation scenario information to the software development kit.
And the software development toolkit performs information fault tolerance and normalization check on the received navigation scene information. Specifically, error information possibly existing in the navigation scene information is corrected according to the voice interaction generation specification, so that the navigation scene information data are guaranteed to meet the generation specification and can be identified and analyzed by the server. In addition, the software development toolkit checks the data of the navigation scene information in the vehicle-mounted map application according to the generation specification of the voice interaction. For example, it is checked whether the attributes of the data are correct, whether the encoding of elements in the data is unique, and the like. And if the attribute configuration is correct, namely the attribute configuration meets the generation specification, releasing the navigation scene information. Otherwise, feedback is given to the vehicle-mounted map application program, for example, an error log is fed back, or a prompt is given on a vehicle-mounted map application program interface.
The parsing of the voice interaction information generally includes two parts of voice recognition and semantic parsing, and the voice recognition may be performed locally, for example, the voice interaction information may be recognized by a text-to-speech module of the vehicle to convert the voice into text. Of course, the voice recognition may also be performed at the server, thereby reducing the processing load on the vehicle-side operating system. Semantic parsing can be completed in a server, and generally, understanding of voice interaction information is achieved through steps of word segmentation, analysis and the like of a text.
The navigation scene information can make the server more definite the current interactive scene of the vehicle when performing semantic analysis, and effectively limits the scope of the semantic analysis. For example, when the vehicle-mounted map application program is in a navigation scene, and a user is in a navigation driving process in the navigation scene with the shortest route, the road condition is gradually congested, so that the user wants to switch to the route with the shortest time, so as to switch the navigation route from the current route to the route with the shortest time and go to a destination, and send out the voice interaction information with the shortest time. And under the condition of synchronously acquiring navigation scene information, the server can judge that the user hopes to control the vehicle to navigate to the destination by taking the calculated route with the shortest time as the navigation route, so that the vehicle is controlled to navigate to the destination by adopting the shortest time route.
Therefore, the intelligence of voice control and the success rate of hitting the real intention can be improved, and the user experience is better.
The information template of the navigation scene information is an information template formed after processing according to functions and contents in a navigation scene graphical user interface uploaded by a vehicle. The information template is stored in the server, so that after the navigation scene information uploaded by the user is received, the server can confirm the information template corresponding to the current navigation scene information through matching with the information template, the current interaction scene of the user is obtained, the intention of the user can be judged according to the voice interaction information, and the voice interaction information can be assisted to analyze the real intention of the user according to the navigation scene interface interacted by the user.
In addition, in the application, a driver can perform voice interaction with the vehicle-mounted map application program at any time in the driving process, such as a driving or parking state, so that the adjustment of the vehicle-mounted map scale is realized. Particularly, in the driving state, the voice input is adopted to replace the manual input of a user to interact with the vehicle-mounted map application program, and the driving safety can be considered.
In this embodiment, the navigation scene information includes control information of a graphical user interface of the navigation scene.
Specifically, in the process of actually using the vehicle-mounted map application program, after a user selects a certain information point in the map, for example, a destination of the current driving, a request for obtaining a route to the information point is initiated, and after calculation, the vehicle-mounted map application program returns to the navigation route result list. The user can select one line in the navigation route list to initiate navigation so as to enter a navigation scene, and further operations such as route switching, 2D/3D navigation interface switching and the like are carried out in the navigation process.
The contents are distributed and displayed by corresponding controls, navigation scene information is also the control information of a graphical user interface in the current navigation scene, and the vehicle-mounted map application program distributes the navigation scene information through the voice interaction control library control, so that a layout data structure capable of being controlled by voice is constructed. In the data structure design process, a control supporting graphic interaction operation needs to be replaced by a control supporting voice interaction, namely a control in a voice interaction control library. For example, the linear layout control LinearLayout in the original structure is replaced by a linear layout control XLinearLayout supporting voice interaction operation packaged by a voice interaction control library. For another example, the text control TextView in the original structure is replaced by the text control XTextView supporting voice interaction operation packaged by the voice interaction control library.
Controls generally include, but are not limited to, the following information: an element identification, an element type, an action type of the element, a phonetic utterance of the element, and the like. Wherein the element identification is unique for each element by which the element can be found. The element types may include groups, text, images, and the like. The action type of an element may include clicking, sliding, and the like. The phonetic interpretation of an element includes waking up a certain operation keyword, etc.
Referring to fig. 3, the control information includes one or more of a traffic information window, a control indicating that navigation is started, a control indicating that a navigation scene exits, a control indicating a special location icon, a control indicating that a navigation route is switched, a control indicating that a via point of the navigation route is set, a control indicating that a navigation view angle is switched, and a control indicating that a navigation route is previewed.
Specifically, the vehicle-mounted map application program may perform layout for a control corresponding to the following navigation scene interaction, for example, a navigation operation, a large map mode, route switching, route point setting, a navigation mode, a route preview mode, and the like.
The "traffic information window" is used to display the current traffic and navigation information.
The control indicating that navigation is initiated may be a "navigation operation" control that may support the initiation or resumption of interrupted navigation through voice interaction.
The control representing exiting the navigation scenario may be an "exit" control that may enable exiting the navigation scenario through voice interaction.
The control for representing the special place graphic can be a 'big picture mode' control which can support voice interaction to close the picture displayed in an enlarged manner on a graphic user interface of the special place such as an intersection, a tunnel portal and the like.
The control representing the navigation route switching can be a 'route switching' control, and the 'route switching' control can support voice interaction to reselect the current navigation scene and initiate navigation according to the route. During the interaction, a voice feedback such as "switched to a route" may be provided.
The control for showing the setting of the route points of the navigation route can be a 'route point setting' control, and the 'route point setting' control can support voice interaction to add the route points, search the route points, and plan the navigation scene again according to the added route points and perform navigation.
The control representing the navigation view angle switching may be a "navigation mode" control that may support voice interaction to switch the view angle of the current map, such as a 2D planar map or a 3D stereoscopic map.
The control representing the navigation route preview can be a 'route preview mode' control which can support voice interaction switching of the current driving route and the preview of the navigation route, and switching of the view angles of the main road and the auxiliary road of the current driving route.
Referring to fig. 4, in some embodiments, the server matches the voice interaction information and the navigation scene information with the information template, and generates an operation instruction according to a matching result. S30 includes:
s31: receiving an execution instruction generated by the server according to successful matching;
s40 includes:
s41: and performing operation corresponding to the execution instruction on the navigation scene.
In some embodiments, S31 may be implemented by the communication module 104 and S41 may be implemented by the control module 106. That is, the communication module 104 is configured to receive the execution instruction generated by the server according to the matching success. The control module 106 is configured to perform an operation corresponding to the execution instruction on the navigation scene.
In some embodiments, the communication element is configured to receive an execution instruction generated by the server upon a successful match. The processor is used for carrying out operation corresponding to the execution instruction on the navigation scene.
Specifically, after the voice assistant is awakened each time, different vehicles upload the voice interaction information and the navigation scene information to the server together. The server can obtain a large amount of historical data of the navigation scene information along with the use of the user, and the collected navigation scene information is supplemented, expanded and sorted in a machine learning or manual labeling mode and the like, so that the understanding of the server on the navigation scene information is enriched, and the sorted content can form a corresponding information template and is stored in the server, so that the accuracy and the recognition efficiency of semantic recognition are improved in the subsequent use process of the user.
In an actual process, if the user uses the voice assistant for the first time, there may not be a pre-stored information template at the server side, and in this case, the server directly assists the voice interaction information to perform semantic recognition according to the navigation scene information. If the voice assistant is not used for the first time, after the server receives the navigation scene information, the current graphical user interface can be identified according to the control information of the navigation scene information, and then an information template corresponding to the control information is called, so that the voice interaction information and the navigation scene information can be matched with the information template to analyze the real intention of the user.
It can be understood that the same user may express the same voice interaction instruction differently in the previous and subsequent implementation processes, and different users may also express the same instruction differently. And the set information template is generalized aiming at each possible expression mode of voice interaction. The richer the content of the information template, the higher the probability and success rate of recognizing the voice interaction instruction
For example, for "route switching" in the navigation scene information, the expression that the user switches the navigation route may be expanded to, for example, switch to the nth route, switch to the alternative route N, help me switch to the nth route, navigate to the nth route, and the like. These representations are stored in the information template.
The speech-to-text conversion module of the vehicle performs speech recognition on the speech interaction information, and of course, the speech recognition may also be performed by the speech-to-text conversion module of the server. And comparing the uploaded information with the information template to realize the analysis of the voice interaction information semantics. And under the condition that the matching is successful, generating an execution instruction corresponding to the interactive information, returning to the vehicle, and executing the execution instruction by the vehicle.
For example, when a user wants to switch a route, voice interaction information such as 'switch to nth route' is sent out, the voice interaction information and navigation scene information are sent to a server together, the server can obtain the state, the structural frame layout and a control which can carry out interaction of the current navigation scene according to the navigation scene information, the voice interaction information and the navigation scene information are matched with an information template, after matching, the semantic meaning of the voice interaction information is confirmed to be that the route which is currently navigated is switched to the nth route in a route list for navigation, an execution instruction for switching the route to the nth route and carrying out navigation is generated, and after the vehicle-mounted map application program receives the execution instruction, the vehicle-mounted map application program switches the navigation route to the nth route and starts navigation.
Referring again to fig. 4, in some embodiments, S30 includes:
s32: receiving a feedback instruction generated by the server according to the matching failure;
s40 includes:
s42: and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
In some embodiments, S32 may be implemented by the communication module 104 and S42 may be implemented by the control module 106. That is, the communication module 104 is configured to receive a feedback instruction generated by the server according to the matching failure. The control module 106 is configured to broadcast the information of the matching failure according to the feedback instruction to prompt the user.
In some embodiments, the communication element is to receive a feedback instruction generated by the server based on the failure to match. And the processor is used for broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
Specifically, for interaction which is not supported in a navigation scene or voice interaction information which cannot be subjected to semantic analysis, the server also gives feedback which cannot be identified, and the application program can broadcast the feedback information in modes of voice, text popup display and the like, so that the user is prompted that input information is invalid.
For the voice interaction information which cannot be identified, the vehicle-mounted map application program can monitor the interaction operation of the user through the graphical interaction interface within the preset time period of the broadcast feedback prompt, and reports the interaction operation to the server, relevant personnel manually detect the interaction operation of the voice interaction information and the graphical user interface, judge whether the voice interaction information and the graphical user interface are related, and if the voice interaction information and the graphical user interface are related, expand the expression of the voice interaction information into an information template corresponding to the execution instruction. And if no association exists, ignoring the reported information.
For example, the user wants to switch the navigation route from the current route to the route with the shortest time and go to the destination. Sending out the voice interaction information with the shortest time, matching the voice interaction information and the navigation scene information with the information template, confirming that the voice interaction information and the navigation scene information cannot be matched with the current information template after matching, generating a feedback instruction, and broadcasting the information which cannot be identified after the vehicle-mounted map application program receives the feedback instruction. The user then manually clicks on the route and initiates navigation. The vehicle-mounted map application program reports the operation of the user to the voice server, and relevant workers judge that the expression of the shortest time is related to the operation of starting navigation, so that the shortest time can be added to an information template of a voice interaction instruction related to route switching.
Referring to fig. 5, in some embodiments, S41 includes:
s411: judging whether the vehicle-mounted map application program intercepts an execution instruction;
s412: and if the vehicle-mounted map application program does not intercept the execution instruction, performing operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program.
In some embodiments, S411, S412 may be implemented by the control module 106. That is, the control module 106 is configured to determine whether the execution instruction is intercepted by the vehicle-mounted map application program, and perform an operation corresponding to the execution instruction on the navigation route through the software development kit of the vehicle-mounted map application program when the execution instruction is not intercepted by the vehicle-mounted map application program.
In some embodiments, the processor determines whether the execution instruction is intercepted by the vehicle-mounted map application program, and is used for performing an operation corresponding to the execution instruction on the navigation route through a software development kit of the vehicle-mounted map application program under the condition that the execution instruction is not intercepted by the vehicle-mounted map application program.
Specifically, an execution instruction is generated after the server is successfully matched, and the execution instruction is returned. According to the business requirement, different objects are usually selected to process the execution instruction. For example, if a relatively simple, single operation is performed, the execution instructions may be processed directly by the software development kit. And if more personalized subsequent operations are needed on the basis of the basic operations, the execution instructions are processed by the vehicle-mounted map application program.
In the specific implementation process, the processing mechanism is preset, and after the vehicle-mounted map application program receives the execution instruction, the vehicle-mounted map application program selects whether to intercept the execution instruction according to different execution instruction processing mechanisms. And if the vehicle-mounted map application program is not intercepted, the execution instruction is processed and executed by a software development kit.
Referring again to fig. 5, in some embodiments, S41 further includes:
s413: if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
s414: and carrying out operation corresponding to the execution instruction on the navigation scene through the vehicle-mounted map application program.
In some embodiments, S413, S414 may be implemented by the control module 106. That is, the control module 106 is configured to pass through the execution instruction to the vehicle-mounted map application program through the software development kit when the vehicle-mounted map application program intercepts the execution instruction, and is configured to perform an operation corresponding to the execution instruction on the navigation scene through the vehicle-mounted map application program.
In some embodiments, the processor is used for transmitting the execution instruction to the vehicle-mounted map application program through the software development kit when the execution instruction is intercepted by the vehicle-mounted map application program, and is used for performing an operation corresponding to the execution instruction on the navigation scene through the vehicle-mounted map application program.
In the specific implementation process, the processing mechanism is preset, and after the execution instruction is received, the vehicle-mounted map application program selects whether to intercept the execution instruction according to different execution instruction processing mechanisms. If the in-vehicle map application intercepts the execution instructions, the software development kit will not process the execution instructions, but instead pass the execution instructions through to the in-vehicle map application, which processes the execution instructions.
In one example, for an "exit navigation" interaction, since the operation is relatively simple and there is generally no subsequent operation, then the setup may be performed by a software development kit. The vehicle-mounted map application program does not intercept an execution instruction related to exit, the software development kit processes the execution instruction and triggers click processing on an exit tag, and therefore the exit navigation scene is achieved.
For the "via point setting" interaction, since the user is usually further recalculating the route including the via point and initiating navigation with the newly calculated route when adding the via point, the setting may be performed by the in-vehicle map application. The vehicle-mounted map application program intercepts an execution instruction related to 'passing point setting', the software development kit does not process, and the vehicle-mounted map application program triggers the setting of adding the passing point, automatically calculates a route reaching a destination and including the passing point and initiates the setting of navigation.
In another example, taking "route switching" interaction, i.e. interaction for switching navigation route as an example, if the application program does not intercept, the software development kit processes the interaction, and triggers the click processing of the route to switch the route, but does not perform the operation of navigating with the switched route, that is, if the user wishes to further initiate navigation, a manual operation is also required.
And if the application program is intercepted and the software development kit does not process, the application program triggers the processing of switching the route and automatically triggers the operation of initiating navigation by the switched route. With better intelligence and operational efficiency.
Referring to fig. 6, the present application further provides an information processing method for processing the voice interaction information sent from the vehicle 100 to the server 200 in the above embodiment. The information processing method comprises the following steps:
s50: receiving navigation scene information uploaded by a vehicle-mounted map application program; and
s60: and processing the navigation scene information to obtain a corresponding information template.
The embodiment of the application provides a server. The server includes a communication element and a processor. The communication element is used for receiving navigation scene information synchronized by the vehicle-mounted map application program through the software development kit. The processor is used for processing the navigation scene information to obtain an information template.
Referring to fig. 7, an embodiment of the present application further provides a server 200, and an information processing method according to the embodiment of the present application may be implemented by the server 200 according to the embodiment of the present application.
Specifically, the server 200 includes a communication module 202 and a processing module 204. S50 may be implemented by the communication module 202, and S60 may be implemented by the processing module 204. Or, the communication module 202 is configured to receive navigation scene information uploaded by the in-vehicle map application. The processing module 204 is configured to process the navigation scene information to obtain a corresponding information template.
Referring to fig. 8, in the process of implementing voice control on the vehicle, the server communicates with the vehicle, and navigation scene information on the vehicle-mounted map application program is synchronized to the server, so that synchronization and consistency between local information and cloud information are realized, the server grasps more information of the vehicle-mounted map application program interface, the possibility of interaction in the navigation scene through voice is provided, and voice interaction is more intelligent.
The server receives navigation scene information sent by different vehicles, and an information template corresponding to the navigation scene is constructed according to control information contained in the navigation scene information.
The information template may include the same element and different elements, or common elements and personalized elements, in the graphical user interface for the same navigation scenario. According to the same element or the common element in the graphical user interface, the server can construct a basic frame of the current navigation scene as a basis of the information template. According to different elements in the graphical user interface, the server can acquire specific information of the current navigation scene, so that the content of the information template is enriched. The information template has the significance of mastering more user interaction information and providing more accurate assistance for voice recognition.
Referring to fig. 9, in some embodiments, S60 includes:
s61: and generalizing an expression mode of information interaction of the navigation scene to obtain an information template.
In some embodiments, S61 may be implemented by the processing module 204, that is, the processing module 204 is configured to generalize the expression of interaction with the navigation scenario information to obtain the information template.
In some embodiments, the processor is configured to generalize the expression of interaction with navigation scenario information to obtain an information template.
In particular, voice interaction refers to generally comprising two parts, an instruction object and a manner of operation. Correspondingly, the instruction object, namely the control in the graphical user interface included in the navigation scene information, corresponds to the information template, and the expression mode of the control is generalized. That is, the same instruction object is generalized, so that different expression modes correspond to the instruction object.
For example, for "route switching", the generalization process may include expressions of a route, a line, an nth route, and the like.
The operation mode is interaction with the control and generalization processing is carried out on the expression mode of the interaction with the control, namely generalization processing is carried out on the same operation mode, so that different expression modes correspond to the interaction operation.
For example, for "route switching", the generalization process may include expressions of looking at the nth item, looking at the nth route, helping me to switch to the nth route, switching to the nth item, navigating to the nth route, going to the route with the shortest time, going to the route with the least traffic lights, going to the route with the shortest route, and the like.
For "start navigation", the generalization process may include expressions of navigation, driving, walking, driving, start navigation, and the like.
After a certain amount of voice interaction information is collected, the information template can be expanded manually, the information template has richer contents, and the same instruction has more expression modes, so that the analysis of the voice interaction information can be better assisted.
Referring to fig. 10, in some embodiments, the information processing method further includes:
s70: receiving voice interaction information aiming at a navigation scene sent by a vehicle;
s80: matching the information template with the information according to the voice interaction information and the navigation scene information;
s90: and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
In some embodiments, S70 may be implemented by communication module 202. S80 may be implemented by the processing module 204, and S90 may be implemented by the communication module 202 and the processing module 204. In other words, the communication module 202 is configured to receive the voice interaction information for the navigation scene sent by the vehicle 100. The processing module 204 matches the information template according to the voice interaction information and the navigation scene information, and is configured to generate an execution instruction or a feedback instruction according to a matching result. The communication module 202 is also used to send execution instructions or feedback instructions to the vehicle 100.
In some embodiments, the communication element is configured to receive voice interaction information for a navigation scenario transmitted by a vehicle. The processor is used for matching the information template with the information according to the voice interaction information and the navigation scene information, and generating an execution instruction or a feedback instruction according to a matching result. The communication element is also used for sending the execution instruction or the feedback instruction to the vehicle.
Specifically, the vehicle sends the voice interaction information to a server at the cloud end, the server matches the voice interaction information and the navigation scene information with the information template, a feedback instruction is generated after the matching is successful and is transmitted back to the vehicle, and then the vehicle executes corresponding operation on the navigation scene according to the execution instruction or prompts a user according to the feedback instruction.
For example, when the user wants to switch the route, voice interaction information such as "switch to nth route" is sent out, and the vehicle uploads the voice interaction information and the navigation scene information to the server together. After receiving the interactive information, the server matches the interactive information and the navigation scene information with the information template, confirms that the semantics of the interactive information is to switch the current driving navigation route to the Nth route in the route list and navigate by the Nth route after matching, thereby generating an execution instruction from the switching route to the Nth route and sending the execution instruction back to the vehicle, and after receiving the execution instruction, the vehicle-mounted map application program switches the navigation scene to the Nth route and starts navigation.
The embodiment of the application also provides a computer readable storage medium. One or more non-transitory computer-readable storage media containing computer-executable instructions that, when executed by one or more processors, cause the processors to perform the method for interacting with or processing information for navigating a scene of an in-vehicle map application of any of the embodiments described above.
It will be understood by those skilled in the art that all or part of the processes of the methods of the above embodiments may be implemented by hardware instructions of a computer program, which may be stored in a non-volatile computer-readable storage medium, and when executed, may include the processes of the embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), or the like.
The above examples only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the present application. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (13)

1. An interaction method for a navigation scene of a vehicle-mounted map application program is characterized in that the vehicle-mounted map application program comprises navigation scene information, and the interaction method comprises the following steps:
acquiring voice interaction information of a user aiming at a navigation scene;
sending the voice interaction information and the navigation scene information to a server;
receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and an information template corresponding to the navigation scene information;
and executing the operation corresponding to the operation instruction.
2. The interaction method according to claim 1, wherein the navigation scene information includes control information of a graphical user interface of the navigation scene.
3. The interaction method according to claim 2, wherein the control information includes one or more of a traffic information window, a control for starting navigation, a control for exiting a navigation scene, a control for displaying a special location diagram, a control for displaying switching of a navigation route, a control for displaying setting of a via point of a navigation route, a control for displaying switching of a navigation view angle, and a control for displaying a preview of a navigation route.
4. The interaction method according to claim 3, wherein the server matches the voice interaction information and the navigation scenario information with the information template, and generates the operation instruction according to a result of the matching, and the receiving the operation instruction generated by the server according to the voice interaction information, the information, and the information template corresponding to the navigation scenario information comprises:
receiving an execution instruction generated by the server according to successful matching;
the executing the operation corresponding to the operation instruction comprises:
and performing operation corresponding to the execution instruction on the navigation scene.
5. The interaction method according to claim 4, wherein the receiving the operation instruction generated by the server according to the voice interaction information, the navigation scenario information and the information template corresponding to the navigation scenario information comprises:
receiving a feedback instruction generated by the server according to the matching failure;
the executing the operation corresponding to the operation instruction comprises:
and broadcasting the information of the matching failure according to the feedback instruction so as to prompt the user.
6. The interaction method according to claim 4, wherein the performing the operation corresponding to the execution instruction on the navigation scene comprises:
judging whether the vehicle-mounted map application program intercepts the execution instruction;
and if the vehicle-mounted map application program does not intercept the execution instruction, performing operation corresponding to the execution instruction on the navigation scene through a software development kit of the vehicle-mounted map application program.
7. The interaction method according to claim 6, wherein performing the operation corresponding to the execution instruction on the navigation scene further comprises:
if the vehicle-mounted map application program intercepts the execution instruction, the execution instruction is transmitted to the vehicle-mounted map application program through the software development kit;
and performing operation corresponding to the execution instruction on the navigation scene through the vehicle-mounted map application program.
8. An information processing method characterized by comprising:
receiving navigation scene information uploaded by a vehicle-mounted map application program; and
and processing the navigation scene information to obtain a corresponding information template.
9. The information processing method of claim 8, wherein the processing the navigation scene information to obtain an information template comprises:
and generalizing an expression mode interacted with the navigation scene information to obtain the information template.
10. The information processing method according to claim 8, characterized by further comprising:
receiving voice interaction information aiming at a navigation scene sent by the vehicle;
matching the information template with the voice interaction information and the navigation scene information;
and generating an execution instruction or a feedback instruction according to the matching result and sending the execution instruction or the feedback instruction to the vehicle.
11. A vehicle, characterized in that an operating system of the vehicle is installed with an in-vehicle map application program including navigation scene information, the vehicle comprising:
the voice acquisition module is used for acquiring voice interaction information of a user aiming at a navigation scene;
the communication module is used for sending the voice interaction information and the navigation scene information to a server and receiving an operation instruction generated by the server according to the voice interaction information, the navigation scene information and an information template corresponding to the navigation scene information;
and the control module is used for executing the operation corresponding to the operation instruction.
12. A server, comprising:
the communication module is used for receiving navigation scene information uploaded by a vehicle-mounted map application program; and
and the processing module is used for processing the navigation scene information to obtain a corresponding information template.
13. A non-transitory computer-readable storage medium of computer-executable instructions, which when executed by one or more processors, cause the processors to perform the method of interacting with the in-vehicle map application navigation scenario of any one of claims 1-7 or the method of information processing of claims 8-10.
CN202010599635.3A 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server Pending CN111722905A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010599635.3A CN111722905A (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010599635.3A CN111722905A (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Publications (1)

Publication Number Publication Date
CN111722905A true CN111722905A (en) 2020-09-29

Family

ID=72569510

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010599635.3A Pending CN111722905A (en) 2020-06-28 2020-06-28 Interaction method, information processing method, vehicle and server

Country Status (1)

Country Link
CN (1) CN111722905A (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112525214A (en) * 2020-11-24 2021-03-19 广州橙行智动汽车科技有限公司 Interaction method and device for map card, vehicle and readable medium
CN113359690A (en) * 2021-06-28 2021-09-07 北京小鹏汽车有限公司 Vehicle map full-view presentation control method and device
CN113436627A (en) * 2021-08-27 2021-09-24 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
CN113436628A (en) * 2021-08-27 2021-09-24 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
CN113450801A (en) * 2021-08-27 2021-09-28 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
CN113506569A (en) * 2021-06-07 2021-10-15 上海七十迈数字科技有限公司 Voice interaction method, system and equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108538291A (en) * 2018-04-11 2018-09-14 百度在线网络技术(北京)有限公司 Sound control method, terminal device, cloud server and system
CN108682419A (en) * 2018-03-30 2018-10-19 京东方科技集团股份有限公司 Sound control method and equipment, computer readable storage medium and equipment
CN111179930A (en) * 2019-12-31 2020-05-19 苏州思必驰信息科技有限公司 Method and system for realizing intelligent voice interaction in driving process

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108682419A (en) * 2018-03-30 2018-10-19 京东方科技集团股份有限公司 Sound control method and equipment, computer readable storage medium and equipment
US20190304466A1 (en) * 2018-03-30 2019-10-03 Boe Technology Group Co., Ltd. Voice control method, voice control device and computer readable storage medium
CN108538291A (en) * 2018-04-11 2018-09-14 百度在线网络技术(北京)有限公司 Sound control method, terminal device, cloud server and system
CN111179930A (en) * 2019-12-31 2020-05-19 苏州思必驰信息科技有限公司 Method and system for realizing intelligent voice interaction in driving process

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
杨前华;万宇;: "基于增强现实技术的老年人导航应用***设计", 电视技术 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112525214A (en) * 2020-11-24 2021-03-19 广州橙行智动汽车科技有限公司 Interaction method and device for map card, vehicle and readable medium
CN112525214B (en) * 2020-11-24 2024-05-28 广州橙行智动汽车科技有限公司 Interaction method and device of map card, vehicle and readable medium
CN113506569A (en) * 2021-06-07 2021-10-15 上海七十迈数字科技有限公司 Voice interaction method, system and equipment
CN113359690A (en) * 2021-06-28 2021-09-07 北京小鹏汽车有限公司 Vehicle map full-view presentation control method and device
CN113436627A (en) * 2021-08-27 2021-09-24 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
CN113436628A (en) * 2021-08-27 2021-09-24 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
CN113450801A (en) * 2021-08-27 2021-09-28 广州小鹏汽车科技有限公司 Voice interaction method, device, system, vehicle and medium
EP4086749A1 (en) * 2021-08-27 2022-11-09 Guangzhou Xiaopeng Motors Technology Co., Ltd. Voice interaction method, apparatus and system, vehicles, and storage medium
EP4086580A1 (en) * 2021-08-27 2022-11-09 Guangzhou Xiaopeng Motors Technology Co., Ltd. Voice interaction method, apparatus and system, vehicle, and storage medium
EP4086579A1 (en) * 2021-08-27 2022-11-09 Guangzhou Xiaopeng Motors Technology Co., Ltd. Speech interaction method, apparatus and system, transportation means, and medium

Similar Documents

Publication Publication Date Title
CN111722905A (en) Interaction method, information processing method, vehicle and server
CN111768779B (en) Interaction method, information processing method, vehicle and server
CN111722825A (en) Interaction method, information processing method, vehicle and server
US20190120649A1 (en) Dialogue system, vehicle including the dialogue system, and accident information processing method
US9922651B1 (en) Avionics text entry, cursor control, and display format selection via voice recognition
US9618352B2 (en) Apparatus and method for controlling navigator via speech dialogue
JP7042240B2 (en) Navigation methods, navigation devices, equipment and media
EP4124532A2 (en) Control method, apparatus, device, and medium for autonomous driving vehicle
EP3044781B1 (en) Vehicle interface system
CN111768777A (en) Voice control method, information processing method, vehicle and server
CN112525214B (en) Interaction method and device of map card, vehicle and readable medium
CN111753039A (en) Adjustment method, information processing method, vehicle and server
CN110767219B (en) Semantic updating method, device, server and storage medium
EP4086749A1 (en) Voice interaction method, apparatus and system, vehicles, and storage medium
CN113113015A (en) Interaction method, information processing method, vehicle and server
CN114327185A (en) Vehicle screen control method and device, medium and electronic equipment
CN113961114A (en) Theme replacement method and device, electronic equipment and storage medium
EP4086580A1 (en) Voice interaction method, apparatus and system, vehicle, and storage medium
KR102371513B1 (en) Dialogue processing apparatus and dialogue processing method
US20200251109A1 (en) Method For Operating And/Or Controlling A Dialog System
US20240210197A1 (en) requesting and receiving reminder instructions in a navigation session
US9858918B2 (en) Root cause analysis and recovery systems and methods
CN113470412B (en) Vehicle map real-time road condition presentation control method and device
CN118173086A (en) Voice interaction method, vehicle, device and storage medium
CN113359690B (en) Vehicle map overview presentation control method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant after: Guangzhou Xiaopeng Automatic Driving Technology Co.,Ltd.

Address before: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant before: Guangzhou Xiaopeng Internet of vehicles Technology Co.,Ltd.

TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20201207

Address after: Room 209, No.8, Songgang street, Cencun, Changxing street, Tianhe District, Guangzhou City, Guangdong Province, 510000

Applicant after: Guangzhou Zhipeng Internet of vehicles Technology Co.,Ltd.

Address before: Room 46, room 406, No.1, Yichuang street, Zhongxin knowledge city, Huangpu District, Guangzhou City, Guangdong Province

Applicant before: Guangzhou Xiaopeng Automatic Driving Technology Co.,Ltd.