CN115424614A - Human-computer interaction method and device, electronic equipment and vehicle - Google Patents

Human-computer interaction method and device, electronic equipment and vehicle Download PDF

Info

Publication number
CN115424614A
CN115424614A CN202211063683.6A CN202211063683A CN115424614A CN 115424614 A CN115424614 A CN 115424614A CN 202211063683 A CN202211063683 A CN 202211063683A CN 115424614 A CN115424614 A CN 115424614A
Authority
CN
China
Prior art keywords
instruction
event
dimension
interactive
target
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202211063683.6A
Other languages
Chinese (zh)
Inventor
吴爱勇
聂春音
王瑛玮
王津生
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Great Wall Motor Co Ltd
Original Assignee
Great Wall Motor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Great Wall Motor Co Ltd filed Critical Great Wall Motor Co Ltd
Priority to CN202211063683.6A priority Critical patent/CN115424614A/en
Publication of CN115424614A publication Critical patent/CN115424614A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/22Interactive procedures; Man-machine interfaces
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention provides a human-computer interaction method and device, electronic equipment and a vehicle, and relates to the technical field of human-computer interaction. Firstly, acquiring interactive instructions of different dimensions input by a user and generating an identification result of the interactive instruction of each dimension; then, determining an event instruction of each dimension according to the identification result of the interactive instruction of each dimension; and finally, integrating the event instructions of each dimension according to the preset control logic and the priority to generate a fusion event instruction, and executing a target control action according to the fusion event instruction. In the invention, the user can input in various modes without limiting the input mode of the user, and fusion output can be performed in various modes, so that various interactive experiences are brought to the user, and the scene perception level of the vehicle and the human-computer interactive experience of the user are improved.

Description

Human-computer interaction method and device, electronic equipment and vehicle
Technical Field
The invention relates to the technical field of human-computer interaction, in particular to a human-computer interaction method and device, electronic equipment and a vehicle.
Background
The HMI (Human Machine Interface) is a product of development of information technology, and the system realizes a conversation function between a person and a vehicle. The man-machine interaction mode is also expanded from the character interface interaction to the graphic interface interaction, gesture recognition, voice interaction, brain-machine interface and the like. The current user interface is often presented in a GUI (graphical user interface) and VUI (voice user interface) form, and with the increase of the machine intelligence level, the VUI (voice user interface) is gradually developed, but the GUI is still the mainstream experience medium of the current human-computer interaction form.
In the related art, the existing interaction modes have respective disadvantages, and the two interaction interfaces have lower scene perception levels, so that the requirements of users cannot be met.
Disclosure of Invention
The embodiment of the invention provides a human-computer interaction method, a human-computer interaction device, electronic equipment and a vehicle, and aims to solve or partially solve the problems in the background art.
In order to solve the technical problem, the invention is realized as follows:
in a first aspect, an embodiment of the present invention provides a human-computer interaction method, where the method includes:
acquiring interactive instructions with different dimensions input by a user;
generating an identification result of the interactive instruction of each dimension;
determining an event instruction of each dimension according to the identification result of the interactive instruction of each dimension;
integrating the event instructions of each dimension according to a preset control logic and a priority level to generate a fusion event instruction;
and executing the target control action according to the fusion event instruction.
Optionally, the step of generating the recognition result of the interactive instruction of each dimension includes:
determining a target identification mode corresponding to the interactive instruction of each dimension;
and respectively obtaining the identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
Optionally, the step of determining the event instruction of each dimension according to the recognition result of the interactive instruction of each dimension includes:
determining a target event source corresponding to the identification result of the interactive instruction of each dimension;
and traversing the mapping dictionary by taking the target event source as an index, and determining the event instruction matched with the target event source.
Optionally, the mapping dictionary is obtained by:
acquiring the hardware support capability of a vehicle, and constructing a mapping relation between an event source and an event instruction according to the hardware support capability of the vehicle;
and storing the mapping relation between the event source and the event instruction in the mapping dictionary.
Optionally, according to the fusion event instruction, the step of executing the target control action includes:
determining a target control program fusing event instructions;
determining target application logic executed in a target control program;
and controlling the physical equipment according to the target application logic.
In a second aspect, an embodiment of the present invention provides a human-computer interaction device, where the device includes:
the acquisition module is used for acquiring interactive instructions with different dimensions input by a user;
the identification module is used for generating an identification result of the interactive instruction of each dimension;
the matching module is used for determining the event instruction of each dimension according to the identification result of the interactive instruction of each dimension;
the fusion module is used for integrating the event instructions of each dimension according to preset control logic and priority to generate fusion event instructions;
and the execution module is used for executing the target control action according to the fusion event instruction.
Optionally, the identification module comprises:
the first identification submodule is used for determining a target identification mode corresponding to the interactive instruction of each dimension;
and the second identification submodule is used for respectively obtaining an identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
Optionally, the matching module comprises:
the determining submodule is used for determining a target event source corresponding to the identification result of the interactive instruction of each dimension;
and the searching submodule is used for traversing the mapping dictionary by taking the target event source as an index and determining the event instruction matched with the target event source.
A third aspect of an embodiment of the present invention provides an electronic device, including:
at least one processor; and a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to cause the at least one processor to perform method steps set forth in the first aspect of embodiments of the invention.
A fourth aspect of an embodiment of the present invention provides a vehicle comprising a processor configured to carry out the method steps as set forth in the first aspect of an embodiment of the present invention when executed.
The embodiment of the invention has the following advantages: firstly, acquiring interactive instructions of different dimensions input by a user and generating an identification result of the interactive instruction of each dimension; then, determining an event instruction of each dimension according to the identification result of the interactive instruction of each dimension; and finally, integrating the event instructions of each dimension according to the preset control logic and the priority to generate a fusion event instruction, and executing the target control action according to the fusion event instruction. In the invention, the user can input in various modes without limiting the input mode of the user, and fusion output can be performed in various modes, so that various interactive experiences are brought to the user, and the scene perception level of the vehicle and the human-computer interactive experience of the user are improved.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings used in the embodiments or the prior art descriptions will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present application, and other drawings can be obtained by those skilled in the art without creative efforts.
FIG. 1 is a flow chart illustrating steps of a human-computer interaction method according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a converged interaction management layer in an embodiment of the invention;
FIG. 3 is a diagram illustrating a converged display management layer in an embodiment of the present invention;
fig. 4 is a schematic block diagram of a human-computer interaction device according to an embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some, but not all, of the embodiments of the present application. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
The basis of human-computer interaction is to realize the fundamental requirements of users. HMI (human machine interface) from text interface interaction to graphical interface interaction, gesture recognition, voice interaction, brain-computer interface. The man-machine interaction gradually expands from the original graphical interface interaction to the directions of voice interaction, gesture recognition, brain-computer interface and the like. Although the use of these interactive modes makes people's lives more convenient, each of the above interactive modes has a considerable disadvantage. The mainstream human-computer interaction mode is mainly graphics interface interaction and assisted by voice interaction. Neither is the man-machine interaction mode perfect, nor is it useful.
For the interaction mode of the graphical interface, the defects are as follows: the visual system of the person and the electronic screen of the equipment are excessively depended on, the interaction mode is single, and fatigue is easily caused. It is difficult for people with impaired vision and limbs to operate. The application is poor in an immersive virtual environment such as AR or VR.
For the interactive mode of voice interaction, the disadvantages exist: the linear input mode is adopted, so that more content cannot be continuously input and output, information identification is easily influenced by the environment, identification is blocked when a user is in a noisy environment, and voice interaction in a public environment has psychological burden.
For the gesture interaction, the interaction mode has the following disadvantages: the hands of a person are difficult to stay in the air for a long time without being supported by foreign objects. And the habit of human beings in long-term natural social contact can not use gesture interaction in a large amount in public occasions such as outdoor, office, family environment of many people living together, and the like, and the touch feedback experience is also lacked.
For the brain-computer interaction mode, the defects are as follows: the low prevalence rate of brain-computer interface devices is present. The second is that the user acceptance is unknown.
Meanwhile, currently, in a human-computer interaction process, two types of interaction interfaces mainly exist, one is GUI (graphical user interface) morphological presentation, and the other is VUI (voice user interface). The two types of interactive interfaces present user interface presentation forms and features as shown in table 1.
Table 1: GUI and VUI user interface presentation forms and features
Figure BDA0003827316830000041
Figure BDA0003827316830000051
Therefore, at present, various interaction modes have respective defects, and the existing user interaction interfaces often support single-dimension interaction. Therefore, the level of the context awareness is low, the user experience is also poor, the vehicle cannot actively serve people, and the man-machine compatibility (MMC) effect cannot be achieved.
Based on this, the inventive concept of the present application is proposed: the multi-mode interactive form is formed by combining various interactive modes, and the machine can actively serve people through context perception, so that human-machine compatibility is achieved.
Firstly, the application provides a Natural User Interface (NUI) applied to a vehicle-mounted machine system, compared with GUI and VUI, the NUI is a more Natural output which integrates multi-sensory experience of various interactive devices, is more intuitive, has no strong dependence and tends to be screenless (a screen is integrated into an environment).
For NUI, the content of its inputs may be: the input mode can be a more natural mode, such as 'gesture', 'voice', 'vision', 'gesture', 'perception', and the like.
The process of a user interacting with a vehicle based on NUI may be divided into two parts. The first part is to identify the intention of a user for a vehicle, an embodiment of the present invention provides a human-computer interaction method, and referring to fig. 1, fig. 1 shows a flowchart of steps of a human-computer interaction method according to an embodiment of the present invention, where the method includes:
s101: and acquiring interactive instructions of different dimensions input by a user.
In this embodiment, the dimension may be understood as an input mode, that is, a voice dimension input through the voice interaction service, a key dimension input through the key interaction service, a touch dimension input through the touch interaction service, and a somatosensory dimension input through the somatosensory interaction service. The implementation of the above dimension depends on the physical device layer of the vehicle, i.e. the kind of the physical devices supported by the physical device layer of the vehicle, and determines the number of input dimensions of the user.
As an example, a user may perform input of interaction instructions with different dimensions through a device physical layer of the intelligent fusion interaction system, for example, input an interaction instruction with a voice dimension through a microphone of a vehicle, input an interaction instruction with a key dimension through a key of the vehicle, input an interaction instruction with a touch dimension through a touch device of the vehicle, and input an interaction instruction with a somatosensory dimension through a somatosensory device such as a vehicle-mounted camera. When inputting, the user may input only one-dimensional interactive instruction, or may input interactive instructions of multiple dimensions.
S102: and generating an identification result of the interactive instruction of each dimension.
In this embodiment, the interactive instructions with different dimensions include an interactive instruction with a voice dimension, an interactive instruction with a key dimension, an interactive instruction with a touch dimension, and an interactive instruction with a somatosensory dimension. The specific steps of obtaining the recognition result of the interactive instruction of each dimension may be:
s102-1: and determining a target identification mode corresponding to the interactive instruction of each dimension.
In this embodiment, the vehicle-mounted machine system of the vehicle is not a human, so in order to make it possible to understand what the user input is specifically expressed, it is necessary to convert the interactive instructions of different dimensions input by the user into a language that the machine can understand, and this process is performed depending on the selected target recognition mode.
As an example, for the interactive instruction of the voice dimension, the schematic diagram of the converged interaction management layer shown in fig. 2 includes: the system comprises a physical device layer, an interactive lexical layer, an interactive grammar layer, an interactive semantic layer and an application interface layer. After an interactive instruction obtained by a voice dimension of a physical device layer is input into an interactive lexical layer, a corresponding target identification mode is determined to be voice identification, after the interactive instruction of a key dimension is input into the interactive lexical layer, the corresponding target identification mode is determined to be key identification, after the interactive instruction of a touch dimension is input into the interactive lexical layer, the corresponding target identification mode is determined to be touch identification, and after the interactive instruction of a somatosensory dimension is input into the interactive lexical layer, the corresponding target identification mode is determined to be somatosensory identification.
S102-2: and respectively obtaining the identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
In the present embodiment: after the target recognition mode corresponding to the interactive instruction of each dimension is determined, the target recognition mode is input into the interactive grammar layer, and then the recognition result of each dimension can be obtained. For the interactive instruction of the voice dimension, the corresponding recognition result is the voice content, for the interactive instruction of the key dimension, the corresponding recognition result is the key content, for the interactive instruction of the touch dimension, the corresponding recognition result is the touch event, and for the somatosensory instruction of the voice dimension, the corresponding recognition result is the somatosensory gesture.
As an example, for the interactive instruction of the voice dimension, the recognized voice content may be: "turn on the air conditioner", for the interactive instruction of the key dimension, the key content obtained by its recognition may be: "turn off the air conditioner", for the interactive instruction of the touch dimension, the recognized touch event may be: for an interactive instruction of a somatosensory dimension, a somatosensory gesture recognized by the "birthday celebration event" may be: "open window".
S103: and determining the event instruction of each dimension according to the identification result of the interactive instruction of each dimension.
In the embodiment, after the vehicle acquires the identification result of each dimension, it can be understood that the vehicle acquires the real intention of the user, that is, the vehicle understands what the user wants to do, and therefore a corresponding response needs to be made to the real intention of the user. The method comprises the following specific steps:
s103-1: and determining a target event source corresponding to the identification result of the interactive instruction of each dimension.
In the embodiment, after the vehicle acquires the identification result of each dimension, the identification result of each dimension is input into the fusion interaction semantic layer, which is a centralized event processing mechanism and allows different interaction components to communicate with each other without mutual dependence. The fusion interaction semantic layer adopts an event bus mode and mainly comprises three parts: events, event buses, and event listeners. The event is an instruction issued by each interactive component of the interactive semantic layer to the fused interactive semantic layer; the event bus is an 'infrastructure' for realizing event instruction propagation and is an event 'scheduling center' of the whole fusion interaction semantic layer; the event listener is used for listening to objects of various interactive instruction events.
The centralized event processing mechanism fusing the interactive semantic layers mainly realizes that an event instruction sender sends an event to an event bus, and an event instruction subscriber subscribes and receives the event to the event bus and then processes the received event in an event-driven mode.
As an example, in the application, the recognition result of each dimension is an instruction issued by each interactive component of the interactive semantic layer to the fused interactive semantic layer, an event instruction sender is a user who inputs interactive instructions of different dimensions, an event instruction subscriber is various application programs on a vehicle, and different application programs are controlled to work according to the interactive instructions of the user.
Therefore, the interactive instruction input by the user is of various types, such as: similarly, the interactive instructions are interactive instructions with voice dimensions, and the corresponding recognition results can be 'turn on the air conditioner', 'please lower the temperature in the vehicle', but the user wants the vehicle to do something is 'start the air conditioner', so that the 'start the air conditioner' can be understood as the target event source corresponding to the recognition result of the interactive instructions with voice dimensions.
S103-2: and traversing the preset mapping dictionary by taking the target event source as an index, and determining the event instruction matched with the target event source.
In this embodiment, continuing with the above example, if it is determined that the target event source corresponding to the recognition result of the interactive command with the voice dimension is "start air conditioner". And searching in a mapping dictionary, wherein the mapping dictionary is used for storing the mapping relation between the target event source and the event instruction, and the mapping relation between the target event source and the event instruction needs to be edited in advance and then is prestored in the mapping dictionary.
As an example, after determining that the target event source is "start air conditioner", searching in the mapping dictionary, then according to a pre-stored mapping relationship, determining that the event instruction corresponding to the target event source being "start air conditioner" is "start power supply of the air conditioner controller, and sending a control instruction for adjusting the air conditioner operation mode to the cooling mode to the air conditioner controller. "for the recognition results of different interactive instructions, if the corresponding target event sources are the same, the corresponding event instructions are the same, and if there are a plurality of target event sources corresponding to the recognition results of different interactive instructions, there are a plurality of corresponding event instructions.
S104: and integrating the event instructions of each dimension according to preset control logic and priority to generate a fusion event instruction.
In this embodiment, after the event instructions of each dimension are obtained, since the event instructions of each dimension are independent and scattered, they need to be integrated, so as to implement a complete control logic. When the integration is performed, the adjustment can be performed according to the priority of the calling application program. In the integration, the adjustment can be performed according to the control logic of the calling application program.
As an example, if the event instruction of the voice dimension is "start the power supply of the air conditioner controller, and send a control instruction for adjusting the air conditioner operation mode to the cooling mode to the air conditioner controller. And if the event instruction of the gesture dimension is 'starting the power supply of the seat controller', and sending a seat intersection angle adjusting control instruction to the seat controller. "therefore, the target program corresponding to the calling is the seat control program, and the priority of the seat control program is lower than that of the air conditioning control program. Thus, when the integration is performed, the event instruction of the gesture dimension is executed before the event instruction of the voice dimension.
As an example, if the event instruction of the voice dimension is "start the power supply of the air conditioner controller, and send a control instruction for adjusting the air conditioner operation mode to the cooling mode to the air conditioner controller. And if the event instruction of the gesture dimension is 'sending a control instruction for adjusting the air conditioner temperature to xx temperature to the air conditioner controller'. "so its corresponding control logic is an adjustment to the temperature. When the air conditioner is controlled, the common control logic is to start the air conditioner and then adjust the temperature, so that the event instruction of the gesture dimension is executed after the event instruction of the voice dimension when the integration is performed.
Therefore, according to the integration logic, the event instructions of each dimension are integrated, so that a plurality of dimensions of fused event instructions are generated. Based on this, the task of the first part of vehicles recognizing the user's intention is completed.
S105: and executing the target control action according to the fusion event instruction.
In this embodiment, after the content of the first part of the human-computer interaction is completed, that is, after the fusion event instruction is generated in the fusion interaction semantic layer, the fusion event instruction needs to be sent to the corresponding control program, and the target control result is achieved, and the specific steps may be:
s105-1: determining a target control program fusing event instructions;
s105-2: determining target application logic executed in a target control program;
s105-3: and controlling the physical equipment according to the target application logic.
In the embodiments of S105-1 to S105-3, the target control program fused with the event instruction is the target control program determined by the event instruction of each dimension.
As an example, a schematic diagram of a converged display management layer as shown in fig. 3 includes: a display semantic layer, a display grammar layer and a physical device layer. If the target control program corresponding to the fusion event instruction A comprises: a fragrance control program, a seat control program, an ambience light control program, and a voice system program. After determining the target control program, inputting the fusion event command to the fusion display semantic layer to obtain a specific control command, such as a display command, an atmosphere lamp display command, and the like, and then inputting the specific control command to the display syntax layer to obtain a specific control logic for the target control program, for example, for the atmosphere lamp control program, the target application logic may be: "light control sequence of atmosphere lamp", for the speech system program, its target application logic may be: "adopt the sound effect of xx, broadcast and congratulate on xx birthday happy". Then, in the physical device, firstly, turning on a power supply of the atmosphere lamp, then controlling the atmosphere lamp according to an atmosphere lamp control time sequence corresponding to the target application logic, and simultaneously playing through the sound equipment of the vehicle: "congratulate on birthday xx" voice content.
In one possible embodiment, the mapping dictionary is obtained by:
acquiring the hardware support capability of the vehicle, and constructing a mapping relation between an event source and an event instruction according to the hardware support capability of the vehicle;
and storing the mapping relation between the event source and the event instruction in the mapping dictionary.
In the present embodiment, before the mapping dictionary is constructed, the hardware support capability of the vehicle needs to be acquired, where the hardware support capability refers to the capability supported by the hardware device of the vehicle, and as an example, if the vehicle supports the smart fragrance function, an event instruction of "turn on the smart fragrance" may be added when the event source is "birthday celebration". Thus, the dimensionality of the event instructions output by the vehicle also depends on the number and variety of hardware devices included in the vehicle. And the user can construct the mapping relation between the event source and the event instruction according to the requirement of the user.
The solution of the present application is described below in a complete embodiment, where the user wants to celebrate a wedding anniversary. Firstly, a user inputs through gestures and keys to generate a plurality of dimensionality interactive instructions, a vehicle identifies the key input and the gesture input of the user, a target event source corresponding to an identification result is determined to be 'celebration wedding anniversary', then a mapping dictionary is searched, a corresponding event instruction is determined to be 'turning on an atmosphere lamp firstly, then turning on a fragrance system, and finally playing voice content input by the user in advance'. And then, issuing the time instruction to each application program of the vehicle, and then executing the action instruction by each application program. Therefore, the input of the user is responded through the output of a plurality of dimensions, the user is provided with all-round experience, the use physical examination of the user is improved, the user can realize complex feedback only through simple operation, and the intelligent level of the intelligent machine is improved.
An embodiment of the present invention further provides a human-computer interaction device, and referring to fig. 4, a functional block diagram of the human-computer interaction device of the present invention is shown, where the device may include the following modules:
an obtaining module 401, configured to obtain interaction instructions with different dimensions input by a user;
an identification module 402, configured to generate an identification result of the interactive instruction for each dimension;
a matching module 403, configured to determine an event instruction for each dimension according to an identification result of the interactive instruction for each dimension;
the fusion module 404 is configured to integrate the event instructions of each dimension according to a preset control logic and a priority, and generate a fusion event instruction;
and the execution module 405 is configured to execute the target control action according to the fusion event instruction.
In one possible implementation, the identification module 402 includes:
the first identification submodule is used for determining a target identification mode corresponding to the interactive instruction of each dimension;
and the second identification submodule is used for respectively obtaining an identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
In one possible implementation, the matching module 403 includes:
the determining submodule is used for determining a target event source corresponding to the identification result of the interactive instruction of each dimension;
and the searching submodule is used for traversing the mapping dictionary by taking the target event source as an index and determining the event instruction matched with the target event source.
In one possible implementation, the execution module 405 includes:
the first determining submodule is used for determining a target control program of the fusion event instruction;
a second determination submodule for determining a target application logic executed in the target control program;
and the control sub-module is used for controlling the physical equipment according to the target application logic.
In a further embodiment provided by the present invention, there is also provided a vehicle comprising a processor configured to implement the method as set forth in the first aspect of the embodiments of the present invention when executed.
Based on the same inventive concept, another embodiment of the present invention provides an electronic device, which includes a processor, a communication interface, a memory, and a communication bus, wherein the processor, the communication interface, and the memory communicate with each other via the communication bus,
a memory for storing a computer program;
and the processor is used for realizing the man-machine interaction method when executing the program stored in the memory.
The communication bus mentioned in the above terminal may be a Peripheral Component Interconnect (PCI) bus or an Extended Industry Standard Architecture (EISA) bus. The communication bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown, but this does not mean that there is only one bus or one type of bus. The communication interface is used for communication between the terminal and other devices. The Memory may include a Random Access Memory (RAM) or a non-volatile Memory (non-volatile Memory), such as at least one disk Memory. Optionally, the memory may also be at least one memory system located remotely from the processor.
The Processor may be a general-purpose Processor, including a Central Processing Unit (CPU), a Network Processor (NP), and the like; the Integrated Circuit may also be a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other Programmable logic device, a discrete Gate or transistor logic device, or a discrete hardware component.
In addition, in order to achieve the above object, an embodiment of the present application further provides a computer-readable storage medium, which stores a computer program, and when the computer program is executed by a processor, the computer program implements the human-computer interaction method according to the embodiment of the present application.
As will be appreciated by one skilled in the art, embodiments of the present invention may be provided as a method, system, or computer program product. Accordingly, embodiments of the present invention may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, embodiments of the present invention may take the form of a computer program product on one or more computer-usable vehicles having computer-usable program code embodied therein (including, but not limited to, disk storage, CD-ROM, optical storage, and the like).
Embodiments of the present invention are described with reference to flowchart illustrations and/or block diagrams of methods, terminal devices (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing terminal to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing terminal, create a system for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including an instruction system which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing terminal to cause a series of operational steps to be performed on the computer or other programmable terminal to produce a computer implemented process such that the instructions which execute on the computer or other programmable terminal provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
Finally, it should also be noted that, herein, relational terms such as first and second, and the like may be used solely to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. "and/or" means that either or both of them can be selected. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or terminal that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or terminal. Without further limitation, an element defined by the phrases "comprising one of \ 8230; \8230;" does not exclude the presence of additional like elements in a process, method, article, or terminal device that comprises the element.
The human-computer interaction method, the human-computer interaction device, the electronic equipment and the vehicle provided by the invention are described in detail, specific examples are applied in the description to explain the principle and the implementation mode of the invention, and the description of the examples is only used for helping to understand the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (10)

1. A human-computer interaction method, the method comprising:
acquiring interactive instructions with different dimensions input by a user;
generating an identification result of the interactive instruction of each dimension;
determining an event instruction of each dimension according to the identification result of the interactive instruction of each dimension;
integrating the event instructions of each dimension according to a preset control logic and a priority level to generate a fusion event instruction;
and executing a target control action according to the fusion event instruction.
2. The human-computer interaction method according to claim 1, wherein the step of generating the recognition result of the interaction instruction for each dimension comprises:
determining a target identification mode corresponding to the interactive instruction of each dimension;
and respectively obtaining the identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
3. The human-computer interaction method according to claim 2, wherein the step of determining the event command for each dimension according to the recognition result of the interaction command for each dimension comprises:
determining a target event source corresponding to the identification result of the interactive instruction of each dimension;
and traversing a mapping dictionary by taking the target event source as an index, and determining the event instruction matched with the target event source.
4. A human-computer interaction method according to claim 3, characterized in that said mapping dictionary is obtained by:
acquiring the hardware support capability of a vehicle, and constructing a mapping relation between an event source and an event instruction according to the hardware support capability of the vehicle;
and storing the mapping relation between the event source and the event instruction in a mapping dictionary.
5. The human-computer interaction method according to claim 1, wherein the step of executing a target control action according to the fusion event instruction comprises:
determining a target control program of the fusion event instruction;
determining a target application logic executed in the target control program;
and controlling the physical equipment according to the target application logic.
6. A human-computer interaction device, characterized in that the device comprises:
the acquisition module is used for acquiring interactive instructions with different dimensions input by a user;
the identification module is used for generating an identification result of the interactive instruction of each dimension;
the matching module is used for determining the event instruction of each dimension according to the identification result of the interactive instruction of each dimension;
the fusion module is used for integrating the event instructions of each dimension according to preset control logic and priority to generate fusion event instructions;
and the execution module is used for executing the target control action according to the fusion event instruction.
7. The human-computer interaction device of claim 6, wherein the recognition module comprises:
the first identification submodule is used for determining a target identification mode corresponding to the interactive instruction of each dimension;
and the second identification submodule is used for respectively obtaining the identification result corresponding to the interactive instruction of each dimension according to the target identification mode.
8. The human-computer interaction device of claim 6, wherein the matching module comprises:
the determining submodule is used for determining a target event source corresponding to the identification result of the interactive instruction of each dimension;
and the searching submodule is used for traversing the mapping dictionary by taking the target event source as an index and determining the event instruction matched with the target event source.
9. An electronic device is characterized by comprising a processor, a communication interface, a memory and a communication bus, wherein the processor and the communication interface are used for realizing mutual communication by the memory through the communication bus;
a memory for storing a computer program;
a processor for implementing the human-computer interaction method according to any one of claims 1 to 5 when executing the program stored in the memory.
10. A vehicle comprising a processor and a memory, the memory storing machine executable instructions executable by the processor, the processor being configured to execute the machine executable instructions to implement the human-computer interaction method of any one of claims 1 to 5.
CN202211063683.6A 2022-08-31 2022-08-31 Human-computer interaction method and device, electronic equipment and vehicle Pending CN115424614A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211063683.6A CN115424614A (en) 2022-08-31 2022-08-31 Human-computer interaction method and device, electronic equipment and vehicle

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211063683.6A CN115424614A (en) 2022-08-31 2022-08-31 Human-computer interaction method and device, electronic equipment and vehicle

Publications (1)

Publication Number Publication Date
CN115424614A true CN115424614A (en) 2022-12-02

Family

ID=84199564

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211063683.6A Pending CN115424614A (en) 2022-08-31 2022-08-31 Human-computer interaction method and device, electronic equipment and vehicle

Country Status (1)

Country Link
CN (1) CN115424614A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116434027A (en) * 2023-06-12 2023-07-14 深圳星寻科技有限公司 Artificial intelligent interaction system based on image recognition

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116434027A (en) * 2023-06-12 2023-07-14 深圳星寻科技有限公司 Artificial intelligent interaction system based on image recognition

Similar Documents

Publication Publication Date Title
JP7198861B2 (en) Intelligent assistant for home automation
JP7354301B2 (en) Detection and/or registration of hot commands to trigger response actions by automated assistants
US9953648B2 (en) Electronic device and method for controlling the same
JP7322076B2 (en) Dynamic and/or context-specific hotwords to launch automated assistants
US20190027129A1 (en) Method, apparatus, device and storage medium for switching voice role
CN112868060B (en) Multimodal interactions between users, automated assistants, and other computing services
JP7341171B2 (en) Dynamic and/or context-specific hotwords to invoke automated assistants
KR102411619B1 (en) Electronic apparatus and the controlling method thereof
JP2023120205A (en) Supplementing voice inputs to automated assistant according to selected suggestions
KR102581347B1 (en) Efficient and low latency automated assistant control of smart devices
EP2933607A1 (en) Navigation system having language category self-adaptive function and method of controlling the system
CN110459222A (en) Sound control method, phonetic controller and terminal device
KR102581348B1 (en) Recommending automated assistant action for inclusion in automated assistant routine
JP7051799B2 (en) Speech recognition control methods, devices, electronic devices and readable storage media
KR102358012B1 (en) Speech control method and apparatus, electronic device, and readable storage medium
CN115424614A (en) Human-computer interaction method and device, electronic equipment and vehicle
CN115547332A (en) Sight attention-based awakening-free intention recall method and system and vehicle
JP2019091444A (en) Processing method of smart interactive, device, facility and computer storage medium
TWI847393B (en) Language data processing system and method and computer program product
CN113241067B (en) Voice interaction method and system and voice interaction equipment
JP2018181250A (en) Device, program, and method for generating interaction scenario corresponding to context
KR20240054021A (en) Electronic device capable of proposing behavioral patterns for each situation and control method therefor
KR101780464B1 (en) Chatting method, chatting server and chatting system for learning language
JP2024077568A (en) Language data processing system, language data processing method, and computer program
CN117912460A (en) Intelligent device control method based on semantic recognition and related device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination