CN114979543A - Intelligent household control method and device - Google Patents
Intelligent household control method and device Download PDFInfo
- Publication number
- CN114979543A CN114979543A CN202110204624.5A CN202110204624A CN114979543A CN 114979543 A CN114979543 A CN 114979543A CN 202110204624 A CN202110204624 A CN 202110204624A CN 114979543 A CN114979543 A CN 114979543A
- Authority
- CN
- China
- Prior art keywords
- user
- terminal
- control
- voiceprint
- call request
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 61
- 238000004891 communication Methods 0.000 claims abstract description 35
- 238000012795 verification Methods 0.000 claims description 13
- 238000004590 computer program Methods 0.000 claims description 12
- 230000007774 longterm Effects 0.000 claims description 2
- 230000008569 process Effects 0.000 abstract description 9
- 230000005540 biological transmission Effects 0.000 abstract description 3
- 230000006870 function Effects 0.000 description 13
- 238000012545 processing Methods 0.000 description 11
- 238000005516 engineering process Methods 0.000 description 8
- 238000010586 diagram Methods 0.000 description 6
- 230000008878 coupling Effects 0.000 description 5
- 238000010168 coupling process Methods 0.000 description 5
- 238000005859 coupling reaction Methods 0.000 description 5
- 239000000203 mixture Substances 0.000 description 4
- 230000004044 response Effects 0.000 description 4
- 238000013507 mapping Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 238000012937 correction Methods 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 238000010295 mobile communication Methods 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 239000000758 substrate Substances 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- 240000007594 Oryza sativa Species 0.000 description 1
- 235000007164 Oryza sativa Nutrition 0.000 description 1
- 238000001604 Rao's score test Methods 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 235000009566 rice Nutrition 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 239000007787 solid Substances 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000013518 transcription Methods 0.000 description 1
- 230000035897 transcription Effects 0.000 description 1
- 238000005406 washing Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/14—Systems for two-way working
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/06—Decision making techniques; Pattern matching strategies
- G10L17/14—Use of phonemic categorisation or speech recognition prior to speaker recognition or verification
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02P—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
- Y02P90/00—Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
- Y02P90/02—Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Game Theory and Decision Science (AREA)
- Business, Economics & Management (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Telephonic Communication Services (AREA)
Abstract
The application provides an intelligent home control method and device. The method comprises the following steps: the method comprises the steps that a control end answers a call request from a terminal, wherein the call request is sent by a user of the terminal through dialing a preset number, and the call request is used for requesting remote control of the intelligent home; the control terminal responds to the call request and calls a pre-configured virtual digital person to carry out video call with the user of the terminal; and acquiring a control instruction of the user to the intelligent home based on the video call. The call request can be carried on resources of an LTE communication system or a 5G communication system. Due to the fact that the dialing process based on the standard of the terminal is based on the LTE standard or the 5G standard, the control of the user on the intelligent home can be finished without downloading special application, and the occupation of the memory of the terminal can be reduced; and the call is borne in LTE or 5G resources, so that the transmission of information such as identity authentication, instructions and the like is safer and more reliable.
Description
Technical Field
The application relates to the technical field of mobile communication, in particular to an intelligent home control method and device.
Background
The intelligent home connects various devices (such as lighting devices, curtains, air conditioners, security systems, digital cinema systems, audio and video servers, network household appliances and the like) in the home through the Internet of things technology, and provides various control functions and means. Currently, the main control means include remote network control, local control, timing control, and one-key scenario control.
However, the existing remote network control needs to download and install special software provided by the smart home manufacturer, such as a special Application (APP), so as to perform the relevant operations. With the increase of the types of smart homes, corresponding special software is increased, which may occupy a large storage space of a user terminal.
Disclosure of Invention
The application provides an intelligent home control method and device, so that remote control over an intelligent home can be completed without downloading an APP, and occupation of a memory of a user terminal is reduced.
In a first aspect, the present application provides a smart home control method, including: receiving a call request from a terminal, wherein the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home; responding to the call request, and calling a pre-configured virtual digital person to carry out video call with a user of the terminal; and acquiring a control instruction of the user to the intelligent home based on the video call.
Based on the scheme, the call request which is sent by the user by dialing the preset number and is based on the self-contained standard (such as the LTE standard or the 5G standard) of the terminal is responded, the pre-configured virtual digital person is called to carry out video call with the user, so that the control instruction of the user to the intelligent home is obtained, the control of the user to the intelligent home can be completed without downloading a special application at the terminal, and the occupation of the memory of the user terminal is reduced. And because the complicated operation interface of APP need not to adopt, the user operation of more making things convenient for.
Optionally, before the obtaining of the control instruction of the user to the smart home based on the video call, the method further includes: prompting the user to input an identification through the virtual digital person; determining whether the user is a registered user based on the identity; under the condition that the user is a registered user, performing identity authentication on the user based on the voiceprint and the face of the user; or prompting the user to perform voiceprint registration and face registration under the condition that the user is an unregistered user.
Optionally, the identity is a mobile phone number of the user.
Optionally, the authenticating the user includes: collecting voiceprints and faces of the user based on the video call; and carrying out identity verification on the user according to the voiceprint recognition, the face recognition, the prestored voiceprint and face of the user.
Optionally, the method further comprises: inquiring the home address of the user and the product model of the smart home through the virtual digital person; acquiring a voiceprint and a face of the user based on the video call; and storing the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
Optionally, the obtaining a control instruction of the user to the smart home based on the video call includes: acquiring a voice instruction of the user based on the video call; and carrying out voice recognition on the voice command so as to determine the smart home which the user wants to control and the control operation of the smart home.
Optionally, the method further comprises: executing the control instruction; and after the control instruction is finished, feeding back the finishing condition of the control instruction to the user through the virtual digital person.
Optionally, the call request and the video callThe speech is carried over resources of a Long Term Evolution (LTE) communication system or fifth generation (5) th generation, 5G) resources of the communication system.
In a second aspect, the present application provides an intelligent home control apparatus, including a unit for implementing the intelligent home control method of the first aspect.
In a third aspect, the present application provides an intelligent home control device, including a processor, where the processor is configured to execute the intelligent home control method according to the first aspect.
In a fourth aspect, there is provided a computer-readable storage medium comprising instructions which, when executed on a computer, cause the computer to carry out the method of any one of the first aspect and the first aspect.
In a fifth aspect, there is provided a computer program product comprising: a computer program (also referred to as code, or instructions), which when executed, causes a computer to perform the method of the first aspect as well as any one of the possible implementations of the first aspect.
It should be understood that the second aspect to the fifth aspect of the present application correspond to the technical solutions of the first aspect of the present application, and the beneficial effects achieved by the aspects and the corresponding possible implementations are similar and will not be described again.
Drawings
Fig. 1 is a network architecture suitable for an intelligent home control method provided in the present application;
fig. 2 is a schematic flow chart of an intelligent home control method provided in the embodiment of the present application;
fig. 3 is a schematic block diagram of an intelligent home control system provided in an embodiment of the present application;
fig. 4 is a schematic block diagram of an intelligent home control device provided in an embodiment of the present application;
fig. 5 is another schematic block diagram of an intelligent home control device provided in an embodiment of the present application.
Detailed Description
The technical solution in the present application will be described below with reference to the accompanying drawings.
The intelligent home is characterized in that a home is used as a platform, facilities related to home life are integrated by utilizing a comprehensive wiring technology, a network communication technology, a safety precaution technology, an automatic control technology and an audio and video technology, an efficient home facility and a home schedule error management system are constructed, home safety, convenience, comfortableness and artistry are improved, and an environment-friendly and energy-saving living environment is realized. Compared with the common home, the intelligent home has the traditional living function, integrates the functions of building, network communication, information household appliance and equipment automation, provides an all-around information interaction function, and even saves funds for various energy expenses.
The technical scheme of the embodiment of the application can be used for controlling various intelligent household devices. The smart home may include, for example, but is not limited to, smart light fixtures, smart washing machines, smart refrigerators, smart rice cookers, smart door locks, smart curtains, central dust extraction systems, smart background music systems, smart video systems, digital cinema systems, smart security alarms, and smart surveillance systems, etc.
The control of smart homes may include, for example, but is not limited to, the control of various smart home devices listed above. For the sake of brevity, this is not to be enumerated here.
To facilitate understanding of the embodiment of the present application, first, a network architecture suitable for the smart home control method provided in the present application is described in detail with reference to fig. 1. As shown in fig. 1, the network architecture 100 may include a terminal 110, a control end 120, and smart home devices 131, 132, and 133. The terminal 110 and the control terminal 120 may communicate with each other through a global system for mobile communications (GSM), and the control terminal 120 may be configured to control the smart home devices 131 to 133.
It should be understood that, in the embodiment of the present application, the terminal 110 may be, for example, a mobile phone or a wearable device; the control end 120 may be, for example, an intelligent home control device having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself, or a control device integrated on an intelligent home having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself. The embodiments of the present application do not limit this. The control end 120 may be connected to a plurality of terminals 110, and the number of the terminals is not limited in the embodiment of the present application.
An intelligent home control method provided by the embodiment of the present application will be described in detail below with reference to fig. 2. As shown in fig. 2, the method 200 may include steps 201 through 214.
In step 201, the control terminal receives a call request from the terminal.
Optionally, the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home. The preset number may be set on the control terminal for connecting the control terminal. The control terminal may be, for example, an intelligent home control device having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself, or a control device integrated on an intelligent home having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself. The embodiments of the present application do not limit this.
Alternatively, the user of the terminal may be a family member or other user who wants to control the smart home.
It should be understood that in the embodiment of the present application, the terminal, the user of the terminal, and the user are used alternately, and the three meanings are the same.
Alternatively, the call request may be to use the fourth generation (4) th generation, 4G) terminal or 5G terminal makes a call. The 4G terminal is also a terminal supporting 4G, or a terminal supporting LTE. A 5G terminal is also a terminal supporting 5G. The call request can be carried on resources of an LTE communication system or resources of a 5G communication system. For example, the call request may be transmitted via a voice over LTE (VoLTE) protocol.
It should be noted that the VoLTE protocol is mainly a high-speed wireless communication standard for mobile phones and data terminals, and compared with a conventional circuit-switched voice network, VoLTE carries voice services in an LTE data carrying network for transmission. It should be understood that the VoLTE protocol is only an example, and the present application does not exclude the possibility of defining other possible protocols in future protocols for carrying voice traffic in future communication systems (such as 5G, etc.).
It should be understood that the above description is only exemplary to show a possible case where the terminal sends the call request, and the network and the mobile phone used by the call request are not limited in the embodiment of the present application.
In step 202, in response to the call request, the control end invokes a pre-configured virtual digital person to perform a video call with the user of the terminal.
Among them, the virtual digital person can be understood as being generated based on an Artificial Intelligence (AI) technology, which is customized according to the user's needs. The virtual digital person can convert the static image-text content into dynamic video explained by the real person based on AI technology. The virtual digital person in the embodiment of the application can perform semantic understanding based on the received voice, generate corresponding reply based on the semantic understanding, and output the reply in an audio and video mode so as to have a conversation with a user. In other words, the virtual digital person may have a video call with the user. It should be understood that the virtual number is only one possible nomenclature and should not be construed as limiting the application in any way. For example, a virtual digital person may also be referred to as a virtual steward, an emulated digital person, and so on, including but not limited to this application.
In the embodiment of the application, the virtual digital person may be configured at the control end in advance, for example, and respond to the call of the virtual digital person to perform a video call with the user of the terminal.
In step 203, based on the video call, the control end obtains the control instruction of the user to the smart home.
It should be understood that the control command is used for controlling the smart home.
It should also be understood that, by receiving a call request from the terminal, in response to the call request, the control end invokes a pre-configured virtual digital person to perform a video call with a user of the terminal, and based on the video call, the control end obtains a control instruction of the user to the smart home. The control of the user to the intelligent home can be completed without downloading the APP at the terminal, so that the occupation of the memory of the user terminal is reduced, and the operation of the user is facilitated.
Optionally, step 204 to step 212 are included before step 203.
In step 204, the user is prompted by the virtual digitizer to enter an identification.
Alternatively, the identity may be a mobile phone number, for example. The mobile phone number may be, for example, the user's own mobile phone number or another mobile phone number entered by the user. This is not a limitation of the present application.
Optionally, the control-end local database is pre-stored with the identity. The pre-stored identity may be used to determine whether the user is present, i.e. may be used to verify whether the user is a registered user.
Optionally, the user voiceprint and the face corresponding to the identity are prestored in the local database of the control terminal. The pre-stored user voiceprint and face can be used for identity verification of the user.
In step 205, it is determined whether the user is a registered user based on the identity.
For example, the identity input by the user may be found in the local database at the control end to determine whether the user is a registered user. If the user is found, the user is a registered user, and steps 206 to 207 can be continuously executed; if not found, it is an unregistered user, and steps 208 to 211 may be continued.
In step 206, voiceprints and faces of the user are collected based on the video call.
Illustratively, the process of capturing the user's voiceprint is as follows:
first, Pulse Code Modulation (PCM) audio data of a user is acquired based on a video call. For example, the PCM audio data is one frame in 25 ms, and each frame has 400 sample data.
The PCM audio data is then front-end processed (active speech detection and speech enhancement, etc.) by filtering various environmental noises and mutes, etc. that may be present in the PCM audio data through the energy value and zero-crossing rate of each frame, for example, to obtain 320 frames of valid PCM audio data.
It should be understood that since the PCM audio data is one frame in 25 milliseconds, 320 frames of valid PCM audio data is 8 seconds of PCM audio data. The 8-second PCM audio data will be described as an example.
Forward-backward (baum-welch) algorithm zeroth order, first order and second order statistics are then extracted for the valid PCM audio data.
For example, the valid PCM audio data may be represented as a data set X ═ X 1 ,x 2 ,……,x s For the s-th segment PCM audio data x in the data set s The extracted acoustic feature sequence may be denoted as o s The t frame feature in the acoustic feature sequence is represented as o s,t Then, according to a Universal Background Model (UBM), a zeroth order statistic N corresponding to the c-th gaussian mixture component of the speech segment can be calculated c,s First order statistic F c,s And second order statistic S c,s As follows:
wherein c represents the c-th Gaussian mixture component in the general background model, diag {. cndot } represents the diagonal operation of the matrix, mu c Is the mean value, gamma, of the corresponding c-th Gaussian mixture component c,s,t The posterior probability of the t frame characteristic of the s speech on the c Gaussian mixture component.
And finally, obtaining 8 seconds of PCM audio data serving as a voiceprint password to be stored in a local database.
Alternatively, the resulting 8 seconds of PCM audio data may be stored in a voiceprint library, which may be included in the local database.
Illustratively, the process of acquiring the face of the user is as follows:
first, video data of a user is acquired based on a video call.
Then, using H.264 decoder, separating the frame data in the video data, taking one frame data, and sending to the face recognition engine.
And finally, if the face recognition engine returns a valid result and the face identification vector, establishing a mapping table of the incoming call user and the face identification vector, and storing the mapping table in a local database.
Alternatively, the mapping table may be stored in a face library, which may be included in the local database.
However, if the face recognition engine returns an invalid result, the time is delayed by one second, and one frame of data is taken again. If n is 1 If the secondary identification is unsuccessful, the registration is returned to fail. Wherein n is 1 For a predefined value, for example, 0 < n can be satisfied 1 <10。
In step 207, the user is authenticated according to the collected voiceprint and face and the pre-stored user voiceprint and face.
Before the identity authentication, a welcome word is played, the user is informed to perform the identity authentication next time, the user is advised to adjust the mobile phone, the head portrait is located at the appointed position of the screen, and the characters prompted by the screen are read aloud.
Optionally, the identity verification comprises voiceprint verification and face verification.
The following exemplarily shows a specific implementation process of voiceprint verification:
firstly, based on video call, PCM audio data of a user is acquired.
Front-end processing of the PCM audio data (including active speech detection and speech enhancement) then results in 4 seconds of valid PCM audio data. The concrete mode is as follows: calculating the energy value and the zero crossing rate of each frame, filtering various environmental noises and silence and the like possibly existing in the PCM audio data, and obtaining 160 frames of valid data.
And then, accessing the voiceprint library by taking the identity of the user as a condition, and extracting the voiceprint password of the user.
One-to-one voiceprint comparison is then accomplished using Probabilistic Linear Discriminant Analysis (PLDA). Wherein the speech contained in the i-vector factorThe human information and channel information exist in two linear subspaces, assuming that the speaker S has R S The corresponding i-vector factor vector set obtained by analyzing and extracting the i-vector factor can be expressed as { w } s,r :r=1,2,...,R s For the r segment i-vector of the speaker S, the modeling expression is:
w s,r =μ+Ψβ s +Γα s,r +ε。
wherein μ represents the mean of the ensemble of training data; Ψ is an identity space, containing substrates that can be used to represent various identities; beta is a s Is the identity of a person (or the location of a person in identity space); Γ is an error space that contains a substrate that can be used to represent different variations of the same identity; alpha is alpha s,r Indicating the position in the space; the final residual noise term epsilon, which is zero mean gaussian, is used to represent what has not yet been explained. The first two items on the right of the equal sign of the formula are only related to the speaker and are not related to a specific certain voice of the speaker, and the items are called signal parts, which describe the difference between the speakers; the last two items to the right of the equal sign describe the difference between different voices of the same speaker, called the noise part. The data structure of a piece of speech is described by such two imaginary variables.
After the i-vector factors are modeled and estimated by the subspace PLDA to obtain model parameters, speakers can be classified based on the model, and the likelihood ratio scores are adopted to score test tasks under the subspace PLDA modeling method.
And finally, comparing the scores, and if the scores exceed a specified threshold value, determining that the voiceprint verification is passed.
The following exemplarily shows a specific implementation process of face verification:
firstly, accessing a face library by taking the identity of a user as a condition to acquire a face identification vector.
Then, based on the video call, the video data is read.
And separating frame data in the video data by using an H.264 decoder, taking one frame of data, and sending the frame of data and the face identification vector to a face identification engine to request 1 for head portrait identification of 1.
And finally, stopping the recognition if the engine returns that the recognition is successful, and returning that the video identity authentication is passed. Or, if the return is unsuccessful, reading the frame data again, repeating the identification, and repeating the identification for n times 2 Next, the process is carried out. Wherein n is 2 Is a predefined value, e.g. 0 < n can be satisfied 2 <10。
In step 208, the user is prompted for voiceprint registration and face registration.
Optionally, the terminal prompts the user that the currently input mobile phone number is an unregistered mobile phone number and prompts the user to register the input mobile phone number.
In the case where the user determines to perform the voiceprint registration and the face registration, step 210 is performed.
In step 209, the user's home address and smart home product model number are queried by the virtual digital person.
In step 210, voiceprints and faces of the user are collected based on the video call.
It should be understood that, based on the video call, the process of collecting the voiceprint and the face of the user is shown in step 206, and the details are not repeated here.
In step 211, the home address of the user, the product model of the smart home, the collected voiceprint and the face are stored in a local database.
It should be understood that the voiceprint and the face stored in the local database are the above-mentioned pre-stored voiceprint and face, and can be used for performing identity verification on the user.
Optionally, the identity of the user is saved. The stored identity is the pre-stored identity, and can be used to determine whether the user is a registered user.
In case the above identity authentication is passed, or the above voiceprint registration and face registration are successful, the above step 203 is executed.
Optionally, after step 203, steps 212 to 214 may be performed.
In step 212, voice recognition is performed on the control command to determine the smart home that the user wishes to control and the control operation of the smart home.
The process of performing voice recognition on the control command is specifically as follows:
first, the audio data of the user is read from the message queue, and the audio data is encapsulated according to the second version of the media resource control protocol (MRCPv 2) and forwarded to the speech recognition engine.
And then after the transcription result of the voice recognition engine is obtained, the intention recognition of the user is completed through pinyin error correction and hotword error correction, and the voice command with the closest similarity is selected.
It should be understood that the accuracy of the control instruction can be improved through the voice recognition process, so that the accuracy and efficiency of the voice control are improved to a certain extent.
In step 213, the control instruction is executed.
In step 214, the completion of the control instruction is fed back to the user through the virtual digital person.
Based on the scheme, the call request sent by the user through dialing the preset number and based on the terminal self-contained standard (such as the LTE standard or the 5G standard) is responded, and the pre-configured virtual digital person is called to carry out video call with the user, so that the control instruction of the user to the intelligent home is obtained, the control of the user to the intelligent home can be completed without downloading the APP at the terminal, and the occupation of the memory of the user terminal is reduced. And because the complicated operation interface of APP need not to adopt, the user operation of more making things convenient for. In addition, since the video call is carried by using the LTE communication resource or the 5G communication resource, the identity information, the instruction, and the like of the user can be carried by the LTE communication resource or the 5G communication resource, so that privacy disclosure can be avoided and the method is safer and more reliable compared with the public internet.
Fig. 3 is a schematic block diagram of an intelligent home control system provided in an embodiment of the present application. As shown in fig. 3, the smart home control system may include a terminal, a control terminal, a local database, a user voiceprint acquisition module, a user voiceprint comparison module, a user head portrait acquisition module, a user head portrait comparison module, a voice instruction recognition module, and a smart home interface gateway module.
Optionally, the local database above may include, for example, a voiceprint library and an avatar library.
It should be understood that the method 200 described in FIG. 2 may be applied to the system shown in FIG. 3, and the modules in FIG. 3 may be used to complete the method described in FIG. 2.
Fig. 4 is a schematic block diagram of an intelligent home control device provided in an embodiment of the present application. As shown in fig. 4, the apparatus 400 may include: a processing module 410 and a communication module 420.
The processing module 410 may be configured to invoke a preconfigured virtual digital person to conduct a video call with a user of the terminal in response to the call request.
The communication module 420 may be configured to receive a call request from a terminal, where the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used to request remote control of the smart home; and the method can also be used for acquiring the control instruction of the user to the intelligent home based on the video call.
Optionally, the communication module 420 may be further configured to prompt the user for an identification through the virtual digital person.
Optionally, the processing module 410 may be further configured to determine whether the user is a registered user based on the identity; and under the condition that the user is a registered user, performing identity verification on the user based on the voiceprint and the face of the user.
Optionally, the communication module 420 may be further configured to prompt the user to perform voiceprint registration and face registration if the user is an unregistered user.
Optionally, the identity is a mobile phone number of the user.
Optionally, the communication module 420 may be further configured to collect a voiceprint and a face of the user based on the video call.
Optionally, the processing module 410 may be further configured to perform identity verification on the user according to the voiceprint recognition and the face recognition, and the pre-stored voiceprint and face of the user.
Optionally, the communication module 420 may be further configured to query, by the virtual digital person, a home address of the user and a product model of a smart home; and acquiring the voiceprint and the face of the user based on the video call.
Optionally, the processing module 410 may be further configured to store the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
Optionally, the communication module 420 may be further configured to obtain a voice instruction of the user based on the video call.
Optionally, the processing module 410 may be further configured to perform voice recognition on the voice command to determine the smart home that the user wishes to control and the control operation of the smart home.
Optionally, the processing module 410 may also be configured to execute the control instructions.
Optionally, the communication module 420 may be further configured to feed back the completion condition of the control instruction to the user through the virtual digital person after the control instruction is completed.
Optionally, the call request and the video call are carried on resources of an LTE communication system or resources of a 5G communication system.
Fig. 5 is another schematic block diagram of an intelligent home control device provided in an embodiment of the present application. The device can be used for realizing the functions of the processing module and the communication module in the method. Wherein the apparatus may be a system-on-a-chip. In the embodiment of the present application, the chip system may be composed of a chip, and may also include a chip and other discrete devices.
As shown in fig. 5, the apparatus 500 may include at least one processor 510. Illustratively, the processor 510 may be configured to invoke a preconfigured virtual digital person to conduct a video call with the user of the terminal in response to the call request. For details, reference is made to the detailed description in the method example, which is not repeated herein.
The apparatus 500 may also include at least one memory 520 for storing program instructions and/or data. The memory 520 is coupled to the processor 510. The coupling in the embodiments of the present application is an indirect coupling or a communication connection between devices, units or modules, and may be an electrical, mechanical or other form for information interaction between the devices, units or modules. The processor 510 may operate in conjunction with the memory 520. Processor 510 may execute program instructions stored in memory 520. At least one of the at least one memory may be included in the processor.
The apparatus 500 may also include a communication interface 530 for communicating with other devices over a transmission medium so that the apparatus used in the apparatus 500 may communicate with other devices. The communication interface 530 may be, for example, a transceiver, an interface, a bus, a circuit, or a device capable of performing a transceiving function. The processor 510 may utilize the communication interface 530 to receive and transmit data and/or information, and is configured to implement a smart home control method described in the corresponding embodiment of fig. 2.
The specific connection medium between the processor 510, the memory 520 and the communication interface 530 is not limited in the embodiments of the present application. In fig. 5, the processor 510, the memory 520, and the communication interface 530 are connected by a bus 540. The bus 540 is shown in fig. 5 by a thick line, and the connection between other components is merely illustrative and not intended to be limiting. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown in FIG. 5, but this is not intended to represent only one bus or type of bus.
The present application further provides a computer program product, the computer program product comprising: a computer program (also referred to as code, or instructions), which when executed, causes an electronic device to perform the method of the embodiment shown in fig. 2.
The present application also provides a computer-readable storage medium having stored thereon a computer program (also referred to as code, or instructions). The computer program, when executed, causes the electronic device to perform the method in the embodiment shown in fig. 2.
It should be understood that the processor in the embodiments of the present application may be an integrated circuit chip having signal processing capability. In implementation, the steps of the above method embodiments may be performed by integrated logic circuits of hardware in a processor or instructions in the form of software. The processor may be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, or discrete hardware components. The various methods, steps, and logic blocks disclosed in the embodiments of the present application may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor. The software modules may be located in ram, flash, rom, prom, or eprom, registers, etc. as is well known in the art. The storage medium is located in a memory, and a processor reads information in the memory and completes the steps of the method in combination with hardware of the processor.
It will also be appreciated that the memory in the embodiments of the subject application can be either volatile memory or nonvolatile memory, or can include both volatile and nonvolatile memory. The non-volatile memory may be a read-only memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an electrically Erasable EPROM (EEPROM), or a flash memory. Volatile memory can be Random Access Memory (RAM), which acts as external cache memory. By way of example, but not limitation, many forms of RAM are available, such as Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), Synchronous Dynamic Random Access Memory (SDRAM), double data rate SDRAM, enhanced SDRAM, SLDRAM, Synchronous Link DRAM (SLDRAM), and direct rambus RAM (DR RAM). It should be noted that the memory of the systems and methods described herein is intended to comprise, without being limited to, these and any other suitable types of memory.
As used in this specification, the terms "unit," "module," and the like are intended to refer to a computer-related entity, either hardware, firmware, a combination of hardware and software, or software in execution.
Those of ordinary skill in the art will appreciate that the various illustrative logical blocks and steps (step) described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application. In the several embodiments provided in the present application, it should be understood that the disclosed apparatus, device and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the units is only one logical division, and other divisions may be realized in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit.
In the above embodiments, the functions of the functional units may be fully or partially implemented by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions (programs). The procedures or functions described in accordance with the embodiments of the present application are generated in whole or in part when the computer program instructions (programs) are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center via wired (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.) means. The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
The functions, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application or portions thereof that substantially contribute to the prior art may be embodied in the form of a software product stored in a storage medium and including instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: various media capable of storing program codes, such as a usb disk, a removable hard disk, a read-only memory (ROM), a Random Access Memory (RAM), a magnetic disk, or an optical disk.
The above description is only for the specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present application, and shall be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.
Claims (11)
1. The intelligent household control method is characterized by comprising the following steps:
receiving a call request from a terminal, wherein the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home;
responding to the call request, and calling a pre-configured virtual digital person to carry out video call with the user of the terminal;
and acquiring a control instruction of the user to the intelligent home based on the video call.
2. The method of claim 1, wherein before the obtaining of the control instruction of the user for the smart home based on the video call, the method further comprises: prompting the user to input an identity through the virtual digital person;
determining whether the user is a registered user based on the identity;
under the condition that the user is a registered user, performing identity verification on the user based on the voiceprint and the face of the user; or
And prompting the user to perform voiceprint registration and face registration under the condition that the user is an unregistered user.
3. The method of claim 2, wherein the identity is a cell phone number of the user.
4. The method of claim 2, wherein said authenticating the user comprises:
collecting voiceprints and faces of the user based on the video call;
and carrying out identity verification on the user according to the voiceprint recognition, the face recognition, the prestored voiceprint and face of the user.
5. The method of claim 2, wherein the method further comprises:
inquiring the home address of the user and the product model of the smart home through the virtual digital person;
acquiring a voiceprint and a face of the user based on the video call;
and storing the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
6. The method of claim 1, wherein the method further comprises:
executing the control instruction;
and after the control instruction is finished, feeding back the finishing condition of the control instruction to the user through the virtual digital person.
7. The method of any of claims 1 to 6, wherein the call request and the video call are carried on resources of a Long Term Evolution (LTE) communication system or resources of a fifth generation 5G communication system.
8. An intelligent home control apparatus, comprising means for performing the method of any of claims 1-7.
9. An intelligent home control device, comprising a processor configured to perform the method of any one of claims 1 to 7.
10. A computer-readable storage medium, characterized in that the storage medium stores instructions that, when executed on a computer, cause the computer to perform the method of any of claims 1 to 7.
11. A computer program product, comprising a computer program which, when executed, causes a computer to perform the method of any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110204624.5A CN114979543B (en) | 2021-02-24 | 2021-02-24 | Smart home control method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110204624.5A CN114979543B (en) | 2021-02-24 | 2021-02-24 | Smart home control method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114979543A true CN114979543A (en) | 2022-08-30 |
CN114979543B CN114979543B (en) | 2024-07-02 |
Family
ID=82972710
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110204624.5A Active CN114979543B (en) | 2021-02-24 | 2021-02-24 | Smart home control method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114979543B (en) |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105978775A (en) * | 2016-07-29 | 2016-09-28 | 镇江惠通电子有限公司 | Speech control system and speech control method |
CN106097495A (en) * | 2016-06-03 | 2016-11-09 | 赵树龙 | A kind of intelligent voice control vocal print face authentication door access control system and method |
CN106782559A (en) * | 2016-12-31 | 2017-05-31 | 广东博意建筑设计院有限公司 | Smart home house keeper central control system and its control method with telecommunication control |
CN106792238A (en) * | 2016-11-17 | 2017-05-31 | Tcl集团股份有限公司 | A kind of interactive approach and interaction systems based on live telecast shopping platform |
CN107092196A (en) * | 2017-06-26 | 2017-08-25 | 广东美的制冷设备有限公司 | The control method and relevant device of intelligent home device |
WO2018113526A1 (en) * | 2016-12-20 | 2018-06-28 | 四川长虹电器股份有限公司 | Face recognition and voiceprint recognition-based interactive authentication system and method |
CN109412910A (en) * | 2018-11-20 | 2019-03-01 | 三星电子(中国)研发中心 | The method and apparatus for controlling smart home device |
CN110248018A (en) * | 2018-03-09 | 2019-09-17 | 深圳市云昱科技有限公司 | The call method and Related product of intelligent secretary |
CN110647636A (en) * | 2019-09-05 | 2020-01-03 | 深圳追一科技有限公司 | Interaction method, interaction device, terminal equipment and storage medium |
CN112291331A (en) * | 2020-10-25 | 2021-01-29 | 湖南云脸智联科技有限公司 | Cloud access control implementation method based on AI intelligent sound box application |
CN112291497A (en) * | 2020-10-28 | 2021-01-29 | 上海赛连信息科技有限公司 | Intelligent video customer service access method and device |
-
2021
- 2021-02-24 CN CN202110204624.5A patent/CN114979543B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106097495A (en) * | 2016-06-03 | 2016-11-09 | 赵树龙 | A kind of intelligent voice control vocal print face authentication door access control system and method |
CN105978775A (en) * | 2016-07-29 | 2016-09-28 | 镇江惠通电子有限公司 | Speech control system and speech control method |
CN106792238A (en) * | 2016-11-17 | 2017-05-31 | Tcl集团股份有限公司 | A kind of interactive approach and interaction systems based on live telecast shopping platform |
WO2018113526A1 (en) * | 2016-12-20 | 2018-06-28 | 四川长虹电器股份有限公司 | Face recognition and voiceprint recognition-based interactive authentication system and method |
CN106782559A (en) * | 2016-12-31 | 2017-05-31 | 广东博意建筑设计院有限公司 | Smart home house keeper central control system and its control method with telecommunication control |
CN107092196A (en) * | 2017-06-26 | 2017-08-25 | 广东美的制冷设备有限公司 | The control method and relevant device of intelligent home device |
CN110248018A (en) * | 2018-03-09 | 2019-09-17 | 深圳市云昱科技有限公司 | The call method and Related product of intelligent secretary |
CN109412910A (en) * | 2018-11-20 | 2019-03-01 | 三星电子(中国)研发中心 | The method and apparatus for controlling smart home device |
CN110647636A (en) * | 2019-09-05 | 2020-01-03 | 深圳追一科技有限公司 | Interaction method, interaction device, terminal equipment and storage medium |
CN112291331A (en) * | 2020-10-25 | 2021-01-29 | 湖南云脸智联科技有限公司 | Cloud access control implementation method based on AI intelligent sound box application |
CN112291497A (en) * | 2020-10-28 | 2021-01-29 | 上海赛连信息科技有限公司 | Intelligent video customer service access method and device |
Also Published As
Publication number | Publication date |
---|---|
CN114979543B (en) | 2024-07-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20220317641A1 (en) | Device control method, conflict processing method, corresponding apparatus and electronic device | |
CN107644638B (en) | Audio recognition method, device, terminal and computer readable storage medium | |
CN107220532B (en) | Method and apparatus for recognizing user identity through voice | |
US12010108B2 (en) | Techniques to provide sensitive information over a voice connection | |
CN107544271A (en) | Terminal control method, device and computer-readable recording medium | |
CN108470568B (en) | Intelligent device control method and device, storage medium and electronic device | |
JP6738867B2 (en) | Speaker authentication method and voice recognition system | |
JP6783339B2 (en) | Methods and devices for processing audio | |
JP2017529711A (en) | Method, apparatus and system for voice confirmation | |
CN105872619A (en) | Video playing record matching method and matching device | |
CN107544272A (en) | terminal control method, device and storage medium | |
CN107205097A (en) | Mobile terminal lookup method, device and computer-readable recording medium | |
CN110858841B (en) | Electronic device and method for registering new user through authentication of registered user | |
CN107808662B (en) | Method and device for updating grammar rule base for speech recognition | |
CN111653284B (en) | Interaction and identification method, device, terminal equipment and computer storage medium | |
CN111736474A (en) | Multi-mode intelligent home control system and method | |
CN109087647A (en) | Application on Voiceprint Recognition processing method, device, electronic equipment and storage medium | |
CN110224904B (en) | Voice processing method, device, computer readable storage medium and computer equipment | |
CN114979543A (en) | Intelligent household control method and device | |
CN108399918B (en) | Intelligent device connection method, intelligent device and terminal | |
CN112417923A (en) | System, method and apparatus for controlling smart devices | |
CN112738344B (en) | Method and device for identifying user identity, storage medium and electronic equipment | |
Dumitrescu | Cassandra smart-home system description | |
CN112449059A (en) | Voice interaction device, method and system for realizing call based on voice interaction device | |
CN111091638A (en) | Storage medium, intelligent door lock and authentication method thereof |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant |