CN114979543A - Intelligent household control method and device - Google Patents

Intelligent household control method and device Download PDF

Info

Publication number
CN114979543A
CN114979543A CN202110204624.5A CN202110204624A CN114979543A CN 114979543 A CN114979543 A CN 114979543A CN 202110204624 A CN202110204624 A CN 202110204624A CN 114979543 A CN114979543 A CN 114979543A
Authority
CN
China
Prior art keywords
user
terminal
control
voiceprint
call request
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110204624.5A
Other languages
Chinese (zh)
Other versions
CN114979543B (en
Inventor
黑昱冬
周靖
田天
李雪丹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China United Network Communications Group Co Ltd
Original Assignee
China United Network Communications Group Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China United Network Communications Group Co Ltd filed Critical China United Network Communications Group Co Ltd
Priority to CN202110204624.5A priority Critical patent/CN114979543B/en
Publication of CN114979543A publication Critical patent/CN114979543A/en
Application granted granted Critical
Publication of CN114979543B publication Critical patent/CN114979543B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/14Systems for two-way working
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/06Decision making techniques; Pattern matching strategies
    • G10L17/14Use of phonemic categorisation or speech recognition prior to speaker recognition or verification
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/02Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Game Theory and Decision Science (AREA)
  • Business, Economics & Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Telephonic Communication Services (AREA)

Abstract

The application provides an intelligent home control method and device. The method comprises the following steps: the method comprises the steps that a control end answers a call request from a terminal, wherein the call request is sent by a user of the terminal through dialing a preset number, and the call request is used for requesting remote control of the intelligent home; the control terminal responds to the call request and calls a pre-configured virtual digital person to carry out video call with the user of the terminal; and acquiring a control instruction of the user to the intelligent home based on the video call. The call request can be carried on resources of an LTE communication system or a 5G communication system. Due to the fact that the dialing process based on the standard of the terminal is based on the LTE standard or the 5G standard, the control of the user on the intelligent home can be finished without downloading special application, and the occupation of the memory of the terminal can be reduced; and the call is borne in LTE or 5G resources, so that the transmission of information such as identity authentication, instructions and the like is safer and more reliable.

Description

Intelligent home control method and device
Technical Field
The application relates to the technical field of mobile communication, in particular to an intelligent home control method and device.
Background
The intelligent home connects various devices (such as lighting devices, curtains, air conditioners, security systems, digital cinema systems, audio and video servers, network household appliances and the like) in the home through the Internet of things technology, and provides various control functions and means. Currently, the main control means include remote network control, local control, timing control, and one-key scenario control.
However, the existing remote network control needs to download and install special software provided by the smart home manufacturer, such as a special Application (APP), so as to perform the relevant operations. With the increase of the types of smart homes, corresponding special software is increased, which may occupy a large storage space of a user terminal.
Disclosure of Invention
The application provides an intelligent home control method and device, so that remote control over an intelligent home can be completed without downloading an APP, and occupation of a memory of a user terminal is reduced.
In a first aspect, the present application provides a smart home control method, including: receiving a call request from a terminal, wherein the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home; responding to the call request, and calling a pre-configured virtual digital person to carry out video call with a user of the terminal; and acquiring a control instruction of the user to the intelligent home based on the video call.
Based on the scheme, the call request which is sent by the user by dialing the preset number and is based on the self-contained standard (such as the LTE standard or the 5G standard) of the terminal is responded, the pre-configured virtual digital person is called to carry out video call with the user, so that the control instruction of the user to the intelligent home is obtained, the control of the user to the intelligent home can be completed without downloading a special application at the terminal, and the occupation of the memory of the user terminal is reduced. And because the complicated operation interface of APP need not to adopt, the user operation of more making things convenient for.
Optionally, before the obtaining of the control instruction of the user to the smart home based on the video call, the method further includes: prompting the user to input an identification through the virtual digital person; determining whether the user is a registered user based on the identity; under the condition that the user is a registered user, performing identity authentication on the user based on the voiceprint and the face of the user; or prompting the user to perform voiceprint registration and face registration under the condition that the user is an unregistered user.
Optionally, the identity is a mobile phone number of the user.
Optionally, the authenticating the user includes: collecting voiceprints and faces of the user based on the video call; and carrying out identity verification on the user according to the voiceprint recognition, the face recognition, the prestored voiceprint and face of the user.
Optionally, the method further comprises: inquiring the home address of the user and the product model of the smart home through the virtual digital person; acquiring a voiceprint and a face of the user based on the video call; and storing the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
Optionally, the obtaining a control instruction of the user to the smart home based on the video call includes: acquiring a voice instruction of the user based on the video call; and carrying out voice recognition on the voice command so as to determine the smart home which the user wants to control and the control operation of the smart home.
Optionally, the method further comprises: executing the control instruction; and after the control instruction is finished, feeding back the finishing condition of the control instruction to the user through the virtual digital person.
Optionally, the call request and the video callThe speech is carried over resources of a Long Term Evolution (LTE) communication system or fifth generation (5) th generation, 5G) resources of the communication system.
In a second aspect, the present application provides an intelligent home control apparatus, including a unit for implementing the intelligent home control method of the first aspect.
In a third aspect, the present application provides an intelligent home control device, including a processor, where the processor is configured to execute the intelligent home control method according to the first aspect.
In a fourth aspect, there is provided a computer-readable storage medium comprising instructions which, when executed on a computer, cause the computer to carry out the method of any one of the first aspect and the first aspect.
In a fifth aspect, there is provided a computer program product comprising: a computer program (also referred to as code, or instructions), which when executed, causes a computer to perform the method of the first aspect as well as any one of the possible implementations of the first aspect.
It should be understood that the second aspect to the fifth aspect of the present application correspond to the technical solutions of the first aspect of the present application, and the beneficial effects achieved by the aspects and the corresponding possible implementations are similar and will not be described again.
Drawings
Fig. 1 is a network architecture suitable for an intelligent home control method provided in the present application;
fig. 2 is a schematic flow chart of an intelligent home control method provided in the embodiment of the present application;
fig. 3 is a schematic block diagram of an intelligent home control system provided in an embodiment of the present application;
fig. 4 is a schematic block diagram of an intelligent home control device provided in an embodiment of the present application;
fig. 5 is another schematic block diagram of an intelligent home control device provided in an embodiment of the present application.
Detailed Description
The technical solution in the present application will be described below with reference to the accompanying drawings.
The intelligent home is characterized in that a home is used as a platform, facilities related to home life are integrated by utilizing a comprehensive wiring technology, a network communication technology, a safety precaution technology, an automatic control technology and an audio and video technology, an efficient home facility and a home schedule error management system are constructed, home safety, convenience, comfortableness and artistry are improved, and an environment-friendly and energy-saving living environment is realized. Compared with the common home, the intelligent home has the traditional living function, integrates the functions of building, network communication, information household appliance and equipment automation, provides an all-around information interaction function, and even saves funds for various energy expenses.
The technical scheme of the embodiment of the application can be used for controlling various intelligent household devices. The smart home may include, for example, but is not limited to, smart light fixtures, smart washing machines, smart refrigerators, smart rice cookers, smart door locks, smart curtains, central dust extraction systems, smart background music systems, smart video systems, digital cinema systems, smart security alarms, and smart surveillance systems, etc.
The control of smart homes may include, for example, but is not limited to, the control of various smart home devices listed above. For the sake of brevity, this is not to be enumerated here.
To facilitate understanding of the embodiment of the present application, first, a network architecture suitable for the smart home control method provided in the present application is described in detail with reference to fig. 1. As shown in fig. 1, the network architecture 100 may include a terminal 110, a control end 120, and smart home devices 131, 132, and 133. The terminal 110 and the control terminal 120 may communicate with each other through a global system for mobile communications (GSM), and the control terminal 120 may be configured to control the smart home devices 131 to 133.
It should be understood that, in the embodiment of the present application, the terminal 110 may be, for example, a mobile phone or a wearable device; the control end 120 may be, for example, an intelligent home control device having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself, or a control device integrated on an intelligent home having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself. The embodiments of the present application do not limit this. The control end 120 may be connected to a plurality of terminals 110, and the number of the terminals is not limited in the embodiment of the present application.
An intelligent home control method provided by the embodiment of the present application will be described in detail below with reference to fig. 2. As shown in fig. 2, the method 200 may include steps 201 through 214.
In step 201, the control terminal receives a call request from the terminal.
Optionally, the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home. The preset number may be set on the control terminal for connecting the control terminal. The control terminal may be, for example, an intelligent home control device having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself, or a control device integrated on an intelligent home having an audio/video call function based on a standard (e.g., LTE standard or 5G standard) of the terminal itself. The embodiments of the present application do not limit this.
Alternatively, the user of the terminal may be a family member or other user who wants to control the smart home.
It should be understood that in the embodiment of the present application, the terminal, the user of the terminal, and the user are used alternately, and the three meanings are the same.
Alternatively, the call request may be to use the fourth generation (4) th generation, 4G) terminal or 5G terminal makes a call. The 4G terminal is also a terminal supporting 4G, or a terminal supporting LTE. A 5G terminal is also a terminal supporting 5G. The call request can be carried on resources of an LTE communication system or resources of a 5G communication system. For example, the call request may be transmitted via a voice over LTE (VoLTE) protocol.
It should be noted that the VoLTE protocol is mainly a high-speed wireless communication standard for mobile phones and data terminals, and compared with a conventional circuit-switched voice network, VoLTE carries voice services in an LTE data carrying network for transmission. It should be understood that the VoLTE protocol is only an example, and the present application does not exclude the possibility of defining other possible protocols in future protocols for carrying voice traffic in future communication systems (such as 5G, etc.).
It should be understood that the above description is only exemplary to show a possible case where the terminal sends the call request, and the network and the mobile phone used by the call request are not limited in the embodiment of the present application.
In step 202, in response to the call request, the control end invokes a pre-configured virtual digital person to perform a video call with the user of the terminal.
Among them, the virtual digital person can be understood as being generated based on an Artificial Intelligence (AI) technology, which is customized according to the user's needs. The virtual digital person can convert the static image-text content into dynamic video explained by the real person based on AI technology. The virtual digital person in the embodiment of the application can perform semantic understanding based on the received voice, generate corresponding reply based on the semantic understanding, and output the reply in an audio and video mode so as to have a conversation with a user. In other words, the virtual digital person may have a video call with the user. It should be understood that the virtual number is only one possible nomenclature and should not be construed as limiting the application in any way. For example, a virtual digital person may also be referred to as a virtual steward, an emulated digital person, and so on, including but not limited to this application.
In the embodiment of the application, the virtual digital person may be configured at the control end in advance, for example, and respond to the call of the virtual digital person to perform a video call with the user of the terminal.
In step 203, based on the video call, the control end obtains the control instruction of the user to the smart home.
It should be understood that the control command is used for controlling the smart home.
It should also be understood that, by receiving a call request from the terminal, in response to the call request, the control end invokes a pre-configured virtual digital person to perform a video call with a user of the terminal, and based on the video call, the control end obtains a control instruction of the user to the smart home. The control of the user to the intelligent home can be completed without downloading the APP at the terminal, so that the occupation of the memory of the user terminal is reduced, and the operation of the user is facilitated.
Optionally, step 204 to step 212 are included before step 203.
In step 204, the user is prompted by the virtual digitizer to enter an identification.
Alternatively, the identity may be a mobile phone number, for example. The mobile phone number may be, for example, the user's own mobile phone number or another mobile phone number entered by the user. This is not a limitation of the present application.
Optionally, the control-end local database is pre-stored with the identity. The pre-stored identity may be used to determine whether the user is present, i.e. may be used to verify whether the user is a registered user.
Optionally, the user voiceprint and the face corresponding to the identity are prestored in the local database of the control terminal. The pre-stored user voiceprint and face can be used for identity verification of the user.
In step 205, it is determined whether the user is a registered user based on the identity.
For example, the identity input by the user may be found in the local database at the control end to determine whether the user is a registered user. If the user is found, the user is a registered user, and steps 206 to 207 can be continuously executed; if not found, it is an unregistered user, and steps 208 to 211 may be continued.
In step 206, voiceprints and faces of the user are collected based on the video call.
Illustratively, the process of capturing the user's voiceprint is as follows:
first, Pulse Code Modulation (PCM) audio data of a user is acquired based on a video call. For example, the PCM audio data is one frame in 25 ms, and each frame has 400 sample data.
The PCM audio data is then front-end processed (active speech detection and speech enhancement, etc.) by filtering various environmental noises and mutes, etc. that may be present in the PCM audio data through the energy value and zero-crossing rate of each frame, for example, to obtain 320 frames of valid PCM audio data.
It should be understood that since the PCM audio data is one frame in 25 milliseconds, 320 frames of valid PCM audio data is 8 seconds of PCM audio data. The 8-second PCM audio data will be described as an example.
Forward-backward (baum-welch) algorithm zeroth order, first order and second order statistics are then extracted for the valid PCM audio data.
For example, the valid PCM audio data may be represented as a data set X ═ X 1 ,x 2 ,……,x s For the s-th segment PCM audio data x in the data set s The extracted acoustic feature sequence may be denoted as o s The t frame feature in the acoustic feature sequence is represented as o s,t Then, according to a Universal Background Model (UBM), a zeroth order statistic N corresponding to the c-th gaussian mixture component of the speech segment can be calculated c,s First order statistic F c,s And second order statistic S c,s As follows:
Figure BDA0002949874220000061
wherein c represents the c-th Gaussian mixture component in the general background model, diag {. cndot } represents the diagonal operation of the matrix, mu c Is the mean value, gamma, of the corresponding c-th Gaussian mixture component c,s,t The posterior probability of the t frame characteristic of the s speech on the c Gaussian mixture component.
And finally, obtaining 8 seconds of PCM audio data serving as a voiceprint password to be stored in a local database.
Alternatively, the resulting 8 seconds of PCM audio data may be stored in a voiceprint library, which may be included in the local database.
Illustratively, the process of acquiring the face of the user is as follows:
first, video data of a user is acquired based on a video call.
Then, using H.264 decoder, separating the frame data in the video data, taking one frame data, and sending to the face recognition engine.
And finally, if the face recognition engine returns a valid result and the face identification vector, establishing a mapping table of the incoming call user and the face identification vector, and storing the mapping table in a local database.
Alternatively, the mapping table may be stored in a face library, which may be included in the local database.
However, if the face recognition engine returns an invalid result, the time is delayed by one second, and one frame of data is taken again. If n is 1 If the secondary identification is unsuccessful, the registration is returned to fail. Wherein n is 1 For a predefined value, for example, 0 < n can be satisfied 1 <10。
In step 207, the user is authenticated according to the collected voiceprint and face and the pre-stored user voiceprint and face.
Before the identity authentication, a welcome word is played, the user is informed to perform the identity authentication next time, the user is advised to adjust the mobile phone, the head portrait is located at the appointed position of the screen, and the characters prompted by the screen are read aloud.
Optionally, the identity verification comprises voiceprint verification and face verification.
The following exemplarily shows a specific implementation process of voiceprint verification:
firstly, based on video call, PCM audio data of a user is acquired.
Front-end processing of the PCM audio data (including active speech detection and speech enhancement) then results in 4 seconds of valid PCM audio data. The concrete mode is as follows: calculating the energy value and the zero crossing rate of each frame, filtering various environmental noises and silence and the like possibly existing in the PCM audio data, and obtaining 160 frames of valid data.
And then, accessing the voiceprint library by taking the identity of the user as a condition, and extracting the voiceprint password of the user.
One-to-one voiceprint comparison is then accomplished using Probabilistic Linear Discriminant Analysis (PLDA). Wherein the speech contained in the i-vector factorThe human information and channel information exist in two linear subspaces, assuming that the speaker S has R S The corresponding i-vector factor vector set obtained by analyzing and extracting the i-vector factor can be expressed as { w } s,r :r=1,2,...,R s For the r segment i-vector of the speaker S, the modeling expression is:
w s,r =μ+Ψβ s +Γα s,r +ε。
wherein μ represents the mean of the ensemble of training data; Ψ is an identity space, containing substrates that can be used to represent various identities; beta is a s Is the identity of a person (or the location of a person in identity space); Γ is an error space that contains a substrate that can be used to represent different variations of the same identity; alpha is alpha s,r Indicating the position in the space; the final residual noise term epsilon, which is zero mean gaussian, is used to represent what has not yet been explained. The first two items on the right of the equal sign of the formula are only related to the speaker and are not related to a specific certain voice of the speaker, and the items are called signal parts, which describe the difference between the speakers; the last two items to the right of the equal sign describe the difference between different voices of the same speaker, called the noise part. The data structure of a piece of speech is described by such two imaginary variables.
After the i-vector factors are modeled and estimated by the subspace PLDA to obtain model parameters, speakers can be classified based on the model, and the likelihood ratio scores are adopted to score test tasks under the subspace PLDA modeling method.
And finally, comparing the scores, and if the scores exceed a specified threshold value, determining that the voiceprint verification is passed.
The following exemplarily shows a specific implementation process of face verification:
firstly, accessing a face library by taking the identity of a user as a condition to acquire a face identification vector.
Then, based on the video call, the video data is read.
And separating frame data in the video data by using an H.264 decoder, taking one frame of data, and sending the frame of data and the face identification vector to a face identification engine to request 1 for head portrait identification of 1.
And finally, stopping the recognition if the engine returns that the recognition is successful, and returning that the video identity authentication is passed. Or, if the return is unsuccessful, reading the frame data again, repeating the identification, and repeating the identification for n times 2 Next, the process is carried out. Wherein n is 2 Is a predefined value, e.g. 0 < n can be satisfied 2 <10。
In step 208, the user is prompted for voiceprint registration and face registration.
Optionally, the terminal prompts the user that the currently input mobile phone number is an unregistered mobile phone number and prompts the user to register the input mobile phone number.
In the case where the user determines to perform the voiceprint registration and the face registration, step 210 is performed.
In step 209, the user's home address and smart home product model number are queried by the virtual digital person.
In step 210, voiceprints and faces of the user are collected based on the video call.
It should be understood that, based on the video call, the process of collecting the voiceprint and the face of the user is shown in step 206, and the details are not repeated here.
In step 211, the home address of the user, the product model of the smart home, the collected voiceprint and the face are stored in a local database.
It should be understood that the voiceprint and the face stored in the local database are the above-mentioned pre-stored voiceprint and face, and can be used for performing identity verification on the user.
Optionally, the identity of the user is saved. The stored identity is the pre-stored identity, and can be used to determine whether the user is a registered user.
In case the above identity authentication is passed, or the above voiceprint registration and face registration are successful, the above step 203 is executed.
Optionally, after step 203, steps 212 to 214 may be performed.
In step 212, voice recognition is performed on the control command to determine the smart home that the user wishes to control and the control operation of the smart home.
The process of performing voice recognition on the control command is specifically as follows:
first, the audio data of the user is read from the message queue, and the audio data is encapsulated according to the second version of the media resource control protocol (MRCPv 2) and forwarded to the speech recognition engine.
And then after the transcription result of the voice recognition engine is obtained, the intention recognition of the user is completed through pinyin error correction and hotword error correction, and the voice command with the closest similarity is selected.
It should be understood that the accuracy of the control instruction can be improved through the voice recognition process, so that the accuracy and efficiency of the voice control are improved to a certain extent.
In step 213, the control instruction is executed.
In step 214, the completion of the control instruction is fed back to the user through the virtual digital person.
Based on the scheme, the call request sent by the user through dialing the preset number and based on the terminal self-contained standard (such as the LTE standard or the 5G standard) is responded, and the pre-configured virtual digital person is called to carry out video call with the user, so that the control instruction of the user to the intelligent home is obtained, the control of the user to the intelligent home can be completed without downloading the APP at the terminal, and the occupation of the memory of the user terminal is reduced. And because the complicated operation interface of APP need not to adopt, the user operation of more making things convenient for. In addition, since the video call is carried by using the LTE communication resource or the 5G communication resource, the identity information, the instruction, and the like of the user can be carried by the LTE communication resource or the 5G communication resource, so that privacy disclosure can be avoided and the method is safer and more reliable compared with the public internet.
Fig. 3 is a schematic block diagram of an intelligent home control system provided in an embodiment of the present application. As shown in fig. 3, the smart home control system may include a terminal, a control terminal, a local database, a user voiceprint acquisition module, a user voiceprint comparison module, a user head portrait acquisition module, a user head portrait comparison module, a voice instruction recognition module, and a smart home interface gateway module.
Optionally, the local database above may include, for example, a voiceprint library and an avatar library.
It should be understood that the method 200 described in FIG. 2 may be applied to the system shown in FIG. 3, and the modules in FIG. 3 may be used to complete the method described in FIG. 2.
Fig. 4 is a schematic block diagram of an intelligent home control device provided in an embodiment of the present application. As shown in fig. 4, the apparatus 400 may include: a processing module 410 and a communication module 420.
The processing module 410 may be configured to invoke a preconfigured virtual digital person to conduct a video call with a user of the terminal in response to the call request.
The communication module 420 may be configured to receive a call request from a terminal, where the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used to request remote control of the smart home; and the method can also be used for acquiring the control instruction of the user to the intelligent home based on the video call.
Optionally, the communication module 420 may be further configured to prompt the user for an identification through the virtual digital person.
Optionally, the processing module 410 may be further configured to determine whether the user is a registered user based on the identity; and under the condition that the user is a registered user, performing identity verification on the user based on the voiceprint and the face of the user.
Optionally, the communication module 420 may be further configured to prompt the user to perform voiceprint registration and face registration if the user is an unregistered user.
Optionally, the identity is a mobile phone number of the user.
Optionally, the communication module 420 may be further configured to collect a voiceprint and a face of the user based on the video call.
Optionally, the processing module 410 may be further configured to perform identity verification on the user according to the voiceprint recognition and the face recognition, and the pre-stored voiceprint and face of the user.
Optionally, the communication module 420 may be further configured to query, by the virtual digital person, a home address of the user and a product model of a smart home; and acquiring the voiceprint and the face of the user based on the video call.
Optionally, the processing module 410 may be further configured to store the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
Optionally, the communication module 420 may be further configured to obtain a voice instruction of the user based on the video call.
Optionally, the processing module 410 may be further configured to perform voice recognition on the voice command to determine the smart home that the user wishes to control and the control operation of the smart home.
Optionally, the processing module 410 may also be configured to execute the control instructions.
Optionally, the communication module 420 may be further configured to feed back the completion condition of the control instruction to the user through the virtual digital person after the control instruction is completed.
Optionally, the call request and the video call are carried on resources of an LTE communication system or resources of a 5G communication system.
Fig. 5 is another schematic block diagram of an intelligent home control device provided in an embodiment of the present application. The device can be used for realizing the functions of the processing module and the communication module in the method. Wherein the apparatus may be a system-on-a-chip. In the embodiment of the present application, the chip system may be composed of a chip, and may also include a chip and other discrete devices.
As shown in fig. 5, the apparatus 500 may include at least one processor 510. Illustratively, the processor 510 may be configured to invoke a preconfigured virtual digital person to conduct a video call with the user of the terminal in response to the call request. For details, reference is made to the detailed description in the method example, which is not repeated herein.
The apparatus 500 may also include at least one memory 520 for storing program instructions and/or data. The memory 520 is coupled to the processor 510. The coupling in the embodiments of the present application is an indirect coupling or a communication connection between devices, units or modules, and may be an electrical, mechanical or other form for information interaction between the devices, units or modules. The processor 510 may operate in conjunction with the memory 520. Processor 510 may execute program instructions stored in memory 520. At least one of the at least one memory may be included in the processor.
The apparatus 500 may also include a communication interface 530 for communicating with other devices over a transmission medium so that the apparatus used in the apparatus 500 may communicate with other devices. The communication interface 530 may be, for example, a transceiver, an interface, a bus, a circuit, or a device capable of performing a transceiving function. The processor 510 may utilize the communication interface 530 to receive and transmit data and/or information, and is configured to implement a smart home control method described in the corresponding embodiment of fig. 2.
The specific connection medium between the processor 510, the memory 520 and the communication interface 530 is not limited in the embodiments of the present application. In fig. 5, the processor 510, the memory 520, and the communication interface 530 are connected by a bus 540. The bus 540 is shown in fig. 5 by a thick line, and the connection between other components is merely illustrative and not intended to be limiting. The bus may be divided into an address bus, a data bus, a control bus, etc. For ease of illustration, only one thick line is shown in FIG. 5, but this is not intended to represent only one bus or type of bus.
The present application further provides a computer program product, the computer program product comprising: a computer program (also referred to as code, or instructions), which when executed, causes an electronic device to perform the method of the embodiment shown in fig. 2.
The present application also provides a computer-readable storage medium having stored thereon a computer program (also referred to as code, or instructions). The computer program, when executed, causes the electronic device to perform the method in the embodiment shown in fig. 2.
It should be understood that the processor in the embodiments of the present application may be an integrated circuit chip having signal processing capability. In implementation, the steps of the above method embodiments may be performed by integrated logic circuits of hardware in a processor or instructions in the form of software. The processor may be a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic device, or discrete hardware components. The various methods, steps, and logic blocks disclosed in the embodiments of the present application may be implemented or performed. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like. The steps of the method disclosed in connection with the embodiments of the present application may be directly implemented by a hardware decoding processor, or implemented by a combination of hardware and software modules in the decoding processor. The software modules may be located in ram, flash, rom, prom, or eprom, registers, etc. as is well known in the art. The storage medium is located in a memory, and a processor reads information in the memory and completes the steps of the method in combination with hardware of the processor.
It will also be appreciated that the memory in the embodiments of the subject application can be either volatile memory or nonvolatile memory, or can include both volatile and nonvolatile memory. The non-volatile memory may be a read-only memory (ROM), a Programmable ROM (PROM), an Erasable PROM (EPROM), an electrically Erasable EPROM (EEPROM), or a flash memory. Volatile memory can be Random Access Memory (RAM), which acts as external cache memory. By way of example, but not limitation, many forms of RAM are available, such as Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), Synchronous Dynamic Random Access Memory (SDRAM), double data rate SDRAM, enhanced SDRAM, SLDRAM, Synchronous Link DRAM (SLDRAM), and direct rambus RAM (DR RAM). It should be noted that the memory of the systems and methods described herein is intended to comprise, without being limited to, these and any other suitable types of memory.
As used in this specification, the terms "unit," "module," and the like are intended to refer to a computer-related entity, either hardware, firmware, a combination of hardware and software, or software in execution.
Those of ordinary skill in the art will appreciate that the various illustrative logical blocks and steps (step) described in connection with the embodiments disclosed herein may be implemented as electronic hardware or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application. In the several embodiments provided in the present application, it should be understood that the disclosed apparatus, device and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the units is only one logical division, and other divisions may be realized in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
The units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one position, or may be distributed on multiple network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit.
In the above embodiments, the functions of the functional units may be fully or partially implemented by software, hardware, firmware, or any combination thereof. When implemented in software, may be implemented in whole or in part in the form of a computer program product. The computer program product includes one or more computer instructions (programs). The procedures or functions described in accordance with the embodiments of the present application are generated in whole or in part when the computer program instructions (programs) are loaded and executed on a computer. The computer may be a general purpose computer, a special purpose computer, a network of computers, or other programmable device. The computer instructions may be stored in a computer readable storage medium or transmitted from one computer readable storage medium to another computer readable storage medium, for example, the computer instructions may be transmitted from one website, computer, server, or data center to another website, computer, server, or data center via wired (e.g., coaxial cable, fiber optic, Digital Subscriber Line (DSL)) or wireless (e.g., infrared, wireless, microwave, etc.) means. The computer-readable storage medium can be any available medium that can be accessed by a computer or a data storage device, such as a server, a data center, etc., that incorporates one or more of the available media. The usable medium may be a magnetic medium (e.g., floppy disk, hard disk, magnetic tape), an optical medium (e.g., DVD), or a semiconductor medium (e.g., Solid State Disk (SSD)), among others.
The functions, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, the technical solution of the present application or portions thereof that substantially contribute to the prior art may be embodied in the form of a software product stored in a storage medium and including instructions for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: various media capable of storing program codes, such as a usb disk, a removable hard disk, a read-only memory (ROM), a Random Access Memory (RAM), a magnetic disk, or an optical disk.
The above description is only for the specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present application, and shall be covered by the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (11)

1. The intelligent household control method is characterized by comprising the following steps:
receiving a call request from a terminal, wherein the call request is a request sent by a user of the terminal by dialing a preset number, and the call request is used for requesting remote control of the smart home;
responding to the call request, and calling a pre-configured virtual digital person to carry out video call with the user of the terminal;
and acquiring a control instruction of the user to the intelligent home based on the video call.
2. The method of claim 1, wherein before the obtaining of the control instruction of the user for the smart home based on the video call, the method further comprises: prompting the user to input an identity through the virtual digital person;
determining whether the user is a registered user based on the identity;
under the condition that the user is a registered user, performing identity verification on the user based on the voiceprint and the face of the user; or
And prompting the user to perform voiceprint registration and face registration under the condition that the user is an unregistered user.
3. The method of claim 2, wherein the identity is a cell phone number of the user.
4. The method of claim 2, wherein said authenticating the user comprises:
collecting voiceprints and faces of the user based on the video call;
and carrying out identity verification on the user according to the voiceprint recognition, the face recognition, the prestored voiceprint and face of the user.
5. The method of claim 2, wherein the method further comprises:
inquiring the home address of the user and the product model of the smart home through the virtual digital person;
acquiring a voiceprint and a face of the user based on the video call;
and storing the home address of the user, the product model of the smart home, the voiceprint and the face of the user in a local database.
6. The method of claim 1, wherein the method further comprises:
executing the control instruction;
and after the control instruction is finished, feeding back the finishing condition of the control instruction to the user through the virtual digital person.
7. The method of any of claims 1 to 6, wherein the call request and the video call are carried on resources of a Long Term Evolution (LTE) communication system or resources of a fifth generation 5G communication system.
8. An intelligent home control apparatus, comprising means for performing the method of any of claims 1-7.
9. An intelligent home control device, comprising a processor configured to perform the method of any one of claims 1 to 7.
10. A computer-readable storage medium, characterized in that the storage medium stores instructions that, when executed on a computer, cause the computer to perform the method of any of claims 1 to 7.
11. A computer program product, comprising a computer program which, when executed, causes a computer to perform the method of any one of claims 1 to 7.
CN202110204624.5A 2021-02-24 2021-02-24 Smart home control method and device Active CN114979543B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110204624.5A CN114979543B (en) 2021-02-24 2021-02-24 Smart home control method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110204624.5A CN114979543B (en) 2021-02-24 2021-02-24 Smart home control method and device

Publications (2)

Publication Number Publication Date
CN114979543A true CN114979543A (en) 2022-08-30
CN114979543B CN114979543B (en) 2024-07-02

Family

ID=82972710

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110204624.5A Active CN114979543B (en) 2021-02-24 2021-02-24 Smart home control method and device

Country Status (1)

Country Link
CN (1) CN114979543B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105978775A (en) * 2016-07-29 2016-09-28 镇江惠通电子有限公司 Speech control system and speech control method
CN106097495A (en) * 2016-06-03 2016-11-09 赵树龙 A kind of intelligent voice control vocal print face authentication door access control system and method
CN106782559A (en) * 2016-12-31 2017-05-31 广东博意建筑设计院有限公司 Smart home house keeper central control system and its control method with telecommunication control
CN106792238A (en) * 2016-11-17 2017-05-31 Tcl集团股份有限公司 A kind of interactive approach and interaction systems based on live telecast shopping platform
CN107092196A (en) * 2017-06-26 2017-08-25 广东美的制冷设备有限公司 The control method and relevant device of intelligent home device
WO2018113526A1 (en) * 2016-12-20 2018-06-28 四川长虹电器股份有限公司 Face recognition and voiceprint recognition-based interactive authentication system and method
CN109412910A (en) * 2018-11-20 2019-03-01 三星电子(中国)研发中心 The method and apparatus for controlling smart home device
CN110248018A (en) * 2018-03-09 2019-09-17 深圳市云昱科技有限公司 The call method and Related product of intelligent secretary
CN110647636A (en) * 2019-09-05 2020-01-03 深圳追一科技有限公司 Interaction method, interaction device, terminal equipment and storage medium
CN112291331A (en) * 2020-10-25 2021-01-29 湖南云脸智联科技有限公司 Cloud access control implementation method based on AI intelligent sound box application
CN112291497A (en) * 2020-10-28 2021-01-29 上海赛连信息科技有限公司 Intelligent video customer service access method and device

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106097495A (en) * 2016-06-03 2016-11-09 赵树龙 A kind of intelligent voice control vocal print face authentication door access control system and method
CN105978775A (en) * 2016-07-29 2016-09-28 镇江惠通电子有限公司 Speech control system and speech control method
CN106792238A (en) * 2016-11-17 2017-05-31 Tcl集团股份有限公司 A kind of interactive approach and interaction systems based on live telecast shopping platform
WO2018113526A1 (en) * 2016-12-20 2018-06-28 四川长虹电器股份有限公司 Face recognition and voiceprint recognition-based interactive authentication system and method
CN106782559A (en) * 2016-12-31 2017-05-31 广东博意建筑设计院有限公司 Smart home house keeper central control system and its control method with telecommunication control
CN107092196A (en) * 2017-06-26 2017-08-25 广东美的制冷设备有限公司 The control method and relevant device of intelligent home device
CN110248018A (en) * 2018-03-09 2019-09-17 深圳市云昱科技有限公司 The call method and Related product of intelligent secretary
CN109412910A (en) * 2018-11-20 2019-03-01 三星电子(中国)研发中心 The method and apparatus for controlling smart home device
CN110647636A (en) * 2019-09-05 2020-01-03 深圳追一科技有限公司 Interaction method, interaction device, terminal equipment and storage medium
CN112291331A (en) * 2020-10-25 2021-01-29 湖南云脸智联科技有限公司 Cloud access control implementation method based on AI intelligent sound box application
CN112291497A (en) * 2020-10-28 2021-01-29 上海赛连信息科技有限公司 Intelligent video customer service access method and device

Also Published As

Publication number Publication date
CN114979543B (en) 2024-07-02

Similar Documents

Publication Publication Date Title
US20220317641A1 (en) Device control method, conflict processing method, corresponding apparatus and electronic device
CN107644638B (en) Audio recognition method, device, terminal and computer readable storage medium
CN107220532B (en) Method and apparatus for recognizing user identity through voice
US12010108B2 (en) Techniques to provide sensitive information over a voice connection
CN107544271A (en) Terminal control method, device and computer-readable recording medium
CN108470568B (en) Intelligent device control method and device, storage medium and electronic device
JP6738867B2 (en) Speaker authentication method and voice recognition system
JP6783339B2 (en) Methods and devices for processing audio
JP2017529711A (en) Method, apparatus and system for voice confirmation
CN105872619A (en) Video playing record matching method and matching device
CN107544272A (en) terminal control method, device and storage medium
CN107205097A (en) Mobile terminal lookup method, device and computer-readable recording medium
CN110858841B (en) Electronic device and method for registering new user through authentication of registered user
CN107808662B (en) Method and device for updating grammar rule base for speech recognition
CN111653284B (en) Interaction and identification method, device, terminal equipment and computer storage medium
CN111736474A (en) Multi-mode intelligent home control system and method
CN109087647A (en) Application on Voiceprint Recognition processing method, device, electronic equipment and storage medium
CN110224904B (en) Voice processing method, device, computer readable storage medium and computer equipment
CN114979543A (en) Intelligent household control method and device
CN108399918B (en) Intelligent device connection method, intelligent device and terminal
CN112417923A (en) System, method and apparatus for controlling smart devices
CN112738344B (en) Method and device for identifying user identity, storage medium and electronic equipment
Dumitrescu Cassandra smart-home system description
CN112449059A (en) Voice interaction device, method and system for realizing call based on voice interaction device
CN111091638A (en) Storage medium, intelligent door lock and authentication method thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant