CN117135266A - Information processing method, device and computer readable storage medium - Google Patents

Information processing method, device and computer readable storage medium Download PDF

Info

Publication number
CN117135266A
CN117135266A CN202311390744.4A CN202311390744A CN117135266A CN 117135266 A CN117135266 A CN 117135266A CN 202311390744 A CN202311390744 A CN 202311390744A CN 117135266 A CN117135266 A CN 117135266A
Authority
CN
China
Prior art keywords
voice
information
preset
feature
current call
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311390744.4A
Other languages
Chinese (zh)
Other versions
CN117135266B (en
Inventor
沈卫东
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
TCL Communication Technology Chengdu Ltd
Original Assignee
TCL Communication Technology Chengdu Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by TCL Communication Technology Chengdu Ltd filed Critical TCL Communication Technology Chengdu Ltd
Priority to CN202311390744.4A priority Critical patent/CN117135266B/en
Publication of CN117135266A publication Critical patent/CN117135266A/en
Application granted granted Critical
Publication of CN117135266B publication Critical patent/CN117135266B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72448User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions
    • H04M1/72454User interfaces specially adapted for cordless or mobile telephones with means for adapting the functionality of the device according to specific conditions according to context-related or environment-related conditions
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/72418User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting emergency services
    • H04M1/72421User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality for supporting emergency services with automatic activation of emergency service functions, e.g. upon sensing an alarm
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones

Landscapes

  • Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Business, Economics & Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Physics & Mathematics (AREA)
  • Emergency Management (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Environmental & Geological Engineering (AREA)
  • Telephone Function (AREA)

Abstract

The embodiment of the application discloses an information processing method, an information processing device and a computer readable storage medium. According to the embodiment of the application, when the conversation state is detected, the current conversation voice is obtained; extracting the characteristics of the current call voice, and extracting the voice characteristics of the current call voice; and when the voice features are matched with the preset voice features, generating prompt information, and sending the prompt information to the preset contact. Therefore, in the call process, the current call voice can be automatically acquired for analysis to obtain corresponding voice characteristics, when the voice characteristics are matched with the preset voice characteristics, the user is in a special situation scene, prompt information is automatically generated for transmission under the noninductive operation, the user operation is not needed, the information processing efficiency is improved, and the user is helped to solve the problem.

Description

Information processing method, device and computer readable storage medium
Technical Field
The present application relates to the field of prompt technologies, and in particular, to an information processing method, an information processing device, and a computer readable storage medium.
Background
Along with the rapid development of terminal technology and communication technology, the functions of the terminal are more and more powerful, and the terminal can solve the demands of people in various aspects such as conversation, social contact, photographing, entertainment, office work and the like, and is an indispensable product in daily life, work and study of people.
When a user encounters a special situation, the user cannot contact a family or other friends by using a telephone, or the user receives an incoming call from the family or friends, but cannot clearly inform a calling party due to the special state, the user is in the special situation and needs to seek help, and the device cannot actively identify the user state and actively provide important information to the family or friends so as to help the user to solve the problem.
Disclosure of Invention
The embodiment of the application provides an information processing method, an information processing device and a computer readable storage medium, which can improve the information processing efficiency and further help users to solve the problems.
In order to solve the technical problems, the embodiment of the application provides the following technical scheme:
an information processing method, comprising:
when the call state is detected, acquiring current call voice;
extracting the characteristics of the current call voice, and extracting the voice characteristics of the current call voice;
and when the voice features are matched with the preset voice features, generating prompt information, and sending the prompt information to the preset contact.
An information processing apparatus comprising:
the acquisition unit is used for acquiring current call voice when the call state is detected;
The extracting unit is used for extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice;
and the sending unit is used for generating prompt information when the voice characteristics are matched with the preset voice characteristics and sending the prompt information to the preset contact person.
In some embodiments, the extraction unit is configured to:
extracting the speech speed characteristics of the current call voice as voice characteristics; or (b)
Extracting volume characteristics of the current call voice as voice characteristics; or (b)
And extracting the tone quality characteristics of the current call voice as voice characteristics.
In some embodiments, the preset speech features include at least a first preset speech sub-feature;
the sending unit is configured to:
when the voice feature is matched with the first preset voice sub-feature, position information is obtained and first reminding information is generated;
starting a camera to shoot, and obtaining image information;
and sending the position information, the image information and the first reminding information to a preset contact person, a current caller and a preset reminder as reminding information.
In some embodiments, the preset speech features further comprise at least a second preset speech sub-feature;
the sending unit is configured to:
when the voice features are matched with the second preset voice sub-features, generating second reminding information and starting timing;
collecting environment information corresponding to preset configuration information;
when the timing end is detected, acquiring first target position information when the timing end is detected, and sending the first target position information, the environment information and the second reminding information to a preset contact person and a preset prompt person as prompt information.
In some embodiments, the apparatus further comprises a shutdown sending unit configured to:
when a shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated is obtained, and the second target position information, the environment information and the reminding information are used as prompt information and sent to a preset contact person and a preset prompt person.
In some embodiments, the preset speech features further comprise at least a third preset speech sub-feature;
the sending unit is configured to:
when the voice feature is matched with the third preset voice sub-feature, generating third reminding information and starting timing;
When the timing end is detected, acquiring third target position information when the timing end is detected, and sending the third target position information and the third reminding information to a preset contact person together as prompt information.
In some embodiments, the apparatus further comprises a cancellation transmitting unit configured to:
and when a cancel instruction is generated in the timing process, ending the timing and canceling the sending of the prompt information.
A computer readable storage medium storing a plurality of instructions adapted to be loaded by a processor to perform the steps of the information processing method described above.
A computer device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, the processor implementing the steps in the information processing method provided above when the computer program is executed.
A computer program product or computer program comprising computer instructions stored in a storage medium, the computer instructions being read from the storage medium by a processor of a computer device, the computer instructions being executed by the processor such that the computer device performs the steps in the information processing method provided above.
The application has the beneficial effects that when the current call voice is obtained when the call state is detected; extracting the characteristics of the current call voice, and extracting the voice characteristics of the current call voice; and when the voice features are matched with the preset voice features, generating prompt information, and sending the prompt information to the preset contact. Therefore, in the call process, the current call voice can be automatically acquired for analysis to obtain corresponding voice characteristics, when the voice characteristics are matched with the preset voice characteristics, the user is in a special situation scene, prompt information is automatically generated for transmission under the noninductive operation, the user operation is not needed, the information processing efficiency is improved, and the user is helped to solve the problem.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings that are needed in the description of the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a schematic diagram of a scenario of an information processing system provided by an embodiment of the present application;
FIG. 2 is a schematic flow chart of an information processing method according to an embodiment of the present application;
FIG. 3 is another flow chart of an information processing method according to an embodiment of the present application;
fig. 4 is a schematic structural view of an information processing apparatus according to an embodiment of the present application;
fig. 5 is a schematic structural diagram of a terminal according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to fall within the scope of the application.
The embodiment of the application provides an information processing method, an information processing device and a computer readable storage medium.
Referring to fig. 1, fig. 1 is a schematic view of a scenario of an information processing system according to an embodiment of the present application, including: the terminal A and the server B can be connected through a communication network, wherein the communication network comprises a wireless network and a wired network, and the wireless network comprises one or a combination of a plurality of wireless wide area networks, wireless local area networks, wireless metropolitan area networks and wireless personal networks. The network includes network entities such as routers, gateways, etc., which are not shown. The terminal a may interact with the server B via a communication network, for example, the terminal a may send a prompt message to the server B, and the server B may forward the prompt message to a different terminal a.
The information processing system may include an information processing device, which may be specifically integrated in a terminal having a storage unit and a microprocessor installed therein and having an operation capability, such as a tablet computer, a mobile phone, a notebook computer, a desktop computer, etc., in fig. 1, the terminal is a terminal a in fig. 1, and various application software required by a user, such as a call application software, a recording application software, or an instant messaging application software, etc., may be installed in the terminal a. The terminal A can be used for acquiring current conversation voice when detecting that the terminal A is in a conversation state; extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice; when the voice features are matched with the preset voice features, prompting information is generated and sent to the preset contact person.
The server B can be an independent physical server, a server cluster or a distributed system formed by a plurality of physical servers, and can also be a cloud server for providing cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, CDNs, basic cloud computing services such as big data and artificial intelligent platforms and the like and forwarding prompt information.
It should be noted that, the schematic view of the scenario of the information processing system shown in fig. 1 is only an example, and the information processing system and the scenario described in the embodiment of the present application are for more clearly describing the technical solution of the embodiment of the present application, and do not constitute a limitation on the technical solution provided by the embodiment of the present application, and those skilled in the art can know that, with the evolution of the information processing system and the appearance of a new service scenario, the technical solution provided by the embodiment of the present application is equally applicable to similar technical problems.
The following will describe in detail.
The embodiment of the application provides an information processing method which can be executed by a terminal.
Referring to fig. 2, fig. 2 is a flowchart illustrating an information processing method according to an embodiment of the application. The information processing method comprises the following steps:
in step 101, when it is detected that the call is in a call state, a current call voice is acquired.
When the call state is detected, in order to detect whether a user of a machine owner encounters a special condition, current call voice can be obtained in real time, the current call voice is generated by communication between the user of the machine owner and a call party, in one embodiment, when the call state is detected, a terminal can automatically start a recording function, record the current call voice, and the recording function needs to be described, in order to prevent other people from perceiving the special condition brought by the recording to the user, the recording is started to be a background noninductive operation, and no perception can be performed on a foreground operation interface.
It will be appreciated that in the specific embodiments of the present application, related data such as current call voice is involved, and when the above embodiments of the present application are applied to specific products or technologies, user permissions or consents need to be obtained, and the collection, use and processing of related data need to comply with related regulations and standards.
In step 102, feature extraction is performed on the current speech, and speech features of the current speech are extracted.
The feature extraction may be performed on the current call voice, for example, the speech speed feature extraction, the volume feature extraction, or the tone quality feature extraction, and the speech speed, the volume, and the tone quality are similar when the user speaks in the normal scene, so that the speech speed, the volume, or the tone quality different from the normal scene may be prerecorded, and whether the owner user is in the special situation may be subsequently determined based on the feature extraction.
Based on this, the voice feature of the current call voice can be a speech speed feature, a volume feature or a tone quality feature, so that whether the user is in a special condition can be rapidly determined according to the voice feature, and the user is prompted under the condition that the user cannot perceive the voice feature.
In some embodiments, the extracting the features of the current call voice, extracting the voice features of the current call voice, includes:
(1) Extracting the speech speed characteristics of the current call voice as voice characteristics; or (b)
(2) Extracting volume characteristics of the current call voice as voice characteristics; or (b)
(3) And extracting the tone quality characteristics of the current call voice as voice characteristics.
The speech speed of the user is in a stable range when the user speaks in a normal scene, so that the user can record the speech speed different from the speech speed of the normal scene in advance, for example, record a preset speech with the speech speed faster than the speech speed when the user speaks in the normal scene, so that the speech speed characteristic extraction can be carried out on the current call speech, the speech speed characteristic of the current call speech is extracted as a speech characteristic, and whether the owner user is in a special condition can be judged according to the speech speed characteristic.
Because the volume is in a stable range when the user speaks in the normal scene, the user can record the preset voice different from the volume of the normal scene in advance, for example, record the volume larger than the volume when the user speaks in the normal scene, so that the volume characteristic extraction can be carried out on the current talking voice, the volume characteristic of the current talking voice is extracted as the voice characteristic, and whether the owner user is in the special condition can be judged according to the volume characteristic.
Because the tone quality is in a stable range when the user speaks in the normal scene, the user can record the tone quality different from the normal scene in advance, for example, record the preset voice with more sharp tone quality than the tone quality when the user speaks in the normal scene, thereby extracting the tone quality characteristic of the current talking voice as the voice characteristic, and then judging whether the owner user is in the special condition according to the tone quality characteristic.
In step 103, when the voice feature matches with the preset voice feature, a prompt message is generated and sent to the preset contact.
The preset voice features are corresponding to preset voice features recorded by the user in advance, the preset voice features are features different from voices speaking in normal scenes, based on the preset voice features, the user can speak voices matched with the preset voice features in the process of speaking, for example, voice with more sharp tone quality is spoken, after the voice is recorded as current speaking voice, the voice features are matched with the preset voice features, and the terminal can sense that the user is in special conditions under the condition of no operation.
Further, in order to achieve the purpose of helping the user of the machine owner to separate from the special situation, prompt information, for example, "the user is in the special situation currently, please catch up with me", can be automatically generated, and the prompt information is sent to the preset contact person.
In one embodiment, three kinds of scenes, namely a high-level scene, a medium-level scene and a low-level scene, can be set, wherein the high-level scene represents that special conditions are met currently, and the special conditions exist and immediate help is needed; the middle-level scene represents that the current situation is not in an extreme special situation, the user feels that the user is in the special situation, but the user has certain processing/rotation capacity and needs to wait for observation to finally decide to send prompt information; or the user is inconvenient to send the prompt information to the outside at present and needs to wait for a period of time to send the prompt information; the low level degree represents just the object of interest or scene of interest, but it takes time to observe for final confirmation.
That is, the user of the mobile phone can set corresponding preset voices in three scenes in advance, for example, a first preset voice is a high-level scene, the first preset voice comprises a first preset voice sub-feature, a second preset voice is a middle-level scene, the second preset voice comprises a second preset voice sub-feature, a third preset voice is a low-level scene, and the third preset voice comprises a third preset voice sub-feature.
Based on this, in some embodiments, when the voice feature matches a preset voice feature, generating a prompt message, and sending the prompt message to a preset contact, including:
(1) When the voice feature is matched with the first preset voice sub-feature, acquiring position information and generating first reminding information;
(2) Starting a camera to shoot, and obtaining image information;
(3) And sending the position information, the image information and the first reminding information to a preset contact person, a current caller and a preset reminder as reminding information.
The terminal can compare the voice characteristic of the current voice of the recorded call with the first preset voice sub-characteristic, the second preset voice sub-characteristic and the third preset voice sub-characteristic stored before, and when the voice characteristic is detected to be matched with the first preset voice sub-characteristic, the user of the machine owner is in the high-level scene, so that the terminal can automatically acquire the current position and generate first reminding information, for example, the first reminding information can be in a very special state, please immediately help me, and the terminal can automatically give permission and automatically acquire the position information in a background without an interface when the voice characteristic is detected to be matched with the first preset voice sub-characteristic.
Furthermore, the camera can be started to shoot under the conditions of no flash and no sound in the background to obtain image information, and in one embodiment, the terminal is generally provided with front and rear cameras, so that the front and rear cameras can shoot in a wheel flow mode to obtain images of the front and rear cameras.
The preset contact person is a contact person preset by a user of the machine owner, such as a friend or a relative, the current call person is a current call object, the preset prompt person can be a help mechanism, and finally, the position information, the image information and the first prompt information are sent to the preset contact person, the current call person and the preset prompt person as prompt information, the whole process is automatically executed under the noninductive operation, key information is acquired, and the user can be better helped to solve the problem in the improvement of the information processing efficiency.
In some embodiments, when the voice feature matches a preset voice feature, generating a prompt message and sending the prompt message to a preset contact, including:
(1.1) when the voice feature matches the second preset voice sub-feature, generating a second reminder and starting timing;
(1.2) collecting environment information corresponding to preset configuration information;
And (1.3) when the timing end is detected, acquiring first target position information at the timing end, and sending the first target position information, the environment information and the second reminding information to a preset contact person and a preset reminder as prompt information.
The terminal can compare the voice characteristic of the current talking voice with the first preset voice sub-characteristic, the second preset voice sub-characteristic and the third preset voice sub-characteristic stored before, and when the voice characteristic is detected to be matched with the second preset voice sub-characteristic, the user is indicated to be in the middle-level scene, so that the terminal can automatically generate second reminding information, for example, the second reminding information can be in a state of 'present me in special condition, please help me', and simultaneously start timing, and the timing time can be preset, for example, 20 seconds.
Further, the user may configure preset configuration information in the middle-level scene, for example, preset configuration information such as setting photographing, acquiring position information, etc., the terminal may perform environmental information acquisition according to the preset configuration information, for example, setting photographing, and then the terminal may automatically acquire an image corresponding to photographing as the environmental information.
The preset contact person is a contact person preset by a user of the machine owner, such as a friend or a relative, the preset prompt person can be a mechanism capable of providing assistance, such as an assistance mechanism, finally, when the timing end is detected, in order to better help the user to solve the problem, first target position information of the timing end can be acquired, and it is required to be noted that authority can be automatically given when the first target position information is acquired, and the first target position information is automatically acquired in a background without an interface. Therefore, the first target position information, the environment information and the second reminding information are sent to the preset contact person and the preset reminder as the reminding information, the whole process is automatically executed under the noninductive operation, the environment information and the position information corresponding to the preset configuration information which are configured in advance are automatically acquired, the information processing efficiency is improved, and the user can be better helped to solve the problem.
In an embodiment, the user may cancel the timing by manually operating the timer before the timing is finished, and then the prompt message is not sent later.
In some embodiments, the method further comprises: when the shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated is obtained, and the second target position information, the environment information and the reminding information are used as the reminding information and sent to the preset contact person and the preset reminder.
The shutdown instruction is an instruction generated before the terminal is shutdown, that is, in the countdown process, if other people shutdown the terminal or automatically shutdown the terminal due to insufficient electric quantity, so that in order to help a user solve the problem, when the shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated can be acquired, and it is required to say that the permission is automatically given when the second target position information is acquired, and the terminal is automatically acquired in a background without an interface. And automatically sending the second target position information, the environment information and the reminding information to a preset contact person and a preset reminder as the reminding information. The problem that the user is in special conditions due to failure in sending prompt information caused by shutdown is avoided.
In some embodiments, when the voice feature matches a preset voice feature, generating a prompt message and sending the prompt message to a preset contact, including:
(2.1) when the voice feature matches the third preset voice sub-feature, generating a third reminding message and starting timing;
and (2.2) when the timing end is detected, acquiring third target position information at the timing end, and sending the third target position information and third reminding information to a preset contact person as prompt information.
The terminal can compare the voice characteristic of the current voice of the recorded call with the first preset voice sub-characteristic, the second preset voice sub-characteristic and the third preset voice sub-characteristic stored before, and when the voice characteristic is detected to be matched with the third preset voice sub-characteristic, the user of the machine owner is in the low-level scene, so that the terminal can automatically generate third reminding information, for example, the third reminding information can be in a state of ' probably being in special conditions of the current me ', probably needing help me ', and simultaneously starts timing, and the timing time can be preset, for example, 30 seconds.
Furthermore, the preset contact is a contact preset by the owner user, such as a friend or a relative, and finally, when the timing end is detected, in order to better help the user to solve the problem, third target position information of the timing end can be obtained. Therefore, the third target position information and the third reminding information are sent to the preset contact person together as the reminding information, the whole process is automatically executed under the noninductive operation, the preset position information is automatically acquired, and the problem can be better solved by the aid of a user in improving the information processing efficiency.
In some embodiments, the method further comprises: and when a cancel instruction is generated in the timing process, ending the timing and canceling the sending of the prompt information.
The user can cancel the timing of the prompt information by manual operation before the timing is finished, namely, a cancel instruction is generated by manually canceling the prompt information, and the sending of the prompt information is canceled after the timing is finished.
As can be seen from the above, in the embodiment of the present application, when a call state is detected, a current call voice is obtained; extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice; when the voice features are matched with the preset voice features, prompting information is generated and sent to the preset contact person. Therefore, in the call process, the current call voice can be automatically acquired for analysis to obtain corresponding voice characteristics, when the voice characteristics are matched with the preset voice characteristics, the user is in a special situation scene, prompt information is automatically generated for transmission under the noninductive operation, the user operation is not needed, the information processing efficiency is improved, and the user is helped to solve the problem.
In this embodiment, a description will be given of an example in which the information processing apparatus is specifically integrated in a terminal, and a description will be further given with specific reference to the following embodiments.
Referring to fig. 3, fig. 3 is another flow chart of the information processing method according to the embodiment of the application. The method flow may include:
in step 201, when the terminal detects that it is in a call state, a current call voice is acquired.
In the embodiment of the application, taking the voice characteristic as the tone characteristic as an example for explanation, a machine owner user can also set three scenes, namely a high-level scene, a middle-level scene and a low-level scene, wherein the high-level scene represents that special conditions are met at present, and the machine owner user needs immediate help in the case of extremely special conditions; the middle-level scene represents that the current scene is not in a high-level scene, the user feels that the user is in a special condition, but the user has certain processing/rotation capacity and needs to wait for observation to finally decide to send prompt information; or the user is inconvenient to send the prompt information to the outside at present and needs to wait for a period of time to give the prompt information; the low level degree represents just the object of interest or scene of interest, but it takes time to observe for final confirmation.
And it should be noted that default actions of the terminal can be automatically set, for example, flash is automatically turned off when photographing, a speaker of the mobile phone is automatically prohibited when recording, permission is automatically given when position information is acquired, and the device is automatically in a non-interface background when executing the actions.
That is, the user of the mobile phone may set corresponding preset voices in three scenes in advance, for example, the first preset voice is a high-level scene, the first preset voice includes a first preset voice sub-feature formed by the first tone quality, the second preset voice is a middle-level scene, the second preset voice includes a second preset voice sub-feature formed by the second tone quality, the third preset voice is a low-level scene, the third preset voice includes a third preset voice sub-feature formed by the third tone quality, the first tone quality, the second tone quality and the third tone quality are different, and in particular, the tone quality sharpness of the first tone quality is smaller than that of the second tone quality, and the tone quality sharpness of the second tone quality is smaller than that of the third tone quality.
Based on this, when the terminal detects that the terminal is in a call state, in order to detect whether the owner user encounters a special situation, the current call voice can be obtained in real time, and the current call voice is generated by the communication between the owner user and the call party.
In step 202, the terminal performs voice quality feature extraction on the current call voice, and extracts the voice quality feature of the current call voice as a voice feature.
The voice quality is in a stable range when the user speaks in a normal scene, so that the user can record the first preset voice, the second preset voice and the third preset voice which are different from the three corresponding voice quality of the normal scene in advance, for example, record three preset voices with more sharp voice quality than the voice quality when the user speaks in the normal scene, thereby extracting the voice quality characteristics of the current talking voice, extracting the voice quality characteristics of the current talking voice as voice characteristics, and judging whether the user of the machine owner is in special conditions according to the voice quality characteristics.
In step 203, when the terminal detects that the voice feature matches with the first preset voice sub-feature, the terminal obtains the position information and generates first reminding information, starts the camera to shoot, obtains image information, and sends the position information, the image information and the first reminding information together as reminding information to the preset contact person, the current caller and the preset reminder.
The terminal can compare the tone quality characteristics of the current talking voice with the first preset voice sub-characteristics, the second preset voice sub-characteristics and the third preset voice sub-characteristics stored before, and when the voice characteristics are detected to be matched with the first preset voice sub-characteristics, the main user is indicated to be in the high-level scene, so that the terminal can automatically acquire the current position and generate first reminding information, for example, the first reminding information can be in a very special state, please immediately help me, and the terminal can automatically give permission and automatically acquire the position information in the background without an interface when acquiring the position information.
Furthermore, the camera can be started to shoot under the conditions of no flash and no sound in the background to obtain image information, and in one embodiment, the terminal is generally provided with front and rear cameras, so that the front and rear cameras can shoot in a wheel flow mode to obtain images of the front and rear cameras.
The preset contact person is a contact person preset by a user of a machine owner, such as a friend or a relative, the current call person is a current call object, the preset prompt person can be a mechanism which can provide help, such as a help mechanism, and finally, the position information, the image information and the first prompt information are sent to the preset contact person, the current call person and the preset prompt person together as prompt information, the whole process is automatically executed under the noninductive operation, key information is acquired, and the user can be better helped to solve the problem while the information processing efficiency is improved.
In step 204, when the terminal detects that the voice feature matches with the second preset voice sub-feature, generating second reminding information, starting timing, collecting environment information corresponding to the preset configuration information, when the terminal detects that the timing is finished, obtaining first target position information when the timing is finished, and sending the first target position information, the environment information and the second reminding information to the preset contact person and the preset reminder together as reminding information.
The terminal can compare the tone quality characteristic of the current talking voice with the first preset voice sub-characteristic, the second preset voice sub-characteristic and the third preset voice sub-characteristic stored before, and when the tone quality characteristic is detected to be matched with the second preset voice sub-characteristic, the user is in the middle-level scene, so that the terminal can automatically generate second reminding information, for example, the second reminding information can be in a special state for 'current me, please help me', and the timing is started at the same time, and the timing time can be preset, for example, 20 seconds.
Further, the user may configure preset configuration information in the middle-level scene, for example, preset configuration information such as setting photographing, acquiring position information, etc., the terminal may perform environmental information acquisition according to the preset configuration information, for example, setting photographing, and then the terminal may automatically acquire an image corresponding to photographing as the environmental information.
The preset contact person is a contact person preset by a user of the machine owner, such as a friend or a relative, the preset prompt person can be a mechanism capable of providing assistance, such as an assistance mechanism, finally, when the timing end is detected, in order to better help the user to solve the problem, first target position information of the timing end can be acquired, and it is required to be noted that authority can be automatically given when the first target position information is acquired, and the first target position information is automatically acquired in a background without an interface. Therefore, the first target position information, the environment information and the second reminding information are sent to the preset contact person and the preset reminder as the reminding information, the whole process is automatically executed under the noninductive operation, the environment information and the position information corresponding to the preset configuration information which are configured in advance are automatically acquired, the information processing efficiency is improved, and the user can be better helped to solve the problem.
In step 205, when the terminal detects that the shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated is obtained, and the second target position information, the environment information and the reminding information are used as the reminding information and sent to the preset contact person and the preset reminder.
The shutdown instruction is an instruction generated before the terminal is shutdown, that is, in the countdown process, if other people shutdown the terminal or automatically shutdown the terminal due to insufficient electric quantity, so that in order to help a user solve the problem, when the shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated can be acquired, and it is required to say that the permission is automatically given when the second target position information is acquired, and the terminal is automatically acquired in a background without an interface. And automatically sending the second target position information, the environment information and the reminding information to a preset contact person and a preset reminder as the reminding information. The problem that the user is in special conditions due to failure in sending prompt information caused by shutdown is avoided.
In step 206, when the terminal detects that the voice feature matches with the third preset voice sub-feature, third reminding information is generated, timing is started, when the timing is detected to be finished, third target position information at the timing is obtained, and the third target position information and the third reminding information are sent to the preset contact person together as prompt information.
In the call process, the user only needs to speak according to the set low-level scene, the terminal will compare the sound quality feature of the current call voice with the first preset voice sub-feature, the second preset voice sub-feature and the third preset voice sub-feature stored previously, and when the matching of the sound quality feature and the third preset voice sub-feature is detected, the user is in the low-level scene, so that the terminal can automatically generate third reminding information, for example, the third reminding information can be in a state of ' probably being in special condition of the current me ', probably needs to help me ', and starts timing at the same time, and the timing time can be preset, for example, 30 seconds.
Furthermore, the preset contact is a contact preset by the owner user, such as a friend or a relative, and finally, when the timing end is detected, in order to better help the user to solve the problem, third target position information of the timing end can be obtained. Therefore, the third target position information and the third reminding information are sent to the preset contact person together as the reminding information, the whole process is automatically executed under the noninductive operation, the preset position information is automatically acquired, and the problem can be better solved by the aid of a user in improving the information processing efficiency.
In step 207, when the terminal detects that a cancel instruction is generated in the timing process, the timing is ended, and the transmission of the prompt message is canceled.
The user can cancel the timing of the prompt information manually by ensuring that the environment is in a state of no special condition before the timing is finished, namely, a cancel instruction is generated by manually canceling the prompt information, so that when the terminal detects that the cancel instruction is generated in the timing process, the timing is finished, and the sending of the prompt information is canceled.
In order to facilitate better implementation of the information processing method provided by the embodiment of the application, the embodiment of the application also provides a device based on the information processing method. Where the meaning of a noun is the same as in the information processing method described above, specific implementation details may be referred to the description in the method embodiment.
Referring to fig. 4, fig. 4 is a schematic structural diagram of an information processing apparatus according to an embodiment of the present application, where the information processing apparatus is applied to a terminal and may include an acquisition unit 301, an extraction unit 302, a transmission unit 303, and the like.
An obtaining unit 301 is configured to obtain a current call voice when it is detected that the call is in a call state.
The extracting unit 302 is configured to perform feature extraction on the current voice, and extract a voice feature of the current voice.
In some embodiments, the extracting unit 302 is configured to:
extracting the speech speed characteristics of the current call voice as voice characteristics; or (b)
Extracting volume characteristics of the current call voice as voice characteristics; or (b)
And extracting the tone quality characteristics of the current call voice as voice characteristics.
And a sending unit 303, configured to generate a prompt message when the voice feature matches with a preset voice feature, and send the prompt message to a preset contact.
In some embodiments, the predetermined speech features include at least a first predetermined speech sub-feature;
the transmitting unit 303 is configured to:
when the voice feature is matched with the first preset voice sub-feature, acquiring position information and generating first reminding information;
starting a camera to shoot, and obtaining image information;
and sending the position information, the image information and the first reminding information to a preset contact person, a current caller and a preset reminder as reminding information.
In some embodiments, the preset speech feature further comprises at least a second preset speech sub-feature;
The transmitting unit 303 is configured to:
when the voice feature is matched with the second preset voice sub-feature, generating second reminding information and starting timing;
collecting environment information corresponding to preset configuration information;
when the timing end is detected, acquiring first target position information when the timing end is detected, and sending the first target position information, the environment information and the second reminding information to a preset contact person and a preset prompt person as prompt information.
In some embodiments, the apparatus further comprises a shutdown sending unit (not identified) for:
when the shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated is obtained, and the second target position information, the environment information and the reminding information are used as the reminding information and sent to the preset contact person and the preset reminder.
In some embodiments, the preset speech features further comprise at least a third preset speech sub-feature;
the transmitting unit 303 is configured to:
when the voice feature is matched with the third preset voice sub-feature, generating third reminding information and starting timing;
when the timing end is detected, acquiring third target position information when the timing end is detected, and sending the third target position information and third reminding information to a preset contact person together as prompt information.
In some embodiments, the apparatus further comprises a cancellation transmission unit (not identified) for:
and when a cancel instruction is generated in the timing process, ending the timing and canceling the sending of the prompt information.
In the foregoing embodiments, the descriptions of the embodiments are focused on, and the portions of an embodiment that are not described in detail in the foregoing embodiments may be referred to in the foregoing detailed description of the information processing method, which is not repeated herein.
The embodiment of the application also provides a computer device, which may be a terminal, as shown in fig. 5, which shows a schematic structural diagram of the terminal according to the embodiment of the application, specifically:
the computer device may include Radio Frequency (RF) circuitry 401, memory 402 including one or more computer readable storage media, input unit 403, display unit 404, sensor 405, audio circuitry 406, wiFi module 407, processor 408 including one or more processing cores, and power supply 409. It will be appreciated by those skilled in the art that the terminal structure shown in fig. 5 is not limiting of the terminal and may include more or fewer components than shown, or may combine certain components, or a different arrangement of components. Wherein:
The radio frequency circuit 401 may be used for receiving and transmitting signals during the process of receiving and transmitting information or communication, in particular, after receiving downlink information of the base station, the downlink information is processed by one or more processors 408; in addition, data relating to uplink is transmitted to the base station. Typically, the radio frequency circuitry 401 includes, but is not limited to, an antenna, at least one amplifier, a tuner, one or more oscillators, a subscriber identity module (SIM, subscriber Identity Module) card, a transceiver, a coupler, a low noise amplifier (LNA, low Noise Amplifier), a duplexer, and the like. In addition, the radio frequency circuitry 401 may also communicate with networks and other devices via wireless communications. The wireless communication may use any communication standard or protocol including, but not limited to, global system for mobile communications (GSM, global System of Mobile communication), general packet radio service (GPRS, general Packet Radio Service), code division multiple access (CDMA, code Division Multiple Access), wideband code division multiple access (WCDMA, wideband Code Division Multiple Access), long term evolution (LTE, long Term Evolution), email, short message service (SMS, short Messaging Service), and the like.
The memory 402 may be used to store software programs and modules, and the processor 408 may execute various functional applications and information processing by executing the software programs and modules stored in the memory 402. The memory 402 may mainly include a storage program area and a storage data area, wherein the storage program area may store an operating system, an application program (such as a sound playing function, an image playing function, etc.) required for at least one function, and the like; the storage data area may store data (such as audio data, phonebook, etc.) created according to the use of the terminal, etc. In addition, memory 402 may include high-speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device. Accordingly, the memory 402 may also include a memory controller to provide access to the memory 402 by the processor 408 and the input unit 403.
The input unit 403 may be used to receive input numeric or character information and to generate keyboard, mouse, joystick, optical or trackball signal inputs related to user settings and function control. In particular, in one particular embodiment, input unit 403 may include a touch-sensitive surface, as well as other input devices. The touch-sensitive surface, also referred to as a touch display screen or a touch pad, may collect touch operations thereon or thereabout by a user (e.g., operations thereon or thereabout by a user using any suitable object or accessory such as a finger, stylus, etc.), and actuate the corresponding connection means according to a predetermined program. Alternatively, the touch-sensitive surface may comprise two parts, a touch detection device and a touch controller. The touch detection device detects the touch azimuth of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch detection device and converts it into touch point coordinates, which are then sent to the processor 408, and can receive commands from the processor 408 and execute them. In addition, touch sensitive surfaces may be implemented in a variety of types, such as resistive, capacitive, infrared, and surface acoustic waves. In addition to the touch-sensitive surface, the input unit 403 may also comprise other input devices. In particular, other input devices may include, but are not limited to, one or more of a physical keyboard, function keys (such as volume control keys, switch keys, etc.), a trackball, mouse, joystick, etc.
The display unit 404 may be used to display information input by a user or information provided to the user and various graphical user interfaces of the terminal, which may be composed of graphics, text, icons, video and any combination thereof. The display unit 404 may include a display panel, which may be optionally configured in the form of a liquid crystal display (LCD, liquid Crystal Display), an Organic Light-Emitting Diode (OLED), or the like. Further, the touch-sensitive surface may overlay a display panel, and upon detection of a touch operation thereon or thereabout, the touch-sensitive surface is communicated to the processor 408 to determine a type of touch event, and the processor 408 then provides a corresponding visual output at the display panel based on the type of touch event. Although in fig. 5 the touch sensitive surface and the display panel are implemented as two separate components for input and output functions, in some embodiments the touch sensitive surface may be integrated with the display panel to implement the input and output functions.
The terminal may also include at least one sensor 405, such as a light sensor, a motion sensor, and other sensors. Specifically, the light sensor may include an ambient light sensor that may adjust the brightness of the display panel according to the brightness of ambient light, and a proximity sensor that may turn off the display panel and/or backlight when the terminal moves to the ear. The gravity acceleration sensor can detect the acceleration in all directions (generally three axes), can detect the gravity and the direction when the mobile phone is stationary, can be used for identifying the gesture of the mobile phone (such as horizontal and vertical screen switching, related games, magnetometer gesture calibration), vibration identification related functions (such as pedometer and knocking), and other sensors such as a gyroscope, a barometer, a hygrometer, a thermometer, an infrared sensor and the like which are also configured by the terminal are not repeated herein.
Audio circuitry 406, speakers, and a microphone may provide an audio interface between the user and the terminal. The audio circuit 406 may transmit the received electrical signal after audio data conversion to a speaker, where the electrical signal is converted to a sound signal for output; on the other hand, the microphone converts the collected sound signals into electrical signals, which are received by the audio circuit 406 and converted into audio data, which are processed by the audio data output processor 408 for transmission to, for example, another terminal via the radio frequency circuit 401, or which are output to the memory 402 for further processing. Audio circuitry 406 may also include an ear bud jack to provide communication of the peripheral ear bud with the terminal.
The WiFi belongs to a short-distance wireless transmission technology, and the terminal can help the user to send and receive e-mail, browse web pages, access streaming media and the like through the WiFi module 407, so that wireless broadband internet access is provided for the user. Although fig. 5 shows a WiFi module 407, it is understood that it does not belong to the essential constitution of the terminal, and can be omitted entirely as required within the scope of not changing the essence of the invention.
The processor 408 is a control center of the terminal, and connects various parts of the entire handset using various interfaces and lines, and performs various functions of the terminal and processes data by running or executing software programs and/or modules stored in the memory 402, and invoking data stored in the memory 402, thereby performing overall monitoring of the handset. Optionally, the processor 408 may include one or more processing cores; preferably, the processor 408 may integrate an application processor that primarily handles operating systems, user interfaces, applications, etc., with a modem processor that primarily handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 408.
The terminal also includes a power supply 409 (e.g., a battery) for powering the various components, which may be logically connected to the processor 408 through a power management system that performs functions such as managing charge, discharge, and power consumption. The power supply 409 may also include one or more of any of a direct current or alternating current power supply, a recharging system, a power failure detection circuit, a power converter or inverter, a power status indicator, and the like.
Although not shown, the terminal may further include a camera, a bluetooth module, etc., which will not be described herein. In this embodiment, the processor 408 in the terminal loads executable files corresponding to the processes of one or more application programs into the memory 402 according to the following instructions, and the processor 408 executes the application programs stored in the memory 402, so as to implement various functions:
when the call state is detected, acquiring current call voice;
extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice;
when the voice features are matched with the preset voice features, prompting information is generated and sent to the preset contact person.
In the foregoing embodiments, the descriptions of the embodiments are focused on, and the portions of an embodiment that are not described in detail in the foregoing embodiments may be referred to in the foregoing detailed description of the information processing method, which is not repeated herein.
Those of ordinary skill in the art will appreciate that all or a portion of the steps of the various methods of the above embodiments may be performed by instructions, or by instructions controlling associated hardware, which may be stored in a computer-readable storage medium and loaded and executed by a processor.
To this end, an embodiment of the present application provides a computer readable storage medium having stored therein a plurality of instructions capable of being loaded by a processor to perform the steps of any one of the information processing methods provided by the embodiment of the present application. For example, the instructions may perform the steps of:
when the call state is detected, acquiring current call voice;
extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice;
when the voice features are matched with the preset voice features, prompting information is generated and sent to the preset contact person.
According to one aspect of the present application, there is provided a computer program product or computer program comprising computer instructions stored in a computer readable storage medium. The processor of the computer device reads the computer instructions from the computer-readable storage medium, and the processor executes the computer instructions, so that the computer device performs the methods provided in the various alternative implementations provided in the above embodiments.
The specific implementation of each operation above may be referred to the previous embodiments, and will not be described herein.
Wherein the computer-readable storage medium may comprise: read Only Memory (ROM), random access Memory (RAM, random Access Memory), magnetic or optical disk, and the like.
Because the instructions stored in the computer readable storage medium may execute the steps in any information processing method provided by the embodiments of the present application, the beneficial effects that any information processing method provided by the embodiments of the present application can be achieved, which are detailed in the previous embodiments and are not described herein.
The foregoing has described in detail the methods, apparatuses and computer readable storage medium for processing information provided by the embodiments of the present application, and specific examples have been applied to illustrate the principles and embodiments of the present application, and the above description of the embodiments is only for aiding in understanding the methods and core ideas of the present application; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in light of the ideas of the present application, the present description should not be construed as limiting the present application.

Claims (10)

1. An information processing method, characterized by comprising:
when the call state is detected, acquiring current call voice;
extracting the characteristics of the current call voice, and extracting the voice characteristics of the current call voice;
and when the voice features are matched with the preset voice features, generating prompt information, and sending the prompt information to the preset contact.
2. The information processing method according to claim 1, wherein the feature extraction of the current call voice, extracting the voice feature of the current call voice, comprises:
extracting the speech speed characteristics of the current call voice as voice characteristics; or (b)
Extracting volume characteristics of the current call voice as voice characteristics; or (b)
And extracting the tone quality characteristics of the current call voice as voice characteristics.
3. The information processing method according to claim 1 or 2, wherein the preset voice features include at least a first preset voice sub-feature;
When the voice feature is matched with a preset voice feature, generating prompt information and sending the prompt information to a preset contact, wherein the method comprises the following steps:
when the voice feature is matched with the first preset voice sub-feature, position information is obtained and first reminding information is generated;
starting a camera to shoot, and obtaining image information;
and sending the position information, the image information and the first reminding information to a preset contact person, a current caller and a preset reminder as reminding information.
4. The information processing method according to claim 3, wherein the preset voice feature further includes at least a second preset voice sub-feature;
when the voice feature is matched with a preset voice feature, generating prompt information and sending the prompt information to a preset contact, wherein the method comprises the following steps:
when the voice features are matched with the second preset voice sub-features, generating second reminding information and starting timing;
collecting environment information corresponding to preset configuration information;
when the timing end is detected, acquiring first target position information when the timing end is detected, and sending the first target position information, the environment information and the second reminding information to a preset contact person and a preset prompt person as prompt information.
5. The information processing method according to claim 4, characterized in that the method further comprises:
when a shutdown instruction is generated in the timing process, second target position information when the shutdown instruction is generated is obtained, and the second target position information, the environment information and the reminding information are used as prompt information and sent to a preset contact person and a preset prompt person.
6. The information processing method according to claim 3, wherein the preset voice feature further includes at least a third preset voice sub-feature;
when the voice feature is matched with a preset voice feature, generating prompt information and sending the prompt information to a preset contact, wherein the method comprises the following steps:
when the voice feature is matched with the third preset voice sub-feature, generating third reminding information and starting timing;
when the timing end is detected, acquiring third target position information when the timing end is detected, and sending the third target position information and the third reminding information to a preset contact person together as prompt information.
7. The information processing method according to claim 6, characterized in that the method further comprises:
and when a cancel instruction is generated in the timing process, ending the timing and canceling the sending of the prompt information.
8. An information processing apparatus, characterized by comprising:
the acquisition unit is used for acquiring current call voice when the call state is detected;
the extracting unit is used for extracting the characteristics of the current call voice and extracting the voice characteristics of the current call voice;
and the sending unit is used for generating prompt information when the voice characteristics are matched with the preset voice characteristics and sending the prompt information to the preset contact person.
9. A computer readable storage medium storing a plurality of instructions adapted to be loaded by a processor to perform the steps in the information processing method of any one of claims 1 to 7.
10. A computer device comprising a memory, a processor and a computer program stored in the memory and executable on the processor, the processor performing the steps in the information processing method of any of claims 1 to 7.
CN202311390744.4A 2023-10-25 2023-10-25 Information processing method, device and computer readable storage medium Active CN117135266B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311390744.4A CN117135266B (en) 2023-10-25 2023-10-25 Information processing method, device and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311390744.4A CN117135266B (en) 2023-10-25 2023-10-25 Information processing method, device and computer readable storage medium

Publications (2)

Publication Number Publication Date
CN117135266A true CN117135266A (en) 2023-11-28
CN117135266B CN117135266B (en) 2024-03-22

Family

ID=88858495

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311390744.4A Active CN117135266B (en) 2023-10-25 2023-10-25 Information processing method, device and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN117135266B (en)

Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR200315289Y1 (en) * 2002-04-06 2003-06-02 한별전자 주식회사 USB based Voice Processing Device
JP2005283972A (en) * 2004-03-30 2005-10-13 Advanced Media Inc Speech recognition method, and information presentation method and information presentation device using the speech recognition method
CN101827161A (en) * 2009-03-02 2010-09-08 鸿富锦精密工业(深圳)有限公司 Call-prompting system and method
JP2011090566A (en) * 2009-10-23 2011-05-06 Ntt Docomo Inc System and method for retrieving voice information
CN105721492A (en) * 2016-03-25 2016-06-29 宇龙计算机通信科技(深圳)有限公司 Voice processing method and apparatus and terminal
CN106412312A (en) * 2016-10-19 2017-02-15 北京奇虎科技有限公司 Method and system for automatically awakening camera shooting function of intelligent terminal, and intelligent terminal
CN106412346A (en) * 2016-10-31 2017-02-15 努比亚技术有限公司 Speech communication method and device
CN107464557A (en) * 2017-09-11 2017-12-12 广东欧珀移动通信有限公司 Call recording method, device, mobile terminal and storage medium
CN107547717A (en) * 2017-08-01 2018-01-05 联想(北京)有限公司 Information processing method, electronic equipment and computer-readable storage medium
CN107872555A (en) * 2017-11-09 2018-04-03 广东欧珀移动通信有限公司 Call recording method, device, terminal device and computer-readable recording medium
CN107948417A (en) * 2017-11-22 2018-04-20 周燕红 A kind of method, apparatus, terminal and the storage medium of voice data monitoring
CN109256151A (en) * 2018-11-21 2019-01-22 努比亚技术有限公司 Call voice regulates and controls method, apparatus, mobile terminal and readable storage medium storing program for executing
CN109905851A (en) * 2018-12-26 2019-06-18 维沃移动通信有限公司 A kind of reminding method and terminal device
CN110660398A (en) * 2019-09-19 2020-01-07 北京三快在线科技有限公司 Voiceprint feature updating method and device, computer equipment and storage medium
JP6647722B1 (en) * 2019-07-08 2020-02-14 株式会社Tmj Information processing apparatus, information processing method, information processing program
CN111034152A (en) * 2017-09-25 2020-04-17 深圳市欢太科技有限公司 Information processing method, device, mobile terminal and computer readable storage medium
CN111083273A (en) * 2019-12-06 2020-04-28 北京搜狗科技发展有限公司 Voice processing method and device and electronic equipment
KR20200101479A (en) * 2019-01-28 2020-08-28 이동수 Home care management system utilizing IOT and voice recognition function
CN112216306A (en) * 2020-09-25 2021-01-12 广东电网有限责任公司佛山供电局 Voiceprint-based call management method and device, electronic equipment and storage medium
CN112751971A (en) * 2020-12-30 2021-05-04 维沃移动通信有限公司 Voice playing method and device and electronic equipment
US20220084529A1 (en) * 2019-01-04 2022-03-17 Matrixed Reality Technology Co., Ltd. Method and apparatus for awakening wearable device

Patent Citations (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR200315289Y1 (en) * 2002-04-06 2003-06-02 한별전자 주식회사 USB based Voice Processing Device
JP2005283972A (en) * 2004-03-30 2005-10-13 Advanced Media Inc Speech recognition method, and information presentation method and information presentation device using the speech recognition method
CN101827161A (en) * 2009-03-02 2010-09-08 鸿富锦精密工业(深圳)有限公司 Call-prompting system and method
JP2011090566A (en) * 2009-10-23 2011-05-06 Ntt Docomo Inc System and method for retrieving voice information
CN105721492A (en) * 2016-03-25 2016-06-29 宇龙计算机通信科技(深圳)有限公司 Voice processing method and apparatus and terminal
CN106412312A (en) * 2016-10-19 2017-02-15 北京奇虎科技有限公司 Method and system for automatically awakening camera shooting function of intelligent terminal, and intelligent terminal
CN106412346A (en) * 2016-10-31 2017-02-15 努比亚技术有限公司 Speech communication method and device
CN107547717A (en) * 2017-08-01 2018-01-05 联想(北京)有限公司 Information processing method, electronic equipment and computer-readable storage medium
CN107464557A (en) * 2017-09-11 2017-12-12 广东欧珀移动通信有限公司 Call recording method, device, mobile terminal and storage medium
CN111034152A (en) * 2017-09-25 2020-04-17 深圳市欢太科技有限公司 Information processing method, device, mobile terminal and computer readable storage medium
CN107872555A (en) * 2017-11-09 2018-04-03 广东欧珀移动通信有限公司 Call recording method, device, terminal device and computer-readable recording medium
CN107948417A (en) * 2017-11-22 2018-04-20 周燕红 A kind of method, apparatus, terminal and the storage medium of voice data monitoring
CN109256151A (en) * 2018-11-21 2019-01-22 努比亚技术有限公司 Call voice regulates and controls method, apparatus, mobile terminal and readable storage medium storing program for executing
CN109905851A (en) * 2018-12-26 2019-06-18 维沃移动通信有限公司 A kind of reminding method and terminal device
US20220084529A1 (en) * 2019-01-04 2022-03-17 Matrixed Reality Technology Co., Ltd. Method and apparatus for awakening wearable device
KR20200101479A (en) * 2019-01-28 2020-08-28 이동수 Home care management system utilizing IOT and voice recognition function
JP6647722B1 (en) * 2019-07-08 2020-02-14 株式会社Tmj Information processing apparatus, information processing method, information processing program
CN110660398A (en) * 2019-09-19 2020-01-07 北京三快在线科技有限公司 Voiceprint feature updating method and device, computer equipment and storage medium
CN111083273A (en) * 2019-12-06 2020-04-28 北京搜狗科技发展有限公司 Voice processing method and device and electronic equipment
CN112216306A (en) * 2020-09-25 2021-01-12 广东电网有限责任公司佛山供电局 Voiceprint-based call management method and device, electronic equipment and storage medium
CN112751971A (en) * 2020-12-30 2021-05-04 维沃移动通信有限公司 Voice playing method and device and electronic equipment

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
谢波;陈岭;陈根才;陈纯;: "普通话语音情感识别的特征选择技术", 浙江大学学报(工学版), no. 11, pages 1816 - 1822 *
阎福智;: "语音信号处理中特征提取方法研究", 中国新通信, no. 21, pages 127 - 128 *

Also Published As

Publication number Publication date
CN117135266B (en) 2024-03-22

Similar Documents

Publication Publication Date Title
CN106791892B (en) Method, device and system for live broadcasting of wheelhouses
WO2016134630A1 (en) Method and device for recognizing malicious call
WO2017008569A1 (en) Message updating method, apparatus, and terminal
CN106371964B (en) Method and device for prompting message
CN108900231B (en) Dynamic antenna adjustment method and related product
CN107204883B (en) Network fault processing method and device
US20170318146A1 (en) Silent Invocation of Emergency Broadcasting Mobile Device
CN107846516B (en) Call scene prompting method and mobile terminal
EP3944595A1 (en) Method and system for adding smart home device to contacts
CN107666583B (en) Call processing method and terminal
CN108551530B (en) Incoming call processing method and device, readable storage medium and mobile terminal
CN109495769B (en) Video communication method, terminal, smart television, server and storage medium
CN104917905B (en) Processing method, terminal and the server of Stranger Calls
CN108432220B (en) Method and terminal for switching call mode
CN108810261B (en) Antenna switching method in call and related product
CN107734153B (en) Call control method, terminal and computer readable storage medium
CN109981890B (en) Reminding task processing method, terminal and computer readable storage medium
CN105991510B (en) Method, device and system for synchronizing session between multiple terminals
CN111580911A (en) Operation prompting method and device for terminal, storage medium and terminal
CN110891262A (en) Bluetooth pairing method, system and terminal equipment
CN111372003A (en) Camera switching method and device and terminal
CN117135266B (en) Information processing method, device and computer readable storage medium
CN106878294B (en) Method and device for recording voice information by interphone
CN112640408A (en) Call prompting method and terminal
CN112019738B (en) Photographing method and terminal equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant