CN116968631A - Control method and device based on audio characteristics, electronic equipment and storage medium - Google Patents

Control method and device based on audio characteristics, electronic equipment and storage medium Download PDF

Info

Publication number
CN116968631A
CN116968631A CN202310409908.7A CN202310409908A CN116968631A CN 116968631 A CN116968631 A CN 116968631A CN 202310409908 A CN202310409908 A CN 202310409908A CN 116968631 A CN116968631 A CN 116968631A
Authority
CN
China
Prior art keywords
characteristic information
information
audio
fusion
atmosphere lamp
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310409908.7A
Other languages
Chinese (zh)
Inventor
樊公辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
SHANGHAI YUZHOU ELECTRICAL EQU
Original Assignee
SHANGHAI YUZHOU ELECTRICAL EQU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SHANGHAI YUZHOU ELECTRICAL EQU filed Critical SHANGHAI YUZHOU ELECTRICAL EQU
Publication of CN116968631A publication Critical patent/CN116968631A/en
Pending legal-status Critical Current

Links

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B60VEHICLES IN GENERAL
    • B60QARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
    • B60Q3/00Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
    • B60Q3/80Circuits; Control arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0016Means for indicating which keys, frets or strings are to be actuated, e.g. using lights or leds
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02BCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
    • Y02B20/00Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
    • Y02B20/40Control techniques providing energy savings, e.g. smart controller or presence detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Mechanical Engineering (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Circuit Arrangement For Electric Light Sources In General (AREA)

Abstract

The application relates to the technical field of atmosphere lamp control, and particularly discloses a control method, a device, electronic equipment and a storage medium based on audio characteristics, wherein the method comprises the following steps: invoking a control box to receive sound signals from the sound acquisition terminal; calling an A2B chip in the control box to convert the sound signal into an analog signal; calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information; calling an analog-to-digital conversion chip to convert analog audio information to obtain audio data; extracting the characteristics of the audio data to obtain an audio characteristic information set; carrying out fusion processing on the audio feature information set to obtain fusion feature information; and generating an atmosphere lamp control instruction according to the fusion characteristic information to control the automobile atmosphere lamp. Therefore, the lamp control instruction is generated on the basis of the fused characteristics by extracting and fusing various characteristics of the audio data, so that the lamp control instruction can be more attached to the change of the original audio, and riding experience of a user is improved.

Description

Control method and device based on audio characteristics, electronic equipment and storage medium
Technical Field
The application relates to the technical field of atmosphere lamp control, in particular to a control method and device based on audio characteristics, electronic equipment and a storage medium.
Background
With the continuous development of the automobile industry, the individualization and the comfort of automobiles are gradually improved, and automobiles are not early simple vehicles, but are moving towards vehicles capable of providing a comfortable and agreeable experience for people. Nowadays, people demand vehicles, except appearance, interior decorations, power and brands, experience when the vehicles are restarted is further added, and an atmosphere lamp is used as a novel lamp for improving the experience when the vehicles are started, so that people are increasingly attracted.
Currently, an atmosphere lamp system generally generates a lamp control command according to a single characteristic of an input audio, or matches a preset lamp control command according to a single characteristic of a selected input audio, so as to perform atmosphere lamp control. However, the single feature cannot fully express the characteristic of the audio, so that the current atmosphere lamp control method causes low matching degree between the actual effect of the lamplight and the audio, and influences riding experience of a user.
Disclosure of Invention
In order to solve the problems in the prior art, the embodiment of the application provides a control method, a control device, electronic equipment and a storage medium based on audio characteristics, which can enable a lamp control instruction to be more fit with the change of original audio, so that riding experience of a driver or a passenger is enhanced.
In a first aspect, an embodiment of the present application provides a control method based on audio features, including:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, the audio feature information set includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In this embodiment, the process of fusing the audio feature information set to obtain fused feature information includes:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In this embodiment, generating an atmosphere lamp control command from the fusion feature information includes:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
In a second aspect, an embodiment of the present application provides a control device based on audio features, including:
the acquisition module is used for calling the control box to receive the sound signal from the sound acquisition terminal, calling the A2B chip in the control box to convert the sound signal into an analog signal, calling the audio acquisition circuit to acquire and process the analog signal to obtain analog audio information, and calling the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module is used for carrying out feature extraction on the audio data to obtain an audio feature information set;
the fusion module is used for carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information;
and the processing module is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, the audio feature information set includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In this embodiment, in performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module is configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In this embodiment, in generating the atmosphere lamp control command according to the fusion characteristic information, the processing module is configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
In a third aspect, an embodiment of the present application provides an electronic device, including: a processor and a memory, wherein the processor is connected to the memory, the memory is used for storing a computer program, and the processor is used for executing the computer program stored in the memory, so that the electronic device executes the method as in the first aspect.
In a fourth aspect, embodiments of the present application provide a computer-readable storage medium storing a computer program, the computer program causing a computer to perform the method as in the first aspect.
In a fifth aspect, embodiments of the present application provide a computer program product comprising a non-transitory computer readable storage medium storing a computer program, the computer being operable to cause a computer to perform a method as in the first aspect.
The implementation of the embodiment of the application has the following beneficial effects:
it can be seen that in the embodiment of the application, by extracting and fusing various features of the audio data, and then generating the lamp control instruction on the basis of the fused features, the lamp control instruction can be more fit with the change of the original audio, so that the riding experience of a driver or a passenger is enhanced.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings required for the description of the embodiments will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present application, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic hardware structure of a control device based on audio features according to an embodiment of the present application;
fig. 2 is a schematic flow chart of a control method based on audio features according to an embodiment of the present application;
fig. 3 is a functional block diagram of a control device based on audio features according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are some, but not all embodiments of the present application. All other embodiments, based on the embodiments of the application, which are apparent to those of ordinary skill in the art without inventive faculty, are intended to be within the scope of the application.
The terms "first," "second," "third," and "fourth" and the like in the description and in the claims and drawings are used for distinguishing between different objects and not necessarily for describing a particular sequential or chronological order. Furthermore, the terms "comprise" and "have," as well as any variations thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those listed steps or elements but may include other steps or elements not listed or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, result, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those skilled in the art will explicitly and implicitly understand that the embodiments described herein may be combined with other embodiments.
First, referring to fig. 1, fig. 1 is a schematic hardware structure of a control device based on audio features according to an embodiment of the present application. The audio feature based control device 100 comprises at least one processor 101, communication lines 102, a memory 103 and at least one communication interface 104.
In this embodiment, the processor 101 may be a general purpose central processing unit (central processing unit, CPU), microprocessor, application-specific integrated circuit (ASIC), or one or more integrated circuits for controlling the execution of the program according to the present application.
Communication line 102 may include a pathway to transfer information between the above-described components.
The communication interface 104, which may be any transceiver-like device (e.g., antenna, etc.), is used to communicate with other devices or communication networks, such as ethernet, RAN, wireless local area network (wireless local area networks, WLAN), etc.
The memory 103 may be, but is not limited to, a read-only memory (ROM) or other type of static storage device that can store static information and instructions, a random access memory (random access memory, RAM) or other type of dynamic storage device that can store information and instructions, or an electrically erasable programmable read-only memory (electrically erasable programmable read-only memory, EEPROM), a compact disc (compact disc read-only memory) or other optical disc storage, optical disc storage (including compact disc, laser disc, optical disc, digital versatile disc, blu-ray disc, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer.
In this embodiment, the memory 103 may be independently provided and connected to the processor 101 via the communication line 102. Memory 103 may also be integrated with processor 101. The memory 103 provided by embodiments of the present application may generally have non-volatility. The memory 103 is used for storing computer-executable instructions for executing the scheme of the present application, and is controlled by the processor 101 to execute the instructions. The processor 101 is configured to execute computer-executable instructions stored in the memory 103 to implement the methods provided in the embodiments of the present application described below.
In alternative embodiments, computer-executable instructions may also be referred to as application code, as the application is not particularly limited.
In alternative embodiments, processor 101 may include one or more CPUs, such as CPU0 and CPU1 in fig. 1.
In alternative embodiments, the audio feature based control device 100 may include multiple processors, such as processor 101 and processor 107 in FIG. 1. Each of these processors may be a single-core (single-CPU) processor or may be a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In an alternative embodiment, if the control device 100 based on the audio feature is a server, for example, it may be a stand-alone server, or may be a cloud server that provides cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, content delivery network (Content Delivery Network, CDN), and basic cloud computing services such as big data and artificial intelligence platform. The audio feature based control apparatus 100 may further comprise an output device 105 and an input device 106. The output device 105 communicates with the processor 101 and may display information in a variety of ways. For example, the output device 105 may be a liquid crystal display (liquid crystal display, LCD), a light emitting diode (light emitting diode, LED) display device, a Cathode Ray Tube (CRT) display device, or a projector (projector), or the like. The input device 106 is in communication with the processor 101 and may receive user input in a variety of ways. For example, the input device 106 may be a mouse, a keyboard, a touch screen device, a sensing device, or the like.
The control apparatus 100 based on the audio characteristics may be a general-purpose device or a special-purpose device. The embodiment of the present application does not limit the type of the control device 100 based on the audio characteristics.
Referring to fig. 2, fig. 2 is a flow chart of a control method based on audio features according to an embodiment of the present application. The control method based on the audio characteristics comprises the following steps:
201: and calling the control box to receive the sound signal from the sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signal into an analog signal.
In this embodiment, the audio data may be collected by a sound collection terminal provided inside the automobile, or may be output by a music playing device integrated in the vehicle, such as an on-board DVD, or may be output by an external handheld device such as a mobile phone that can be wired or wirelessly connected to the vehicle.
For example, the sound collection terminal may be a microphone, and for example, a microphone may be disposed on each door beside each seat in the vehicle, so as to collect audio data in the vehicle. After the microphone collects the sound in the car, such as human voice, the collected sound can be transmitted to the control box through the car audio bus (A2B), and the audio data is converted into analog signals after passing through the A2B chip.
202: and calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data.
In the embodiment, the analog signal is processed by an analog-to-digital conversion chip to form audio data, and the audio data is transmitted to a micro control unit (Microcontroller Unit, MCU) for feature extraction.
203: and extracting the characteristics of the audio data to obtain an audio characteristic information set.
In this embodiment, the audio feature information set may include: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information of the audio data. For example, the emotion feature information, rhythm feature information, dynamics feature information, and speed feature information may all be represented by a one-dimensional array.
Illustratively, taking emotion feature information as an example, pre-defining emotion features includes: happiness, sadness, pain, fear, anger, happiness and calm, and represent these emotional characteristics by numerals or symbols, respectively. For example, happy with 1, sad with 2, painful with 3, fear with 4, anger with 5, happy with 6 and calm with 7.
In the present embodiment, since audio data is acquired through a microphone provided in the vehicle, the audio data may include both human voice and music. Based on the above, when emotion recognition is performed, voice and music in the audio data can be split in advance, emotion characteristics of the voice part and the music part are extracted respectively, and finally the overall emotion of the audio data is determined based on the two emotion characteristics.
Specifically, in the human voice section, the speech, intonation, speech speed, etc. of one person speaking will change correspondingly with the change of emotion. Exemplary: in a happy state, a person can secrete more adrenal hormones than usual, and the excessive adrenal hormones can lead to the person in the happy state always being in an excited state. At the same time, the state of hyperactivity accelerates blood flow and nerve conduction, which in turn results in adequate blood and oxygen supply to the brain, faster heart beat, faster thinking, in turn exhibiting faster thinking, more fluent expression, louder sounds, and faster speech.
In addition, when a person is in different emotions, the speaking sound also fluctuates due to the influence of the emotion, for example: high decibels for happiness or anger, low decibels for sadness or pain, etc. Based on this, in an alternative embodiment, the db value variation characteristic of the voice portion may also be determined, and then the emotion characteristic corresponding to the voice portion is determined in combination with the db value variation characteristic.
By way of example, a weight table of different emotions of a person at different sound decibel values can be established by analytical learning of a large amount of historical data. For example: in the interval of 65-85 dB, through analysis and learning of a large number of real conversation scenes under the dB, the weight of a person in the scene in a happy emotion is 0.2, the weight of the person in a sad emotion is 0.05, the weight of the person in an anger emotion is 0.6 and the weight of the person in a calm emotion is 0.02. Therefore, after the human voice part is obtained, the weight of each emotion corresponding to the decibel value interval can be obtained as the characteristic of sound loudness by determining the decibel value of the sound and then inquiring the table according to the decibel value.
Thus, in this embodiment, the speech, intonation, speech speed, sound loudness, etc. of a person speaking in different emotions are changed accordingly. Therefore, the characteristics of the voice, the intonation, the speech speed, the accent distribution and the loudness of the voice can be obtained by extracting the characteristics of the voice, the intonation, the speech speed, the accent distribution and the loudness of the voice part, and then the characteristics are analyzed correspondingly, so that the emotion characteristics corresponding to the voice part are obtained.
Specifically, the probability that the voice part corresponds to each emotion can be obtained through analysis of characteristics such as voice, intonation, speech speed, accent distribution and the like, then the product of the probability and weight values corresponding to different emotions under the decibel is carried out, and then the product result is used as the final result of audio emotion recognition. By way of example, assuming that the dialogue has a decibel of 75 and falls within the interval of 75 to 85 decibels, by the feature analysis, it is obtained that the probability of being a happy emotion is 0.1, the probability of being a sad emotion is 0.1, the probability of being an angry emotion is 0.8, and the probability of being a calm emotion is 0. Based on the above, the final probability can be obtained by multiplying the weight corresponding to each emotion under the interval: the probability of a happy emotion is 0.1X0.2 =0.02, the probability of a sad emotion is 0.1X0.05 =0.005, the probability of an angry emotion is 0.8X0.6 =0.48, and the probability of a calm emotion is 0X0.02 =0. Thus, the probability distribution is taken as an emotional characteristic of the vocal part.
Meanwhile, in an alternative implementation mode, text conversion can be carried out on the content of the voice part, talking content is determined, then the reason of the dialogue is analyzed according to the talking content, and then the emotion recognition precision of the voice part is further improved.
And for the music part, the probability distribution of each emotion corresponding to the music can be determined by extracting the characteristics of the music part such as the wind, the rhythm, the tone and the like, and the probability distribution is used as the emotion characteristics of the music part.
Finally, in the present embodiment, the weights of the human voice part and the music part may be determined according to the decibel value of the human voice part. Specifically, the greater the decibel value of the vocal portion, the less affected the speaker is by the music, which is about the content of the conversation whose dominant emotion has been played by the speaker. Therefore, the greater the decibel value of the human voice part, the greater the weight corresponding to the human voice part. And according to the determined weight, weighting and summing each emotion probability distribution of the voice part and each emotion probability distribution of the music part according to the corresponding relation of each emotion to obtain the comprehensive probability distribution of each emotion.
Illustratively, the probability of a happy emotion in the vocal part is 0.5 and the probability in the music part is 0.8. Through weight calculation, the weight of the vocal part is 0.8, and the weight of the music part is 0.2. The overall probability of happy emotion is: 0.5x0.8+0.8x0.2=0.56. Thus, the emotion having the highest overall probability among the overall probability distributions of the emotions can be used as the emotion feature of the audio data.
In this embodiment, for emotion feature information, each item of data in the emotion feature information may be defined in advance, emotion features corresponding to each second of audio data, and the order of data in the emotion feature information is arranged in the chronological order of the audio data.
Specifically, as for emotion feature information a= [1,1,1,2,2,5,5,7,6], it is shown that in the audio data, the emotion feature of the first 3 seconds is happy, the emotion feature of the 4 th and 5 th seconds is sad, the emotion feature of the 6 th and 7 th seconds is cheerful, the emotion feature of the 8 th second is calm, and the emotion feature of the 9 th second is happy.
Based on the above, the rhythm feature information, the dynamics feature information, and the speed feature information may be quantized according to a preset quantization rule, so as to obtain a one-dimensional array in time sequence.
204: and carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information.
In this embodiment, emotion feature information, rhythm feature information, dynamics feature information, and speed feature information may be sequentially subjected to longitudinal splicing processing to obtain fusion feature information.
Illustratively, for emotion feature information a= [1,1,1,2,2,5,5,7,6], rhythm feature information b= [1,5,6,8,4,2,2,6,5], dynamics feature information c= [2,2,2,2,1,1,2,2,1] and speed feature information d= [2,2,2,2,2,2,3,3,1]. After longitudinal stitching, a 4×9 matrix E can be obtained:
in this embodiment, the matrix E is the fused feature information after fusion, where the j-th column of the matrix E is used to represent emotion, rhythm, strength and speed parameters of the atmosphere lamp corresponding to the audio data in the j-th second. For example, for the first columnIndicating that the emotion parameter corresponding to the first second of the audio data is 1, the rhythm parameter is 1, the dynamics parameter is 2, and the speed parameter is 2.
205: and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, after the fusion feature information is obtained, the light color change information may be determined according to the first line of the fusion feature information; determining brightness change information according to the second row of the fusion characteristic information; determining switch change information according to a third row of the fusion characteristic information; and determining saturation change information according to the fourth line of the fusion characteristic information.
Therefore, the light color change information, the brightness change information, the switching change information and the saturation change information are used as atmosphere lamp control instructions to control the automobile atmosphere lamps to display light colors corresponding to the saturation and the brightness, and to control the switching of the automobile atmosphere lamps.
For example, the light color corresponding to emotion feature information may be predefined, for example, with red for happiness, blue for sadness, gray for pain, purple for fear, yellow for cheerful, orange for happiness, and green for calm. Therefore, according to the first line of the fusion characteristic information representing the emotion characteristics, the photochromic corresponding to each second of the audio data can be obtained, and then the photochromic change information is obtained.
Similarly, the brightness corresponding to the rhythm feature information, the switching condition corresponding to the dynamics feature information and the light color saturation corresponding to the speed feature information can be predefined, and then the corresponding atmosphere lamp control instruction is determined according to the fusion feature information.
Specifically, the first column of the matrix E is followedThe first column indicates that the color of the audio data corresponding to the first second is red, the brightness is low, the on condition is on, and the saturation is 20%.
It can be seen that in the embodiment of the application, by extracting and fusing various features of the audio data, and then generating the lamp control instruction on the basis of the fused features, the lamp control instruction can be more fit with the change of the original audio, so that the riding experience of a driver or a passenger is enhanced.
Referring to fig. 3, fig. 3 is a functional block diagram of a control device based on audio features according to an embodiment of the present application. As shown in fig. 3, the audio feature-based control apparatus 300 includes:
the acquisition module 301 is configured to invoke the control box to receive a sound signal from the sound acquisition terminal, invoke an A2B chip in the control box to convert the sound signal into an analog signal, invoke the audio acquisition circuit to acquire and process the analog signal to obtain analog audio information, and invoke the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module 302 is configured to perform feature extraction on the audio data to obtain an audio feature information set;
the fusion module 303 is configured to perform fusion processing on the audio feature information set to obtain fusion feature information;
and the processing module 304 is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In an embodiment of the present application, the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In the embodiment of the present application, in terms of performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module 303 is specifically configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In the embodiment of the present application, in generating the atmosphere lamp control command according to the fusion characteristic information, the processing module 304 is specifically configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
Referring to fig. 4, fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application. As shown in fig. 4, the electronic device 400 includes a transceiver 401, a processor 402, and a memory 403. Which are connected by a bus 404. The memory 403 is used for storing computer programs and data, and the data stored in the memory 403 can be transferred to the processor 402.
The processor 402 is configured to read a computer program in the memory 403 to perform the following operations:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In an embodiment of the present application, the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In an embodiment of the present application, the processor 402 is specifically configured to perform the following operations in terms of performing fusion processing on the audio feature information set to obtain fusion feature information:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In an embodiment of the present application, the processor 402 is specifically configured to perform the following operations in generating the atmosphere lamp control command according to the fusion characteristic information:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
It should be understood that the control device based on the audio feature in the present application may include a smart Phone (such as an Android mobile Phone, an iOS mobile Phone, a Windows Phone mobile Phone, etc.), a tablet computer, a palm computer, a notebook computer, a mobile internet device MID (Mobile Internet Devices, abbreviated as MID) or a wearable device, etc. The above-described audio feature-based control devices are merely examples and are not intended to be exhaustive, including but not limited to the above-described audio feature-based control devices. In practical applications, the control device based on the audio features may further include: intelligent vehicle terminals, computer devices, etc.
From the above description of embodiments, it will be apparent to those skilled in the art that the present application may be implemented in software in combination with a hardware platform. With such understanding, all or part of the technical solution of the present application contributing to the background art may be embodied in the form of a software product, which may be stored in a storage medium, such as ROM/RAM, a magnetic disk, an optical disk, etc., including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform the methods described in the various embodiments or parts of the embodiments of the present application.
Accordingly, the present application also provides a computer-readable storage medium storing a computer program that is executed by a processor to implement some or all of the steps of any one of the audio feature-based control methods described in the above method embodiments. For example, the storage medium may include a hard disk, a floppy disk, an optical disk, a magnetic tape, a magnetic disk, a flash memory, etc.
Embodiments of the present application also provide a computer program product comprising a non-transitory computer-readable storage medium storing a computer program operable to cause a computer to perform part or all of the steps of any one of the audio feature-based control methods described in the method embodiments above.
It should be noted that, for simplicity of description, the foregoing method embodiments are all described as a series of acts, but it should be understood by those skilled in the art that the present application is not limited by the order of acts described, as some steps may be performed in other orders or concurrently in accordance with the present application. Further, those skilled in the art will also appreciate that the embodiments described in the specification are alternative embodiments, and that the acts and modules involved are not necessarily required for the present application.
In the foregoing embodiments, the descriptions of the embodiments are focused on, and for those portions of one embodiment that are not described in detail, reference may be made to the related descriptions of other embodiments.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus may be implemented in other manners. For example, the apparatus embodiments described above are merely illustrative, such as the division of the units, merely a logical function division, and there may be additional divisions when actually implemented, such as multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, or may be in electrical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units described above may be implemented either in hardware or in software program modules.
The integrated units, if implemented in the form of software program modules, may be stored in a computer-readable memory for sale or use as a stand-alone product. Based on this understanding, the technical solution of the present application may be embodied essentially or partly in the form of a software product or all or part of the technical solution, which is stored in a memory, and includes several instructions for causing a computer device (which may be a personal computer, a server, a network device, or the like) to perform all or part of the steps of the method according to the embodiments of the present application. And the aforementioned memory includes: a U-disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a removable hard disk, a magnetic disk, or an optical disk, or other various media capable of storing program codes.
Those of ordinary skill in the art will appreciate that all or a portion of the steps in the various methods of the above embodiments may be implemented by a program that instructs associated hardware, and the program may be stored in a computer readable memory, and the memory may include: flash disk, read-Only Memory (ROM), random access Memory (Random Access Memory, RAM), magnetic disk or optical disk.
The foregoing has outlined rather broadly the more detailed description of the embodiments of the application in order that the detailed description of the principles and embodiments of the application may be implemented in conjunction with the detailed description of the embodiments that follows, the claims being merely intended to facilitate the understanding of the method and concepts underlying the application; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in accordance with the ideas of the present application, the present description should not be construed as limiting the present application in view of the above.

Claims (10)

1. A control method based on audio characteristics, the control method comprising:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
2. The control method according to claim 1, wherein,
the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
3. The control method according to claim 2, wherein the fusing the audio feature information set to obtain fused feature information includes:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain the fusion characteristic information.
4. The control method according to claim 3, wherein generating an atmosphere lamp control command according to the fusion characteristic information comprises:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to the third row of the fusion characteristic information;
determining saturation change information according to the fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as the atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
5. A control device based on audio characteristics, the control device comprising:
the acquisition module is used for calling the control box to receive sound signals from the sound acquisition terminal, calling the A2B chip in the control box to convert the sound signals into analog signals, calling the audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module is used for carrying out feature extraction on the audio data to obtain an audio feature information set;
the fusion module is used for carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information;
and the processing module is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
6. The control device according to claim 5, wherein,
the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
7. The control device according to claim 6, wherein in the aspect of performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module is configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain the fusion characteristic information.
8. The control device according to claim 7, wherein in the generating an atmosphere lamp control instruction according to the fusion characteristic information, the fusion module is configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to the third row of the fusion characteristic information;
determining saturation change information according to the fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as the atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
9. An electronic device comprising a processor, a memory, a communication interface, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the processor, the one or more programs comprising instructions for performing the steps in the method of any of claims 1-4.
10. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program, which is executed by a processor to implement the method of any of claims 1-4.
CN202310409908.7A 2022-04-28 2023-04-17 Control method and device based on audio characteristics, electronic equipment and storage medium Pending CN116968631A (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2022104566785 2022-04-28
CN202210456678 2022-04-28

Publications (1)

Publication Number Publication Date
CN116968631A true CN116968631A (en) 2023-10-31

Family

ID=88477327

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202310409908.7A Pending CN116968631A (en) 2022-04-28 2023-04-17 Control method and device based on audio characteristics, electronic equipment and storage medium

Country Status (1)

Country Link
CN (1) CN116968631A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117828537A (en) * 2024-03-04 2024-04-05 北京建筑大学 Music emotion recognition method and device based on CBA model

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117828537A (en) * 2024-03-04 2024-04-05 北京建筑大学 Music emotion recognition method and device based on CBA model
CN117828537B (en) * 2024-03-04 2024-05-17 北京建筑大学 Music emotion recognition method and device based on CBA model

Similar Documents

Publication Publication Date Title
CN110838286B (en) Model training method, language identification method, device and equipment
CN112863547B (en) Virtual resource transfer processing method, device, storage medium and computer equipment
US20190394057A1 (en) Device and method for adaptively providing meeting
US20200126566A1 (en) Method and apparatus for voice interaction
CN107392124A (en) Emotion identification method, apparatus, terminal and storage medium
WO2021004481A1 (en) Media files recommending method and device
US20190371319A1 (en) Method for human-machine interaction, electronic device, and computer-readable storage medium
CN113010138B (en) Article voice playing method, device and equipment and computer readable storage medium
CN113160819B (en) Method, apparatus, device, medium, and product for outputting animation
CN109754792A (en) Voice interface device and the voice interface method for applying it
CN109102802A (en) System for handling user spoken utterances
CN107545029A (en) Voice feedback method, equipment and the computer-readable recording medium of smart machine
CN108281138A (en) Age discrimination model training and intelligent sound exchange method, equipment and storage medium
CN116968631A (en) Control method and device based on audio characteristics, electronic equipment and storage medium
KR20190005103A (en) Electronic device-awakening method and apparatus, device and computer-readable storage medium
CN109712644A (en) Method based on speech recognition emotional change control LED display effect, the apparatus and system for controlling LED display effect
CN109377979B (en) Method and system for updating welcome language
CN113327580A (en) Speech synthesis method, device, readable medium and electronic equipment
CN108701333A (en) Electronic equipment for providing content and its control method
CN110728983A (en) Information display method, device, equipment and readable storage medium
CN113643684B (en) Speech synthesis method, device, electronic equipment and storage medium
CN106372203A (en) Information response method and device for smart terminal and smart terminal
CN112644375B (en) Mood perception-based in-vehicle atmosphere lamp adjusting method, system, medium and terminal
EP4293664A1 (en) Voiceprint recognition method, graphical interface, and electronic device
CN108877795B (en) Method and apparatus for presenting information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication