CN116968631A - Control method and device based on audio characteristics, electronic equipment and storage medium - Google Patents
Control method and device based on audio characteristics, electronic equipment and storage medium Download PDFInfo
- Publication number
- CN116968631A CN116968631A CN202310409908.7A CN202310409908A CN116968631A CN 116968631 A CN116968631 A CN 116968631A CN 202310409908 A CN202310409908 A CN 202310409908A CN 116968631 A CN116968631 A CN 116968631A
- Authority
- CN
- China
- Prior art keywords
- characteristic information
- information
- audio
- fusion
- atmosphere lamp
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 44
- 230000004927 fusion Effects 0.000 claims abstract description 72
- 230000008859 change Effects 0.000 claims abstract description 63
- 230000005236 sound signal Effects 0.000 claims abstract description 16
- 238000007499 fusion processing Methods 0.000 claims abstract description 12
- 230000008569 process Effects 0.000 claims abstract description 11
- 238000006243 chemical reaction Methods 0.000 claims abstract description 10
- 230000008451 emotion Effects 0.000 claims description 76
- 230000033764 rhythmic process Effects 0.000 claims description 27
- 238000012545 processing Methods 0.000 claims description 17
- 238000004590 computer program Methods 0.000 claims description 13
- 238000004891 communication Methods 0.000 claims description 11
- 239000003086 colorant Substances 0.000 claims description 7
- 238000000605 extraction Methods 0.000 claims description 7
- 238000003491 array Methods 0.000 claims description 6
- 238000009826 distribution Methods 0.000 description 10
- 230000003287 optical effect Effects 0.000 description 7
- 238000010586 diagram Methods 0.000 description 4
- 239000011159 matrix material Substances 0.000 description 4
- 230000001755 vocal effect Effects 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 230000008909 emotion recognition Effects 0.000 description 3
- 230000001919 adrenal effect Effects 0.000 description 2
- 230000002996 emotional effect Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 229940088597 hormone Drugs 0.000 description 2
- 239000005556 hormone Substances 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 230000003068 static effect Effects 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- QVGXLLKOCUKJST-UHFFFAOYSA-N atomic oxygen Chemical compound [O] QVGXLLKOCUKJST-UHFFFAOYSA-N 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000008280 blood Substances 0.000 description 1
- 210000004369 blood Anatomy 0.000 description 1
- 230000017531 blood circulation Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000005034 decoration Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000005281 excited state Effects 0.000 description 1
- 230000001747 exhibiting effect Effects 0.000 description 1
- 208000013403 hyperactivity Diseases 0.000 description 1
- 230000007830 nerve conduction Effects 0.000 description 1
- 229910052760 oxygen Inorganic materials 0.000 description 1
- 239000001301 oxygen Substances 0.000 description 1
- 230000037361 pathway Effects 0.000 description 1
- 238000013139 quantization Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60Q—ARRANGEMENT OF SIGNALLING OR LIGHTING DEVICES, THE MOUNTING OR SUPPORTING THEREOF OR CIRCUITS THEREFOR, FOR VEHICLES IN GENERAL
- B60Q3/00—Arrangement of lighting devices for vehicle interiors; Lighting devices specially adapted for vehicle interiors
- B60Q3/80—Circuits; Control arrangements
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10H—ELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
- G10H1/00—Details of electrophonic musical instruments
- G10H1/0008—Associated control or indicating means
- G10H1/0016—Means for indicating which keys, frets or strings are to be actuated, e.g. using lights or leds
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02B—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO BUILDINGS, e.g. HOUSING, HOUSE APPLIANCES OR RELATED END-USER APPLICATIONS
- Y02B20/00—Energy efficient lighting technologies, e.g. halogen lamps or gas discharge lamps
- Y02B20/40—Control techniques providing energy savings, e.g. smart controller or presence detection
Landscapes
- Engineering & Computer Science (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Circuit Arrangement For Electric Light Sources In General (AREA)
Abstract
The application relates to the technical field of atmosphere lamp control, and particularly discloses a control method, a device, electronic equipment and a storage medium based on audio characteristics, wherein the method comprises the following steps: invoking a control box to receive sound signals from the sound acquisition terminal; calling an A2B chip in the control box to convert the sound signal into an analog signal; calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information; calling an analog-to-digital conversion chip to convert analog audio information to obtain audio data; extracting the characteristics of the audio data to obtain an audio characteristic information set; carrying out fusion processing on the audio feature information set to obtain fusion feature information; and generating an atmosphere lamp control instruction according to the fusion characteristic information to control the automobile atmosphere lamp. Therefore, the lamp control instruction is generated on the basis of the fused characteristics by extracting and fusing various characteristics of the audio data, so that the lamp control instruction can be more attached to the change of the original audio, and riding experience of a user is improved.
Description
Technical Field
The application relates to the technical field of atmosphere lamp control, in particular to a control method and device based on audio characteristics, electronic equipment and a storage medium.
Background
With the continuous development of the automobile industry, the individualization and the comfort of automobiles are gradually improved, and automobiles are not early simple vehicles, but are moving towards vehicles capable of providing a comfortable and agreeable experience for people. Nowadays, people demand vehicles, except appearance, interior decorations, power and brands, experience when the vehicles are restarted is further added, and an atmosphere lamp is used as a novel lamp for improving the experience when the vehicles are started, so that people are increasingly attracted.
Currently, an atmosphere lamp system generally generates a lamp control command according to a single characteristic of an input audio, or matches a preset lamp control command according to a single characteristic of a selected input audio, so as to perform atmosphere lamp control. However, the single feature cannot fully express the characteristic of the audio, so that the current atmosphere lamp control method causes low matching degree between the actual effect of the lamplight and the audio, and influences riding experience of a user.
Disclosure of Invention
In order to solve the problems in the prior art, the embodiment of the application provides a control method, a control device, electronic equipment and a storage medium based on audio characteristics, which can enable a lamp control instruction to be more fit with the change of original audio, so that riding experience of a driver or a passenger is enhanced.
In a first aspect, an embodiment of the present application provides a control method based on audio features, including:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, the audio feature information set includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In this embodiment, the process of fusing the audio feature information set to obtain fused feature information includes:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In this embodiment, generating an atmosphere lamp control command from the fusion feature information includes:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
In a second aspect, an embodiment of the present application provides a control device based on audio features, including:
the acquisition module is used for calling the control box to receive the sound signal from the sound acquisition terminal, calling the A2B chip in the control box to convert the sound signal into an analog signal, calling the audio acquisition circuit to acquire and process the analog signal to obtain analog audio information, and calling the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module is used for carrying out feature extraction on the audio data to obtain an audio feature information set;
the fusion module is used for carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information;
and the processing module is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, the audio feature information set includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In this embodiment, in performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module is configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In this embodiment, in generating the atmosphere lamp control command according to the fusion characteristic information, the processing module is configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
In a third aspect, an embodiment of the present application provides an electronic device, including: a processor and a memory, wherein the processor is connected to the memory, the memory is used for storing a computer program, and the processor is used for executing the computer program stored in the memory, so that the electronic device executes the method as in the first aspect.
In a fourth aspect, embodiments of the present application provide a computer-readable storage medium storing a computer program, the computer program causing a computer to perform the method as in the first aspect.
In a fifth aspect, embodiments of the present application provide a computer program product comprising a non-transitory computer readable storage medium storing a computer program, the computer being operable to cause a computer to perform a method as in the first aspect.
The implementation of the embodiment of the application has the following beneficial effects:
it can be seen that in the embodiment of the application, by extracting and fusing various features of the audio data, and then generating the lamp control instruction on the basis of the fused features, the lamp control instruction can be more fit with the change of the original audio, so that the riding experience of a driver or a passenger is enhanced.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present application, the drawings required for the description of the embodiments will be briefly described below, and it is obvious that the drawings in the following description are some embodiments of the present application, and other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a schematic hardware structure of a control device based on audio features according to an embodiment of the present application;
fig. 2 is a schematic flow chart of a control method based on audio features according to an embodiment of the present application;
fig. 3 is a functional block diagram of a control device based on audio features according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are some, but not all embodiments of the present application. All other embodiments, based on the embodiments of the application, which are apparent to those of ordinary skill in the art without inventive faculty, are intended to be within the scope of the application.
The terms "first," "second," "third," and "fourth" and the like in the description and in the claims and drawings are used for distinguishing between different objects and not necessarily for describing a particular sequential or chronological order. Furthermore, the terms "comprise" and "have," as well as any variations thereof, are intended to cover a non-exclusive inclusion. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those listed steps or elements but may include other steps or elements not listed or inherent to such process, method, article, or apparatus.
Reference herein to "an embodiment" means that a particular feature, result, or characteristic described in connection with the embodiment may be included in at least one embodiment of the application. The appearances of such phrases in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. Those skilled in the art will explicitly and implicitly understand that the embodiments described herein may be combined with other embodiments.
First, referring to fig. 1, fig. 1 is a schematic hardware structure of a control device based on audio features according to an embodiment of the present application. The audio feature based control device 100 comprises at least one processor 101, communication lines 102, a memory 103 and at least one communication interface 104.
In this embodiment, the processor 101 may be a general purpose central processing unit (central processing unit, CPU), microprocessor, application-specific integrated circuit (ASIC), or one or more integrated circuits for controlling the execution of the program according to the present application.
Communication line 102 may include a pathway to transfer information between the above-described components.
The communication interface 104, which may be any transceiver-like device (e.g., antenna, etc.), is used to communicate with other devices or communication networks, such as ethernet, RAN, wireless local area network (wireless local area networks, WLAN), etc.
The memory 103 may be, but is not limited to, a read-only memory (ROM) or other type of static storage device that can store static information and instructions, a random access memory (random access memory, RAM) or other type of dynamic storage device that can store information and instructions, or an electrically erasable programmable read-only memory (electrically erasable programmable read-only memory, EEPROM), a compact disc (compact disc read-only memory) or other optical disc storage, optical disc storage (including compact disc, laser disc, optical disc, digital versatile disc, blu-ray disc, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer.
In this embodiment, the memory 103 may be independently provided and connected to the processor 101 via the communication line 102. Memory 103 may also be integrated with processor 101. The memory 103 provided by embodiments of the present application may generally have non-volatility. The memory 103 is used for storing computer-executable instructions for executing the scheme of the present application, and is controlled by the processor 101 to execute the instructions. The processor 101 is configured to execute computer-executable instructions stored in the memory 103 to implement the methods provided in the embodiments of the present application described below.
In alternative embodiments, computer-executable instructions may also be referred to as application code, as the application is not particularly limited.
In alternative embodiments, processor 101 may include one or more CPUs, such as CPU0 and CPU1 in fig. 1.
In alternative embodiments, the audio feature based control device 100 may include multiple processors, such as processor 101 and processor 107 in FIG. 1. Each of these processors may be a single-core (single-CPU) processor or may be a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In an alternative embodiment, if the control device 100 based on the audio feature is a server, for example, it may be a stand-alone server, or may be a cloud server that provides cloud services, cloud databases, cloud computing, cloud functions, cloud storage, network services, cloud communication, middleware services, domain name services, security services, content delivery network (Content Delivery Network, CDN), and basic cloud computing services such as big data and artificial intelligence platform. The audio feature based control apparatus 100 may further comprise an output device 105 and an input device 106. The output device 105 communicates with the processor 101 and may display information in a variety of ways. For example, the output device 105 may be a liquid crystal display (liquid crystal display, LCD), a light emitting diode (light emitting diode, LED) display device, a Cathode Ray Tube (CRT) display device, or a projector (projector), or the like. The input device 106 is in communication with the processor 101 and may receive user input in a variety of ways. For example, the input device 106 may be a mouse, a keyboard, a touch screen device, a sensing device, or the like.
The control apparatus 100 based on the audio characteristics may be a general-purpose device or a special-purpose device. The embodiment of the present application does not limit the type of the control device 100 based on the audio characteristics.
Referring to fig. 2, fig. 2 is a flow chart of a control method based on audio features according to an embodiment of the present application. The control method based on the audio characteristics comprises the following steps:
201: and calling the control box to receive the sound signal from the sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signal into an analog signal.
In this embodiment, the audio data may be collected by a sound collection terminal provided inside the automobile, or may be output by a music playing device integrated in the vehicle, such as an on-board DVD, or may be output by an external handheld device such as a mobile phone that can be wired or wirelessly connected to the vehicle.
For example, the sound collection terminal may be a microphone, and for example, a microphone may be disposed on each door beside each seat in the vehicle, so as to collect audio data in the vehicle. After the microphone collects the sound in the car, such as human voice, the collected sound can be transmitted to the control box through the car audio bus (A2B), and the audio data is converted into analog signals after passing through the A2B chip.
202: and calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data.
In the embodiment, the analog signal is processed by an analog-to-digital conversion chip to form audio data, and the audio data is transmitted to a micro control unit (Microcontroller Unit, MCU) for feature extraction.
203: and extracting the characteristics of the audio data to obtain an audio characteristic information set.
In this embodiment, the audio feature information set may include: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information of the audio data. For example, the emotion feature information, rhythm feature information, dynamics feature information, and speed feature information may all be represented by a one-dimensional array.
Illustratively, taking emotion feature information as an example, pre-defining emotion features includes: happiness, sadness, pain, fear, anger, happiness and calm, and represent these emotional characteristics by numerals or symbols, respectively. For example, happy with 1, sad with 2, painful with 3, fear with 4, anger with 5, happy with 6 and calm with 7.
In the present embodiment, since audio data is acquired through a microphone provided in the vehicle, the audio data may include both human voice and music. Based on the above, when emotion recognition is performed, voice and music in the audio data can be split in advance, emotion characteristics of the voice part and the music part are extracted respectively, and finally the overall emotion of the audio data is determined based on the two emotion characteristics.
Specifically, in the human voice section, the speech, intonation, speech speed, etc. of one person speaking will change correspondingly with the change of emotion. Exemplary: in a happy state, a person can secrete more adrenal hormones than usual, and the excessive adrenal hormones can lead to the person in the happy state always being in an excited state. At the same time, the state of hyperactivity accelerates blood flow and nerve conduction, which in turn results in adequate blood and oxygen supply to the brain, faster heart beat, faster thinking, in turn exhibiting faster thinking, more fluent expression, louder sounds, and faster speech.
In addition, when a person is in different emotions, the speaking sound also fluctuates due to the influence of the emotion, for example: high decibels for happiness or anger, low decibels for sadness or pain, etc. Based on this, in an alternative embodiment, the db value variation characteristic of the voice portion may also be determined, and then the emotion characteristic corresponding to the voice portion is determined in combination with the db value variation characteristic.
By way of example, a weight table of different emotions of a person at different sound decibel values can be established by analytical learning of a large amount of historical data. For example: in the interval of 65-85 dB, through analysis and learning of a large number of real conversation scenes under the dB, the weight of a person in the scene in a happy emotion is 0.2, the weight of the person in a sad emotion is 0.05, the weight of the person in an anger emotion is 0.6 and the weight of the person in a calm emotion is 0.02. Therefore, after the human voice part is obtained, the weight of each emotion corresponding to the decibel value interval can be obtained as the characteristic of sound loudness by determining the decibel value of the sound and then inquiring the table according to the decibel value.
Thus, in this embodiment, the speech, intonation, speech speed, sound loudness, etc. of a person speaking in different emotions are changed accordingly. Therefore, the characteristics of the voice, the intonation, the speech speed, the accent distribution and the loudness of the voice can be obtained by extracting the characteristics of the voice, the intonation, the speech speed, the accent distribution and the loudness of the voice part, and then the characteristics are analyzed correspondingly, so that the emotion characteristics corresponding to the voice part are obtained.
Specifically, the probability that the voice part corresponds to each emotion can be obtained through analysis of characteristics such as voice, intonation, speech speed, accent distribution and the like, then the product of the probability and weight values corresponding to different emotions under the decibel is carried out, and then the product result is used as the final result of audio emotion recognition. By way of example, assuming that the dialogue has a decibel of 75 and falls within the interval of 75 to 85 decibels, by the feature analysis, it is obtained that the probability of being a happy emotion is 0.1, the probability of being a sad emotion is 0.1, the probability of being an angry emotion is 0.8, and the probability of being a calm emotion is 0. Based on the above, the final probability can be obtained by multiplying the weight corresponding to each emotion under the interval: the probability of a happy emotion is 0.1X0.2 =0.02, the probability of a sad emotion is 0.1X0.05 =0.005, the probability of an angry emotion is 0.8X0.6 =0.48, and the probability of a calm emotion is 0X0.02 =0. Thus, the probability distribution is taken as an emotional characteristic of the vocal part.
Meanwhile, in an alternative implementation mode, text conversion can be carried out on the content of the voice part, talking content is determined, then the reason of the dialogue is analyzed according to the talking content, and then the emotion recognition precision of the voice part is further improved.
And for the music part, the probability distribution of each emotion corresponding to the music can be determined by extracting the characteristics of the music part such as the wind, the rhythm, the tone and the like, and the probability distribution is used as the emotion characteristics of the music part.
Finally, in the present embodiment, the weights of the human voice part and the music part may be determined according to the decibel value of the human voice part. Specifically, the greater the decibel value of the vocal portion, the less affected the speaker is by the music, which is about the content of the conversation whose dominant emotion has been played by the speaker. Therefore, the greater the decibel value of the human voice part, the greater the weight corresponding to the human voice part. And according to the determined weight, weighting and summing each emotion probability distribution of the voice part and each emotion probability distribution of the music part according to the corresponding relation of each emotion to obtain the comprehensive probability distribution of each emotion.
Illustratively, the probability of a happy emotion in the vocal part is 0.5 and the probability in the music part is 0.8. Through weight calculation, the weight of the vocal part is 0.8, and the weight of the music part is 0.2. The overall probability of happy emotion is: 0.5x0.8+0.8x0.2=0.56. Thus, the emotion having the highest overall probability among the overall probability distributions of the emotions can be used as the emotion feature of the audio data.
In this embodiment, for emotion feature information, each item of data in the emotion feature information may be defined in advance, emotion features corresponding to each second of audio data, and the order of data in the emotion feature information is arranged in the chronological order of the audio data.
Specifically, as for emotion feature information a= [1,1,1,2,2,5,5,7,6], it is shown that in the audio data, the emotion feature of the first 3 seconds is happy, the emotion feature of the 4 th and 5 th seconds is sad, the emotion feature of the 6 th and 7 th seconds is cheerful, the emotion feature of the 8 th second is calm, and the emotion feature of the 9 th second is happy.
Based on the above, the rhythm feature information, the dynamics feature information, and the speed feature information may be quantized according to a preset quantization rule, so as to obtain a one-dimensional array in time sequence.
204: and carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information.
In this embodiment, emotion feature information, rhythm feature information, dynamics feature information, and speed feature information may be sequentially subjected to longitudinal splicing processing to obtain fusion feature information.
Illustratively, for emotion feature information a= [1,1,1,2,2,5,5,7,6], rhythm feature information b= [1,5,6,8,4,2,2,6,5], dynamics feature information c= [2,2,2,2,1,1,2,2,1] and speed feature information d= [2,2,2,2,2,2,3,3,1]. After longitudinal stitching, a 4×9 matrix E can be obtained:
in this embodiment, the matrix E is the fused feature information after fusion, where the j-th column of the matrix E is used to represent emotion, rhythm, strength and speed parameters of the atmosphere lamp corresponding to the audio data in the j-th second. For example, for the first columnIndicating that the emotion parameter corresponding to the first second of the audio data is 1, the rhythm parameter is 1, the dynamics parameter is 2, and the speed parameter is 2.
205: and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In this embodiment, after the fusion feature information is obtained, the light color change information may be determined according to the first line of the fusion feature information; determining brightness change information according to the second row of the fusion characteristic information; determining switch change information according to a third row of the fusion characteristic information; and determining saturation change information according to the fourth line of the fusion characteristic information.
Therefore, the light color change information, the brightness change information, the switching change information and the saturation change information are used as atmosphere lamp control instructions to control the automobile atmosphere lamps to display light colors corresponding to the saturation and the brightness, and to control the switching of the automobile atmosphere lamps.
For example, the light color corresponding to emotion feature information may be predefined, for example, with red for happiness, blue for sadness, gray for pain, purple for fear, yellow for cheerful, orange for happiness, and green for calm. Therefore, according to the first line of the fusion characteristic information representing the emotion characteristics, the photochromic corresponding to each second of the audio data can be obtained, and then the photochromic change information is obtained.
Similarly, the brightness corresponding to the rhythm feature information, the switching condition corresponding to the dynamics feature information and the light color saturation corresponding to the speed feature information can be predefined, and then the corresponding atmosphere lamp control instruction is determined according to the fusion feature information.
Specifically, the first column of the matrix E is followedThe first column indicates that the color of the audio data corresponding to the first second is red, the brightness is low, the on condition is on, and the saturation is 20%.
It can be seen that in the embodiment of the application, by extracting and fusing various features of the audio data, and then generating the lamp control instruction on the basis of the fused features, the lamp control instruction can be more fit with the change of the original audio, so that the riding experience of a driver or a passenger is enhanced.
Referring to fig. 3, fig. 3 is a functional block diagram of a control device based on audio features according to an embodiment of the present application. As shown in fig. 3, the audio feature-based control apparatus 300 includes:
the acquisition module 301 is configured to invoke the control box to receive a sound signal from the sound acquisition terminal, invoke an A2B chip in the control box to convert the sound signal into an analog signal, invoke the audio acquisition circuit to acquire and process the analog signal to obtain analog audio information, and invoke the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module 302 is configured to perform feature extraction on the audio data to obtain an audio feature information set;
the fusion module 303 is configured to perform fusion processing on the audio feature information set to obtain fusion feature information;
and the processing module 304 is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In an embodiment of the present application, the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In the embodiment of the present application, in terms of performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module 303 is specifically configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In the embodiment of the present application, in generating the atmosphere lamp control command according to the fusion characteristic information, the processing module 304 is specifically configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
Referring to fig. 4, fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present application. As shown in fig. 4, the electronic device 400 includes a transceiver 401, a processor 402, and a memory 403. Which are connected by a bus 404. The memory 403 is used for storing computer programs and data, and the data stored in the memory 403 can be transferred to the processor 402.
The processor 402 is configured to read a computer program in the memory 403 to perform the following operations:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
In an embodiment of the present application, the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
In an embodiment of the present application, the processor 402 is specifically configured to perform the following operations in terms of performing fusion processing on the audio feature information set to obtain fusion feature information:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain fusion characteristic information.
In an embodiment of the present application, the processor 402 is specifically configured to perform the following operations in generating the atmosphere lamp control command according to the fusion characteristic information:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to a third row of the fusion characteristic information;
determining saturation change information according to a fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
It should be understood that the control device based on the audio feature in the present application may include a smart Phone (such as an Android mobile Phone, an iOS mobile Phone, a Windows Phone mobile Phone, etc.), a tablet computer, a palm computer, a notebook computer, a mobile internet device MID (Mobile Internet Devices, abbreviated as MID) or a wearable device, etc. The above-described audio feature-based control devices are merely examples and are not intended to be exhaustive, including but not limited to the above-described audio feature-based control devices. In practical applications, the control device based on the audio features may further include: intelligent vehicle terminals, computer devices, etc.
From the above description of embodiments, it will be apparent to those skilled in the art that the present application may be implemented in software in combination with a hardware platform. With such understanding, all or part of the technical solution of the present application contributing to the background art may be embodied in the form of a software product, which may be stored in a storage medium, such as ROM/RAM, a magnetic disk, an optical disk, etc., including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform the methods described in the various embodiments or parts of the embodiments of the present application.
Accordingly, the present application also provides a computer-readable storage medium storing a computer program that is executed by a processor to implement some or all of the steps of any one of the audio feature-based control methods described in the above method embodiments. For example, the storage medium may include a hard disk, a floppy disk, an optical disk, a magnetic tape, a magnetic disk, a flash memory, etc.
Embodiments of the present application also provide a computer program product comprising a non-transitory computer-readable storage medium storing a computer program operable to cause a computer to perform part or all of the steps of any one of the audio feature-based control methods described in the method embodiments above.
It should be noted that, for simplicity of description, the foregoing method embodiments are all described as a series of acts, but it should be understood by those skilled in the art that the present application is not limited by the order of acts described, as some steps may be performed in other orders or concurrently in accordance with the present application. Further, those skilled in the art will also appreciate that the embodiments described in the specification are alternative embodiments, and that the acts and modules involved are not necessarily required for the present application.
In the foregoing embodiments, the descriptions of the embodiments are focused on, and for those portions of one embodiment that are not described in detail, reference may be made to the related descriptions of other embodiments.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus may be implemented in other manners. For example, the apparatus embodiments described above are merely illustrative, such as the division of the units, merely a logical function division, and there may be additional divisions when actually implemented, such as multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, or may be in electrical or other forms.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, each unit may exist alone physically, or two or more units may be integrated in one unit. The integrated units described above may be implemented either in hardware or in software program modules.
The integrated units, if implemented in the form of software program modules, may be stored in a computer-readable memory for sale or use as a stand-alone product. Based on this understanding, the technical solution of the present application may be embodied essentially or partly in the form of a software product or all or part of the technical solution, which is stored in a memory, and includes several instructions for causing a computer device (which may be a personal computer, a server, a network device, or the like) to perform all or part of the steps of the method according to the embodiments of the present application. And the aforementioned memory includes: a U-disk, a Read-Only Memory (ROM), a random access Memory (RAM, random Access Memory), a removable hard disk, a magnetic disk, or an optical disk, or other various media capable of storing program codes.
Those of ordinary skill in the art will appreciate that all or a portion of the steps in the various methods of the above embodiments may be implemented by a program that instructs associated hardware, and the program may be stored in a computer readable memory, and the memory may include: flash disk, read-Only Memory (ROM), random access Memory (Random Access Memory, RAM), magnetic disk or optical disk.
The foregoing has outlined rather broadly the more detailed description of the embodiments of the application in order that the detailed description of the principles and embodiments of the application may be implemented in conjunction with the detailed description of the embodiments that follows, the claims being merely intended to facilitate the understanding of the method and concepts underlying the application; meanwhile, as those skilled in the art will have variations in the specific embodiments and application scope in accordance with the ideas of the present application, the present description should not be construed as limiting the present application in view of the above.
Claims (10)
1. A control method based on audio characteristics, the control method comprising:
calling a control box to receive sound signals from a sound acquisition terminal, and calling an A2B chip in the control box to convert the sound signals into analog signals;
calling an audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling an analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
extracting the characteristics of the audio data to obtain an audio characteristic information set;
carrying out fusion processing on the audio feature information set to obtain fusion feature information;
and generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
2. The control method according to claim 1, wherein,
the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
3. The control method according to claim 2, wherein the fusing the audio feature information set to obtain fused feature information includes:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain the fusion characteristic information.
4. The control method according to claim 3, wherein generating an atmosphere lamp control command according to the fusion characteristic information comprises:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to the third row of the fusion characteristic information;
determining saturation change information according to the fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as the atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
5. A control device based on audio characteristics, the control device comprising:
the acquisition module is used for calling the control box to receive sound signals from the sound acquisition terminal, calling the A2B chip in the control box to convert the sound signals into analog signals, calling the audio acquisition circuit to acquire and process the analog signals to obtain analog audio information, and calling the analog-to-digital conversion chip to convert the analog audio information to obtain audio data;
the feature extraction module is used for carrying out feature extraction on the audio data to obtain an audio feature information set;
the fusion module is used for carrying out fusion processing on the audio characteristic information set to obtain fusion characteristic information;
and the processing module is used for generating an atmosphere lamp control instruction according to the fusion characteristic information so as to control the automobile atmosphere lamp.
6. The control device according to claim 5, wherein,
the set of audio feature information includes: emotion characteristic information, rhythm characteristic information, dynamics characteristic information and speed characteristic information;
the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information are all one-dimensional arrays.
7. The control device according to claim 6, wherein in the aspect of performing fusion processing on the audio feature information set to obtain fusion feature information, the fusion module is configured to:
and carrying out longitudinal splicing processing on the emotion characteristic information, the rhythm characteristic information, the dynamics characteristic information and the speed characteristic information in sequence to obtain the fusion characteristic information.
8. The control device according to claim 7, wherein in the generating an atmosphere lamp control instruction according to the fusion characteristic information, the fusion module is configured to:
determining light color change information according to the first row of the fusion characteristic information;
determining brightness change information according to the second row of the fusion characteristic information;
determining switch change information according to the third row of the fusion characteristic information;
determining saturation change information according to the fourth line of the fusion characteristic information;
and taking the light color change information, the brightness change information, the switching change information and the saturation change information as the atmosphere lamp control instructions, controlling the automobile atmosphere lamp to display light colors corresponding to the saturation and the brightness, and controlling the switching of the automobile atmosphere lamp.
9. An electronic device comprising a processor, a memory, a communication interface, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the processor, the one or more programs comprising instructions for performing the steps in the method of any of claims 1-4.
10. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program, which is executed by a processor to implement the method of any of claims 1-4.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2022104566785 | 2022-04-28 | ||
CN202210456678 | 2022-04-28 |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116968631A true CN116968631A (en) | 2023-10-31 |
Family
ID=88477327
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202310409908.7A Pending CN116968631A (en) | 2022-04-28 | 2023-04-17 | Control method and device based on audio characteristics, electronic equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116968631A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117828537A (en) * | 2024-03-04 | 2024-04-05 | 北京建筑大学 | Music emotion recognition method and device based on CBA model |
-
2023
- 2023-04-17 CN CN202310409908.7A patent/CN116968631A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117828537A (en) * | 2024-03-04 | 2024-04-05 | 北京建筑大学 | Music emotion recognition method and device based on CBA model |
CN117828537B (en) * | 2024-03-04 | 2024-05-17 | 北京建筑大学 | Music emotion recognition method and device based on CBA model |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110838286B (en) | Model training method, language identification method, device and equipment | |
CN112863547B (en) | Virtual resource transfer processing method, device, storage medium and computer equipment | |
US20190394057A1 (en) | Device and method for adaptively providing meeting | |
US20200126566A1 (en) | Method and apparatus for voice interaction | |
CN107392124A (en) | Emotion identification method, apparatus, terminal and storage medium | |
WO2021004481A1 (en) | Media files recommending method and device | |
US20190371319A1 (en) | Method for human-machine interaction, electronic device, and computer-readable storage medium | |
CN113010138B (en) | Article voice playing method, device and equipment and computer readable storage medium | |
CN113160819B (en) | Method, apparatus, device, medium, and product for outputting animation | |
CN109754792A (en) | Voice interface device and the voice interface method for applying it | |
CN109102802A (en) | System for handling user spoken utterances | |
CN107545029A (en) | Voice feedback method, equipment and the computer-readable recording medium of smart machine | |
CN108281138A (en) | Age discrimination model training and intelligent sound exchange method, equipment and storage medium | |
CN116968631A (en) | Control method and device based on audio characteristics, electronic equipment and storage medium | |
KR20190005103A (en) | Electronic device-awakening method and apparatus, device and computer-readable storage medium | |
CN109712644A (en) | Method based on speech recognition emotional change control LED display effect, the apparatus and system for controlling LED display effect | |
CN109377979B (en) | Method and system for updating welcome language | |
CN113327580A (en) | Speech synthesis method, device, readable medium and electronic equipment | |
CN108701333A (en) | Electronic equipment for providing content and its control method | |
CN110728983A (en) | Information display method, device, equipment and readable storage medium | |
CN113643684B (en) | Speech synthesis method, device, electronic equipment and storage medium | |
CN106372203A (en) | Information response method and device for smart terminal and smart terminal | |
CN112644375B (en) | Mood perception-based in-vehicle atmosphere lamp adjusting method, system, medium and terminal | |
EP4293664A1 (en) | Voiceprint recognition method, graphical interface, and electronic device | |
CN108877795B (en) | Method and apparatus for presenting information |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication |