CN115410544B - Sound effect processing method and device and electronic equipment - Google Patents

Sound effect processing method and device and electronic equipment Download PDF

Info

Publication number
CN115410544B
CN115410544B CN202211037097.4A CN202211037097A CN115410544B CN 115410544 B CN115410544 B CN 115410544B CN 202211037097 A CN202211037097 A CN 202211037097A CN 115410544 B CN115410544 B CN 115410544B
Authority
CN
China
Prior art keywords
song
processed
songs
sound effect
audio
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202211037097.4A
Other languages
Chinese (zh)
Other versions
CN115410544A (en
Inventor
夏妍
林锋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Mgjia Beijing Technology Co ltd
Original Assignee
Mgjia Beijing Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Mgjia Beijing Technology Co ltd filed Critical Mgjia Beijing Technology Co ltd
Priority to CN202211037097.4A priority Critical patent/CN115410544B/en
Publication of CN115410544A publication Critical patent/CN115410544A/en
Application granted granted Critical
Publication of CN115410544B publication Critical patent/CN115410544B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0091Means for obtaining special acoustic effects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/65Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/60Information retrieval; Database structures therefor; File system structures therefor of audio data
    • G06F16/68Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/683Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/685Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using automatically derived transcript of audio data, e.g. lyrics

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Library & Information Science (AREA)
  • Databases & Information Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Acoustics & Sound (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The invention discloses a sound effect processing method, a sound effect processing device and electronic equipment, wherein the sound effect processing method comprises the following steps of: acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; classifying the styles of the songs to be processed according to text information for representing the types of the songs and an audio byte array of the songs to be processed; determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result; acquiring audio target characteristics of songs to be processed; performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and storing the attribute information of the to-be-processed song, the compensating sound effect of the to-be-processed song and the style sound effect of the to-be-processed song in an associated mode.

Description

Sound effect processing method and device and electronic equipment
Technical Field
The invention relates to the technical field of intelligent sound effects, in particular to a sound effect processing method and device and electronic equipment.
Background
When selecting sound effects, the current intelligent sound effect technology generally uses a simpler rule, such as matching corresponding sound effects for song genre information provided by a song producer, and because the song genre information is very general, the sound effects of the intelligent sound effect system matched with music based on genre are very rough, even the actual output result of the intelligent sound effects is possibly inaccurate, completely unsuitable sound effects can be appointed for songs, and the sound effect adding effect is affected.
Disclosure of Invention
Therefore, the invention aims to overcome the defect of inaccurate effective allocation of songs Qu Yin in the prior art, and further provides a sound effect processing method, a sound effect processing device and electronic equipment.
According to a first aspect, an embodiment of the present invention discloses an audio processing method, including: acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte arrays of the songs to be processed; determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result; acquiring audio target characteristics of songs to be processed; performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and storing the attribute information of the to-be-processed song, the compensating sound effect of the to-be-processed song and the style sound effect of the to-be-processed song in an associated mode.
Optionally, the method further comprises: and when receiving the sound effect application request, responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played.
Optionally, when receiving the sound effect application request, responding to an adding operation of the style sound effect and/or the compensation sound effect of the song to be played, including: when receiving the audio effect application request, performing song matching operation according to attribute information of the song to be played; when the corresponding song is matched, the similarity comparison is carried out between the audio of the song to be played and the matched song audio; and when the similarity is larger than a preset threshold, adding corresponding sound effects to the song to be played.
Optionally, the method further comprises: and displaying the identification information of the currently added sound effect on the client side while responding to the adding operation of the sound effect to the song to be played.
Optionally, when the matching is to the corresponding song, the similarity comparison is performed between the audio of the song to be played and the matched song audio, including: obtaining a first frequency distribution vector according to the frequency distribution of the audio of the song to be played in the target duration; determining a second frequency distribution vector matched with the corresponding song audio within the target duration; and performing similarity comparison on the first frequency distribution vector and the second frequency distribution vector.
According to a second aspect, an embodiment of the present invention further discloses an audio processing apparatus, including: the first acquisition module is used for acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; the classification module is used for classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte arrays of the songs to be processed; the first determining module is used for determining the style sound effect of the song to be processed by utilizing a preset sound effect decision method according to the style classification result; the second acquisition module is used for acquiring the audio target characteristics of the songs to be processed; the second determining module is used for performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and the first storage module is used for carrying out associated storage on the attribute information of the songs to be processed, the compensating sound effects of the songs to be processed and the style sound effects of the songs to be processed.
Optionally, the apparatus further comprises: and the first response module is used for responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played when the sound effect application request is received.
Optionally, the first response module includes: the matching sub-module is used for carrying out song matching operation according to the attribute information of the song to be played when receiving the audio application request; the comparison sub-module is used for comparing the similarity between the audio of the song to be played and the matched song audio when the corresponding song is matched; and the application submodule is used for adding corresponding sound effects to the song to be played when the similarity is larger than a preset threshold value.
According to a third aspect, an embodiment of the present invention further discloses an electronic device, including: at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor to cause the at least one processor to perform the steps of the sound effect processing method according to the first aspect or any alternative implementation of the first aspect.
According to a fourth aspect, an embodiment of the present invention also discloses a computer-readable storage medium, on which a computer program is stored, which computer program, when being executed by a processor, implements the steps of the sound effect processing method according to the first aspect or any of the alternative embodiments of the first aspect.
The technical scheme of the invention has the following advantages:
the sound effect processing method/device provided by the invention comprises the following steps: acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; classifying the styles of the songs to be processed according to text information for representing the types of the songs and an audio byte array of the songs to be processed; determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result; acquiring audio target characteristics of songs to be processed; performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and storing attribute information of the to-be-processed song, compensating sound effect of the to-be-processed song and style sound effect of the to-be-processed song in an associated mode. According to the method, the songs to be processed are classified through the audio byte array of the songs to be processed and the text information representing the types of the songs, and then the style sound effect of the audio to be processed is determined through the preset sound effect decision method by the classification result, so that the classification of the songs is richer, the obtained style sound effect is better, the audio compensation sound effect of the songs to be processed is obtained through the sound effect compensation operation on the audio target characteristics of the songs to be processed, the obtained compensation sound effect is applied to the songs to be processed, and the problem of poor effect of the songs to be processed can be compensated. And the compensating sound effect and the style sound effect of the song to be processed are stored in an associated mode, and then a better effect can be obtained when the sound effect is applied, so that user experience is improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings that are needed in the description of the embodiments or the prior art will be briefly described, and it is obvious that the drawings in the description below are some embodiments of the present invention, and other drawings can be obtained according to the drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flowchart of a specific example of an audio processing method according to an embodiment of the present invention;
FIG. 2 is a schematic block diagram of a specific example of an audio processing device according to an embodiment of the present invention;
fig. 3 is a schematic diagram of an electronic device according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made apparent and fully in view of the accompanying drawings, in which some, but not all embodiments of the invention are shown. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
In the description of the present invention, it should be noted that the directions or positional relationships indicated by the terms "center", "upper", "lower", "left", "right", "vertical", "horizontal", "inner", "outer", etc. are based on the directions or positional relationships shown in the drawings, are merely for convenience of describing the present invention and simplifying the description, and do not indicate or imply that the devices or elements referred to must have a specific orientation, be configured and operated in a specific orientation, and thus should not be construed as limiting the present invention. Furthermore, the terms "first," "second," and "third" are used for descriptive purposes only and are not to be construed as indicating or implying relative importance.
In the description of the present invention, it should be noted that, unless explicitly specified and limited otherwise, the terms "mounted," "connected," and "connected" are to be construed broadly, and may be either fixedly connected, detachably connected, or integrally connected, for example; can be mechanically or electrically connected; the two components can be directly connected or indirectly connected through an intermediate medium, or can be communicated inside the two components, or can be connected wirelessly or in a wired way. The specific meaning of the above terms in the present invention will be understood in specific cases by those of ordinary skill in the art.
In addition, the technical features of the different embodiments of the present invention described below may be combined with each other as long as they do not collide with each other.
The embodiment of the invention discloses an audio processing method, as shown in fig. 1, comprising the following steps:
step 101, obtaining text information which corresponds to the song to be processed and is used for representing the type of the song.
For example, the text information corresponding to the song to be processed may include, but is not limited to, lyrics, popular comments, an affiliated song list, and the like, and the cloud acquires the text information corresponding to the song to be processed and used for characterizing the song type.
And 102, classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte array of the songs to be processed.
Illustratively, the cloud processes text information to be used to characterize the song type and an array of audio bytes of the song to determine a genre classification for the song to be processed. Specifically, the genre classification of the song to be processed is determined together from the song style classification obtained by inputting an array of audio bytes of the song into the convolutional neural network (Convolutional Neural Networks, CNN) and the text information characterizing the song type, the genre classification includes genre classification including popularity, rock, classical, bruise, etc., the instrument classification includes piano, guitar, drum, violin, etc., the BPM classification includes slow, general tempo, joy, mood classification includes agitation, happiness, calm, anxiety, etc.
And step 103, determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result.
For example, the preset sound effect decision method may be a method for determining a corresponding sound effect for a song by its genre classification. Specifically, the preset sound effect decision method comprises the step of performing sound effect decision on songs to be processed based on classification of each dimension through a random forest.
Step 104, obtaining the audio target characteristics of the song to be processed. Illustratively, the audio target characteristics of the song to be processed include frequency distribution of the song audio, reverberant magnitude, sound field width.
And 105, performing sound effect compensation operation according to the audio target characteristics of the song to be processed to obtain the audio compensation sound effect of the song to be processed. Illustratively, according to the obtained audio target characteristics of the song to be processed, performing an audio effect compensation operation on the audio of the song to be processed, wherein the audio effect compensation operation comprises the steps of applying increasing low frequency, reducing low frequency, widening a sound field and adding reverberation, performing the sound field widening and adding reverberation operation when the voice reverberation of the song audio is small and the sound field of the song is narrow, performing the low frequency increasing operation when the low frequency energy of the song audio is small, and performing the low frequency reducing operation when the low frequency energy of the song audio is too high, so as to obtain the audio effect compensation of the song to be processed.
And 106, storing the attribute information of the to-be-processed song, the compensating sound effect of the to-be-processed song and the style sound effect of the to-be-processed song in an associated mode. The attribute information of the song to be processed includes, for example, song audio, song title, artist, album title. In the embodiment of the application, the audio effect of the audio to be processed finally stored in the cloud is superposition of the audio compensation audio effect of the song to be processed and the corresponding style audio effect. The current intelligent sound effect system can apply the same sound effect to a whole song, and can cause poor hearing to songs with huge style differences among different sections.
The sound effect processing method provided by the invention comprises the following steps: acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; classifying the styles of the songs to be processed according to text information for representing the types of the songs and an audio byte array of the songs to be processed; determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result; acquiring audio target characteristics of songs to be processed; performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and storing the attribute information of the to-be-processed song, the compensating sound effect of the to-be-processed song and the style sound effect of the to-be-processed song in an associated mode. According to the method, the songs to be processed are classified through the audio byte array of the songs to be processed and the text information representing the types of the songs, and then the style sound effect of the audio to be processed is determined through the preset sound effect decision method by the classification result, so that the classification of the songs is richer, the obtained style sound effect is better, the audio compensation sound effect of the songs to be processed is obtained through the sound effect compensation operation on the audio target characteristics of the songs to be processed, the obtained compensation sound effect is applied to the songs to be processed, and the problem of poor effect of the songs to be processed can be compensated. And the compensating sound effect and the style sound effect of the song to be processed are stored in an associated mode, and then a better effect can be obtained when the sound effect is applied, so that user experience is improved.
As an optional embodiment of the present invention, the method further comprises: and when receiving the sound effect application request, responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played.
Illustratively, when a sound effect application request sent by a client is received, a corresponding sound effect is added to a song to be played. Specifically, when a song to be played hits a certain song stored in the cloud, the corresponding sound effects are applied during playing, the applied sound effects are segmented, and the sound effects among different segments can be smoothly switched.
As an optional implementation manner of the invention, when receiving the sound effect application request, the adding operation of responding to the style sound effect and/or the compensation sound effect of the song to be played comprises the following steps: when receiving the audio effect application request, performing song matching operation according to attribute information of the song to be played; when the corresponding song is matched, the similarity comparison is carried out between the audio of the song to be played and the matched song audio; and when the similarity is larger than a preset threshold, adding corresponding sound effects to the song to be played.
The attribute information of the song to be played includes a song name, a singer and an album name, when the song is matched with the corresponding song, the similarity of the song to be played and the matched song is compared, and when the similarity is larger than a preset threshold, corresponding sound effects are added to the song to be played, wherein the preset threshold is an experience value calculated and accumulated through a large amount of data.
In the prior art, the identification of songs mainly depends on song names, singers and album names, so that the fact that the audio version of each large music platform in actual life sometimes has no sound is ignored, and the user can not easily determine the audio version of the song when playing the music by using the USB equipment is omitted.
As an optional embodiment of the present invention, the method further comprises: and displaying the identification information of the currently added sound effect on the client side while responding to the adding operation of the sound effect to the song to be played.
For example, when a song to be played has added a corresponding sound effect, the identification information of the previously added sound effect is displayed on the client, and the identification information of the currently added sound effect includes the name and description information of the sound effect.
The existing intelligent sound effect is presented to the user, the sound effect result is a complete black box, the user cannot intuitively sense what effect the intelligent sound effect actually plays, the identification information of the currently added sound effect is displayed on the client, the sound effect and the specific description of the specific application can be displayed on the user interface, and the user feel is better.
As an optional embodiment of the present invention, when matching to a corresponding song, the similarity comparison between the audio of the song to be played and the matched song audio includes: obtaining a first frequency distribution vector according to the frequency distribution of the audio of the song to be played in the target duration; determining a second frequency distribution vector matched with the corresponding song audio in the target duration; and performing similarity comparison on the first frequency distribution vector and the second frequency distribution vector.
The method includes the steps of comparing a first frequency distribution vector of audio of a song to be played in a target duration with a first frequency distribution vector of audio of a matched song in the target duration, and further determining similarity of the two songs.
Specifically, when a user plays a certain song, firstly, the song name, the singer and the album name are used for matching, when the matching of the song is achieved, a histogram of the frequency distribution of the audio of 20 seconds before the audio of the song to be played is calculated and converted into a frequency distribution vector, the frequency distribution vector of the audio of 20 seconds before the audio of the song to be played and the frequency distribution vector of 20 seconds before the matched audio of the corresponding song are subjected to similarity comparison, and the audio effect is applied to the played song only if the similarity is larger than a certain threshold value, and the specific audio effect description of the applied audio effect can be displayed on an interface.
The embodiment of the invention also discloses an audio processing device, as shown in fig. 2, which comprises: comprising the following steps: a first obtaining module 201, configured to obtain text information corresponding to a song to be processed and used for characterizing a type of the song; the classification module 202 is configured to classify the to-be-processed song according to the text information for characterizing the song type and the audio byte array of the to-be-processed song; a first determining module 203, configured to determine, according to the style classification result, a style sound effect of the song to be processed by using a preset sound effect decision method; a second obtaining module 204, configured to obtain an audio target feature of the song to be processed; a second determining module 205, configured to perform an audio compensation operation according to the audio target feature of the song to be processed to obtain an audio compensation audio of the song to be processed; the first storage module 206 is configured to store attribute information of the to-be-processed song, a compensating sound effect of the to-be-processed song, and a style sound effect of the to-be-processed song in an associated manner.
The sound effect processing device provided by the invention comprises: the first acquisition module is used for acquiring text information which corresponds to the song to be processed and is used for representing the type of the song; the classification module is used for classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte arrays of the songs to be processed; the first determining module is used for determining the style sound effect of the song to be processed by utilizing a preset sound effect decision method according to the style classification result; the second acquisition module is used for acquiring the audio target characteristics of the songs to be processed; the second determining module is used for performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed; and the first storage module is used for carrying out associated storage on the attribute information of the songs to be processed, the compensating sound effects of the songs to be processed and the style sound effects of the songs to be processed. According to the device, the songs to be processed are classified through the audio byte array of the songs to be processed and the text information representing the types of the songs, and then the style sound effect of the audio to be processed is determined through the preset sound effect decision method by the classification result, so that the classification of the songs is richer, the obtained style sound effect is better, the audio compensation sound effect of the songs to be processed is obtained through the sound effect compensation operation on the audio target characteristics of the songs to be processed, the obtained compensation sound effect is applied to the songs to be processed, and the problem of poor effect of the songs to be processed can be compensated. And the compensating sound effect and the style sound effect of the song to be processed are stored in an associated mode, and then a better effect can be obtained when the sound effect is applied, so that user experience is improved.
As an alternative embodiment of the present invention, the apparatus further comprises: and the first response module is used for responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played when the sound effect application request is received.
As an optional embodiment of the present invention, the first response module includes: the matching sub-module is used for carrying out song matching operation according to the attribute information of the song to be played when receiving the audio application request; the comparison sub-module is used for comparing the similarity between the audio of the song to be played and the matched song audio when the corresponding song is matched; and the application submodule is used for adding corresponding sound effects to the song to be played when the similarity is larger than a preset threshold value.
As an alternative embodiment of the present invention, the apparatus further comprises: and the second response module is used for displaying the identification information of the currently added sound effect on the client side when the song to be played responds to the adding operation of the sound effect.
As an alternative embodiment of the present invention, the alignment sub-module includes: the first determining submodule is used for obtaining a first frequency distribution vector according to the frequency distribution of the audio of the song to be played in the target duration; a second determining sub-module, configured to determine a second frequency distribution vector that matches the corresponding song audio at the target duration; and the third determining submodule is used for comparing the similarity between the first frequency distribution vector and the second frequency distribution vector.
The embodiment of the present invention further provides an electronic device, as shown in fig. 3, which may include a processor 401 and a memory 402, where the processor 401 and the memory 402 may be connected by a bus or other means, and in fig. 3, the connection is exemplified by a bus.
The processor 401 may be a central processing unit (Central Processing Unit, CPU). The processor 401 may also be other general purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), field programmable gate arrays (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, or combinations thereof.
The memory 402 is used as a non-transitory computer readable storage medium for storing non-transitory software programs, non-transitory computer executable programs, and modules, such as program instructions/modules corresponding to the sound effect processing method in the embodiment of the present invention. The processor 401 executes various functional applications of the processor and data processing, i.e., implements the sound effect processing method in the above-described method embodiment, by running non-transitory software programs, instructions, and modules stored in the memory 402.
Memory 402 may include a storage program area that may store an operating system, at least one application program required for functionality, and a storage data area; the storage data area may store data created by the processor 401, or the like. In addition, memory 402 may include high-speed random access memory, and may also include non-transitory memory, such as at least one magnetic disk storage device, flash memory device, or other non-transitory solid state storage device. In some embodiments, memory 402 may optionally include memory located remotely from processor 401, such remote memory being connectable to processor 401 through a network. Examples of such networks include, but are not limited to, the internet, intranets, local area networks, mobile communication networks, and combinations thereof.
The one or more modules are stored in the memory 402 and when executed by the processor 401, perform the sound effect processing method in the embodiment shown in fig. 1.
The specific details of the electronic device may be understood correspondingly with respect to the corresponding related descriptions and effects in the embodiment shown in fig. 1, which are not repeated herein.
It will be appreciated by those skilled in the art that implementing all or part of the above-described embodiment method may be implemented by a computer program to instruct related hardware, where the program may be stored in a computer readable storage medium, and the program may include the above-described embodiment method when executed. Wherein the storage medium may be a magnetic Disk, an optical Disk, a Read-Only Memory (ROM), a random access Memory (RandomAccessMemory, RAM), a Flash Memory (Flash Memory), a Hard Disk (HDD), a Solid State Drive (SSD), or the like; the storage medium may also comprise a combination of memories of the kind described above.
Although embodiments of the present invention have been described in connection with the accompanying drawings, various modifications and variations may be made by those skilled in the art without departing from the spirit and scope of the invention, and such modifications and variations are within the scope of the invention as defined by the appended claims.

Claims (10)

1. A sound effect processing method, characterized by comprising:
acquiring text information which corresponds to the song to be processed and is used for representing the type of the song;
classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte arrays of the songs to be processed;
the method for classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte array of the songs to be processed comprises the following steps: inputting the audio byte array of the song to be processed into the convolutional neural network model, so that the convolutional neural network model outputs the song style classification corresponding to the song to be processed;
determining the style classification of the songs to be processed based on the song style classification and the text information for representing the types of the songs;
determining the style sound effect of the song to be processed by using a preset sound effect decision method according to the style classification result;
acquiring audio target characteristics of a song to be processed, wherein the audio target characteristics of the song to be processed comprise frequency distribution, reverberation size and sound field width of the song to be processed;
performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain audio compensation sound effects of the songs to be processed;
and storing the attribute information of the to-be-processed song, the compensating sound effect of the to-be-processed song and the style sound effect of the to-be-processed song in an associated mode.
2. The method according to claim 1, wherein the method further comprises:
and when receiving the sound effect application request, responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played.
3. The method according to claim 2, wherein the adding operation of the style sound effect and/or the compensation sound effect to the song to be played in response to the receiving of the sound effect application request comprises:
when receiving the audio effect application request, performing song matching operation according to attribute information of the song to be played;
when the corresponding song is matched, the similarity comparison is carried out between the audio of the song to be played and the matched song audio;
and when the similarity is larger than a preset threshold, adding corresponding sound effects to the song to be played.
4. A method according to claim 3, characterized in that the method further comprises:
and displaying the identification information of the currently added sound effect on the client side while responding to the adding operation of the sound effect to the song to be played.
5. The method of claim 4, wherein the similarity comparison between the audio of the song to be played and the matched song audio when the corresponding song is matched, comprises:
obtaining a first frequency distribution vector according to the frequency distribution of the audio of the song to be played in the target duration;
determining a second frequency distribution vector matched with the corresponding song audio in the target duration;
and performing similarity comparison on the first frequency distribution vector and the second frequency distribution vector.
6. An audio processing apparatus, comprising:
the first acquisition module is used for acquiring text information which corresponds to the song to be processed and is used for representing the type of the song;
the classification module is used for classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte arrays of the songs to be processed;
the method for classifying the styles of the songs to be processed according to the text information for representing the types of the songs and the audio byte array of the songs to be processed comprises the following steps: inputting the audio byte array of the song to be processed into the convolutional neural network model, so that the convolutional neural network model outputs the song style classification corresponding to the song to be processed;
determining the style classification of the songs to be processed based on the song style classification and the text information for representing the types of the songs;
the first determining module is used for determining the style sound effect of the song to be processed by utilizing a preset sound effect decision method according to the style classification result;
the second acquisition module is used for acquiring the audio target characteristics of the songs to be processed;
the second determining module is used for performing sound effect compensation operation according to the audio target characteristics of the songs to be processed to obtain the audio compensation sound effect of the songs to be processed, wherein the audio target characteristics of the songs to be processed comprise frequency distribution, reverberation size and sound field width of the songs to be processed;
the first storage module is used for carrying out association storage on the attribute information of the songs to be processed, the compensating sound effects of the songs to be processed and the style sound effects of the songs to be processed;
and the sending module is used for storing the attribute information of the songs to be processed and the corresponding style sound effects.
7. The apparatus of claim 6, wherein the apparatus further comprises:
and the first response module is used for responding to the adding operation of the style sound effect and/or the compensation sound effect of the song to be played when the sound effect application request is received.
8. The apparatus of claim 7, wherein the first response module comprises:
the matching sub-module is used for carrying out song matching operation according to the attribute information of the song to be played when receiving the audio application request;
the comparison sub-module is used for comparing the similarity between the audio of the song to be played and the matched song audio when the corresponding song is matched;
and the application submodule is used for adding corresponding sound effects to the song to be played when the similarity is larger than a preset threshold value.
9. An electronic device, comprising: at least one processor; and a memory communicatively coupled to the at least one processor; wherein the memory stores instructions executable by the at least one processor to cause the at least one processor to perform the steps of the sound effect processing method of any one of claims 1-5.
10. A computer-readable storage medium, on which a computer program is stored, characterized in that the computer program, when being executed by a processor, implements the steps of the sound effect processing method according to any one of claims 1-5.
CN202211037097.4A 2022-08-26 2022-08-26 Sound effect processing method and device and electronic equipment Active CN115410544B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211037097.4A CN115410544B (en) 2022-08-26 2022-08-26 Sound effect processing method and device and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211037097.4A CN115410544B (en) 2022-08-26 2022-08-26 Sound effect processing method and device and electronic equipment

Publications (2)

Publication Number Publication Date
CN115410544A CN115410544A (en) 2022-11-29
CN115410544B true CN115410544B (en) 2024-01-30

Family

ID=84162067

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211037097.4A Active CN115410544B (en) 2022-08-26 2022-08-26 Sound effect processing method and device and electronic equipment

Country Status (1)

Country Link
CN (1) CN115410544B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5542000A (en) * 1993-03-19 1996-07-30 Yamaha Corporation Karaoke apparatus having automatic effector control
CN104978962A (en) * 2014-04-14 2015-10-14 安徽科大讯飞信息科技股份有限公司 Query by humming method and system
CN112002296A (en) * 2020-08-24 2020-11-27 广州小鹏汽车科技有限公司 Music playing method, vehicle, server and storage medium
CN113421585A (en) * 2021-05-10 2021-09-21 云境商务智能研究院南京有限公司 Audio fingerprint database generation method and device
CN113641329A (en) * 2021-08-10 2021-11-12 广州艾美网络科技有限公司 Sound effect configuration method and device, intelligent sound box, computer equipment and storage medium
WO2021248964A1 (en) * 2020-06-09 2021-12-16 广东美的制冷设备有限公司 Home appliance and control method therefor, and computer-readable storage medium
CN114661939A (en) * 2022-03-24 2022-06-24 杭州网易云音乐科技有限公司 Song matching method, medium, device and computing equipment
CN114842820A (en) * 2022-05-18 2022-08-02 北京地平线信息技术有限公司 K song audio processing method and device and computer readable storage medium

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5542000A (en) * 1993-03-19 1996-07-30 Yamaha Corporation Karaoke apparatus having automatic effector control
CN104978962A (en) * 2014-04-14 2015-10-14 安徽科大讯飞信息科技股份有限公司 Query by humming method and system
WO2021248964A1 (en) * 2020-06-09 2021-12-16 广东美的制冷设备有限公司 Home appliance and control method therefor, and computer-readable storage medium
CN112002296A (en) * 2020-08-24 2020-11-27 广州小鹏汽车科技有限公司 Music playing method, vehicle, server and storage medium
CN113421585A (en) * 2021-05-10 2021-09-21 云境商务智能研究院南京有限公司 Audio fingerprint database generation method and device
CN113641329A (en) * 2021-08-10 2021-11-12 广州艾美网络科技有限公司 Sound effect configuration method and device, intelligent sound box, computer equipment and storage medium
CN114661939A (en) * 2022-03-24 2022-06-24 杭州网易云音乐科技有限公司 Song matching method, medium, device and computing equipment
CN114842820A (en) * 2022-05-18 2022-08-02 北京地平线信息技术有限公司 K song audio processing method and device and computer readable storage medium

Also Published As

Publication number Publication date
CN115410544A (en) 2022-11-29

Similar Documents

Publication Publication Date Title
US11017010B2 (en) Intelligent playing method and apparatus based on preference feedback
JP4199097B2 (en) Automatic music classification apparatus and method
US20140037111A1 (en) Semantic audio track mixer
WO2017028704A1 (en) Method and device for providing accompaniment music
CN106898339B (en) Song chorusing method and terminal
WO2019137392A1 (en) File classification processing method and apparatus, terminal, server, and storage medium
KR20160069784A (en) Method and device for generating music playlist
US20120300950A1 (en) Management of a sound material to be stored into a database
US11511200B2 (en) Game playing method and system based on a multimedia file
WO2020015411A1 (en) Method and device for training adaptation level evaluation model, and method and device for evaluating adaptation level
CN107767850A (en) A kind of singing marking method and system
CN111785238A (en) Audio calibration method, device and storage medium
CN111046226B (en) Tuning method and device for music
CN107371075A (en) Microphone
JP2021101252A (en) Information processing method, information processing apparatus, and program
CN115410544B (en) Sound effect processing method and device and electronic equipment
JP6288197B2 (en) Evaluation apparatus and program
CN109410972A (en) Generate the method, apparatus and storage medium of sound effect parameters
KR20160056104A (en) Analyzing Device and Method for User's Voice Tone
JP5034599B2 (en) Music introduction sentence generation device, narration addition device, and program
JP2013213907A (en) Evaluation apparatus
JP6589521B2 (en) Singing standard data correction device, karaoke system, program
CN113032616B (en) Audio recommendation method, device, computer equipment and storage medium
CN113282509B (en) Tone recognition, live broadcast room classification method, device, computer equipment and medium
CN109710797B (en) Audio file pushing method and device, electronic device and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant