CN112037738B - Music data processing method and device and computer storage medium - Google Patents

Music data processing method and device and computer storage medium Download PDF

Info

Publication number
CN112037738B
CN112037738B CN202010906667.3A CN202010906667A CN112037738B CN 112037738 B CN112037738 B CN 112037738B CN 202010906667 A CN202010906667 A CN 202010906667A CN 112037738 B CN112037738 B CN 112037738B
Authority
CN
China
Prior art keywords
track data
sound
music
target
sound source
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202010906667.3A
Other languages
Chinese (zh)
Other versions
CN112037738A (en
Inventor
闫震海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Music Entertainment Technology Shenzhen Co Ltd
Original Assignee
Tencent Music Entertainment Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Music Entertainment Technology Shenzhen Co Ltd filed Critical Tencent Music Entertainment Technology Shenzhen Co Ltd
Priority to CN202010906667.3A priority Critical patent/CN112037738B/en
Publication of CN112037738A publication Critical patent/CN112037738A/en
Application granted granted Critical
Publication of CN112037738B publication Critical patent/CN112037738B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0091Means for obtaining special acoustic effects
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/02Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos
    • G10H1/04Means for controlling the tone frequencies, e.g. attack or decay; Means for producing special musical effects, e.g. vibratos or glissandos by additional modulation
    • G10H1/043Continuous modulation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/307Frequency adjustment, e.g. tone control
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/155Musical effects
    • G10H2210/265Acoustic effect simulation, i.e. volume, spatial, resonance or reverberation effects added to a musical sound, usually by appropriate filtering or delays
    • G10H2210/281Reverberation or echo
    • G10H2210/291Reverberator using both direct, i.e. dry, and indirect, i.e. wet, signals or waveforms, indirect signals having sustained one or more virtual reflections
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/155Musical effects
    • G10H2210/265Acoustic effect simulation, i.e. volume, spatial, resonance or reverberation effects added to a musical sound, usually by appropriate filtering or delays
    • G10H2210/295Spatial effects, musical uses of multiple audio channels, e.g. stereo
    • G10H2210/305Source positioning in a soundscape, e.g. instrument positioning on a virtual soundstage, stereo panning or related delay or reverberation changes; Changing the stereo width of a musical source
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/01Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Electrophonic Musical Instruments (AREA)

Abstract

The embodiment of the invention discloses a music data processing method, a device and a computer storage medium, wherein the method comprises the following steps: acquiring sound signals of music to be processed, and carrying out track separation processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects; setting a sound field position of a sound source object in a virtual space based on the sound source object; and modulating the first direction of the sound track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music with the playing effect of the band in the real space. The invention obtains a plurality of sound track data based on the track separation processing of the sound signals of music, and modulates the plurality of sound track data to obtain the target music with the music band playing effect in the real space, thereby realizing the intelligent management of the music data.

Description

Music data processing method and device and computer storage medium
Technical Field
The present application relates to the field of computer technologies, and in particular, to a method and apparatus for processing music data, and a computer storage medium.
Background
In the fast-paced information age, music is an important tool for mass emotion hosting. The music can relax the body and relieve the pressure; the music can break open and close hearts, and relieve the mood of depression and distress; music can stimulate brain and activate brain cells; the music can promote creativity, planning force and stimulate the right brain; music can help sleep, improve immunity, increase nerve conduction velocity, and enhance memory and attention; the melody of the music can lead the infant to breathe calmly and the heartbeat to slow down, and lead the infant not to cry and become restless. Music is a currently used tool with a high frequency. Thus, processing techniques for music have also been developed.
At present, a music processing mode is often to separate background music or human voice from the music or to perform angle modulation on stereo music to achieve a surrounding effect, and in general, the processing means are single, and the achieved music effect is single. Therefore, how to realize more effect processing on music becomes a problem to be solved.
Disclosure of Invention
The embodiment of the invention provides a processing method, a processing device and a computer storage medium for music data, wherein a plurality of sound track data are obtained based on track dividing processing of sound signals of music, and target music with a music band playing effect in a real space is obtained by modulating the plurality of sound track data, so that intelligent management of the music data is realized.
In a first aspect, the present invention discloses a music data processing method, including:
Acquiring a sound signal of music to be processed;
carrying out track division processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is a positive integer greater than or equal to 2;
Determining the sound field position of the sound source object in the virtual space;
and carrying out first azimuth modulation on the sound track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music, wherein the target music has the music effect that a plurality of sound source objects are played in different space positions.
In a second aspect, an embodiment of the present invention discloses a processing apparatus for music data, including:
The acquisition module is used for acquiring sound signals of the music to be processed;
The processing module is used for carrying out track separation processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is a positive integer greater than or equal to 2;
the processing module is used for setting the sound field position of the sound source object in the virtual space;
The processing module is further configured to perform first azimuth modulation on audio track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space, so as to obtain target music, where the target music has a music effect that a plurality of sound source objects perform in different space positions.
In a third aspect, an embodiment of the present invention provides a terminal, including a processor, an input device, an output device, and a memory, where the processor, the input device, the output device, and the memory are connected to each other, and the memory is configured to store a computer program, where the computer program includes program instructions, and where the processor is configured to invoke the program instructions to perform the method according to the first aspect.
In a fourth aspect, embodiments of the present invention provide a computer readable storage medium storing a computer program comprising program instructions which, when executed by a processor, cause the processor to perform the method of the first aspect.
In the embodiment of the invention, a terminal acquires sound signals of music to be processed, performs track separation processing on the sound signals to obtain N pieces of track data, wherein different track data are used for representing different sound source objects, sound field positions of the sound source objects in a virtual space are set based on the sound source objects, and first direction modulation is performed on the track data corresponding to the sound source objects according to the sound field positions of the sound source objects in the virtual space so as to obtain target music with a music band playing effect in a real space.
Drawings
In order to more clearly illustrate the embodiments of the invention or the technical solutions in the prior art, the drawings that are required in the embodiments or the description of the prior art will be briefly described, it being obvious that the drawings in the following description are only some embodiments of the invention, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
Fig. 1 is a flowchart of a method for processing music data according to an embodiment of the present invention;
fig. 2 is a schematic view of an azimuth effect of music data according to an embodiment of the present invention;
fig. 3 is a flowchart illustrating another method for processing music data according to an embodiment of the present invention;
fig. 4 is a flowchart of another method for processing music data according to an embodiment of the present invention;
Fig. 5 is a schematic diagram of a processing device for music data according to an embodiment of the present invention;
fig. 6 is a schematic structural diagram of a terminal according to an embodiment of the present invention.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present invention, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
The method for processing the music data is implemented in a terminal, wherein the terminal comprises an intelligent mobile phone, a tablet personal computer, a digital audio/video player, an electronic reader, a handheld game machine, a smart watch and other electronic devices, and the terminal can support interface display and can receive various operations of a user side, such as clicking operations, input operations or sliding operations.
The music data processing method provided by the embodiment of the invention mainly comprises the following steps: the terminal equipment acquires sound signals of the music to be processed, and performs track separation processing on the sound signals of the music to be processed by using a machine learning method to obtain N pieces of sound track data, wherein different sound track data correspond to different sound source objects, for example, sound sources corresponding to human voice, piano sound and the like are different; further, after the N audio track data are obtained, the sound field positions of the sound source objects corresponding to the different audio track data in the virtual space are determined, and then the audio track data corresponding to the sound source objects are subjected to first azimuth modulation according to the sound field positions of the sound source objects in the virtual space, so that target music with the music band playing effect in the real space is obtained. The playing effect is embodied in that different track data of target music are played at different spatial positions, such as singers standing in the middle of a stage, and guitar and piano accompanist are respectively located at two sides of the stage. Wherein N is an integer greater than or equal to 2.
In the embodiment of the invention, a processing method of music data is provided, the method mainly aims at the condition that target type of music track data exists in N pieces of music track data, namely, a terminal acquires music to be processed, acquires sound signals of the music to be processed, then carries out track dividing processing on the sound signals to obtain N pieces of music track data, different pieces of music track data correspond to different sound source objects, analyzes the N pieces of music track data, and if the fact that the target type of music track data exists in the N pieces of music track data is determined according to analysis results, invokes a tone changing tool to carry out tone changing processing on the target type of music track data to obtain N pieces of tone changing music track data, and according to the N pieces of tone changing music track data and other music track data except the target type of music track data in the N pieces of music track data, the method can only carry out processing on the target type of music track data to obtain target music with different effects. The track data of the target type may refer to one or more of human voice, piano voice, guitar voice, or the like, among others. If the voice is the voice, only the voice is subjected to tone changing processing, and other audio track data are not adjusted.
The specific flow of the music data processing method provided by the embodiment of the invention can be specifically described through the following flow chart.
Referring to fig. 1, fig. 1 is a flowchart of a music data processing method according to an embodiment of the present invention, which specifically includes the following steps:
step S101, the terminal acquires a sound signal of the music to be processed.
Specifically, the terminal acquires the music to be processed, where the music to be processed also refers to a music file to be processed, including a music file that has been uploaded to the server and authorized, and the user side may perform playing operation, downloading operation, and the like on the music file through the terminal interface. After the terminal obtains the music file to be processed, the sound signal of the music file to be processed is obtained, for example, the music file to be processed can be processed through Python, MATLAB and other tools to obtain the sound signal.
Here, the sound signal refers to a sound wave signal whose frequency and amplitude are regularly changed. The three elements of sound are pitch, intensity and timbre. There are three important parameters of sound waves or sine waves: the frequency ω0, the amplitude An and the phase ψn determine the characteristics of the sound signal.
Step S102, the terminal performs track separation processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is an integer greater than or equal to 2.
Wherein different audio track data correspond to different sound source objects, for example: track 1 corresponds to the track data of the voice, and the sound source object is the voice; the audio track 2 corresponds to audio track data of piano sound, and the sound source object is piano, and may further include audio track data of drum sound, audio track data of bass sound, audio track data of other sounds, and the like. The number of tracks is customizable, but is not limited to at least 2. It should be noted that, although the separation result is to separate the audio track data of the human voice and the audio track data of the background music, the audio track processing is relatively simple, and in order to achieve the more complex music processing effect, the embodiment of the present application may separate more various audio track data.
One manifestation of the soundtrack data is parallel "tracks" of one piece in the sequencer software. Each track has its own attributes such as tone color, tone color library, channel number, input/output port, volume, etc.
In one possible implementation, the audio signal is subjected to a track-dividing process to obtain N audio track data, specifically: the sound signals are input into a trained sound signal processing machine model for separation to obtain N sound track data, wherein the setting of the number of sound tracks can be input by a user based on a terminal interface, the input mode comprises a typing mode or a voice input mode, namely the terminal interface comprises an icon control for setting the number of sound tracks, and the setting of the number of sound tracks is realized by correspondingly typing operation or voice input operation on the icon control.
Step S103, the terminal sets the sound field position of the sound source object in the virtual space.
In one possible implementation manner, the terminal sets a sound field position of the sound source object in the virtual space, specifically: the terminal displays a target interface, wherein the target interface comprises icon elements for representing sound source objects corresponding to the audio track data, and the sound field positions input by a user for the sound source objects corresponding to the icon elements through the target interface are set as the sound field positions of the sound source objects corresponding to the icon elements in the virtual space.
In a possible implementation manner, the terminal determines the sound field position of the sound source object in the virtual space, specifically: and the terminal responds to an operation instruction of dragging the icon element of the sound source object from the first position to the second position of the target interface by a user, and determines the sound field position corresponding to the second position as the sound field position of the sound source object in the virtual space. For any sound source object, the user side can drag the icon element corresponding to the sound source object to determine the sound field position, so as to achieve the azimuth effect diagram shown in fig. 2.
In another possible implementation manner, the target interface displays icon elements of the sound source objects corresponding to each audio track data, and clicks on an icon element of one of the sound source objects, so that one angle information can be input for the audio track data corresponding to the sound source object, or the icon elements of all the sound source objects can be clicked respectively and then input corresponding angles. The terminal responds to angle input operation of a user on icon elements of the sound source object, the sound field position of the sound source object in the virtual space is determined according to angles corresponding to the angle input operation, icon elements of the sound source object corresponding to each audio track data are displayed on a target display interface, and the angles corresponding to each audio track data can click the icon elements to view. For example, assume that five audio track data are provided, including a human audio track data, a piano audio track data, bass audio track data, a drum audio track data and other audio track data, and angles of icons corresponding to sound source objects of the five audio track data are set on a target interface, for example, the human audio track data are set to be 5 degrees on the left, the piano audio track data are set to be 10 degrees on the left, the bass audio track data are set to be 30 degrees on the left, the drum audio track data are set to be 30 degrees on the right, the other audio track data are set to be 10 degrees on the right, and after receiving the settings, a terminal determines sound field positions of icon elements corresponding to each audio track data in a virtual space according to angle information, so as to obtain an azimuth effect schematic diagram of a music file as shown in fig. 2.
In the upper half of the azimuth effect diagram shown in fig. 2, each element corresponds to one sound source object, and the sound source objects are set according to the number of audio track data, that is, the number of audio track data is several, and the number of sound source objects is several.
Step S104, the terminal modulates the first azimuth of the sound track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music, wherein the target music has a music band playing effect in the real space.
The azimuth modulation of the sound source object is mainly realized by a Head-Related transfer function (Head-Related TransferFunction, HRTF). HRTF is used to describe the frequency response system where any point source in space reaches the eardrum. The system may be affected by auricles, head shape, shoulders, etc. The human ear determines the sound source location based primarily on binaural cues provided by the system.
The specific application of the function is as follows: assuming that any one of the N audio track data x (N) is obtained by the track division processing, HRTF functions of the target angles are respectively: HLeft (n) and HRight (n). The HRTF functions of the CIPIC database in the united states may be used, or the data of other open source databases may be used, wherein the transfer functions of the non-measured points are generated using a fit with weights of the distances between four adjacent points. The azimuth modulation method is to convolve the track data x (n) with HRTF functions, namely:
Left channel signal:
Right channel signal:
Wherein the symbols are Representing a convolution operation.
In one possible implementation manner, the terminal performs first azimuth modulation on the audio track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space, and the combining angle function specifically may include: the terminal takes angle information of the sound field position of the sound source object in the virtual space as input of an angle function of the sound track data, and obtains output of the angle function after the angle function is processed, wherein the angle function corresponds to HLeft (n) and HRight (n). After the output of the angle function is obtained, the output of the angle function and the track data (any one of the N track data corresponds to x (N)) are convolved to obtain the modulated data modulated by the first direction of the track data corresponding to the sound source object.
Taking voice track data as an example, receiving an angle setting instruction for the voice track data at a target interface, obtaining corresponding angle information after determining the sound field position of the voice track data, and carrying out angle convolution on the voice track data at each moment by using an HRTF (high-temperature transfer function) to realize the effect that the voice rotates in the whole music file playing process, wherein the effect is equivalent to the continuous conversion position of singers from the perspective of a concert. The surrounding rotation effect of the audio track data in the playing process can be realized in a mode of setting the angular speed.
The target music with the music band playing effect in the real space and the music to be processed are different, wherein the space sound field position of the sound track data is changed, and a plurality of left channel data and right channel data are formed after the sound track data are subjected to angle modulation of the HRTF function. All left channel data may be blended together and all right channel data may be blended together in the same manner. The processed music signal is different from the original music signal in that the spatial sound field position of each instrument is changed according to the user's will.
In the embodiment of the invention, a terminal acquires a sound signal of music to be processed, performs track separation processing on the sound signal to obtain N pieces of track data, wherein different track data are used for representing different sound source objects, a sound field position of the sound source object in a virtual space is set based on the sound source object, and first direction modulation is performed on the track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music with a music band playing effect in a real space.
Referring to fig. 3, fig. 3 is a flowchart illustrating another method for processing music files according to an embodiment of the present invention, where the method includes the steps of:
in step S301, the terminal acquires a sound signal of music to be processed.
This step is described in step S101, and will not be described in detail here.
In step S302, if the music to be processed is stereo music, the terminal converts the sound signal of the music to be processed into a mono sound signal.
In one possible implementation manner, after acquiring the sound signal of the music to be processed, the terminal may perform detection on the music type of the music to be processed to determine the type of the music to be processed, where the music type includes: stereo (binaural) music type, mono music type. If the music type is detected to be the stereo music type, converting the sound signal of the music to be processed into a sound signal of a mono music type. The binaural music text includes a left channel (denoted L) and a right channel (denoted R). Methods of converting a stereo music type sound signal into a mono music type sound signal include a variety of methods. For example, a mono music file may be obtained by a channel signal averaging method, i.e., (l+r)/2, and, for example, a stereo type sound signal may be input into a FFmpeg (Fast Forward Mpeg for short) program, by which the sound signals of the left and right channels are combined into one signal from the frequency domain perspective. FFmpeg is a set of open source computer programs that can be used to process digital audio and video.
In step S303, the terminal performs track-dividing processing on the sound signal to obtain N track data, where different track data are used to represent different sound source objects, and N is an integer greater than or equal to 2.
This step is described in step S102, and will not be described in detail here.
Step S304, the terminal sets the sound field position of the sound source object in the virtual space.
This step is described in step S103, and will not be described in detail here.
Step S305, performing first azimuth modulation on the audio track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space.
This step is described in step S104, and will not be described in detail here.
In step S306, if the N audio track data includes the audio track data of the target type, the terminal performs the tone-changing process on the audio track data of the target type to obtain the tone-changing audio track data, and synthesizes the tone-changing audio track data with the audio track data modulated by the first direction, so as to obtain the target music with the music band playing in the real space and different tone-mixing effects.
In one possible implementation manner, after acquiring N audio track data, the terminal analyzes the N audio track data, and if it is determined that the audio track data of the target type exists in the N audio track data according to the analysis result, invokes a tuning tool to perform tuning processing on the audio track data of the target type, so as to obtain tuning audio track data. The tone-changing track data is obtained by tone-changing the target type of track data, and the track data before tone change still exists. The tone-changing tool refers to tone-changing and speed-changing, namely, changing the size of the fundamental frequency of sound, and keeping the speech speed and the semantics unchanged. If the track data of the target type includes track data of a human voice, the track data of the human voice is processed as follows: a process may adjust the sound signals in a first frequency range to sound signals in a second frequency range, such as the effect of modulating male sound track data to female sound track data or modulating female sound track data to male sound track data; a process is capable of performing a plurality of different amplitude pitch-shifting processes on track data of a target type to obtain track data in a plurality of different frequency ranges.
For example, the tuning track data may be in several specific forms, such as calling a tuning tool to perform an up-tuning process on a male sound track data to obtain corresponding female sound track data, or calling a tuning tool to perform a down-tuning process on a female sound track data to obtain male sound track data. And then synthesizing the track-changing data and the track data modulated by the first direction to obtain target music with the antiphonal singing effect of men and women, or target music with the double singing effect, or target music with the chorus effect of multiple persons.
After the tone-changing processing is carried out on the track data of the target type to obtain the tone-changing track data, the terminal synthesizes the tone-changing track data and the track data modulated by the first direction, so as to obtain target music which is played in the real space by the band and has different tone-changing effects.
In a possible implementation manner, the terminal determines the target playing time period corresponding to the target type of audio track data and the tone-changing audio track data according to the playing time of each audio track data, and by controlling the playing time, the stage music effect of singing a male or female (for example, "a male sentence, a female sentence") can be realized, and the stage music effect of chorus can also be realized according to the control of the playing time period.
The terminal can also carry out second azimuth modulation of the relative change of sound field positions on the target type of sound track data and the tone-changing sound track data in the target playing period, and the second azimuth modulation is mainly aimed at the target sound track data and the tone-changing sound track data, so that music with different sound harmony and different sound opposite position change effects can be realized after the second azimuth modulation is carried out.
If the terminal synthesizes the second azimuth modulation and the audio track data obtained after the first azimuth modulation, target music with the effects of playing a plurality of sound source objects at different spatial positions, playing different sound modulation ensemble and changing the relative positions of different sound modulation can be obtained. For example, the audio track data obtained by the first azimuth modulation further includes other audio track data besides the audio track data of the target type, such as piano sound, guitar sound, etc., in which case, the synthesizing of the audio track data obtained by the second azimuth modulation and the audio track data obtained by the first azimuth modulation will obtain a performance effect of the band in the real space: the musical instrument is located at a different position of the stage, and if the audio track data of the type targeting the human voice, the singing effect that different singers are in an ensemble and the singers change in the position of the stage can be achieved, for example, the singing effect that the singers of men and women sing on the stage exchange the positions.
If the obtained tone-changing track data is greater than two, the effect of changing the position of the plurality of sounds on the stage can be achieved through the second azimuth modulation processing, or the stage music effect of the multiple persons singing in a singing mode can be achieved through controlling the playing time period. It should be noted that, the audio track data of a plurality of varios needs to be mixed after a certain time delay on the time axis, thereby forming an effect of multiple people's chorus. Since it is difficult for human ears to distinguish similar sounds within 10ms, the duration of the delay may be 10ms.
Step S307, obtaining a pulse signal corresponding to the target scene from the database, and performing convolution processing on the pulse signal corresponding to the target scene and the pulse signal of the target music to obtain the target music playing effect in the target scene.
In one possible implementation manner, the target music may be obtained by modulating the audio track data corresponding to the sound source object in the first direction, or may be obtained by synthesizing the variable-tone audio track data and the audio track data modulated in the first direction. In both cases, the terminal may obtain the pulse signal corresponding to the target scene from the database. The database stores pulse signals of different scenes, and specific scenes include, but are not limited to, stadiums, theaters, concert halls, gymnasiums, libraries and the like, and a user can select through a terminal interface. And carrying out convolution processing on the pulse signals corresponding to the target scene and the pulse signals of the updated music file target music so as to obtain target music with the playing effect of the target scene. If a specific music file in the concert hall scene is to be realized, the terminal directly acquires a pulse signal corresponding to the concert hall scene from the database, and then convolves the pulse signal with a pulse signal of target music to obtain the target music of the concert hall scene. According to the method, special effect music of different scenes can be simulated so as to meet different requirements of users. Alternatively, simulating special effects music of different scenes can also be implemented by using an artificial reverberator (which is one of the most important devices for beautifying sound), mainly set by a user according to a terminal interface.
In one possible implementation manner, on a display interface of the terminal, the mute switch corresponding to each audio track data is included, specifically, each audio track data has a corresponding audio track data identifier, each audio track data identifier corresponds to one mute switch, if the terminal receives an on instruction of the mute switch through the target interface, the on instruction of the mute switch of a certain audio track data is sent, and the on instruction indicates the corresponding audio track data to perform mute processing. And other audio track data which do not receive the opening instruction of the mute switch are normally played, so that the autonomous control of the user on the music file can be realized. The user can clearly know the role played by each track data in one music file and whether the music effect obtained by the track data exists or not. For example, if a music file including a plurality of audio track data is mute-processed on other audio track data than the piano audio track data and the human audio track data, the playing music effect of the piano can be achieved. In addition, the music adaptation effect of different styles can be realized through other combinations.
In one possible implementation manner, the terminal determines respective energy adjustment multiples of each track data, and adjusts the energy of each track data of the target music according to the respective energy adjustment multiples of each track data to obtain the target music of the recomposition style. Specifically, the target interface of the terminal may include an energy specific gravity setting control corresponding to all audio track data, and the user may set the energy specific gravity through the target interface, so as to achieve a music effect that the rhythm sense becomes stronger or weaker. The effect of muting can also be achieved by means of an energy setting, i.e. setting the energy of the track data to be muted to 0%. The target music may be obtained by modulating the audio track data corresponding to the sound source object in the first direction, or may be obtained by synthesizing the variable audio track data and the audio track data modulated in the first direction, or may be obtained by performing convolution processing on the pulse signal corresponding to the target scene and the pulse signal of the target music.
In one possible implementation manner, the terminal displays a playing interface of the terminal, where the playing interface includes icon elements corresponding to each audio track data of the target music, and in the process of playing the target music in the recomposition style, the terminal can detect real-time energy of each audio track data of the target music in the recomposition style and determine the real-time energy. If the real-time energy of the audio track data is larger than the first energy value, carrying out prompt processing on the icon element of the audio track data corresponding to the real-time energy, wherein the prompt processing can comprise various processing such as coloring processing, amplifying effect processing and the like on the icon element, the color of the coloring processing can be self-defined, and the rendered color can indicate that the audio track data is in a playing or singing state; if the real-time energy of the audio track data is smaller than the second energy value, performing non-prompting processing on the icon element of the audio track data corresponding to the real-time energy, wherein the non-prompting processing can include gray processing on the icon element, and enabling the size of the icon element to be unchanged or reduced, and the icon element corresponding to the audio track data is kept static, so that the audio track data is in an intermittent state at the moment. As shown in fig. 2, each audio track data corresponds to one icon element, and in the playing process of the target music, each audio track data has corresponding energy fluctuation, if the energy fluctuation exists, the terminal interface displays the effect of amplifying the icon element, and the greater the energy, the more obvious the amplifying effect. Generally, during the target music playing, neither the instrument nor the human voice is always in a playing or singing state, and thus it can be checked by real-time energy. By the method, the coordination effect of the sound and the picture can be achieved in the playing process of the target music, and the user can intuitively see which musical instruments are played in which specific time periods in the playing process of the music.
In the embodiment of the invention, besides the process shown in fig. 1 of the music file to be processed, the terminal can perform tone-changing processing on the target music to realize alternate playing effects of different tones or simultaneous playing effects of different tones, obtain special effect music in the target scene through scene mixing processing, and realize picture rendering during music playing through picture rendering processing. Through the embodiment, the intelligent processing of various kinds of music can be realized to a great extent, so that diversified requirements of users for hearing music with different sound effects are met.
Referring to fig. 4, fig. 4 is a flowchart of another music data processing method according to an embodiment of the invention, the method includes the following steps:
in step S401, the terminal acquires a sound signal of music to be processed.
This step is described in step S101, and will not be described in detail here.
In step S402, the terminal performs track-dividing processing on the sound signal to obtain N track data, where different track data are used to represent different sound source objects, and N is an integer greater than or equal to 2.
This step is described in step S102, and will not be described in detail here.
Step S403, if the N audio track data comprise the audio track data of the target type, performing the tone changing processing on the audio track data of the target type to obtain the tone-changed audio track data.
Specifically, the terminal analyzes the N audio track data, and if it is determined that the audio track data of the target type exists in the N audio track data according to the analysis result, the tone-changing tool is called to perform tone-changing processing on the audio track data of the target type, so as to obtain tone-changing audio track data. The number of the tone-changing track data is n, and n is an integer greater than or equal to 2.
In a possible implementation manner, the terminal analyzes the N audio track data, detects whether the audio track data of the target type exists, if yes, calls a transposition tool to carry out transposition processing on the audio track data of the target type to obtain N transposition audio track data, N is an integer greater than or equal to 2, and if still is one, the transposition processing is meaningless. The n pieces of pitch-changed track data may include female track data obtained by pitch-changing processing of male track data, male track data obtained by pitch-changing processing of female track data, or a plurality of track data obtained by pitch-changing processing of arbitrary person track data. The target audio track data may be audio track data of a human voice type, or audio track data of other voice types.
And step S404, the terminal obtains target music according to the change track data and other track data except the track data of the target type in the N track data.
In one possible implementation, the sound characteristics of the other track data than the target type of track data are kept unchanged, and only the key of the target type of track data is changed, thereby generating target music for singing by men and women or chorus. Specifically, the characteristics of other track data of original music, including the position of the other track data, etc., are not changed, and only the track data of the target type is changed, such as changing the voice of one person to the voice of two persons or more.
In one possible implementation manner, the above-mentioned N pieces of track data with variable tone may be subjected to azimuth modulation processing or play time control processing, and the terminal generates the processed target music according to the N pieces of track data with variable tone obtained by the azimuth modulation processing or the play time control processing and other track data except for the track data of the target type in the N pieces of track data. Because the track-changing data is modulated, the sound signals corresponding to the processed music files can show alternate playing music effects or playing music effects at different angles in the playing process, and the experience presented to the clients is the antiphonal effect of the human voice in the music or the surrounding effect of the human voice.
Note that in this method, the angle modulation may be performed or may not be performed on other track data than the target type of track data among the N track data.
In the implementation of the invention, a terminal acquires sound signals of music to be processed, carries out track separation processing on the sound signals to obtain N sound track data, uses different sound track data to represent different sound source objects, analyzes the N sound track data, and if the sound track data of a target type exists in the N sound track data according to analysis results, calls a tone changing tool to carry out tone changing processing on the sound track data of the target type to obtain N tone changing sound track data, and obtains processed target music according to the N tone changing sound track data and other sound track data except the sound track data of the target type in the N sound track data.
Referring to fig. 5, fig. 5 is a schematic diagram of a processing apparatus for music data according to an embodiment of the present invention, where the apparatus 50 includes: the acquisition module 501 and the processing module 502. Wherein:
An acquisition module 501, configured to acquire a sound signal of music to be processed;
the processing module 502 is configured to perform a track-dividing process on the sound signal to obtain N track data, where different track data are used to represent different sound source objects, and N is an integer greater than or equal to 2;
The processing module 502 is configured to set a sound field position of the sound source object in a virtual space;
The processing module 502 is further configured to perform first azimuth modulation on audio track data corresponding to the sound source object according to a sound field position of the sound source object in the virtual space, so as to obtain target music, where the target music has a performance effect of a band in the real space.
In a possible implementation manner, the processing module 502 is configured to perform, according to a sound field position of the sound source object in the virtual space, first azimuth modulation on audio track data corresponding to the sound source object, where the first azimuth modulation includes:
Taking angle information of the sound field position of the sound source object in the virtual space as input of an angle function of sound track data corresponding to the sound source object, and obtaining an output result of the angle function;
And carrying out convolution operation on the output result of the angle function and the sound track data so as to realize first direction modulation of the sound track data corresponding to the sound source object.
In a possible implementation manner, the processing module 502 is configured to set a sound field position of the sound source object in the virtual space, and includes:
Displaying a target interface, wherein the target interface comprises an icon element for representing a sound source object corresponding to the audio track data;
and setting the sound field position input by the user based on the icon element as the sound field position of the sound source object corresponding to the icon element in the virtual space.
In a possible implementation manner, the processing module 502 is configured to set a sound field position input by a user based on the icon element to a sound field position of a sound source object corresponding to the icon element in a virtual space, where the setting includes:
And responding to the operation that a user drags the icon element of the sound source object from the first position to the second position of the target interface, and determining the sound field position corresponding to the second position as the sound field position of the sound source object in the virtual space.
In a possible implementation manner, the processing module 502 is configured to, after performing the first azimuth modulation on the audio track data corresponding to the sound source object according to the sound field position of the sound source object, further be configured to:
If the N audio track data comprise the audio track data of the target type, performing tone changing processing on the audio track data of the target type to obtain tone changing audio track data;
Synthesizing the tuning track data and the first-direction modulated sound track data to obtain target music with music bands playing in real space and different sound tuning effects
In a possible implementation manner, the processing module 502 is configured to synthesize the track change data with the track data modulated by the first direction, and specifically includes:
determining the track data of the target type and the target playing time period corresponding to the tone-changing track data according to the playing time of the track data;
performing second azimuth modulation of sound field position relative change on the target type sound track data and the tone-changing sound track data in the target playing period;
and synthesizing the second azimuth modulation and the audio track data modulated by the first azimuth modulation to obtain target music with the effects of playing the band in real space, playing different sound modulation modes and changing the relative positions of different sound modulation modes.
In a possible implementation manner, the obtaining module 501 is further configured to obtain pulse signals corresponding to a target scene from a database, where the database stores pulse signals corresponding to different scenes;
The processing module 502 is further configured to convolve the pulse signal corresponding to the target scene with the pulse signal of the target music, so as to obtain the target music with the playing effect in the target scene.
In one possible implementation, the processing module 502 is further configured to:
Displaying the target interface, wherein the target interface comprises a mute switch corresponding to the audio track data;
and if an opening instruction of the mute switch is received through the target interface, performing mute processing on the corresponding audio track data indicated by the opening instruction.
In one possible implementation, as shown in fig. 5, the apparatus may further include a determining module 503 and a detecting module 504.
A determining module 503, configured to determine a respective energy adjustment multiple of each of the audio track data;
The processing module 502 is further configured to adjust energy of each of the audio track data of the target music according to respective energy adjustment multiples of each of the audio track data, so as to obtain target music of an adaptation style.
In a possible implementation manner, the processing module 502 is configured to display a playing interface, where the playing interface includes icon elements corresponding to each audio track data of the target music;
the detection module 504 is configured to detect real-time energy of each audio track data of the adapted target music during the playing process of the adapted target music;
The processing module 502 is configured to:
If the real-time energy of the audio track data is larger than a first energy value, prompting the icon elements of the audio track data corresponding to the real-time energy;
And if the real-time energy of the audio track data is smaller than the second energy value, performing non-prompt processing on the icon elements of the audio track data corresponding to the real-time energy.
In the embodiment of the invention, a terminal acquires sound signals of music to be processed, performs track separation processing on the sound signals to obtain N pieces of track data, wherein different track data are used for representing different sound source objects, sound field positions of the sound source objects in a virtual space are set based on the sound source objects, and first direction modulation is performed on the track data corresponding to the sound source objects according to the sound field positions of the sound source objects in the virtual space so as to obtain target music with a music band playing effect in a real space.
Referring to fig. 6, fig. 6 is a schematic structural diagram of a terminal according to an embodiment of the present invention, as shown in fig. 6, the terminal device includes: at least one processor 601, an input device 603, an output device 604, a memory 605, at least one communication bus 602. Wherein the communication bus 602 is used to enable connected communications between these components. The input device 603 may be a control panel or a microphone, and the output device 604 may be a display screen. The memory 605 may be a high-speed RAM memory or a non-volatile memory (non-volatilememory), such as at least one magnetic disk memory. The memory 605 may also optionally be at least one storage device located remotely from the processor 601. Wherein the processor 601 may be described in connection with fig. 5, a set of program code is stored in the memory 605, and the processor 601, the input device 603, the output device 604 call the program code stored in the memory 605 for:
Acquiring a sound signal of music to be processed;
Carrying out track division processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is an integer greater than or equal to 2;
setting the sound field position of the sound source object in the virtual space;
And carrying out first azimuth modulation on the sound track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music, wherein the target music has a music band playing effect in the real space.
In a possible implementation manner, the processor 601 is configured to perform first azimuth modulation on audio track data corresponding to the sound source object according to a sound field position of the sound source object in a virtual space, where the first azimuth modulation includes:
Taking angle information of the sound field position of the sound source object in the virtual space as input of an angle function of sound track data corresponding to the sound source object, and obtaining an output result of the angle function;
And carrying out convolution operation on the output result of the angle function and the sound track data so as to realize first direction modulation of the sound track data corresponding to the sound source object.
In a possible implementation manner, the processor 601 is configured to set a sound field position of the sound source object in the virtual space, including (for):
Displaying a target interface, wherein the target interface comprises an icon element for representing a sound source object corresponding to the audio track data;
and setting the sound field position input by the user based on the icon element as the sound field position of the sound source object corresponding to the icon element in the virtual space.
In a possible implementation manner, the processor 601 is configured to set a sound field position input by a user based on the icon element to a sound field position of a sound source object corresponding to the icon element in a virtual space, where the sound field position includes (for):
Responding to the operation that a user drags the icon element of the sound source object from a first position to a second position of the target interface, and determining the sound field position corresponding to the second position as the sound field position of the sound source object in a virtual space; or alternatively
And responding to the angle input operation of the user on the icon element of the sound source object, and determining the sound field position corresponding to the angle input operation as the sound field position of the sound source object in the virtual space.
In a possible implementation manner, after the processor 601 is configured to perform first azimuth modulation on the audio track data corresponding to the sound source object according to the sound field position of the sound source object, the processor is further configured to:
If the N audio track data comprise the audio track data of the target type, performing tone changing processing on the audio track data of the target type to obtain tone changing audio track data;
And synthesizing the tuning track data and the track data modulated by the first direction to obtain target music with the music band playing in the real space and different sound tuning effects.
In a possible implementation manner, the processor 601 is configured to synthesize the track change data with the track data modulated by the first direction, and includes:
determining the track data of the target type and the target playing time period corresponding to the tone-changing track data according to the playing time of the track data;
performing second azimuth modulation of sound field position relative change on the target type sound track data and the tone-changing sound track data in the target playing period;
and synthesizing the second azimuth modulation and the audio track data modulated by the first azimuth modulation to obtain target music with the effects of playing the band in real space, playing different sound modulation modes and changing the relative positions of different sound modulation modes.
In a possible implementation, the processor 601 is further configured to:
acquiring pulse signals corresponding to target scenes from a database, wherein the database stores the pulse signals corresponding to different scenes;
and carrying out convolution processing on the pulse signals corresponding to the target scene and the pulse signals of the target music so as to obtain the target music with the playing effect in the target scene.
In a possible implementation, the processor 601 is further configured to:
Displaying the target interface, wherein the target interface comprises a mute switch corresponding to the audio track data;
and if an opening instruction of the mute switch is received through the target interface, performing mute processing on the corresponding audio track data indicated by the opening instruction.
In a possible implementation, the processor 601 is further configured to:
determining respective energy adjustment multiples of each of the audio track data;
and adjusting the energy of each audio track data of the target music according to the respective energy adjustment multiple of each audio track data to obtain the target music in the recomposition style.
In a possible embodiment, the processor 601 is further configured to display a playing interface, where the playing interface includes icon elements corresponding to each track data of the target music;
Detecting real-time energy of each track data of the target music of the recomposition style in the playing process of the target music of the recomposition style;
If the real-time energy of the audio track data is larger than a first energy value, lifting icon elements of the audio track data corresponding to the real-time energy;
And if the real-time energy of the audio track data is smaller than the second energy value, performing non-prompt processing on the icon elements of the audio track data corresponding to the real-time energy.
In the embodiment of the present invention, the processor 601 of the terminal obtains a plurality of audio track data based on the track division processing of the sound signal of the music to be processed, and modulates the plurality of audio track data to obtain updated diversified target music, so as to realize intelligent management of various kinds of music. The method also comprises the tone changing processing of the target type of the audio track data so as to realize more diversified hearing effects and meet the diversified requirements of users.
In an embodiment of the present invention, there is further provided a computer readable storage medium storing a computer program, where the computer program includes program instructions, and when the program instructions are executed by a processor, the steps performed in the embodiments of fig. 1, fig. 3, and fig. 4 are performed.
It should be appreciated that in embodiments of the present invention, the Processor 601 may be a central processing module (Central Processing Unit, CPU), which may also be other general purpose processors, digital signal processors (DIGITAL SIGNAL processors, DSPs), application SPECIFIC INTEGRATED Circuits (ASICs), off-the-shelf Programmable gate arrays (Field-Programmable GATE ARRAY, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc. A general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
Those skilled in the art will appreciate that all or part of the processes in the methods of the embodiments described above may be implemented by hardware associated with computer program instructions, where the program may be stored on a computer readable storage medium, where the program, when executed, may include processes in embodiments of the methods described above. The storage medium may be a magnetic disk, an optical disk, a Read-Only Memory (ROM), a random-access Memory (Random Access Memory, RAM), or the like.
References herein to "a plurality" means two or more. "and/or", describes an association relationship of an association object, and indicates that there may be three relationships, for example, a and/or B, and may indicate: a exists alone, A and B exist together, and B exists alone. The character "/" generally indicates that the context-dependent object is an "or" relationship.
The above disclosure is only a few examples of the present invention, and it is not intended to limit the scope of the present invention, but it is understood by those skilled in the art that all or a part of the above embodiments may be implemented and equivalents thereof may be modified according to the scope of the present invention.

Claims (13)

1. A method of processing music data, the method comprising:
Acquiring a sound signal of music to be processed;
Carrying out track division processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is an integer greater than or equal to 2;
setting the sound field position of the sound source object in the virtual space;
and carrying out first azimuth modulation on the sound track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space so as to obtain target music, wherein the target music has the music effect that a plurality of sound source objects are played in different space positions.
2. The method according to claim 1, wherein the performing the first azimuth modulation on the track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space includes:
Taking angle information of the sound field position of the sound source object in the virtual space as input of an angle function of sound track data corresponding to the sound source object, and obtaining an output result of the angle function;
And carrying out convolution operation on the output result of the angle function and the sound track data so as to realize first direction modulation of the sound track data corresponding to the sound source object.
3. The method of claim 1, wherein the setting the sound field position of the sound source object in the virtual space comprises:
Displaying a target interface, wherein the target interface comprises an icon element for representing a sound source object corresponding to the audio track data;
and setting the sound field position input by the user based on the icon element as the sound field position of the sound source object corresponding to the icon element in the virtual space.
4. The method according to claim 3, wherein the setting the sound field position input by the user based on the icon element as the sound field position of the sound source object corresponding to the icon element in the virtual space includes:
And responding to the operation that a user drags the icon element of the sound source object from the first position to the second position of the target interface, and determining the sound field position corresponding to the second position as the sound field position of the sound source object in the virtual space.
5. The method according to claim 1, wherein after the first azimuth modulation is performed on the track data corresponding to the sound source object according to the sound field position of the sound source object, the method further comprises:
If the N audio track data comprise the audio track data of the target type, performing tone changing processing on the audio track data of the target type to obtain tone changing audio track data;
And synthesizing the tuning track data and the track data modulated by the first direction to obtain target music with the music band playing in the real space and different sound tuning effects.
6. The method of claim 5, wherein the synthesizing the change track data with the first-direction modulated track data comprises:
determining the track data of the target type and the target playing time period corresponding to the tone-changing track data according to the playing time of the track data;
performing second azimuth modulation of sound field position relative change on the target type sound track data and the tone-changing sound track data in the target playing period;
And synthesizing the second azimuth modulation and the audio track data modulated by the first azimuth modulation to obtain target music with the effects of playing a plurality of sound source objects at different spatial positions, performing different tone ensemble and changing the positions of the different tone ensemble.
7. The method according to any one of claims 1-6, further comprising:
acquiring pulse signals corresponding to target scenes from a database, wherein the database stores the pulse signals corresponding to different scenes;
and carrying out convolution processing on the pulse signals corresponding to the target scene and the pulse signals of the target music so as to obtain the target music with the playing effect in the target scene.
8. The method according to any one of claims 1-6, further comprising:
displaying a target interface, wherein the target interface comprises a mute switch corresponding to audio track data;
and if an opening instruction of the mute switch is received through the target interface, performing mute processing on the audio track data corresponding to the opening instruction.
9. The method according to any one of claims 1-6, further comprising:
determining respective energy adjustment multiples of each of the audio track data;
and adjusting the energy of each audio track data of the target music according to the respective energy adjustment multiple of each audio track data to obtain the target music in the recomposition style.
10. The method according to claim 9, wherein the method further comprises:
Displaying a playing interface, wherein the playing interface comprises icon elements corresponding to each audio track data of the target music;
Detecting real-time energy of each track data of the target music of the recomposition style in the playing process of the target music of the recomposition style;
If the real-time energy of the audio track data is larger than a first energy value, prompting the icon elements of the audio track data corresponding to the real-time energy;
And if the real-time energy of the audio track data is smaller than the second energy value, performing non-prompt processing on the icon elements of the audio track data corresponding to the real-time energy.
11. A processing apparatus for music data, the apparatus comprising:
The acquisition module is used for acquiring sound signals of the music to be processed;
The processing module is used for carrying out track separation processing on the sound signals to obtain N sound track data, wherein different sound track data are used for representing different sound source objects, and N is a positive integer greater than or equal to 2;
the processing module is used for setting the sound field position of the sound source object in the virtual space;
The processing module is further configured to perform first azimuth modulation on audio track data corresponding to the sound source object according to the sound field position of the sound source object in the virtual space, so as to obtain target music, where the target music has a music effect that a plurality of sound source objects perform in different space positions.
12. A terminal comprising a processor, an input device, an output device and a memory, the processor, the input device, the output device and the memory being interconnected, wherein the memory is adapted to store a computer program comprising program instructions, the processor being configured to invoke the program instructions to perform the method of processing music data according to any of claims 1-10.
13. A computer readable storage medium, characterized in that the computer readable storage medium stores a computer program comprising program instructions which, when executed by a processor, cause the processor to perform the method of processing music data according to any one of claims 1-10.
CN202010906667.3A 2020-08-31 2020-08-31 Music data processing method and device and computer storage medium Active CN112037738B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010906667.3A CN112037738B (en) 2020-08-31 2020-08-31 Music data processing method and device and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010906667.3A CN112037738B (en) 2020-08-31 2020-08-31 Music data processing method and device and computer storage medium

Publications (2)

Publication Number Publication Date
CN112037738A CN112037738A (en) 2020-12-04
CN112037738B true CN112037738B (en) 2024-05-28

Family

ID=73592219

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010906667.3A Active CN112037738B (en) 2020-08-31 2020-08-31 Music data processing method and device and computer storage medium

Country Status (1)

Country Link
CN (1) CN112037738B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115278350A (en) * 2021-04-29 2022-11-01 华为技术有限公司 Rendering method and related equipment
CN113365191A (en) * 2021-06-04 2021-09-07 广州酷狗计算机科技有限公司 Music playing method, device, equipment and storage medium
CN113747337B (en) * 2021-09-03 2024-05-10 杭州网易云音乐科技有限公司 Audio processing method, medium, device and computing equipment
CN113707165A (en) * 2021-09-07 2021-11-26 联想(北京)有限公司 Audio processing method and device, electronic equipment and storage medium
CN114286275A (en) * 2021-12-20 2022-04-05 Oppo广东移动通信有限公司 Audio processing method and device and storage medium
CN113986191B (en) * 2021-12-27 2022-06-07 广州酷狗计算机科技有限公司 Audio playing method and device, terminal equipment and storage medium
CN114598985B (en) * 2022-03-07 2024-05-03 安克创新科技股份有限公司 Audio processing method and device
CN115174959B (en) * 2022-06-21 2024-01-30 咪咕文化科技有限公司 Video 3D sound effect setting method and device

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1797538A (en) * 2004-12-01 2006-07-05 创新科技有限公司 System and method for forming and rendering 3d midi messages
CN102118670A (en) * 2011-03-17 2011-07-06 杭州赛利科技有限公司 Earphone capable of generating three-dimensional stereophonic sound effect
KR20110119339A (en) * 2010-04-27 2011-11-02 (주)아이지솔루션즈 Music synthesis technique for synchroning with rhythm and it's service method
CN106952637A (en) * 2017-03-15 2017-07-14 北京时代拓灵科技有限公司 The creative method and experience apparatus of a kind of interactive music
CN108919953A (en) * 2018-06-29 2018-11-30 咪咕文化科技有限公司 A kind of music method of adjustment, device and storage medium
WO2019002909A1 (en) * 2017-06-26 2019-01-03 Latypov Ray A method of providing to user an interactive music composition
WO2019057530A1 (en) * 2017-09-20 2019-03-28 Nokia Technologies Oy An apparatus and associated methods for audio presented as spatial audio
CN110111761A (en) * 2019-03-28 2019-08-09 深圳芒果未来教育科技有限公司 The method followed in real time and Related product are played to musical sound
CN110170170A (en) * 2019-05-30 2019-08-27 维沃移动通信有限公司 A kind of information display method and terminal device
CN110992970A (en) * 2019-12-13 2020-04-10 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method and related device

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1797538A (en) * 2004-12-01 2006-07-05 创新科技有限公司 System and method for forming and rendering 3d midi messages
KR20110119339A (en) * 2010-04-27 2011-11-02 (주)아이지솔루션즈 Music synthesis technique for synchroning with rhythm and it's service method
CN102118670A (en) * 2011-03-17 2011-07-06 杭州赛利科技有限公司 Earphone capable of generating three-dimensional stereophonic sound effect
CN106952637A (en) * 2017-03-15 2017-07-14 北京时代拓灵科技有限公司 The creative method and experience apparatus of a kind of interactive music
WO2019002909A1 (en) * 2017-06-26 2019-01-03 Latypov Ray A method of providing to user an interactive music composition
CN110915240A (en) * 2017-06-26 2020-03-24 雷.拉蒂波夫 Method for providing interactive music composition to user
WO2019057530A1 (en) * 2017-09-20 2019-03-28 Nokia Technologies Oy An apparatus and associated methods for audio presented as spatial audio
CN108919953A (en) * 2018-06-29 2018-11-30 咪咕文化科技有限公司 A kind of music method of adjustment, device and storage medium
CN110111761A (en) * 2019-03-28 2019-08-09 深圳芒果未来教育科技有限公司 The method followed in real time and Related product are played to musical sound
CN110170170A (en) * 2019-05-30 2019-08-27 维沃移动通信有限公司 A kind of information display method and terminal device
CN110992970A (en) * 2019-12-13 2020-04-10 腾讯音乐娱乐科技(深圳)有限公司 Audio synthesis method and related device

Also Published As

Publication number Publication date
CN112037738A (en) 2020-12-04

Similar Documents

Publication Publication Date Title
CN112037738B (en) Music data processing method and device and computer storage medium
Emmerson et al. Electro-acoustic music
Begault et al. 3-D sound for virtual reality and multimedia
WO2021103314A1 (en) Listening scene constructing method and related device
US11437004B2 (en) Audio performance with far field microphone
US10924875B2 (en) Augmented reality platform for navigable, immersive audio experience
WO2020224322A1 (en) Method and device for processing music file, terminal and storage medium
Müller et al. The boomRoom: mid-air direct interaction with virtual sound sources
US9967693B1 (en) Advanced binaural sound imaging
CN113823250B (en) Audio playing method, device, terminal and storage medium
JP3624805B2 (en) Sound image localization device
CN110915240A (en) Method for providing interactive music composition to user
Worrall Space in sound: sound of space
Cohen et al. Spatial soundscape superposition and multimodal interaction
CN114501297B (en) Audio processing method and electronic equipment
WO2022163137A1 (en) Information processing device, information processing method, and program
CA3044260A1 (en) Augmented reality platform for navigable, immersive audio experience
WO2022143530A1 (en) Audio processing method and apparatus, computer device, and storage medium
Hinde Concurrency in auditory displays for connected television
JP2004088608A (en) Mixing device
Munoz Space Time Exploration of Musical Instruments
Lorenz Impact of Head-Tracking on the listening experience of binaural music
CN113318432B (en) Music control method in game, nonvolatile storage medium and electronic device
JP2005051801A (en) Sound image localization apparatus
McDonnell Development of Open Source tools for creative and commercial exploitation of spatial audio

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant