CN112634841A - Guitar music automatic generation method based on voice recognition - Google Patents

Guitar music automatic generation method based on voice recognition Download PDF

Info

Publication number
CN112634841A
CN112634841A CN202011392002.1A CN202011392002A CN112634841A CN 112634841 A CN112634841 A CN 112634841A CN 202011392002 A CN202011392002 A CN 202011392002A CN 112634841 A CN112634841 A CN 112634841A
Authority
CN
China
Prior art keywords
music
chord
guitar
information
playing
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011392002.1A
Other languages
Chinese (zh)
Other versions
CN112634841B (en
Inventor
刘强
陈盛
马文亮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ai Li Chi Technology Beijing Co ltd
Original Assignee
Ai Li Chi Technology Beijing Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ai Li Chi Technology Beijing Co ltd filed Critical Ai Li Chi Technology Beijing Co ltd
Priority to CN202011392002.1A priority Critical patent/CN112634841B/en
Publication of CN112634841A publication Critical patent/CN112634841A/en
Application granted granted Critical
Publication of CN112634841B publication Critical patent/CN112634841B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/24Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being the cepstrum
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
    • G10L25/30Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/90Pitch determination of speech signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/056Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for extraction or identification of individual instrumental parts, e.g. melody, chords, bass; Identification or separation of instrumental parts by their characteristic voices or timbres
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/031Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal
    • G10H2210/066Musical analysis, i.e. isolation, extraction or identification of musical elements or musical parameters from a raw acoustic signal or from an encoded audio signal for pitch analysis as part of wider processing for musical purposes, e.g. transcription, musical performance evaluation; Pitch recognition, e.g. in polyphonic sounds; Estimation or use of missing fundamental
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/101Music Composition or musical creation; Tools or processes therefor
    • G10H2210/111Automatic composing, i.e. using predefined musical rules
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/90Pitch determination of speech signals
    • G10L2025/906Pitch tracking

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Theoretical Computer Science (AREA)
  • Auxiliary Devices For Music (AREA)

Abstract

The invention discloses a guitar music automatic generation method based on voice recognition, which comprises the following steps of: step 201, performing framing operation on an input guitar playing audio file to be identified, and extracting the Mel cepstrum coefficient characteristics of each frame; step 202, calculating the classification of the Mel cepstrum coefficient characteristics of each frame in the step 201, and obtaining the pitch and playing mode information of each frame of the tested audio according to the output layer label of the model; step 203, analyzing chord, tone and rhythm information of the music based on the pitch information in the music score file in the step 202; and step 204, integrating the pitch, playing mode, chord, tone and rhythm information output in the steps 202 and 203, and outputting a music score file. The method can automatically generate the guitar music, replaces manual music picking work, and effectively improves the work efficiency.

Description

Guitar music automatic generation method based on voice recognition
Technical Field
The invention belongs to the technical field of sound, and particularly relates to a music score generation method.
Background
In the field of music score, the current technical research mainly focuses on directions of music retrieval, music performance evaluation, automatic composition and the like; the guitar music is processed by focusing on the picture music, and music score is generated by identifying the picture music, and automatic playing and playing can be performed.
In the actual guitar learning process, people often encounter guitar playing of interest, but no music score in any form can be referred to, and in this case, the music score is determined according to the played sound. In reality, the process is generally called as music score scratching, repeated listening and comparison are needed for music score scratching work, and finally the tone, rhythm and chord configuration and decomposition or chord sweeping details of the music score are determined; the staffs taking music must train for a long time to have sufficient ability to distinguish the ears, and also have to master sufficient music theory and deep knowledge of guitar principle and playing technique, so to speak, it is the high-level skill in guitar learning.
It is difficult for the ordinary guitar fans to do. If the guitar music can be directly generated from the playing sound, the learning difficulty of a beginner can be reduced, convenience can be provided for professionals, and only the music score generated automatically needs to be optimized.
The guitar is a stringed musical instrument, the vibration of strings gives out sound, the vibration frequency of the strings determines the pitch, the vibration frequency of the strings is influenced by the diameters and the vibration lengths of the strings, the different diameters of the strings are different, the different grades on the fingerboard determine the vibration lengths of the strings, and therefore the combination of the strings and the grades determines the pitch of the playing. In addition, the guitar belongs to the standard tuning musical instrument, and the standard tones of one string to six strings are E, B, G, D, A, E in turn, so that different strings and taste combinations may give the same pitch.
The chord is a musical concept and is composed of a group of sounds satisfying a certain interval relationship, wherein the lowest pitch of the sounds is a root and is the basis of the chord, and the intervals between different tones determine the color of the chord, which jointly form a chord. The chord is played mainly by two types of chord decomposition and chord sweeping, the chord decomposition is to pop out the chord in turn to form the sound, and the chord sweeping sweeps a plurality of strings simultaneously to produce a plurality of chord tones.
The playing of the guitar can be divided into accompaniment and solo according to the function, the accompaniment generally only plays the chord, provides harmony for the melody, and the solo plays the melody while playing the chord. Whether the music accompaniment or the solo, the pitch of each note is firstly determined when the music score is played, and if the music score is swept, the chord and the swept string are determined, which all need to be distinguished by human ears.
With the development of artificial intelligence technology, sound processing based on deep learning has a great progress, and the method achieves ideal effects in scenes such as voice recognition, scene classification and the like, and is widely applied. The strong learning ability of the neural network is utilized, the work of identifying the pitch of the note can be completed instead of human ears, all the sounds which can be played by the guitar are used as parameters of training data to train a neural network model, unknown sounds are classified by using the model, the pitch information in playing can be determined, and the first step of scratching the music score is also completed.
After the pitch of the music is determined, the musical instruments are divided into bars, and the chord used by each bar is further determined, usually the root of the chord is played first, and the first beat of the bar is the repeated beat, and then the chord trend of the music is determined by combining the combined sound information in the chord.
Finally, the key and the rhythm of the music are determined, the key is generally determined by the composition of the chord, and the rhythm reflects the weight distribution and can be determined by analyzing the energy distribution of the music.
Disclosure of Invention
The invention aims to improve the prior technical problem, namely the invention aims to provide a guitar spectrum automatic generation method based on voice recognition. The technical scheme of the invention is as follows:
a guitar music automatic generation method based on voice recognition is characterized by comprising a process of picking up music of a guitar playing audio file to be recognized, wherein the process comprises the following steps:
step 201, performing framing operation on an input guitar playing audio file to be identified, and extracting the Mel cepstrum coefficient characteristics of each frame;
step 202, calculating the classification of the Mel cepstrum coefficient characteristics of each frame in the step 201, and obtaining the pitch and playing mode information of each frame of the tested audio according to the output layer label of the model;
step 203, analyzing chord, tone and rhythm information of the music based on the pitch information in the music score file in the step 202;
and step 204, integrating the pitch, the playing mode, the chord, the tone, the rhythm and the information output in the steps 202 and 203, and outputting a music score file.
Further, the step 203 comprises the following sub-steps:
step 2031, determining the positions of all chord root sounds in the music, calculating the time intervals, selecting the interval duration with the most frequent occurrence, preliminarily dividing the music into uniform time segments, comparing the sounds appearing in each time segment with the chord template, calculating the similarity, selecting the chord with the highest similarity as the chord of the time segment, and writing the chord information into the music score file;
step 2032, comparing the chord tables of the guitar according to the chord information obtained in step 2031, comparing the number of the neutralized strings in the music in each tune, selecting the tune with the most number as the tonality of the music, and writing the tonality in the music file;
step 2033, calculating the energy of each chord root, calculating the average value of all the energies as a threshold, selecting the root with the energy larger than the threshold, calculating the time interval between adjacent roots, and selecting the most frequently occurring time interval as the bar of the music; calculating the number of notes of each measure according to the duration of the measure, taking the root as accent, selecting notes with the energy close to that of the root from other notes as accent beats, determining which one of four beats, four beats and three beats or eight beats is selected to obtain rhythm information, and writing the rhythm information into a music score file;
step 2034, calculating the duration of a quarter note according to the bar duration and the rhythm information to obtain the playing speed of the music, and writing the playing speed into the music file.
Further, in step 201, the classification to which the mel-frequency cepstrum coefficient feature of each frame belongs is calculated based on a deep neural network model, and the building of the deep neural network model includes the following steps:
step 101, collecting guitar playing sound information by using a recording device, wherein the sound information comprises the grade string plucking sound of each string and all chord sweeping sounds, and generating an audio file;
102, marking the collected guitar sound information, wherein the marking content is a pitch and a playing mode, the pitch may include one or more than one, and the playing mode is divided into string plucking and string sweeping;
103, performing framing operation on the audio file in the step 101, and extracting the Mel cepstrum coefficient characteristics of each frame;
and step 104, training the deep neural network by using a back propagation algorithm based on the characteristic file obtained in the step 103 and the labeled file obtained in the step 102 to obtain a deep neural network model.
Based on the automatic generation method of the guitar music, for guitar playing without picture music, the guitar music can be automatically generated, manual music picking work is replaced, and the working efficiency can be effectively improved for professionals; the method reduces the difficulty of picking up the music score, even a beginner can independently finish the music score picking, and the guitar learning work is convenient. In addition, the method utilizes a deep learning technology, training data can be continuously added according to the characteristics of the neural network aiming at the weaknesses of the neural network, the resolving power exceeding that of human ears is finally obtained, and the method is finally greatly superior to the traditional method in effect and efficiency by combining the strong computing power of a computer.
Detailed Description
The following description is presented to disclose the invention so as to enable any person skilled in the art to practice the invention. The preferred embodiments in the following description are given by way of example only, and other obvious variations will occur to those skilled in the art.
For ease of understanding, before describing embodiments of the present invention, the features of the guitar sound will be described first.
The guitar is a stringed musical instrument, the vibration of strings gives out sound, the vibration frequency of the strings determines the pitch, the vibration frequency of the strings is influenced by the diameters and the vibration lengths of the strings, the different diameters of the strings are different, the different grades on the fingerboard determine the vibration lengths of the strings, and therefore the combination of the strings and the grades determines the pitch of the playing. In addition, the guitar belongs to the standard tuning musical instrument, and the standard tones of one string to six strings are E, B, G, D, A, E in turn, so that different strings and taste combinations may give the same pitch.
The chord is a musical concept and is composed of a group of sounds satisfying a certain interval relationship, wherein the lowest pitch of the sounds is a root and is the basis of the chord, and the intervals between different tones determine the color of the chord, which jointly form a chord. The chord is played mainly by two types of chord decomposition and chord sweeping, the chord decomposition is to pop out the chord in turn to form the sound, and the chord sweeping sweeps a plurality of strings simultaneously to produce a plurality of chord tones.
The playing of the guitar can be divided into accompaniment and solo according to the function, the accompaniment generally only plays the chord, provides harmony for the melody, and the solo plays the melody while playing the chord. Whether the music accompaniment or the solo, the pitch of each note is firstly determined when the music score is played, and if the music score is swept, the chord and the swept string are determined, which all need to be distinguished by human ears.
The invention utilizes the strong learning ability of the neural network to replace the human ear to finish the work of identifying the pitch of the note, takes all the sounds played by the guitar as the parameters of training data to train the neural network model, and uses the model to classify the unknown sounds, so that the pitch information in playing can be determined, and the first step of scratching the music score is finished. After the pitch of the music is determined, the musical instruments are divided into bars, and the chord used by each bar is further determined, usually the root of the chord is played first, and the first beat of the bar is the repeated beat, and then the chord trend of the music is determined by combining the combined sound information in the chord. Finally, the key and the rhythm of the music are determined, the key is generally determined by the composition of the chord, and the rhythm reflects the weight distribution and can be determined by analyzing the energy distribution of the music.
The guitar spectrum automatic generation method based on the voice recognition provided by the embodiment of the invention comprises two steps which are sequentially executed: and establishing a neural network model, and performing score scratching on the guitar playing audio file to be identified.
The neural network model building method comprises the following steps:
step 101, collecting guitar playing sound information by using a recording device, wherein the sound information comprises the grade string-plucking sound of each string and all chord string-sweeping sounds, and generating an audio file. Here, the guitar playing sound may come from a finger or a pick, and the audio file is preferably saved in 16k16bit PCM format.
And 102, marking the collected sound information of the guitar, wherein the marking content is a pitch and a playing mode, the pitch may contain one or more than one, and the playing mode is divided into string plucking and string sweeping. The output result of this step is a label file in text format.
And 103, performing framing operation on the audio file in the step 101, extracting the Mel cepstrum coefficient characteristics of each frame, wherein the output result of the step is a binary characteristic file.
And step 104, training the deep neural network by using a back propagation algorithm based on the characteristic file obtained in the step 103 and the labeled file obtained in the step 102 to obtain a deep neural network model. The output result of this step is a binary model file.
Establishing the deep neural network model, namely training the neural network model by using guitar playing sound and labels by using a deep learning technology; the characteristics of the vibration of different pitches on the frequency domain are learned by converting the audio frequency of the time domain into the frequency domain and extracting relevant characteristics and utilizing the strong learning capacity of the neural network, so that the human ear can be replaced, and the work of distinguishing different pitches is completed. According to the characteristics of the neural network, training data can be continuously added aiming at the weaknesses of the neural network, the resolving power exceeding that of human ears is finally obtained, and the effect and the efficiency are greatly superior to those of the traditional method by combining the strong computing power of a computer.
The method specifically comprises the following steps of:
step 201, performing framing operation on an input guitar playing audio file to be identified, and extracting the mel cepstrum coefficient characteristics of each frame. The output result of this step is a binary signature file.
Step 202, calculating the classification of the Mel cepstrum coefficient characteristics of each frame in the step 201 by using the deep neural network model generated in the step 104, obtaining the pitch and the playing mode information of each frame of the tested audio according to the output layer label of the model, newly building a binary music score file, and writing the corresponding pitch information;
step 203, analyzing the chord, tone and rhythm information of the music based on the pitch information in the music file in step 202, including the following steps:
step 2031, determining the positions of all chord root sounds in the music, calculating the time intervals, selecting the interval duration with the most frequent occurrence, preliminarily dividing the music into uniform time segments, comparing the sounds appearing in each time segment with the chord template, calculating the similarity, selecting the chord with the highest similarity as the chord of the time segment, and writing the chord information into the music score file;
step 2032, comparing the chord tables of the guitar according to the chord information obtained in step 2031, comparing the number of the neutralized strings in the music in each tune, selecting the tune with the most number as the tonality of the music, and writing the tonality in the music file;
step 2033, calculating the energy of each chord root, calculating the average value of all the energies as a threshold, selecting the root with the energy larger than the threshold, calculating the time interval between adjacent roots, and selecting the most frequently occurring time interval as the bar of the music; calculating the number of notes of each measure according to the duration of the measure, taking the root as accent, selecting notes with the energy close to that of the root from other notes as accent beats, determining which one of four beats, four beats and three beats or eight beats is selected to obtain rhythm information, and writing the rhythm information into a music score file;
step 2034, calculating the duration of a quarter note according to the bar duration and the rhythm information to obtain the playing speed of the music, and writing the playing speed into a music score file;
and 204, integrating the previous pitch, playing mode, chord, tone and rhythm information, and outputting a music score file in a text form.
The automatic guitar music generation method directly generates the guitar music from the playing sound of the guitar, and has the following characteristics and effects:
1. compared with a method for obtaining music scores from picture music, the method can automatically generate the guitar music for playing the guitar without the picture music, replaces the manual music picking work, and can effectively improve the working efficiency for professionals; the method reduces the difficulty of picking up the music score, even a beginner can independently finish the music score picking, and the guitar learning work is convenient.
2. Training a neural network model by using guitar playing sound and labels by utilizing a deep learning technology; the characteristics of the vibration of different pitches on the frequency domain are learned by converting the audio frequency of the time domain into the frequency domain and extracting relevant characteristics and utilizing the strong learning capacity of the neural network, so that the human ear can be replaced, and the work of distinguishing different pitches is completed. According to the characteristics of the neural network, training data can be continuously added aiming at the weaknesses of the neural network, the resolving power exceeding that of human ears is finally obtained, and the effect and the efficiency are greatly superior to those of the traditional method by combining the strong computing power of a computer.
3. After the neural network identifies the pitch information of each frame of audio, the information of chord distribution, tone, rhythm, speed and the like of the music score is obtained through analysis by combining the pitch composition of the chord and the characteristics of guitar playing. These are all done automatically, do not need manual participation, help the guitar fan to do guitar study faster better.
The foregoing shows and describes the general principles, essential features, and advantages of the invention. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are given by way of illustration of the principles of the present invention, and that various changes and modifications may be made without departing from the spirit and scope of the invention. Such changes and modifications are intended to be within the scope of the claimed invention. The scope of the invention is defined by the appended claims and equivalents thereof.

Claims (4)

1. A guitar music automatic generation method based on voice recognition is characterized by comprising a process of picking up music of a guitar playing audio file to be recognized, wherein the process comprises the following steps:
step 201, performing framing operation on an input guitar playing audio file to be identified, and extracting the Mel cepstrum coefficient characteristics of each frame;
step 202, calculating the classification of the Mel cepstrum coefficient characteristics of each frame in the step 201, and obtaining the pitch and playing mode information of each frame of the tested audio according to the output layer label of the model;
step 203, analyzing chord, tone and rhythm information of the music based on the pitch information in the music score file in the step 202;
and step 204, integrating the pitch, playing mode, chord, tone and rhythm information output in the steps 202 and 203, and outputting a music score file.
2. A method for automatic generation of guitar spectrum based on voice recognition according to claim 1, characterized by the step 203 comprising the sub-steps of:
step 2031, determining the positions of all chord root sounds in the music, calculating the time intervals, selecting the interval duration with the most frequent occurrence, preliminarily dividing the music into uniform time segments, comparing the sounds appearing in each time segment with the chord template, calculating the similarity, selecting the chord with the highest similarity as the chord of the time segment, and writing the chord information into the music score file;
step 2032, comparing the chord tables of the guitar according to the chord information obtained in step 2031, comparing the number of the neutralized strings in the music in each tune, selecting the tune with the most number as the tonality of the music, and writing the tonality in the music file;
step 2033, calculating the energy of each chord root, calculating the average value of all the energies as a threshold, selecting the root with the energy larger than the threshold, calculating the time interval between adjacent roots, and selecting the most frequently occurring time interval as the bar of the music; calculating the number of notes of each measure according to the duration of the measure, taking the root as accent, selecting notes with the energy close to that of the root from other notes as accent beats, determining which one of four beats, four beats and three beats or eight beats is selected to obtain rhythm information, and writing the rhythm information into a music score file;
step 2034, calculating the duration of a quarter note according to the bar duration and the rhythm information to obtain the playing speed of the music, and writing the playing speed into the music file.
3. The method for automatically generating guitar spectrum based on voice recognition according to claim 1 or 2, wherein in the step 201, the classification to which the mel cepstral coefficient feature of each frame belongs is calculated based on a deep neural network model, and the deep neural network model is established by the following steps:
step 101, collecting guitar playing sound information by using a recording device, wherein the sound information comprises the grade string plucking sound of each string and all chord sweeping sounds, and generating an audio file;
102, marking the collected guitar sound information, wherein the marking content is pitch and playing mode;
103, performing framing operation on the audio file in the step 101, and extracting the Mel cepstrum coefficient characteristics of each frame;
and step 104, training the deep neural network by using a back propagation algorithm based on the characteristic file obtained in the step 103 and the labeled file obtained in the step 102 to obtain a deep neural network model.
4. The method of automatic generation of guitar spectrum based on voice recognition according to claim 3, characterized in that the guitar playing voice information comes from finger or pick playing.
CN202011392002.1A 2020-12-02 2020-12-02 Guitar music automatic generation method based on voice recognition Active CN112634841B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011392002.1A CN112634841B (en) 2020-12-02 2020-12-02 Guitar music automatic generation method based on voice recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011392002.1A CN112634841B (en) 2020-12-02 2020-12-02 Guitar music automatic generation method based on voice recognition

Publications (2)

Publication Number Publication Date
CN112634841A true CN112634841A (en) 2021-04-09
CN112634841B CN112634841B (en) 2022-11-29

Family

ID=75307435

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011392002.1A Active CN112634841B (en) 2020-12-02 2020-12-02 Guitar music automatic generation method based on voice recognition

Country Status (1)

Country Link
CN (1) CN112634841B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113707112A (en) * 2021-08-13 2021-11-26 陕西师范大学 Recursive jump connection deep learning music automatic generation method based on layer standardization
CN113763913A (en) * 2021-09-16 2021-12-07 腾讯音乐娱乐科技(深圳)有限公司 Music score generation method, electronic device and readable storage medium

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040231498A1 (en) * 2003-02-14 2004-11-25 Tao Li Music feature extraction using wavelet coefficient histograms
CN102723079A (en) * 2012-06-07 2012-10-10 天津大学 Music and chord automatic identification method based on sparse representation
JP3185768U (en) * 2013-05-23 2013-09-05 二郎 須澤 Score for shakuhachi
CN103714806A (en) * 2014-01-07 2014-04-09 天津大学 Chord recognition method combining SVM with enhanced PCP
CN104992712A (en) * 2015-07-06 2015-10-21 成都云创新科技有限公司 Music reorganization-based music score automatic formation method
CN106205570A (en) * 2016-08-10 2016-12-07 秦桂芳 A kind of music score recording method
US10008188B1 (en) * 2017-01-31 2018-06-26 Kyocera Document Solutions Inc. Musical score generator
CN110310621A (en) * 2019-05-16 2019-10-08 平安科技(深圳)有限公司 Sing synthetic method, device, equipment and computer readable storage medium
CN111898753A (en) * 2020-08-05 2020-11-06 字节跳动有限公司 Music transcription model training method, music transcription method and corresponding device

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040231498A1 (en) * 2003-02-14 2004-11-25 Tao Li Music feature extraction using wavelet coefficient histograms
CN102723079A (en) * 2012-06-07 2012-10-10 天津大学 Music and chord automatic identification method based on sparse representation
JP3185768U (en) * 2013-05-23 2013-09-05 二郎 須澤 Score for shakuhachi
CN103714806A (en) * 2014-01-07 2014-04-09 天津大学 Chord recognition method combining SVM with enhanced PCP
CN104992712A (en) * 2015-07-06 2015-10-21 成都云创新科技有限公司 Music reorganization-based music score automatic formation method
CN106205570A (en) * 2016-08-10 2016-12-07 秦桂芳 A kind of music score recording method
US10008188B1 (en) * 2017-01-31 2018-06-26 Kyocera Document Solutions Inc. Musical score generator
CN110310621A (en) * 2019-05-16 2019-10-08 平安科技(深圳)有限公司 Sing synthetic method, device, equipment and computer readable storage medium
CN111898753A (en) * 2020-08-05 2020-11-06 字节跳动有限公司 Music transcription model training method, music transcription method and corresponding device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王峰等: "基于CRFs和MPCP特征的和弦识别研究", 《计算机工程与应用》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113707112A (en) * 2021-08-13 2021-11-26 陕西师范大学 Recursive jump connection deep learning music automatic generation method based on layer standardization
CN113707112B (en) * 2021-08-13 2024-05-28 陕西师范大学 Automatic generation method of recursion jump connection deep learning music based on layer standardization
CN113763913A (en) * 2021-09-16 2021-12-07 腾讯音乐娱乐科技(深圳)有限公司 Music score generation method, electronic device and readable storage medium
WO2023040332A1 (en) * 2021-09-16 2023-03-23 腾讯音乐娱乐科技(深圳)有限公司 Method for generating musical score, electronic device, and readable storage medium
CN113763913B (en) * 2021-09-16 2024-06-18 腾讯音乐娱乐科技(深圳)有限公司 Music score generating method, electronic equipment and readable storage medium

Also Published As

Publication number Publication date
CN112634841B (en) 2022-11-29

Similar Documents

Publication Publication Date Title
Barbancho et al. Automatic transcription of guitar chords and fingering from audio
Lee et al. Acoustic chord transcription and key extraction from audio using key-dependent HMMs trained on synthesized audio
Barbedo et al. Automatic genre classification of musical signals
Hung et al. Frame-level instrument recognition by timbre and pitch
CN112382257B (en) Audio processing method, device, equipment and medium
Lee et al. A Unified System for Chord Transcription and Key Extraction Using Hidden Markov Models.
WO2013080210A1 (en) Method for extracting representative segments from music
CN112634841B (en) Guitar music automatic generation method based on voice recognition
JP2010054802A (en) Unit rhythm extraction method from musical acoustic signal, musical piece structure estimation method using this method, and replacing method of percussion instrument pattern in musical acoustic signal
CN113192471B (en) Musical main melody track recognition method based on neural network
Chordia Segmentation and Recognition of Tabla Strokes.
Paulus Signal processing methods for drum transcription and music structure analysis
EP2342708B1 (en) Method for analyzing a digital music audio signal
Ramirez et al. Automatic performer identification in commercial monophonic jazz performances
Kum et al. Pseudo-label transfer from frame-level to note-level in a teacher-student framework for singing transcription from polyphonic music
Lerch Software-based extraction of objective parameters from music performances
CN110134823B (en) MIDI music genre classification method based on normalized note display Markov model
Setragno et al. Feature-based characterization of violin timbre
Jensen et al. Binary decision tree classification of musical sounds
Nichols et al. Automatically discovering talented musicians with acoustic analysis of youtube videos
MA et al. Four-way classification of tabla strokes with models adapted from Automatic Drum Transcription
Zhao et al. Transfer learning for violinist identification
Camurri et al. An experiment on analysis and synthesis of musical expressivity
Gogineni et al. Mridangam artist identification from taniavartanam audio
Zhao et al. Violinist identification using note-level timbre feature distributions

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant