US20040244565A1 - Method of creating music file with main melody and accompaniment - Google Patents

Method of creating music file with main melody and accompaniment Download PDF

Info

Publication number
US20040244565A1
US20040244565A1 US10/250,141 US25014103A US2004244565A1 US 20040244565 A1 US20040244565 A1 US 20040244565A1 US 25014103 A US25014103 A US 25014103A US 2004244565 A1 US2004244565 A1 US 2004244565A1
Authority
US
United States
Prior art keywords
accompaniment
music
main melody
tracks
note
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/250,141
Inventor
Wen-Ni Cheng
Chun-Bin Kuo
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
BenQ Corp
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US10/250,141 priority Critical patent/US20040244565A1/en
Assigned to BENQ CORPORATION reassignment BENQ CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHENG, WEN-NI, KUO, CHUN-BIN
Priority to TW093116006A priority patent/TWI263975B/en
Priority to CNA2004100492619A priority patent/CN1573915A/en
Publication of US20040244565A1 publication Critical patent/US20040244565A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/36Accompaniment arrangements
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/005Musical accompaniment, i.e. complete instrumental rhythm synthesis added to a performed melody, e.g. as output by drum machines
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/101Music Composition or musical creation; Tools or processes therefor
    • G10H2210/105Composing aid, e.g. for supporting creation, edition or modification of a piece of music
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/101Music Composition or musical creation; Tools or processes therefor
    • G10H2210/151Music Composition or musical creation; Tools or processes therefor using templates, i.e. incomplete musical sections, as a basis for composing
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/056MIDI or other note-oriented file format

Definitions

  • the present invention relates to creating music files with multiple tracks, and more specifically, to a method for creating music files containing a main melody and an accompaniment.
  • MIDI Musical Instrument Digital Interface
  • Each MIDI file can contain multiple tracks, and each track may contain music for a different instrument. Often one track of the MIDI file is used for storing a main melody and other tracks are used for storing an accompaniment to the main melody.
  • FIG. 1 is a diagram showing a basic structure of a MIDI file 30 according to the prior art.
  • the MIDI file 30 is composed of a series of bytes of data, each represented in hexadecimal format in FIG. 1.
  • the MIDI file 30 shown in FIG. 1 contains a file header 32 , a first track 36 , a second track 38 , and a third track 40 .
  • the file header 32 includes a track number indicator 34 for indicating a total number of tracks included in the MIDI file 30 .
  • the track number indicator 34 contains a value of “3” since there are three tracks.
  • Each of the tracks 36 , 38 , and 40 can be used for storing the notes of a different instrument, so the MIDI file 30 shown in FIG. 1 may contain music for three different instruments.
  • the file header 32 also contains a quarter note tick indicator 35 for indicating how many clock ticks a quarter note receives. In this case, 78 (measured in hexadecimal; equal to 120 decimal) clock ticks will be equal to the duration of a quarter note.
  • the first track 36 is used for storing,text and other information.
  • the first track 36 contains a tempo indicator 37 for indicating the duration of a quarter note.
  • the tempo indicator 37 contains six bytes.
  • the first three bytes “FF 51 03” make up an event type indicator.
  • the event type indicator shows that the following three bytes “09 27 C0” (equal to 600,000 decimal) is how many microseconds the duration of a quarter note should be. In this case, the duration of each quarter note will be 0.6 seconds.
  • the second track 38 and the third track 40 of the MIDI file 30 are examples of two different music tracks.
  • the second track 38 could represent a main melody and the third track 40 could represent an accompaniment track. Additional accompaniment tracks could also be added to the MIDI file 30 , according to the wishes of the user.
  • a method of creating a music file comprising a plurality of tracks to be played simultaneously when the music file is played.
  • the method includes creating a main melody track by selecting a respective pitch and duration of a plurality of notes, selecting a style of accompaniment music, retrieving accompaniment tracks for the selected style of accompaniment music from a memory, and combining the main melody track and the accompaniment tracks to create the music file.
  • FIG. 1 is a diagram showing a basic structure of a MIDI file according to the prior art.
  • FIG. 2 is a diagram illustrating a main melody entered by a user according to the present invention.
  • FIG. 3 is a detailed diagram of a second track of the MIDI file shown in FIG. 1.
  • FIG. 4 is a chart showing timing of each event in the second track.
  • FIG. 5 is a diagram illustrating the main melody of FIG. 2 being divided into measures.
  • FIG. 6 is a chart of an event buffer showing all note-on events shown in FIG. 4.
  • FIG. 7 illustrates assigning keys to measures of the main melody for changing a key of the accompaniment.
  • FIG. 8 illustrates shifting a key of the accompaniment according to the present invention.
  • FIG. 9 is a diagram of shifting the key of accompaniment tracks according to the present invention.
  • FIG. 10 is a chart illustrating the offsets of different keys from the key of C.
  • FIG. 11 is a flowchart illustrating creating the MIDI file according to the present invention method.
  • FIG. 12 is a flowchart further illustrating calculating the total number of measures in the main melody (step 150 in the flowchart of FIG. 11) according to the present invention method.
  • FIG. 13 is a flowchart further illustrating combining the main melody with accompaniment tracks (step 200 in the flowchart of FIG. 11) according to the present invention method.
  • the present invention simplifies the process of creating a MIDI file by automatically adding accompaniment tracks to a main melody track created by the user.
  • the user may music editing software on a cellular phone or computer, for example, to create the MIDI files according to the present invention.
  • FIG. 2 is a diagram illustrating a main melody 60 entered by a user according to the present invention.
  • FIG. 2 shows the first seven notes of the children's, song “Twinkle, Twinkle Little Star” as an example for the main melody 60 .
  • a user would be presented with an interface allowing the user to select a type of note (such as a whole note, half note, quarter note, etc.) and a pitch of the note (such as A, C, G, etc.). The user could add notes one note at a time until the main melody 60 shown in FIG. 2 is complete.
  • the main melody 60 can then be converted into a standard MIDI track format.
  • the MIDI file 30 shown in FIG. 1 contains the first track 36 , the second track 38 , and the third track 40 .
  • the second track 38 will be used as an example.
  • FIG. 3 is a detailed diagram of the second track 38 of the MIDI file 30 shown in FIG. 1.
  • FIG. 4 is a chart showing timing of each event in the second track 38 .
  • the second track 38 contains the main melody 60 created by the user.
  • the present invention first involves analyzing the main melody 60 for creating the second track 38 based on the main melody 60 .
  • the second track 38 contains a track header 50 , a plurality of delta times 52 , a plurality of non-note events 54 , and a plurality of note-events 56 .
  • the delta time 52 is placed before each non-note event 54 and note-event 56 for indicating a period of elapsed time before that event. Since the non-note events 54 do not play any notes in the second track 38 , the delta time 52 before each non-note event 54 is equal to “00”.
  • the delta time 52 is varied to change the duration of notes that are specified in the note-events 56 . For instance, each quarter note would have a delta time 52 of 78 (measured in hexadecimal; equal to 120 decimal) clock ticks.
  • All of the non-note events 54 and note-events 56 are shown in rows of FIG. 4. Seven columns in FIG. 4 show an event number given for reference, the delta time 52 value, a play sequence indicator, the byte representation of the event, a period of the event, a type of note played, and the event type.
  • the delta time 52 value shows the amount of time that elapses between the previous event and the current event.
  • the event period shows how long each event is valid for.
  • Three different event types are shown in FIG. 4.
  • the non-note events 54 do not affect audible notes, the note-on events are the start of new notes, and the note-off events are the endings of notes.
  • the first six events will be briefly described.
  • the first two events are non-note events, each having a delta time of “0x00” (hexadecimal) preceding it.
  • the third event is a note-on event having a delta time of “0x00” preceding it.
  • the byte representation for this event is “90 3C 64”, wherein the “3C” byte represents a pitch of the note being played and the “64” byte represents a volume of the note.
  • the fourth event is a note-off event having a delta time of “0x78” preceding it.
  • the byte representation for this event is “90 3C 00”, meaning that the volume of the previous note has now been set to “00”, which is zero volume.
  • the fifth event is a note-on event having a delta time of “0x00” preceding it.
  • the sixth event is a note-off event having a delta time of “0x78” preceding it.
  • FIG. 5 is a diagram illustrating the main melody 60 of FIG. 2 being divided into measures. Since 4/4 time is the most popular timing for songs used in electronic devices, 4/4 time will be used to break the main melody 60 into a first measure 62 and a second measure 64 .
  • the first measure 62 contains four quarter notes and the second measure 64 contains two quarter notes and a half note.
  • FIG. 6 is a chart of an event buffer showing all of the note-on events shown in FIG. 4.
  • the each note will be added to an event buffer.
  • Each note-on event is stored along with its event period, and the measure that the note is placed in.
  • the first note has a tone of “3C”, which is converted into “60”in decimal.
  • the event period for the first note is “0x78”, which is the same as 600 ms.
  • the event buffer for the first measure will hold four quarter notes and the event buffer for the second measure will hold two quarter notes and one half note.
  • FIG. 7 illustrates assigning keys to measures of the main melody 60 for changing a key of the accompaniment.
  • the first measure 62 is assigned an accompaniment key of D
  • the second measure 64 is assigned an accompaniment key of E.
  • FIG. 8 illustrates shifting a key of the accompaniment according to the present invention.
  • An accompaniment database 74 stored in a memory 72 contains accompaniment measures for each available style of accompaniment music, and feeds these accompaniment measures to a key shifter 70 .
  • the key shifter 70 is a device used to shift a key of the accompaniment music based on a measure key input to the key shifter 70 . For instance, to change a key of the accompaniment from C to D, an increase of two half steps is required. Therefore, a value of “2” could be added to the pitch of all notes in the accompaniment measures retrieved from the database.
  • FIG. 9 is a diagram of shifting the key of accompaniment tracks according to the present invention.
  • the first measure 62 of the main melody 60 is shown as having a key of D selected for the accompaniment chord therefore the accompaniment needs to be shifted from the key of C to the key of D.
  • a value of “2” is then added to the pitch of each note in the accompaniment tracks.
  • FIG. 10 is a chart illustrating the offsets of different keys from the key of C. To go from the key of C to the key of A, for example, a value of “9” could be added to the pitch of each note or a value of “3” could be subtracted from the pitch of each note, depending on the desired octave.
  • FIG. 11 is a flowchart illustrating creating the MIDI file 30 according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step 140 Start;
  • Step 142 The user edits the notes of the main melody 60 by selecting a duration and pitch of each note;
  • Step 144 Determine if the user is finished editing the main melody 60 ; if so, go to step 150 ; if not, go back to step 142 ;
  • Step 150 Calculate the total number of measures of the main melody 60 ; go to step 194 ;
  • Step 194 The user edits the accompaniment key corresponding to each measure of the main melody 60 ;
  • Step 196 Determine if the user is finished editing the accompaniment keys; if so, go to step 198 ; if not, go back to step 194 ;
  • Step 198 The user selects the style of music for the accompaniment such as jazz, dance, etc;
  • Step 200 Combine the main melody 60 with the accompaniment measure-by-measure based on the selected style and key of the accompaniment, and output the MIDI file 30 ; go to step 250 ; and
  • Step 250 End.
  • FIG. 12 is a flowchart further illustrating calculating the total number of measures in the main melody 60 (step 150 in the flowchart of FIG. 11) according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step 152 Start;
  • Step 154 Calculate the total period of a measure based on the period of a quarter note
  • Step 156 Read the main melody track
  • Step 158 Determine if the end of the main melody track has been reached; if so, go to step 176 ; if not, go to step 160 ;
  • Step 160 Read next delta time
  • Step 162 Read next track event
  • Step 164 Determine if this event is a note-on event; if so, go to step 168 ; if not, go to step 166 ;
  • Step 166 Adjust the timer by adding up all previous delta times; go to step 158 ;
  • Step 168 Calculate the period of this event
  • Step 170 Determine if this event is over the period of the current measure; if so, go to step 172 ; if not, go to step 174 ;
  • Step 172 Create a buffer for the next measure
  • Step 174 Put this event into the corresponding measure buffer; go to step 166 ; and
  • Step 176 End.
  • FIG. 13 is a flowchart further illustrating combining the main melody. 60 with accompaniment tracks (step 200 in the flowchart of FIG. 11) according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step 202 Start;
  • Step 204 Open the MIDI file 30 for writing
  • Step 206 Write the midi file header 32 ;
  • Step 208 Determine if all tracks have been written to the MIDI file 30 ; if yes, go to step 220 ; if not, go to step 210 ;
  • Step 210 Write the track header for the current track
  • Step 212 Determine if all data for all measures has been written for the current track; if so, go back to step 208 ; if not, go to step 214 ;
  • Step 214 Read the style and key for the accompaniment corresponding to the current measure
  • Step 216 Shift the key of the accompaniment for this measure based on the selected key
  • Step 218 Write the data for this measure into the MIDI file 30 ; go back to step 212 ;
  • Step 220 Close the file to finish the writing process.
  • Step 222 End.
  • the present invention method allows users to create a MIDI file by simply editing a main melody, selecting an accompaniment key for each measure of the main melody, and selecting a style of the accompaniment.
  • This improved process for creating MIDI files allows users to create their own songs quickly and easily. Moreover, even users with no knowledge of music theory can still create sophisticated music files.

Abstract

A method of creating a music file comprising a plurality of tracks to be played simultaneously when the music file is played. The method includes creating a main melody track by selecting a respective pitch and duration of a plurality of notes, selecting a style of accompaniment music, retrieving accompaniment tracks for the selected style of accompaniment music from a memory, and combining the main melody track and the accompaniment tracks to create the music file.

Description

    BACKGROUND OF INVENTION
  • 1. Field of the Invention [0001]
  • The present invention relates to creating music files with multiple tracks, and more specifically, to a method for creating music files containing a main melody and an accompaniment. [0002]
  • 2. Description of the Prior Art [0003]
  • With the popularity of electronic devices such as cellular phones, users enjoy personalizing their electronic devices with unique songs or tunes. One popular format for creating music files is a Musical Instrument Digital Interface (MIDI) file. Each MIDI file can contain multiple tracks, and each track may contain music for a different instrument. Often one track of the MIDI file is used for storing a main melody and other tracks are used for storing an accompaniment to the main melody. [0004]
  • Please refer to FIG. 1. FIG. 1 is a diagram showing a basic structure of a [0005] MIDI file 30 according to the prior art. The MIDI file 30 is composed of a series of bytes of data, each represented in hexadecimal format in FIG. 1. The MIDI file 30 shown in FIG. 1 contains a file header 32, a first track 36, a second track 38, and a third track 40. The file header 32 includes a track number indicator 34 for indicating a total number of tracks included in the MIDI file 30. In this case, the track number indicator 34 contains a value of “3” since there are three tracks. Each of the tracks 36, 38, and 40 can be used for storing the notes of a different instrument, so the MIDI file 30 shown in FIG. 1 may contain music for three different instruments. The file header 32 also contains a quarter note tick indicator 35 for indicating how many clock ticks a quarter note receives. In this case, 78 (measured in hexadecimal; equal to 120 decimal) clock ticks will be equal to the duration of a quarter note. In the example shown in FIG. 1, the first track 36 is used for storing,text and other information. The first track 36 contains a tempo indicator 37 for indicating the duration of a quarter note. The tempo indicator 37 contains six bytes. The first three bytes “FF 51 03” make up an event type indicator. The event type indicator shows that the following three bytes “09 27 C0” (equal to 600,000 decimal) is how many microseconds the duration of a quarter note should be. In this case, the duration of each quarter note will be 0.6 seconds.
  • The [0006] second track 38 and the third track 40 of the MIDI file 30 are examples of two different music tracks. For instance, the second track 38 could represent a main melody and the third track 40 could represent an accompaniment track. Additional accompaniment tracks could also be added to the MIDI file 30, according to the wishes of the user.
  • Unfortunately, the prior art method of creating the [0007] MIDI file 30 is a long and tedious process. The user has to create individual notes not only for the main melody track, but also for each additional accompaniment track. Not many people have the musical knowledge necessary to compose a main melody and an acceptable group of accompaniment tracks. In addition, those who are capable of composing may feel overwhelmed by the amount of time needed for creating many tracks, and may give up before completion.
  • SUMMARY OF INVENTION
  • It is therefore a primary objective of the claimed invention to provide a method for creating a main melody and accompaniment tracks in a music file in order to solve the above-mentioned problems. [0008]
  • According to the claimed invention, a method of creating a music file comprising a plurality of tracks to be played simultaneously when the music file is played is introduced. The method includes creating a main melody track by selecting a respective pitch and duration of a plurality of notes, selecting a style of accompaniment music, retrieving accompaniment tracks for the selected style of accompaniment music from a memory, and combining the main melody track and the accompaniment tracks to create the music file. [0009]
  • It is an advantage of the claimed invention that users can create a MIDI file with a main melody and accompaniment by simply editing the main melody and selecting a style of accompaniment music. This allows users to create their own songs quickly and easily, and no significant knowledge of music is required of the user. [0010]
  • These and other objectives of the claimed invention will no doubt become obvious to those of ordinary skill in the art after reading the following detailed description of the preferred embodiment, which is illustrated in the various figures and drawings.[0011]
  • BRIEF DESCRIPTION OF DRAWINGS
  • FIG. 1 is a diagram showing a basic structure of a MIDI file according to the prior art. [0012]
  • FIG. 2 is a diagram illustrating a main melody entered by a user according to the present invention. [0013]
  • FIG. 3 is a detailed diagram of a second track of the MIDI file shown in FIG. 1. [0014]
  • FIG. 4 is a chart showing timing of each event in the second track. [0015]
  • FIG. 5 is a diagram illustrating the main melody of FIG. 2 being divided into measures. [0016]
  • FIG. 6 is a chart of an event buffer showing all note-on events shown in FIG. 4. [0017]
  • FIG. 7 illustrates assigning keys to measures of the main melody for changing a key of the accompaniment. [0018]
  • FIG. 8 illustrates shifting a key of the accompaniment according to the present invention. [0019]
  • FIG. 9 is a diagram of shifting the key of accompaniment tracks according to the present invention. [0020]
  • FIG. 10 is a chart illustrating the offsets of different keys from the key of C. [0021]
  • FIG. 11 is a flowchart illustrating creating the MIDI file according to the present invention method. [0022]
  • FIG. 12 is a flowchart further illustrating calculating the total number of measures in the main melody ([0023] step 150 in the flowchart of FIG. 11) according to the present invention method.
  • FIG. 13 is a flowchart further illustrating combining the main melody with accompaniment tracks ([0024] step 200 in the flowchart of FIG. 11) according to the present invention method.
  • DETAILED DESCRIPTION
  • The present invention simplifies the process of creating a MIDI file by automatically adding accompaniment tracks to a main melody track created by the user. The user may music editing software on a cellular phone or computer, for example, to create the MIDI files according to the present invention. [0025]
  • Please refer to FIG. 2. FIG. 2 is a diagram illustrating a [0026] main melody 60 entered by a user according to the present invention. FIG. 2 shows the first seven notes of the children's, song “Twinkle, Twinkle Little Star” as an example for the main melody 60. For creating the main melody 60, a user would be presented with an interface allowing the user to select a type of note (such as a whole note, half note, quarter note, etc.) and a pitch of the note (such as A, C, G, etc.). The user could add notes one note at a time until the main melody 60 shown in FIG. 2 is complete. Once the main melody 60 is entered, the main melody 60 can then be converted into a standard MIDI track format.
  • Please refer back to FIG. 1. The [0027] MIDI file 30 shown in FIG. 1 contains the first track 36, the second track 38, and the third track 40. For showing how the main melody 60 can be converted into a MIDI track of the MIDI file 30, the second track 38 will be used as an example. Please refer to FIG. 3 and FIG. 4. FIG. 3 is a detailed diagram of the second track 38 of the MIDI file 30 shown in FIG. 1. FIG. 4 is a chart showing timing of each event in the second track 38. Suppose that the second track 38 contains the main melody 60 created by the user. The present invention first involves analyzing the main melody 60 for creating the second track 38 based on the main melody 60. The second track 38 contains a track header 50, a plurality of delta times 52, a plurality of non-note events 54, and a plurality of note-events 56. The delta time 52 is placed before each non-note event 54 and note-event 56 for indicating a period of elapsed time before that event. Since the non-note events 54 do not play any notes in the second track 38, the delta time 52 before each non-note event 54 is equal to “00”. The delta time 52 is varied to change the duration of notes that are specified in the note-events 56. For instance, each quarter note Would have a delta time 52 of 78 (measured in hexadecimal; equal to 120 decimal) clock ticks.
  • All of the [0028] non-note events 54 and note-events 56 are shown in rows of FIG. 4. Seven columns in FIG. 4 show an event number given for reference, the delta time 52 value, a play sequence indicator, the byte representation of the event, a period of the event, a type of note played, and the event type. The delta time 52 value shows the amount of time that elapses between the previous event and the current event. The event period shows how long each event is valid for. Three different event types are shown in FIG. 4. The non-note events 54 do not affect audible notes, the note-on events are the start of new notes, and the note-off events are the endings of notes.
  • To further illustrate the events shown in FIG. 4, the first six events will be briefly described. The first two events are non-note events, each having a delta time of “0x00” (hexadecimal) preceding it. [0029]
  • The third event is a note-on event having a delta time of “0x00” preceding it. [0030]
  • The byte representation for this event is “90 [0031] 3C 64”, wherein the “3C” byte represents a pitch of the note being played and the “64” byte represents a volume of the note. By looking at the delta time 52 for the following event, which is “0x78”, we can determine that the event period for this event is equal to “0x78”, meaning that this is a quarter note.
  • The fourth event is a note-off event having a delta time of “0x78” preceding it. The byte representation for this event is “90 [0032] 3C 00”, meaning that the volume of the previous note has now been set to “00”, which is zero volume.
  • Since the [0033] delta time 52 immediately following this note-off event is equal to “0x00”, this event has a period of 0.
  • The fifth event is a note-on event having a delta time of “0x00” preceding it. [0034]
  • The following [0035] delta time 52 is “0x78”, making the fifth event another quarter note. In fact, the fifth event plays the same note as the previous note immediately after the previous note has stopped playing.
  • The sixth event is a note-off event having a delta time of “0x78” preceding it. [0036]
  • The sixth event terminates the note that was begun in the fifth event. [0037]
  • Therefore, so far a total of two notes have been played, with each note having the same pitch and same duration. This is equal to playing the first two notes shown in FIG. 2. [0038]
  • Please refer to FIG. 5. FIG. 5 is a diagram illustrating the [0039] main melody 60 of FIG. 2 being divided into measures. Since 4/4 time is the most popular timing for songs used in electronic devices, 4/4 time will be used to break the main melody 60 into a first measure 62 and a second measure 64. The first measure 62 contains four quarter notes and the second measure 64 contains two quarter notes and a half note.
  • Please refer to FIG. 6. FIG. 6 is a chart of an event buffer showing all of the note-on events shown in FIG. 4. After the user creates the [0040] main melody 60, the each note will be added to an event buffer. Each note-on event is stored along with its event period, and the measure that the note is placed in. For example, the first note has a tone of “3C”, which is converted into “60”in decimal. The event period for the first note is “0x78”, which is the same as 600 ms. The event buffer for the first measure will hold four quarter notes and the event buffer for the second measure will hold two quarter notes and one half note.
  • Once the [0041] main melody 60 has been divided into measures and written to a track of the MIDI file 30 (in this case, the second track 38), the user is prompted to enter a desired key of the accompaniment tracks for each measure of the main melody 60. If there was a key change in the main melody-60, the key of the accompaniment could easily be changed by specifying a different key for those corresponding measures of the accompaniment. Please refer to FIG. 7. FIG. 7 illustrates assigning keys to measures of the main melody 60 for changing a key of the accompaniment. As the example in FIG. 7 shows, the first measure 62 is assigned an accompaniment key of D, and the second measure 64 is assigned an accompaniment key of E.
  • In addition to specifying the key of the accompaniment corresponding to each measure of the [0042] main melody 60, the user is also asked to select a style of music such as jazz, dance, etc. Based on the style selection made by the user, accompaniment measures will be retrieved from a database. For simplicity, the database only stores accompaniment measures in the key of C. Any other accompaniment keys will be generated by shifting from the key of C. Please refer to FIG. 8. FIG. 8 illustrates shifting a key of the accompaniment according to the present invention. An accompaniment database 74 stored in a memory 72 contains accompaniment measures for each available style of accompaniment music, and feeds these accompaniment measures to a key shifter 70. The key shifter 70 is a device used to shift a key of the accompaniment music based on a measure key input to the key shifter 70. For instance, to change a key of the accompaniment from C to D, an increase of two half steps is required. Therefore, a value of “2” could be added to the pitch of all notes in the accompaniment measures retrieved from the database.
  • Please refer to FIG. 9. FIG. 9 is a diagram of shifting the key of accompaniment tracks according to the present invention. The [0043] first measure 62 of the main melody 60 is shown as having a key of D selected for the accompaniment chord therefore the accompaniment needs to be shifted from the key of C to the key of D. A value of “2” is then added to the pitch of each note in the accompaniment tracks. Please refer to FIG. 10. FIG. 10 is a chart illustrating the offsets of different keys from the key of C. To go from the key of C to the key of A, for example, a value of “9” could be added to the pitch of each note or a value of “3” could be subtracted from the pitch of each note, depending on the desired octave.
  • Please refer to FIG. 11. FIG. 11 is a flowchart illustrating creating the [0044] MIDI file 30 according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step [0045] 140: Start;
  • Step [0046] 142: The user edits the notes of the main melody 60 by selecting a duration and pitch of each note;
  • Step [0047] 144: Determine if the user is finished editing the main melody 60; if so, go to step 150; if not, go back to step 142;
  • Step [0048] 150: Calculate the total number of measures of the main melody 60; go to step 194;
  • Step [0049] 194: The user edits the accompaniment key corresponding to each measure of the main melody 60;
  • Step [0050] 196: Determine if the user is finished editing the accompaniment keys; if so, go to step 198; if not, go back to step 194;
  • Step [0051] 198: The user selects the style of music for the accompaniment such as jazz, dance, etc;
  • Step [0052] 200: Combine the main melody 60 with the accompaniment measure-by-measure based on the selected style and key of the accompaniment, and output the MIDI file 30; go to step 250; and
  • Step [0053] 250: End.
  • Please refer to FIG. 12. FIG. 12 is a flowchart further illustrating calculating the total number of measures in the main melody [0054] 60 (step 150 in the flowchart of FIG. 11) according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step [0055] 152: Start;
  • Step [0056] 154: Calculate the total period of a measure based on the period of a quarter note;
  • Step [0057] 156: Read the main melody track;
  • Step [0058] 158: Determine if the end of the main melody track has been reached; if so, go to step 176; if not, go to step 160;
  • Step [0059] 160: Read next delta time;
  • Step [0060] 162: Read next track event;
  • Step [0061] 164: Determine if this event is a note-on event; if so, go to step 168; if not, go to step 166;
  • Step [0062] 166: Adjust the timer by adding up all previous delta times; go to step 158;
  • Step [0063] 168: Calculate the period of this event;
  • Step [0064] 170: Determine if this event is over the period of the current measure; if so, go to step 172; if not, go to step 174;
  • Step [0065] 172: Create a buffer for the next measure;
  • Step [0066] 174: Put this event into the corresponding measure buffer; go to step 166; and
  • Step [0067] 176: End.
  • Please refer to FIG. 13. FIG. 13 is a flowchart further illustrating combining the main melody. [0068] 60 with accompaniment tracks (step 200 in the flowchart of FIG. 11) according to the present invention method. Steps contained in the flowchart will be explained below.
  • Step [0069] 202: Start;
  • Step [0070] 204: Open the MIDI file 30 for writing;
  • Step [0071] 206: Write the midi file header 32;
  • Step [0072] 208: Determine if all tracks have been written to the MIDI file 30; if yes, go to step 220; if not, go to step 210;
  • Step [0073] 210: Write the track header for the current track;
  • Step [0074] 212: Determine if all data for all measures has been written for the current track; if so, go back to step 208; if not, go to step 214;
  • Step [0075] 214: Read the style and key for the accompaniment corresponding to the current measure;
  • Step [0076] 216: Shift the key of the accompaniment for this measure based on the selected key;
  • Step [0077] 218: Write the data for this measure into the MIDI file 30; go back to step 212;
  • Step [0078] 220: Close the file to finish the writing process; and
  • Step [0079] 222: End.
  • Compared to the prior art, the present invention method allows users to create a MIDI file by simply editing a main melody, selecting an accompaniment key for each measure of the main melody, and selecting a style of the accompaniment. This improved process for creating MIDI files allows users to create their own songs quickly and easily. Moreover, even users with no knowledge of music theory can still create sophisticated music files. [0080]
  • Those skilled in the art will readily observe that numerous modifications and alterations of the device may be made while retaining the teachings of the invention. Accordingly, the above disclosure should be construed as limited only by the metes and bounds of the appended claims. [0081]

Claims (10)

What is claimed is:
1. A method of creating a music file comprising a plurality of tracks to be played simultaneously when the music file is played, the method comprising:
creating a main melody track by selecting a respective pitch and duration of a plurality of notes;
selecting a style of accompaniment music;
retrieving accompaniment tracks for the selected style of accompaniment music from a memory; and
combining the main melody track and the accompaniment tracks to create the music file.
2. The method of claim 1 further comprising selecting a key of the accompaniment music.
3. The method of claim 1 wherein the accompaniment tracks are retrieved from a database stored in the memory according to the selected style of accompaniment music.
4. The method of claim 1 further comprising dividing the main melody track into a plurality of measures according to the duration of the notes.
5. The method of claim 4 further comprising selecting a key of the accompaniment music for each measure of the main melody track.
6. The method of claim 5 wherein the key of each measure of the accompaniment tracks is shifted to match the selected key.
7. The method of claim 1 wherein the music file is a Musical Instrument Digital Interface (MIDI) file.
8. A music editing device for implementing the method of claim 1.
9. A computing device for creating a music file comprising a plurality of tracks to be played simultaneously when the music file is played, the computing device comprising: a plurality of input keys used for inputting a main melody track by selecting a respective pitch and duration of a plurality of notes and for selecting a style of accompaniment music;
a memory for storing accompaniment tracks of various styles of accompaniment music; and
a processor for combining the main melody track and the selected style of accompaniment tracks to create the music file.
10. The computing device of claim 9 wherein the music file is a Musical Instrument Digital Interface (MIDI) file.
US10/250,141 2003-06-06 2003-06-06 Method of creating music file with main melody and accompaniment Abandoned US20040244565A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
US10/250,141 US20040244565A1 (en) 2003-06-06 2003-06-06 Method of creating music file with main melody and accompaniment
TW093116006A TWI263975B (en) 2003-06-06 2004-06-03 Method and apparatus of creating a music file with main melody and accompaniment
CNA2004100492619A CN1573915A (en) 2003-06-06 2004-06-07 Method of creating music file with main melody and accompaniment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US10/250,141 US20040244565A1 (en) 2003-06-06 2003-06-06 Method of creating music file with main melody and accompaniment

Publications (1)

Publication Number Publication Date
US20040244565A1 true US20040244565A1 (en) 2004-12-09

Family

ID=33489131

Family Applications (1)

Application Number Title Priority Date Filing Date
US10/250,141 Abandoned US20040244565A1 (en) 2003-06-06 2003-06-06 Method of creating music file with main melody and accompaniment

Country Status (3)

Country Link
US (1) US20040244565A1 (en)
CN (1) CN1573915A (en)
TW (1) TWI263975B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050223879A1 (en) * 2004-01-20 2005-10-13 Huffman Eric C Machine and process for generating music from user-specified criteria
US20090107320A1 (en) * 2007-10-24 2009-04-30 Funk Machine Inc. Personalized Music Remixing
US20090217805A1 (en) * 2005-12-21 2009-09-03 Lg Electronics Inc. Music generating device and operating method thereof
US7705231B2 (en) 2007-09-07 2010-04-27 Microsoft Corporation Automatic accompaniment for vocal melodies
US20100162879A1 (en) * 2008-12-29 2010-07-01 International Business Machines Corporation Automated generation of a song for process learning
CN112420002A (en) * 2019-08-21 2021-02-26 北京峰趣互联网信息服务有限公司 Music generation method, device, electronic equipment and computer readable storage medium

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2013028315A1 (en) * 2011-07-29 2013-02-28 Music Mastermind Inc. System and method for producing a more harmonious musical accompaniment and for applying a chain of effects to a musical composition
CN105096922A (en) * 2014-05-07 2015-11-25 风彩创意有限公司 Composing method, composing program product, and composing system
CN106652655B (en) * 2015-10-29 2019-11-26 施政 A kind of musical instrument of track replacement
CN109872708B (en) * 2019-01-23 2023-04-28 平安科技(深圳)有限公司 Music generation method and device based on DCGAN
CN113076967B (en) * 2020-12-08 2022-09-23 无锡乐骐科技股份有限公司 Image and audio-based music score dual-recognition system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5859381A (en) * 1996-03-12 1999-01-12 Yamaha Corporation Automatic accompaniment device and method permitting variations of automatic performance on the basis of accompaniment pattern data
US6046396A (en) * 1998-08-25 2000-04-04 Yamaha Corporation Stringed musical instrument performance information composing apparatus and method
US20020007720A1 (en) * 2000-07-18 2002-01-24 Yamaha Corporation Automatic musical composition apparatus and method
US20020189425A1 (en) * 2001-03-06 2002-12-19 Yamaha Corporation Apparatus and method for automatically determining notational symbols based on musical composition data

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5859381A (en) * 1996-03-12 1999-01-12 Yamaha Corporation Automatic accompaniment device and method permitting variations of automatic performance on the basis of accompaniment pattern data
US6046396A (en) * 1998-08-25 2000-04-04 Yamaha Corporation Stringed musical instrument performance information composing apparatus and method
US20020007720A1 (en) * 2000-07-18 2002-01-24 Yamaha Corporation Automatic musical composition apparatus and method
US20020189425A1 (en) * 2001-03-06 2002-12-19 Yamaha Corporation Apparatus and method for automatically determining notational symbols based on musical composition data

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7394011B2 (en) * 2004-01-20 2008-07-01 Eric Christopher Huffman Machine and process for generating music from user-specified criteria
US20050223879A1 (en) * 2004-01-20 2005-10-13 Huffman Eric C Machine and process for generating music from user-specified criteria
US20090217805A1 (en) * 2005-12-21 2009-09-03 Lg Electronics Inc. Music generating device and operating method thereof
US20100192755A1 (en) * 2007-09-07 2010-08-05 Microsoft Corporation Automatic accompaniment for vocal melodies
US7705231B2 (en) 2007-09-07 2010-04-27 Microsoft Corporation Automatic accompaniment for vocal melodies
US7985917B2 (en) 2007-09-07 2011-07-26 Microsoft Corporation Automatic accompaniment for vocal melodies
US20090107320A1 (en) * 2007-10-24 2009-04-30 Funk Machine Inc. Personalized Music Remixing
US8173883B2 (en) * 2007-10-24 2012-05-08 Funk Machine Inc. Personalized music remixing
US20120210844A1 (en) * 2007-10-24 2012-08-23 Funk Machine Inc. Personalized music remixing
US8513512B2 (en) * 2007-10-24 2013-08-20 Funk Machine Inc. Personalized music remixing
US20140157970A1 (en) * 2007-10-24 2014-06-12 Louis Willacy Mobile Music Remixing
US20100162879A1 (en) * 2008-12-29 2010-07-01 International Business Machines Corporation Automated generation of a song for process learning
US7977560B2 (en) * 2008-12-29 2011-07-12 International Business Machines Corporation Automated generation of a song for process learning
CN112420002A (en) * 2019-08-21 2021-02-26 北京峰趣互联网信息服务有限公司 Music generation method, device, electronic equipment and computer readable storage medium

Also Published As

Publication number Publication date
TWI263975B (en) 2006-10-11
TW200428357A (en) 2004-12-16
CN1573915A (en) 2005-02-02

Similar Documents

Publication Publication Date Title
JP5813913B2 (en) A system for music composition
US5990407A (en) Automatic improvisation system and method
US5243123A (en) Music reproducing device capable of reproducing instrumental sound and vocal sound
KR100200290B1 (en) Automatic playing apparatus substituting available pattern for absent pattern
CN106708894B (en) Method and device for configuring background music for electronic book
JPS59189392A (en) Automatic transformer
CN101796587A (en) Automatic accompaniment for vocal melodies
WO2006118405A1 (en) Internet music composition application with pattern-combination method
US20040244565A1 (en) Method of creating music file with main melody and accompaniment
US7053291B1 (en) Computerized system and method for building musical licks and melodies
Butler Unlocking the groove: Rhythm, meter, and musical design in electronic dance music
JP4182613B2 (en) Karaoke equipment
JP4447524B2 (en) Karaoke equipment characterized by medley music selection processing with uniform tempo
JP4179063B2 (en) Performance setting data selection device and program
JPS58220189A (en) Automatic performer
US20040206227A1 (en) Method of playing a game according to events in a selected track of a music file
JP2940449B2 (en) Automatic performance device
JPH04168493A (en) Electronic musical sound reproducing device
JP3752940B2 (en) Automatic composition method, automatic composition device and recording medium
JPS5995591A (en) Rom cartridge type electronic musical instrument
US6188009B1 (en) Electronic musical instrument with help function
KR100666010B1 (en) Arrangement system of a music using the internet, the method thereof
EP2793222B1 (en) Method for implementing an automatic music jam session
JP3820620B2 (en) Arrangement data storage device and arrangement performance device
JPH08314484A (en) Automatic playing device

Legal Events

Date Code Title Description
AS Assignment

Owner name: BENQ CORPORATION, TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:CHENG, WEN-NI;KUO, CHUN-BIN;REEL/FRAME:013709/0772

Effective date: 20030519

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION