US5308915A - Electronic musical instrument utilizing neural net - Google Patents

Electronic musical instrument utilizing neural net Download PDF

Info

Publication number
US5308915A
US5308915A US07/779,110 US77911091A US5308915A US 5308915 A US5308915 A US 5308915A US 77911091 A US77911091 A US 77911091A US 5308915 A US5308915 A US 5308915A
Authority
US
United States
Prior art keywords
neural net
parameter
output
net device
musical
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
US07/779,110
Inventor
Kenichi Ohya
Hirofumi Mukaino
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yamaha Corp
Original Assignee
Yamaha Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from JP2282854A external-priority patent/JP2605477B2/en
Priority claimed from JP2293058A external-priority patent/JP2663705B2/en
Application filed by Yamaha Corp filed Critical Yamaha Corp
Assigned to YAMAHA CORPORATION A CORP. OF JAPAN reassignment YAMAHA CORPORATION A CORP. OF JAPAN ASSIGNMENT OF ASSIGNORS INTEREST. Assignors: MUKAINO, HIROFUMI, OHYA, KENICHI
Application granted granted Critical
Publication of US5308915A publication Critical patent/US5308915A/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H7/00Instruments in which the tones are synthesised from a data store, e.g. computer organs
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2250/00Aspects of algorithms or signal processing methods without intrinsic musical character, yet specifically adapted for or used in electrophonic musical processing
    • G10H2250/311Neural networks for electrophonic musical instruments or musical processing, e.g. for musical recognition or control, automatic composition or improvisation
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S706/00Data processing: artificial intelligence
    • Y10S706/902Application using ai with detail of the ai system

Definitions

  • the present invention relates to an electronic musical instrument utilizing neural nets more particularly, to an electronic musical instrument to generate musical patterns, such as a rhythm pattern and a bass pattern, using a neural net.
  • the patterns to be generated are previously stored in a memory.
  • the pattern is read from the memory and supplied to a musical tone generating circuit.
  • the conventional electronic musical instruments have only had a memory to generate musical patterns, such as a rhythm pattern and a bass pattern, so that available patterns are limited. Therefore, the musical representations have been scanty.
  • an electronic musical instrument utilizing neural nets comprises parameter input means for inputting a parameter, a neural net device for utilizing the parameter inputted from the parameter input means with internal organization, and change means for changing output data from the neural net device into musical pattern signal.
  • the neural net device is in advance learning, therefore, any input parameter results in a proper output by interpolation.
  • FIG. 1 is a block diagram of a rhythm pattern generating instrument embodying the present invention.
  • FIG. 2 shows correlation between the first series' neurons and the rhythm pattern.
  • FIG. 3 shows correlation between the second series' neurons and the rhythm pattern.
  • FIG. 4 is a block diagram of another rhythm pattern generating instrument embodying the present invention.
  • FIG. 5 is a graph showing change of the rhythm pattern in use of random numbers generator.
  • FIG. 6 shows correlation between the neurons and the bass pattern.
  • FIG. 7 is a block diagram of a bas pattern generating instrument embodying the present invention.
  • rhythm pattern generating instrument embodying the present invention is disclosed in detail as follows.
  • This rhythm pattern generating instrument is provided with a parameter designation operator 1, a normalization part 2, a neural net 3, weighting data memory 4 for storing various weighting data, a weighting data selector 5 for selecting weighting data in the weighting data memory 4, an interpreter 6, an interpretation knowledge memory 7 for storing various interpretation knowledge, an interpretation knowledge selector 8 for selecting interpretation knowledge, an output modifier 9, a modification knowledge memory 10 for storing various modification knowledge, a modification knowledge selector 11 for selecting modification knowledge, a musical playing data synthesizer 12, a key code designation switch 13, and a musical playing part 14.
  • the parameter designation operator 1 has four volumes, each of which sets a musical parameter.
  • the musical parameters depend on the learning mode of the neural net 3. In the learning mode previously performed, a plurality of data sets of parameters and output data are supplied successively to the neural net 3. It is unnecessary to give basic musical sense to the parameters in the learning mode of the neural net 3.
  • the learning process is usually carried out by use of a back-propagation method.
  • the parameter designation operator includes an analog-digital converter to output digital values.
  • the normalization part 2 normalizes the output of the parameter designation operator 1 to use it as input data to the neural net 3.
  • the normalized data is given to each neuron of an input layer of the neural net 3.
  • the neural net 3 consists of three layers, the input layer, a middle layer, and an output layer. Each neuron of the layers is combined with an adjacent neuron at a certain weighting factor.
  • the number of the neurons of the input layer is equal to the number of parameters of the parameter designation operator 1.
  • the number of the neurons of the middle layer is decided depending on a degree of the learning. In this example, the number of the neurons of the middle layer is twenty.
  • the weighting data memory 4 stores a plurality of weighting data to comply with different music genre.
  • the weighting data selector 5 is operated by the performer to select a weighting data in the memory 4.
  • the interpreter 6 is used to interpret the value output from the neural net 3, thereby changing the value to musical feeling data, using the interpretation knowledge stated later.
  • each output neuron is independent, not combined with the other output neurons.
  • the interpretation knowledge stored in the interpretation knowledge memory 7 is used to realize adjustment to some musical genres.
  • a plurality of sets of the interpretation knowledge is stored in the memory 7.
  • the interpretation knowledge selector 8 is used to select one interpretation knowledge from the memory 7.
  • the selector is operated by a performer.
  • the output modifier 9 is used to modify the output value of the interpreter 6 using the modification knowledge in the memory 10, thereby changing musically unacceptable values to musically acceptable values.
  • the modification knowledge stored in the modification knowledge memory 10 is used to realize adjustment to the musical genres.
  • a plurality of modification knowledge sets is stored in the memory 10.
  • the musical playing data synthesizer 12 generates the actual musical playing data according to the output data from the output modifier 9.
  • the key code designation switch 13 is used to assign a key code to the output data from the output modifier 9.
  • the musical playing part 14 is an output device, such as an MIDI device, to actually output the musical playing data.
  • the performer arbitrarily inputs parameters in the parameter designation operator 1 by use of the four volumes.
  • the learning mode is, at this time, already carried out by the neural net 3.
  • the performer operates the weighting data selector 5 to input any weighting data in the memory 4 to the neural net 3 so as to match the output rhythm pattern to the playing song played by another instrument.
  • the performer also operates the interpretation knowledge selector 8 to input any interpretation knowledge in the memory 7 to the interpreter 6 so as to match the output rhythm pattern to the playing data played by another instrument.
  • the performer operates the modification knowledge selector 11 to input any modification knowledge in the memory 10 to the output modifier 9 so as to match the output rhythm pattern to the playing data played by another instrument.
  • the parameters inputted by the parameter designation operator 1 are normalized with the normalization part 2, and transferred to the input neurons of the input layer in the neural net 3.
  • the values of the neurons of the middle layer are calculated using the weighting data specified by the selector 5.
  • the values of the neurons of the output layer are calculated using the values of the middle layer. ##EQU1##
  • the values of the output neurons are modified using the selected interpretation knowledge so as to have the musical sense.
  • the output neurons of the output layer are interpreted on time series. In the example, a set of sixteen neurons beginning from the first neuron in the output layer forms the first series, the remainder of another sixteen neurons forming the second series. Output data of each neuron corresponds to a sixteenth note.
  • the value of the neurons should be a real number of 0 to 1. However, the real number value is changed to integer value of 0 to 127 for convenience of calculation.
  • the output neurons' values are interpreted as follows:
  • the first series (see FIG. 2):
  • the velocity of the hi-hat, snare drum, and each tom are decided according to the neuron's value.
  • FIG. 2 and FIG. 3 show correspondences between the first series and the rhythm pattern, and the second series and the rhythm pattern, respectively.
  • the numbers 0 to 31 represent the output neuron's number.
  • the interpreted data (value) output from the interpreter 6 is modified to the value which can be musically accepted using the selected modification knowledge in the memory 10. For example, if a tone is generated at the timing corresponding to the back beat of sixteenth beat in an eight beat music score, the modification is done so that the back beat is released. Furthermore, If the hi-hat will keep open state after interpretation, the hi-hat is closed without open.
  • the modified data output from the output modifier 9 is represented with velocity value of the tone color (i.e., an instrument name, such as hi-hat, bass drum).
  • the key code switch 13 gives a key code of the tone color to the synthesizer 12 to change it to the musical playing data which can be actually performed.
  • the musical playing part 14 receives the data from the synthesizer 12 and performs the musical playing data.
  • adjusting the volume of the parameter designation operator 1 allows various rhythm patterns to be outputted.
  • FIG. 4 is another example of the present invention.
  • the rhythm pattern generating instrument shown in FIG. 4 is provided with a group of random numbers generators 1a.
  • the rhythm pattern generating instrument differs from the example shown in FIG. 1 in that this instrument is provided with a group of random numbers generators 1a, a random numbers selector 1b for selecting the random numbers generator, a previous parameter memory 1c, and an adder 1d.
  • the group of random numbers generators 1a is configured with a plurality of random numbers generator each of which generates digital random numbers with different distribution.
  • the random numbers selector 1b is provided for selecting one random numbers generator in the group of the random numbers generators 1a.
  • the previous parameter memory 1c stores previously used parameters which were used as input data to the neural net 3.
  • the adder 1d is used to add the value of the previous parameter memory 1c to the output value of the random numbers selector 1b to form a new parameter. This new parameter is stored into the previous parameter memory 1c as a previous parameter for the next time.
  • the normalization part 2 and the other parts, such as the neural net 3, and weighting data memory 4, are the same as the instrument in FIG. 1.
  • a performer arbitrarily inputs initial parameters into the previous parameter memory 1c, and then, selects a random numbers generator to get rhythm patterns changing as expected.
  • the performer operates the weighting data selector 5 to input any weighting data in the memory 4 to the neural net 3 so as to match the output rhythm pattern to the playing song played by another instrument.
  • the performer also operates the interpretation knowledge selector 8 to input any interpretation knowledge in the memory 7 to the interpreter 6 so as to match the output rhythm pattern to the playing data played by another instrument.
  • the performer operates the modification knowledge selector 11 to input any modification knowledge in the memory 10 to the output modifier 9 so as to match the output rhythm pattern to the playing data played by another instrument.
  • the output of the adder in which the value of the previous parameter memory 1c is added to the output value of the random numbers selector 1b is fed to the normalization part 2 to normalize it.
  • the output is also supplied to the previous parameter memory 1c to be stored as a previous parameter for next time. Therefore, if the selected random numbers generator distributes numbers non-uniformly, the parameter outputted from the adder 1d is shifted gradually from the first parameter given by the performer.
  • rhythm output pattern once initialized, is automatically changed without a performer's operation because the input patterns change with the random numbers, so that various trends of the rhythm patterns can be made using random numbers having different characteristics (distribution). For example, if random numbers distributed between -4 and +3 are added successively to the previous parameter in the memory 1c for every bar, parameter (number) is gradually decreased. In experiment, the parameter approximately decides a property of rhythm as follows,
  • the parameter is gradually decreased, and then the rhythm pattern becomes a less tones pattern, the phenomenon of the rhythm pattern's change images a performer who is tired from playing a drum. While, if the parameter reaches "0" and underflow occurs, the parameter is increased gradually, the phenomenon of the rhythm pattern change images a performer who is getting well.
  • FIG. 5 shows this state. If another type of random numbers is used, another pattern characteristic is obtained.
  • the neural net 3 is learned so that the input parameters correspond to bass patterns, and the other elements, such as the interpretation knowledge, are properly configured.
  • FIG. 6 shows a score according to the abovelisted correlation.
  • the output modifier 9 is used to modify the output data from the interpreter 6 so as to be musically accepted. For example, any discordant tone is deleted or modified, and the rhythm is modified.
  • FIG. 7 shows a block diagram of the above-mentioned example. In this diagram, the chord designation switch is different from the switch in FIG. 1.
  • the key code can be inputted in a real time mode from the key code designation switch 13.
  • one bar of four beats is divided into sixteen, a bass tone being outputted at each timing. It is possible to generate bass patterns fully musical in an instrument arranged so as to be able to output bass tones for two bars, i.e., at each timing of thirty two timings.
  • the neural net 3 not only plays back the learned patterns, but also generates middle patterns between two learned patterns, resulting in output patterns that give variety. Also, selecting the weighting data makes variation of velocity or the like, so that generated patterns give variety of rising and falling pitch, diminuendo and crescendo, and so on.

Abstract

An electronic musical instrument utilizing neural nets comprises parameter input device for inputting a parameter and a neural net device for calculating the parameter. The neural net device is in advance learning therefore, any input parameter results in proper output by interpolation. The instrument further comprises a weighting data memory, the weighting data being provided to the neural net device. The output of the neural net device is interpreted by an interpreter using interpretation knowledge stored in a memory, thereby the output of the neural net device being changed to musical values. Further, the musical values are modified by an output modifier using modification knowledge stored in another memory so as to be accepted musically. The weighting data, the interpretation knowledge and the modification knowledge can be selected by use of selectors, thus use of the selectors expands musical variation.

Description

BACKGROUND OF THE INVENTION
1. Field of the Invention
The present invention relates to an electronic musical instrument utilizing neural nets more particularly, to an electronic musical instrument to generate musical patterns, such as a rhythm pattern and a bass pattern, using a neural net.
2. Description of the Prior Art
In a conventional electronic musical instrument having the function of an automatic rhythm pattern generation or an automatic accompaniment pattern generation, the patterns to be generated are previously stored in a memory. When any pattern is selected by a performer, the pattern is read from the memory and supplied to a musical tone generating circuit.
As described above, the conventional electronic musical instruments have only had a memory to generate musical patterns, such as a rhythm pattern and a bass pattern, so that available patterns are limited. Therefore, the musical representations have been scanty.
SUMMARY OF THE INVENTION
It is therefore an object of the present invention to provide an electronic musical instrument which allows itself to generate more musical patterns by use of a neural net.
In accordance with the present invention, an electronic musical instrument utilizing neural nets comprises parameter input means for inputting a parameter, a neural net device for utilizing the parameter inputted from the parameter input means with internal organization, and change means for changing output data from the neural net device into musical pattern signal.
In the above-mentioned instrument, the neural net device is in advance learning, therefore, any input parameter results in a proper output by interpolation.
BRIEF DESCRIPTION OF THE DRAWINGS
FIG. 1 is a block diagram of a rhythm pattern generating instrument embodying the present invention.
FIG. 2 shows correlation between the first series' neurons and the rhythm pattern.
FIG. 3 shows correlation between the second series' neurons and the rhythm pattern.
FIG. 4 is a block diagram of another rhythm pattern generating instrument embodying the present invention.
FIG. 5 is a graph showing change of the rhythm pattern in use of random numbers generator.
FIG. 6 shows correlation between the neurons and the bass pattern.
FIG. 7 is a block diagram of a bas pattern generating instrument embodying the present invention.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
Referring to the drawings, a rhythm pattern generating instrument embodying the present invention is disclosed in detail as follows.
This rhythm pattern generating instrument is provided with a parameter designation operator 1, a normalization part 2, a neural net 3, weighting data memory 4 for storing various weighting data, a weighting data selector 5 for selecting weighting data in the weighting data memory 4, an interpreter 6, an interpretation knowledge memory 7 for storing various interpretation knowledge, an interpretation knowledge selector 8 for selecting interpretation knowledge, an output modifier 9, a modification knowledge memory 10 for storing various modification knowledge, a modification knowledge selector 11 for selecting modification knowledge, a musical playing data synthesizer 12, a key code designation switch 13, and a musical playing part 14.
The parameter designation operator 1 has four volumes, each of which sets a musical parameter. The musical parameters depend on the learning mode of the neural net 3. In the learning mode previously performed, a plurality of data sets of parameters and output data are supplied successively to the neural net 3. It is unnecessary to give basic musical sense to the parameters in the learning mode of the neural net 3. The learning process is usually carried out by use of a back-propagation method.
The parameter designation operator includes an analog-digital converter to output digital values.
The normalization part 2 normalizes the output of the parameter designation operator 1 to use it as input data to the neural net 3. The normalized data is given to each neuron of an input layer of the neural net 3.
The neural net 3 consists of three layers, the input layer, a middle layer, and an output layer. Each neuron of the layers is combined with an adjacent neuron at a certain weighting factor. The number of the neurons of the input layer is equal to the number of parameters of the parameter designation operator 1. The number of the neurons of the middle layer is decided depending on a degree of the learning. In this example, the number of the neurons of the middle layer is twenty.
The number of the neurons of the output layer is decided depending on time resolution of the neural net 3. In the case of M bars output at Nth note notch, the number of notes is equal to N*M per one channel. In the example, notes of a bass drum tone color and a hi-hat tone color are generated at the first channel: notes of a snare drum tone color and a tom tom tone color are generated at the second channel. As time resolution is in a notes pattern of one bar with sixteenth note are generated, the output layer needs 32 neurons (16*1*2(channels)=32).
The weighting data memory 4 stores a plurality of weighting data to comply with different music genre.
The weighting data selector 5 is operated by the performer to select a weighting data in the memory 4.
The interpreter 6 is used to interpret the value output from the neural net 3, thereby changing the value to musical feeling data, using the interpretation knowledge stated later. In this example, each output neuron is independent, not combined with the other output neurons.
The interpretation knowledge stored in the interpretation knowledge memory 7 is used to realize adjustment to some musical genres. In this example, a plurality of sets of the interpretation knowledge is stored in the memory 7.
The interpretation knowledge selector 8 is used to select one interpretation knowledge from the memory 7. The selector is operated by a performer.
The output modifier 9 is used to modify the output value of the interpreter 6 using the modification knowledge in the memory 10, thereby changing musically unacceptable values to musically acceptable values.
The modification knowledge stored in the modification knowledge memory 10 is used to realize adjustment to the musical genres. In this example, a plurality of modification knowledge sets is stored in the memory 10.
The musical playing data synthesizer 12 generates the actual musical playing data according to the output data from the output modifier 9.
The key code designation switch 13 is used to assign a key code to the output data from the output modifier 9.
The musical playing part 14 is an output device, such as an MIDI device, to actually output the musical playing data.
The following is a description of the operation of the above-mentioned rhythm pattern generating instrument.
Step 1--Initializing
The performer arbitrarily inputs parameters in the parameter designation operator 1 by use of the four volumes. Of course, the learning mode is, at this time, already carried out by the neural net 3.
The performer operates the weighting data selector 5 to input any weighting data in the memory 4 to the neural net 3 so as to match the output rhythm pattern to the playing song played by another instrument. The performer also operates the interpretation knowledge selector 8 to input any interpretation knowledge in the memory 7 to the interpreter 6 so as to match the output rhythm pattern to the playing data played by another instrument. Further, the performer operates the modification knowledge selector 11 to input any modification knowledge in the memory 10 to the output modifier 9 so as to match the output rhythm pattern to the playing data played by another instrument.
Step 2--Input to neural net
The parameters inputted by the parameter designation operator 1 are normalized with the normalization part 2, and transferred to the input neurons of the input layer in the neural net 3.
Step 3--Calculating in neural net
First, the values of the neurons of the middle layer are calculated using the weighting data specified by the selector 5. Next, the values of the neurons of the output layer are calculated using the values of the middle layer. ##EQU1##
Step 4--Interpretation of output neurons
The values of the output neurons are modified using the selected interpretation knowledge so as to have the musical sense. The output neurons of the output layer are interpreted on time series. In the example, a set of sixteen neurons beginning from the first neuron in the output layer forms the first series, the remainder of another sixteen neurons forming the second series. Output data of each neuron corresponds to a sixteenth note. In neural net theory, the value of the neurons should be a real number of 0 to 1. However, the real number value is changed to integer value of 0 to 127 for convenience of calculation.
For example, the output neurons' values are interpreted as follows:
The first series (see FIG. 2):
______________________________________                                    
Output neurons' value:                                                    
                   Interpreted value:                                     
______________________________________                                    
0 to 5             ungenerating tone                                      
 6 to 31           hi-hat-close                                           
32 to 56           hi-hat-open                                            
57 to 63           bass drum (weak)                                       
64 to 69           bass drum (strong)                                     
70 to 95           hi-hat-close                                           
                   +bass drum (strong)                                    
 96 to 127         hi-hat-open                                            
                   +bass drum (strong)                                    
______________________________________                                    
The second series (see FIG. 3):
______________________________________                                    
Output neurons' value:                                                    
                    Interpreted value:                                    
______________________________________                                    
 0 to 18            ungenerating tone                                     
19 to 37            low tom                                               
38 to 41            snare drum (weak)                                     
42 to 60            middle tom                                            
61 to 64            snare drum (weak)                                     
65 to 83            high tom                                              
84 to 87            snare drum (weak)                                     
 88 to 127          snare drum (strong)                                   
______________________________________                                    
The velocity of the hi-hat, snare drum, and each tom are decided according to the neuron's value.
FIG. 2 and FIG. 3 show correspondences between the first series and the rhythm pattern, and the second series and the rhythm pattern, respectively. The numbers 0 to 31 represent the output neuron's number.
Step 5--Output modification
In this step, the interpreted data (value) output from the interpreter 6, is modified to the value which can be musically accepted using the selected modification knowledge in the memory 10. For example, if a tone is generated at the timing corresponding to the back beat of sixteenth beat in an eight beat music score, the modification is done so that the back beat is released. Furthermore, If the hi-hat will keep open state after interpretation, the hi-hat is closed without open.
Step 6--Synthesizing playing data
The modified data output from the output modifier 9 is represented with velocity value of the tone color (i.e., an instrument name, such as hi-hat, bass drum). The key code switch 13 gives a key code of the tone color to the synthesizer 12 to change it to the musical playing data which can be actually performed. The musical playing part 14 receives the data from the synthesizer 12 and performs the musical playing data.
As mentioned above, adjusting the volume of the parameter designation operator 1 allows various rhythm patterns to be outputted.
FIG. 4 is another example of the present invention.
The rhythm pattern generating instrument shown in FIG. 4 is provided with a group of random numbers generators 1a. The rhythm pattern generating instrument differs from the example shown in FIG. 1 in that this instrument is provided with a group of random numbers generators 1a, a random numbers selector 1b for selecting the random numbers generator, a previous parameter memory 1c, and an adder 1d.
The group of random numbers generators 1a is configured with a plurality of random numbers generator each of which generates digital random numbers with different distribution. The random numbers selector 1b is provided for selecting one random numbers generator in the group of the random numbers generators 1a. The previous parameter memory 1c stores previously used parameters which were used as input data to the neural net 3.
The adder 1d is used to add the value of the previous parameter memory 1c to the output value of the random numbers selector 1b to form a new parameter. This new parameter is stored into the previous parameter memory 1c as a previous parameter for the next time.
The normalization part 2 and the other parts, such as the neural net 3, and weighting data memory 4, are the same as the instrument in FIG. 1.
The following is a description of the process of the above-mentioned instrument.
Step 1--Initialization
A performer arbitrarily inputs initial parameters into the previous parameter memory 1c, and then, selects a random numbers generator to get rhythm patterns changing as expected.
The performer operates the weighting data selector 5 to input any weighting data in the memory 4 to the neural net 3 so as to match the output rhythm pattern to the playing song played by another instrument. The performer also operates the interpretation knowledge selector 8 to input any interpretation knowledge in the memory 7 to the interpreter 6 so as to match the output rhythm pattern to the playing data played by another instrument. Further, the performer operates the modification knowledge selector 11 to input any modification knowledge in the memory 10 to the output modifier 9 so as to match the output rhythm pattern to the playing data played by another instrument.
Step 2--Input to neural net
The output of the adder in which the value of the previous parameter memory 1c is added to the output value of the random numbers selector 1b is fed to the normalization part 2 to normalize it. The output is also supplied to the previous parameter memory 1c to be stored as a previous parameter for next time. Therefore, if the selected random numbers generator distributes numbers non-uniformly, the parameter outputted from the adder 1d is shifted gradually from the first parameter given by the performer.
The process in the neural net 3 and the other processes in the example are the same previously stated in step 3 to 6.
This example is characterized in that the rhythm output pattern, once initialized, is automatically changed without a performer's operation because the input patterns change with the random numbers, so that various trends of the rhythm patterns can be made using random numbers having different characteristics (distribution). For example, if random numbers distributed between -4 and +3 are added successively to the previous parameter in the memory 1c for every bar, parameter (number) is gradually decreased. In experiment, the parameter approximately decides a property of rhythm as follows,
______________________________________                                    
parameter (number)  rhythm                                                
______________________________________                                    
0 to 40             eight beats                                           
50 to 70            sixteen beats                                         
80 to 100           sixteen back beats                                    
______________________________________                                    
so that if the process is advanced using the random number distributed between -4 and +3, i.e., the random number offset to a minus, after "100" is stored in the memory 1c as the first parameter, the parameter is gradually decreased, and then the rhythm pattern becomes a less tones pattern, the phenomenon of the rhythm pattern's change images a performer who is tired from playing a drum. While, if the parameter reaches "0" and underflow occurs, the parameter is increased gradually, the phenomenon of the rhythm pattern change images a performer who is getting well. FIG. 5 shows this state. If another type of random numbers is used, another pattern characteristic is obtained.
As another example, it is possible to input parameters for outputting a bass pattern from the parameter designation operator 1. In this case, the neural net 3 is learned so that the input parameters correspond to bass patterns, and the other elements, such as the interpretation knowledge, are properly configured.
The correlation between the output value of the neural net 3 and the bass tone is as follows:
______________________________________                                    
Output neuron's value (0 to 1):                                           
                    Bass tone:                                            
______________________________________                                    
0.00 to 0.35        ungenerating tone (keep                               
                    previous tone)                                        
0.35 to 0.45        root tone (C)                                         
0.45 to 0.55        third tone (E)                                        
0.55 to 0.65        fourth tone (F)                                       
0.65 to 0.75        fifth tone (G)                                        
0.75 to 0.85        sixth tone (A)                                        
0.85 to 0.95        seventh tone (B)                                      
0.95 to 1.0         octave (C)                                            
______________________________________                                    
FIG. 6 shows a score according to the abovelisted correlation. The output modifier 9 is used to modify the output data from the interpreter 6 so as to be musically accepted. For example, any discordant tone is deleted or modified, and the rhythm is modified.
As the bass pattern outputted from the output modifier 9 is represented with an interval from a root tone of a chord, or with a tone pitch in "C" chord, it is necessary to change the tone pitch of the bass pattern according to the chord progress of music. This change performance is carried out by the musical playing data synthesizer 12 and a chord designation switch 13. FIG. 7 shows a block diagram of the above-mentioned example. In this diagram, the chord designation switch is different from the switch in FIG. 1.
Even if there is the same image music, an ideal bass pattern or a rule of the music is different depending on the music type. Therefore, the weighting data, the interpretation knowledge, and the modification knowledge are manually or automatically switched according to the music type. The key code can be inputted in a real time mode from the key code designation switch 13.
In this example, one bar of four beats is divided into sixteen, a bass tone being outputted at each timing. It is possible to generate bass patterns fully musical in an instrument arranged so as to be able to output bass tones for two bars, i.e., at each timing of thirty two timings.
As mentioned above, the neural net 3 not only plays back the learned patterns, but also generates middle patterns between two learned patterns, resulting in output patterns that give variety. Also, selecting the weighting data makes variation of velocity or the like, so that generated patterns give variety of rising and falling pitch, diminuendo and crescendo, and so on.

Claims (33)

What is claimed is:
1. An electronic musical instrument utilizing neural nets, comprising:
parameter input means for providing a parameter;
a neural net device for utilizing said parameter provided by said parameter input means, said neural net device having a layer of output neurons for providing output data;
change means for changing said output data from said neural net device into a musical pattern signal, said change means including:
a. interpretation knowledge memory means for storing interpretation knowledge for interpreting said output data from said neural net device;
b. interpreter means for interpreting said output data from said neural net device as musical values using said interpretation knowledge;
c. modification knowledge memory means for storing modification knowledge for modifying said musical values from said interpreter means; and
d. output modification means using said modification knowledge for modifying said musical vales from said interpreter means so as to be musically acceptable.
2. An electronic musical instrument utilizing neural nets according to claim 1, wherein said change means comprises modification knowledge selecting means for selecting modification knowledge in said modification knowledge memory.
3. An electronic musical instrument utilizing neural nets according to claim 1, wherein said change means comprises a musical playing data synthesizer for synthesizing actual musical playing data based on data output from said output modification means.
4. An electronic musical instrument utilizing neural nets according to claim 3, wherein said change means comprises a key code designation switch for providing to said musical playing data synthesizer a key code designated by a performer.
5. An electronic musical instrument utilizing neural nets according to claim 3, wherein said change means comprises a chord designation switch for providing to said musical playing data synthesizer a chord designated by a performer.
6. An electronic musical instrument utilizing neural nets, comprising:
parameter input means for providing a parameter;
a neural net device for utilizing said parameter provided by said parameter input means, said neural net device having a layer of output neurons for providing output data;
change means for changing said output data from said neural net device into a musical pattern signal, said change means including:
a. a weighting data memory for storing weighting data to be fed to said neural net device as a weighting factor for weighting said output data as musical values of said neural net;
b. weighting data selecting means for selecting said weighting data in said weighting data memory;
c. a modification knowledge memory for storing modification knowledge for modifying said musical values from said neural net; and
d. output modification means using said modification knowledge for modifying said musical vales from said neural net so as to be musically acceptable.
7. An electronic musical instrument utilizing neural nets, comprising:
parameter input means for providing a parameter;
neural net device for utilizing the parameter provided by said parameter input means, said neural net device having a layer of output neurons for providing output data;
change means for changing said output data from said neural net device into a musical pattern signal; and
modification means for storing modification knowledge to modify said musical pattern signal, and for modifying said musical pattern signal using said modification knowledge so as to be musically acceptable.
8. An electronic musical instrument utilizing neural nets according to claim 7, wherein said musical pattern signal is a rhythm pattern signal.
9. An electronic musical instrument utilizing neural nets according to claim 7, wherein said musical pattern signal is a bass pattern signal.
10. An electronic musical instrument utilizing neural nets according to claim 7, further comprising normalization means for normalizing said parameter provided by said parameter input means.
11. An electronic musical instrument utilizing neural nets according to claim 7, further comprising a weighting data memory for storing weighting data to be fed to said neural net device as a weighting factor of said neural net.
12. An electronic musical instrument utilizing neural nets according to claim 11, further comprising weighting data selecting means for selecting said weighting data in said weighting data memory.
13. An electronic musical instrument utilizing neural nets according to claim 7, wherein said change means comprises interpretation knowledge memory means for storing interpretation knowledge for interpreting said output data from said neural net device, and interpreter means for interpreting said output data from said neural net device as musical values using said interpretation knowledge.
14. An electronic musical instrument utilizing neural nets according to claim 13, wherein said change means comprises interpretation knowledge selecting means for selecting said interpretation knowledge in said interpretation knowledge memory means.
15. An electronic musical instrument utilizing neural nets, comprising:
parameter input means for providing a parameter, said parameter input means including a selectable random numbers generator for generating a random number as said parameter;
a neural net device for utilizing the parameter provided by said parameter input means, said neural net device having a layer of output neurons for providing output data; and
change means for changing said output data from said neural net device into a musical pattern signal.
16. An electronic musical instrument utilizing neural nets according to claim 15, further comprising previous parameter memory means for storing a previous parameter, and adder means for adding said previous parameter in said previous parameter memory to a random number outputed from said random numbers generator, said adder having adder output being supplied to said neural net device and said previous parameter memory means to be stored as a next previous parameter.
17. An electronic musical instrument utilizing neural nets according to claim 15, wherein said random numbers generator generates random numbers with an offset distribution having a minus or a plus direction.
18. An electronic musical instrument utilizing neural nets, comprising:
parameter input means for providing a parameter;
a neural net device for utilizing the parameter provided by said parameter input means, said neural net device having a layer of output neurons for providing output data, said layer of output neurons comprising a plurality of output layer neurons respectively corresponding to tone generation timings of a set of musical tones to be generated;
change means for changing said output data from said neural net device into a musical pattern signal; and
modification means for storing modification knowledge to modify said musical pattern signal, and for modifying said musical pattern signal using said modification knowledge so as to be musically acceptable.
19. An electronic musical instrument utilizing neural nets according to claim 18, wherein said output data of said plurality of neurons respectively represent tone colors of a set of musical tones to be generated.
20. An electronic musical instrument utilizing neural nets according to claim 18, wherein said output data of said plurality of neurons respectively represent tone pitches of a set of bass tones to be generated.
21. An electronic musical instrument utilizing neural nets according to claim 13, wherein said interpretation knowledge stored in said interpretation knowledge memory means indicates correspondence between a plurality of neurons in said neural net device and tone colors of a set of musical tones to be generated.
22. A method using an electronic musical instrument utilizing neural nets, comprising the steps of:
providing a parameter;
utilizing said provided parameter in a neural net device having a layer of output neurons;
providing output data from said layer of output neurons;
changing said output data from said neural net device into a musical pattern signal;
storing modification knowledge to modify said musical pattern signal; and
modifying said musical pattern signal using said stored modification knowledge so as to be musically acceptable.
23. A method according to claim 22, wherein said musical pattern signal is a rhythm pattern signal.
24. A method according to claim 22, wherein said musical pattern signal is a bass pattern signal.
25. A method according to claim 22, further comprising the step of normalizing said provided parameter.
26. A method according to claim 22, further comprising the step of storing weighting data to be fed to said neural net device as a weighting factor of said neural net device.
27. A method according to claim 26, further comprising the step of selecting said stored weighting data.
28. A method according to claim 22, further comprises the steps of:
storing interpretation knowledge for interpreting said output data from said neural net device; and
interpreting said output data from said neural net device as musical values using said stored interpretation knowledge.
29. A method according to claim 28, further comprising the step of selecting said interpretation knowledge from said stored interpretation knowledge.
30. A method according to claim 28, wherein said stored interpretation knowledge indicates a correspondence between a plurality of neurons in said neural net device and tone colors of a set of musical tones to be generated.
31. A method of using an electronic musical instrument utilizing neural nets, comprising the steps of:
selectably generating a random number as a parameter;
utilizing said parameter in a neural net device having a layer of output neurons for providing output data;
producing output data from said layer of output neurons; and
changing said output data from said neural net device into a musical pattern signal.
32. A method according to claim 31, further comprising the steps of:
adding a stored previous parameter to said selectably generated random number to produce an adder output;
supplying said adder output as said parameter used by said neural net device; and
storing said adder output as said previous parameter.
33. A method according to claim 31, further comprising the step of offsetting, in a minus or a plus direction, a distribution of said selectably generated random number.
US07/779,110 1990-10-19 1991-10-18 Electronic musical instrument utilizing neural net Expired - Fee Related US5308915A (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
JP2282854A JP2605477B2 (en) 1990-10-19 1990-10-19 Base pattern generation device and base pattern generation method
JP2-282854 1990-10-19
JP2293058A JP2663705B2 (en) 1990-10-29 1990-10-29 Rhythm pattern generator
JP2-293058 1990-10-29

Publications (1)

Publication Number Publication Date
US5308915A true US5308915A (en) 1994-05-03

Family

ID=26554803

Family Applications (1)

Application Number Title Priority Date Filing Date
US07/779,110 Expired - Fee Related US5308915A (en) 1990-10-19 1991-10-18 Electronic musical instrument utilizing neural net

Country Status (1)

Country Link
US (1) US5308915A (en)

Cited By (25)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5486646A (en) * 1992-01-16 1996-01-23 Roland Corporation Rhythm creating system for creating a rhythm pattern from specifying input data
US5495073A (en) * 1992-05-18 1996-02-27 Yamaha Corporation Automatic performance device having a function of changing performance data during performance
DE4430628A1 (en) * 1994-08-29 1996-03-14 Hoehn Marcus Dipl Wirtsch Ing Intelligent music accompaniment synthesis method with learning capability
US5541356A (en) * 1993-04-09 1996-07-30 Yamaha Corporation Electronic musical tone controller with fuzzy processing
US5581658A (en) * 1993-12-14 1996-12-03 Infobase Systems, Inc. Adaptive system for broadcast program identification and reporting
WO1997015914A1 (en) * 1995-10-23 1997-05-01 The Regents Of The University Of California Control structure for sound synthesis
US5696883A (en) * 1992-01-24 1997-12-09 Mitsubishi Denki Kabushiki Kaisha Neural network expressing apparatus including refresh of stored synapse load value information
US5736666A (en) * 1996-03-20 1998-04-07 California Institute Of Technology Music composition
US5824937A (en) * 1993-12-18 1998-10-20 Yamaha Corporation Signal analysis device having at least one stretched string and one pickup
US5850051A (en) * 1996-08-15 1998-12-15 Yamaha Corporation Method and apparatus for creating an automatic accompaniment pattern on the basis of analytic parameters
US6292791B1 (en) * 1998-02-27 2001-09-18 Industrial Technology Research Institute Method and apparatus of synthesizing plucked string instruments using recurrent neural networks
US20040089138A1 (en) * 2002-11-12 2004-05-13 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070051229A1 (en) * 2002-01-04 2007-03-08 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070071205A1 (en) * 2002-01-04 2007-03-29 Loudermilk Alan R Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070075971A1 (en) * 2005-10-05 2007-04-05 Samsung Electronics Co., Ltd. Remote controller, image processing apparatus, and imaging system comprising the same
US20070116299A1 (en) * 2005-11-01 2007-05-24 Vesco Oil Corporation Audio-visual point-of-sale presentation system and method directed toward vehicle occupant
US20070186752A1 (en) * 2002-11-12 2007-08-16 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070227338A1 (en) * 1999-10-19 2007-10-04 Alain Georges Interactive digital music recorder and player
US20070280270A1 (en) * 2004-03-11 2007-12-06 Pauli Laine Autonomous Musical Output Using a Mutually Inhibited Neuronal Network
US20080156178A1 (en) * 2002-11-12 2008-07-03 Madwares Ltd. Systems and Methods for Portable Audio Synthesis
US20090272251A1 (en) * 2002-11-12 2009-11-05 Alain Georges Systems and methods for portable audio synthesis
US20170103740A1 (en) * 2015-10-12 2017-04-13 International Business Machines Corporation Cognitive music engine using unsupervised learning
US9818386B2 (en) 1999-10-19 2017-11-14 Medialab Solutions Corp. Interactive digital music recorder and player
EP3857538A4 (en) * 2018-09-25 2022-06-22 Reactional Music Group AB Real-time music generation engine for interactive systems
US11842710B2 (en) 2021-03-31 2023-12-12 DAACI Limited Generative composition using form atom heuristics

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4941122A (en) * 1989-01-12 1990-07-10 Recognition Equipment Incorp. Neural network image processing system
US4953099A (en) * 1988-06-07 1990-08-28 Massachusetts Institute Of Technology Information discrimination cell
US5033006A (en) * 1989-03-13 1991-07-16 Sharp Kabushiki Kaisha Self-extending neural-network
US5138928A (en) * 1989-07-21 1992-08-18 Fujitsu Limited Rhythm pattern learning apparatus
US5138924A (en) * 1989-08-10 1992-08-18 Yamaha Corporation Electronic musical instrument utilizing a neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4953099A (en) * 1988-06-07 1990-08-28 Massachusetts Institute Of Technology Information discrimination cell
US4941122A (en) * 1989-01-12 1990-07-10 Recognition Equipment Incorp. Neural network image processing system
US5033006A (en) * 1989-03-13 1991-07-16 Sharp Kabushiki Kaisha Self-extending neural-network
US5138928A (en) * 1989-07-21 1992-08-18 Fujitsu Limited Rhythm pattern learning apparatus
US5138924A (en) * 1989-08-10 1992-08-18 Yamaha Corporation Electronic musical instrument utilizing a neural network

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
"A Connectionist Approach to Algorithmic Composition", P. Todd, 13 Computer Music J. 27 (Winter 1989).
"The Representations of Pitch in a Neural Net Model of Chord Classification", B. Laden and D. Keefe, 13 Computer Music J. 12 (Winter 1989).
A Connectionist Approach to Algorithmic Composition , P. Todd, 13 Computer Music J. 27 (Winter 1989). *
Johnson, Margaret L. "Toward an Expert System for Expressive Musical Performance", Computer, Jul. 1991, pp. 30-34.
Johnson, Margaret L. Toward an Expert System for Expressive Musical Performance , Computer, Jul. 1991, pp. 30 34. *
The Representations of Pitch in a Neural Net Model of Chord Classification , B. Laden and D. Keefe, 13 Computer Music J. 12 (Winter 1989). *

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5486646A (en) * 1992-01-16 1996-01-23 Roland Corporation Rhythm creating system for creating a rhythm pattern from specifying input data
US5696883A (en) * 1992-01-24 1997-12-09 Mitsubishi Denki Kabushiki Kaisha Neural network expressing apparatus including refresh of stored synapse load value information
US5495073A (en) * 1992-05-18 1996-02-27 Yamaha Corporation Automatic performance device having a function of changing performance data during performance
US5541356A (en) * 1993-04-09 1996-07-30 Yamaha Corporation Electronic musical tone controller with fuzzy processing
US5581658A (en) * 1993-12-14 1996-12-03 Infobase Systems, Inc. Adaptive system for broadcast program identification and reporting
US5824937A (en) * 1993-12-18 1998-10-20 Yamaha Corporation Signal analysis device having at least one stretched string and one pickup
DE4430628A1 (en) * 1994-08-29 1996-03-14 Hoehn Marcus Dipl Wirtsch Ing Intelligent music accompaniment synthesis method with learning capability
DE4430628C2 (en) * 1994-08-29 1998-01-08 Hoehn Marcus Dipl Wirtsch Ing Process and setup of an intelligent, adaptable music accompaniment for electronic sound generators
WO1997015914A1 (en) * 1995-10-23 1997-05-01 The Regents Of The University Of California Control structure for sound synthesis
US5736666A (en) * 1996-03-20 1998-04-07 California Institute Of Technology Music composition
US5850051A (en) * 1996-08-15 1998-12-15 Yamaha Corporation Method and apparatus for creating an automatic accompaniment pattern on the basis of analytic parameters
US6292791B1 (en) * 1998-02-27 2001-09-18 Industrial Technology Research Institute Method and apparatus of synthesizing plucked string instruments using recurrent neural networks
US9818386B2 (en) 1999-10-19 2017-11-14 Medialab Solutions Corp. Interactive digital music recorder and player
US8704073B2 (en) 1999-10-19 2014-04-22 Medialab Solutions, Inc. Interactive digital music recorder and player
US20110197741A1 (en) * 1999-10-19 2011-08-18 Alain Georges Interactive digital music recorder and player
US20070227338A1 (en) * 1999-10-19 2007-10-04 Alain Georges Interactive digital music recorder and player
US7847178B2 (en) 1999-10-19 2010-12-07 Medialab Solutions Corp. Interactive digital music recorder and player
US7504576B2 (en) 1999-10-19 2009-03-17 Medilab Solutions Llc Method for automatically processing a melody with sychronized sound samples and midi events
US20090241760A1 (en) * 1999-10-19 2009-10-01 Alain Georges Interactive digital music recorder and player
US8989358B2 (en) 2002-01-04 2015-03-24 Medialab Solutions Corp. Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070051229A1 (en) * 2002-01-04 2007-03-08 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070071205A1 (en) * 2002-01-04 2007-03-29 Loudermilk Alan R Systems and methods for creating, modifying, interacting with and playing musical compositions
US8674206B2 (en) 2002-01-04 2014-03-18 Medialab Solutions Corp. Systems and methods for creating, modifying, interacting with and playing musical compositions
US20110192271A1 (en) * 2002-01-04 2011-08-11 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US7807916B2 (en) 2002-01-04 2010-10-05 Medialab Solutions Corp. Method for generating music with a website or software plug-in using seed parameter values
US20080156178A1 (en) * 2002-11-12 2008-07-03 Madwares Ltd. Systems and Methods for Portable Audio Synthesis
US6979767B2 (en) * 2002-11-12 2005-12-27 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US20090272251A1 (en) * 2002-11-12 2009-11-05 Alain Georges Systems and methods for portable audio synthesis
US20080053293A1 (en) * 2002-11-12 2008-03-06 Medialab Solutions Llc Systems and Methods for Creating, Modifying, Interacting With and Playing Musical Compositions
US7928310B2 (en) 2002-11-12 2011-04-19 MediaLab Solutions Inc. Systems and methods for portable audio synthesis
US20040089138A1 (en) * 2002-11-12 2004-05-13 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070186752A1 (en) * 2002-11-12 2007-08-16 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US8153878B2 (en) 2002-11-12 2012-04-10 Medialab Solutions, Corp. Systems and methods for creating, modifying, interacting with and playing musical compositions
US8247676B2 (en) 2002-11-12 2012-08-21 Medialab Solutions Corp. Methods for generating music using a transmitted/received music data file
US9065931B2 (en) 2002-11-12 2015-06-23 Medialab Solutions Corp. Systems and methods for portable audio synthesis
US7655855B2 (en) 2002-11-12 2010-02-02 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US20070280270A1 (en) * 2004-03-11 2007-12-06 Pauli Laine Autonomous Musical Output Using a Mutually Inhibited Neuronal Network
US20070075971A1 (en) * 2005-10-05 2007-04-05 Samsung Electronics Co., Ltd. Remote controller, image processing apparatus, and imaging system comprising the same
US20070116299A1 (en) * 2005-11-01 2007-05-24 Vesco Oil Corporation Audio-visual point-of-sale presentation system and method directed toward vehicle occupant
US20170103740A1 (en) * 2015-10-12 2017-04-13 International Business Machines Corporation Cognitive music engine using unsupervised learning
US9715870B2 (en) * 2015-10-12 2017-07-25 International Business Machines Corporation Cognitive music engine using unsupervised learning
US10360885B2 (en) 2015-10-12 2019-07-23 International Business Machines Corporation Cognitive music engine using unsupervised learning
US11562722B2 (en) 2015-10-12 2023-01-24 International Business Machines Corporation Cognitive music engine using unsupervised learning
EP3857538A4 (en) * 2018-09-25 2022-06-22 Reactional Music Group AB Real-time music generation engine for interactive systems
US11842710B2 (en) 2021-03-31 2023-12-12 DAACI Limited Generative composition using form atom heuristics
US11887568B2 (en) 2021-03-31 2024-01-30 DAACI Limited Generative composition with defined form atom heuristics

Similar Documents

Publication Publication Date Title
US5308915A (en) Electronic musical instrument utilizing neural net
US5033352A (en) Electronic musical instrument with frequency modulation
JPH02504432A (en) pitch control system
US4433601A (en) Orchestral accompaniment techniques
US4713996A (en) Automatic rhythm apparatus with tone level dependent timbres
US4682526A (en) Accompaniment note selection method
US5406022A (en) Method and system for producing stereophonic sound by varying the sound image in accordance with tone waveform data
US3637914A (en) Automatic rhythm sound producing device with volume control
US6946595B2 (en) Performance data processing and tone signal synthesizing methods and apparatus
US4616547A (en) Improviser circuit and technique for electronic musical instrument
US4685370A (en) Automatic rhythm playing apparatus having plurality of rhythm patterns for a rhythm sound
US5218157A (en) Auto-accompaniment instrument developing chord sequence based on inversion variations
US4440058A (en) Digital tone generation system with slot weighting of fixed width window functions
US4160404A (en) Electronic musical instrument
US4205577A (en) Implementation of multiple voices in an electronic musical instrument
US4215616A (en) Asynchronous tone generator
US5403967A (en) Electronic musical instrument having melody correction capabilities
US6657115B1 (en) Method for transforming chords
US5521327A (en) Method and apparatus for automatically producing alterable rhythm accompaniment using conversion tables
JPH06180588A (en) Electronic musical instrument
US4526079A (en) Automatic rhythm performance device for electronic musical instruments
US4178825A (en) Musical tone synthesizer for generating a marimba effect
US4183277A (en) Rhythm accent circuit
US4458572A (en) Tone color changes in an electronic musical instrument
JP2743919B2 (en) Rhythm pattern generator

Legal Events

Date Code Title Description
AS Assignment

Owner name: YAMAHA CORPORATION A CORP. OF JAPAN, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST.;ASSIGNORS:OHYA, KENICHI;MUKAINO, HIROFUMI;REEL/FRAME:005945/0912

Effective date: 19911126

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

REMI Maintenance fee reminder mailed
LAPS Lapse for failure to pay maintenance fees
STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20020503