WO2001086628A2 - Automated generation of sound sequences - Google Patents

Automated generation of sound sequences Download PDF

Info

Publication number
WO2001086628A2
WO2001086628A2 PCT/GB2001/001985 GB0101985W WO0186628A2 WO 2001086628 A2 WO2001086628 A2 WO 2001086628A2 GB 0101985 W GB0101985 W GB 0101985W WO 0186628 A2 WO0186628 A2 WO 0186628A2
Authority
WO
WIPO (PCT)
Prior art keywords
audio
server
clients
engine
client
Prior art date
Application number
PCT/GB2001/001985
Other languages
French (fr)
Other versions
WO2001086628A3 (en
Inventor
John Tim Cole
Murray Peter Cole
Jeremy Louis Leach
Paul Alexander Blampied
Nicholas John Barefoot
Original Assignee
Sseyo Limited
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from GB0010967A external-priority patent/GB0010967D0/en
Priority claimed from GB0010969A external-priority patent/GB0010969D0/en
Priority claimed from GB0011178A external-priority patent/GB0011178D0/en
Priority claimed from GB0022164A external-priority patent/GB0022164D0/en
Priority claimed from GB0030833A external-priority patent/GB0030833D0/en
Application filed by Sseyo Limited filed Critical Sseyo Limited
Priority to AU2001254939A priority Critical patent/AU2001254939A1/en
Publication of WO2001086628A2 publication Critical patent/WO2001086628A2/en
Publication of WO2001086628A3 publication Critical patent/WO2001086628A3/en

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0008Associated control or indicating means
    • G10H1/0025Automatic or semi-automatic music composition, e.g. producing random music, applying rules from music theory or modifying a musical piece
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H1/00Details of electrophonic musical instruments
    • G10H1/0033Recording/reproducing or transmission of music for electrophonic musical instruments
    • G10H1/0041Recording/reproducing or transmission of music for electrophonic musical instruments in coded form
    • G10H1/0058Transmission between separate instruments or between individual components of a musical system
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2210/00Aspects or methods of musical processing having intrinsic musical character, i.e. involving musical theory or musical parameters or relying on musical knowledge, as applied in electrophonic musical tools or instruments
    • G10H2210/021Background music, e.g. for video sequences, elevator music
    • G10H2210/026Background music, e.g. for video sequences, elevator music for games, e.g. videogames
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2230/00General physical, ergonomic or hardware implementation of electrophonic musical tools or instruments, e.g. shape or architecture
    • G10H2230/005Device type or category
    • G10H2230/015PDA [personal digital assistant] or palmtop computing devices used for musical purposes, e.g. portable music players, tablet computers, e-readers or smart phones in which mobile telephony functions need not be used
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/056MIDI or other note-oriented file format
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/011Files or data streams containing coded musical information, e.g. for transmission
    • G10H2240/046File format, i.e. specific or non-standard musical file format used in or adapted for electrophonic musical instruments, e.g. in wavetables
    • G10H2240/061MP3, i.e. MPEG-1 or MPEG-2 Audio Layer III, lossy audio compression
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/171Transmission of musical instrument data, control or status information; Transmission, remote access or control of music data for electrophonic musical instruments
    • G10H2240/281Protocol or standard connector for transmission of analog or digital data to or from an electrophonic musical instrument
    • G10H2240/295Packet switched network, e.g. token ring
    • G10H2240/305Internet or TCP/IP protocol use for any electrophonic musical instrument data or musical parameter transmission purposes
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10HELECTROPHONIC MUSICAL INSTRUMENTS; INSTRUMENTS IN WHICH THE TONES ARE GENERATED BY ELECTROMECHANICAL MEANS OR ELECTRONIC GENERATORS, OR IN WHICH THE TONES ARE SYNTHESISED FROM A DATA STORE
    • G10H2240/00Data organisation or data communication aspects, specifically adapted for electrophonic musical tools or instruments
    • G10H2240/171Transmission of musical instrument data, control or status information; Transmission, remote access or control of music data for electrophonic musical instruments
    • G10H2240/281Protocol or standard connector for transmission of analog or digital data to or from an electrophonic musical instrument
    • G10H2240/311MIDI transmission

Definitions

  • This invention relates to methods and systems for automated generation of sound sequences, and especially (though not exclusively) of sound sequences in the form of music. It relates in the preferred embodiment to a Generative Music System (GMS).
  • GMS Generative Music System
  • Non-generative systems include deterministic systems which will produce the same sequences every time, along with systems that simply replay (perhaps in a random or other order) pre-composed sections of music.
  • Generative Music Systems may be considerably more complex. Such systems generate musical content, typically note by note, on the basis of a higher-level of musical knowledge. Such systems either explicitly or implicitly are aware of a variety of musical rules which are used to control or influence the generation of the music. In some systems, the rules may operate purely on the individual- notes being generated, without imposing any form of higher order musical structure on the output; in such systems, any musical order that arises will be of an emergent nature. More sophisticated systems may include higher-level rules which can influence the overall musical structure.
  • Generative Music Systems will normally create musical content "on the fly", in other words the musical sequences are built up note by note and phrase by phrase, starting at the beginning and finishing at the end. This means that - in contrast with some of the non-generative systems - the musical content can be generated and played in real time: there is no need for example for the whole of the phrase to be generated before the first few notes of the phrase can be played.
  • a generative music system For our present purposes, the essential features of a generative music system are that it generates musical content in a non-deterministic way, based upon a plurality of musical rules (which may either be implicit within the software or which may be explicitly specified by either the program writer or the user of the program).
  • a generative sound system produces non-deterministic sound sequences based upon sound-generation rules.
  • a system for the composition of audio sequences comprising a server and a generative audio engine on the server controlled by a plurality of remote clients in communication with the server, the server being arranged to transmit a completed composed audio sequence to one or more of the clients.
  • a method for the composition of audio sequences comprising composing an audio sequence at a central location, under control of one or more clients at remote locations, and transmitting the completed audio sequence to one or more of the clients.
  • a generative audio engine arranged to compose audio sequences under the control of instructions received from remote clients, the engine including means for transmitting a completed audio sequence to one or more of the clients.
  • the invention further extends to a computer game using a system or method as previously defined.
  • composition of a sequence of music or other sounds is effected within, or otherwise in conjunction with, one or more servers.
  • the method and system of the invention are applicable to the use of generative music (or other sound) systems within a server architecture, and in this regard have many aspects by which such an engine can render audio output.
  • the audio output may be streamed in real-time to networked digital devices, or may be composed on the server as a batch-process for subsequent download to such devices.
  • the server-based audio generation of the present invention may be arranged to deliver the audio by streaming.
  • the generated audio may in this case be streamed from the generative engine on the server, directly to the client device from an interactive session being conducted by the user.
  • Each stream rendered within the server system might go to a single client, or to a number of clients that all share the same stream. This facilitates server-side creation of generative interactive audio streams, which might be delivered uniquely to one client (for a highly personal audio experience), or shared with multiple clients (for shared audio experiences). Users might be charged for this service.
  • the download of the server-generated audio work may be deferred.
  • the generated audio may be saved to audio files (for example, as an MP3 or WAV formatted file) on the server, to be made available to client devices after sufficient server computation time has passed.
  • the download which may be regarded as providing a recorded generative session, may take place on any client device (for example, a personal computer). More especially, the device receiving the download may not necessarily be the original client device from which the request for audio generation was made. This latter feature is of benefit where there is not sufficient local audio processing power on the original client device (for example, a WAP phone) to accept an audio stream from a server, or perhaps to produce the music in real time.
  • users of the deferred download service like those of the streaming service, might be charged for using the service.
  • Networked servers may be used according to invention to generate many different streams and/or files simultaneously.
  • the facility for generation of audio over multiple servers is of advantage in the circumstances in which the amount of audio generation required exceeds the processing power of a single server.
  • the method and system of the invention may include provision for interaction by clients in creating instructions or rules for generating audio, and in general may be adapted to allow individual clients to manipulate or influence directly the audio that is created on the server. Also, the method and system may adapted to allow group interactive sessions, where many clients share and influence the same composition in either a simultaneous manner, or through deferred processing for later listening by group members.
  • the general context of the invention is thus that of a generative music or other sound system in which data presented or interpreted in a musical or other sound context is integrated and an output reflecting this integration is generated.
  • the generative system may include multiple interdependent musical or other sound agents that generate a stream of coordinated musical or other sound events which can in turn affect the way in which incoming data is interpreted and integrated, and which can act as an input to processing units within and external to the system.
  • the music or other sound may be delivered either in a real-time generated audio stream, or by being written to a storage medium for deferred delivery. In either case the method and system of the invention may be part of operation and structure of a larger method and system for generating musical works, audio, sounds and sound environments in real-time.
  • Figure 1 is a schematic representation of an embodiment of the present invention
  • Figure 2 is illustrative of objects that are involved in a component of the system of Figure 1;
  • Figure 3 is illustrative of operation of the method and system of the invention in relation to scale and harmony rules;
  • Figure 4 illustrates operation of the method and system of the preferred embodiment in relation to the triggering of note sequences and their integration into a musical work as currently being composed and played;
  • Figures 5 and 6 illustrate aspects of operation of the method and system of the preferred embodiment in relation to audio generation with rendering at a server.
  • the method and system to be described are for automated generation of sound sequences and to integrate data presented or interpreted in a musical context for generating an output reflecting this integration. Operation is within the context of generation of musical works, audio, sounds and sound environments in real- time. More especially, the method and system function in the manner of a "generative music system' operating in real-time to enable user-interaction to be incorporated into the composition on-the-fly.
  • the overall construction of the system is shown in Figure 1 and will now be described.
  • the system involves four high-level layers, namely, an applications layer I comprising software components 1 to 5, a layer II formed by an application programmer's interface (API) 6 for interfacing with a music (or, more generally, an audio or compositional) engine SKME (SSEYO Koan Music Engine) that is manifest in objects or components 7 to 14 of a layer III, and a hardware device layer IV comprising hardware components 15 to 19 that interact with the music engine SKME of layer III.
  • an applications layer I comprising software components 1 to 5
  • SKME an audio or compositional engine
  • SKME SSEYO Koan Music Engine
  • the applications layer I determines the look, feel and physical instantiation of the music engine SKME.
  • Users can interact with the music/audio/compositional engine SKME through web applications 1, or through desktop computer applications 2 such as those marketed by the Applicants under their Registered Trade Mark KOAN " as KOAN PRO and KOAN X; the music engine SKME may itself be such as marketed by the Applicants under the Registered Trade Mark KOAN.
  • Interaction with the engine SKME may also be through applications on other diverse platforms 3 such as, for example through mobile telephones or electronic toys. All applications 1 to 3 ultimately communicate with the music engine SKME via the API 6 which protects the internals of the music engine SKME from the outside world and controls the way in which the applications can interact with it.
  • the instructions sent to the API 6 from the applications 1 to 3 consist of commands that instruct the music engine SKME to carry out certain tasks, for example starting the composition and playback, and changing the settings of certain parameters (which may affect the way in which the music is composed/played).
  • communication with the API 6 may be direct or via an intermediate API.
  • the SKME according to the present invention resides in one or more servers, and the music or sound generation is effected at the server, for transmittal to one or more remote clients.
  • the music engine SKME which is held in memory within the system, comprises eight main components 7 to 14.
  • SSFIO SSEYO file input/output 7
  • SSFIO SSEYO file input/output 7
  • a soundscape 8 is created in memory and this is responsible for creating a composer 10, conductor 12 and all the individual compositional objects 9 relating to the description of the piece as recorded in the SSFIO 7.
  • the compositional objects are referred to by the composer 10 (which runs in a background loop) to decide what notes to compose next.
  • the composed notes are stored in a number of buffers 11 along with a time-stamp which specifies when they should be played.
  • the conductor 12 keeps time, by receiving accurate time information from a timer device 19 of level IV.
  • the relevant notes are removed from the buffers 11 and the information they contain (such as concerning pitch, amplitude, play time, the instrument to be used, etc.) is passed to the appropriate rendering objects 13.
  • the rendering objects 13 determine how to play this information, in particular whether via a MIDI output device 17, or as an audio sample via an audio-out device 18, or via a synthesiser engine 14 which generates complex wave-forms for audio output directly, adding effects as needed.
  • the hardware devices layer IV includes in addition to the devices 17 to 19, a file system 15 that stores complete descriptions of rules and parameters used for individual compose/playback sessions in the system; each of these descriptions is stored as an 'SSfile', and many of these files may be stored by the file system 15.
  • a MIDI in device 16 is included in layer IV to allow note and other musical-event information triggered by an external hardware object (such as a musical keyboard) to be passed into the music engine SKME and influence the composition in progress.
  • the system can be described as having essentially two operative states, one, a "dynamic' state, in which it is composing and the other, a "static' state, in which it is not composing.
  • a dynamic state in which it is composing
  • static state the system allows modification of the rules that are used by the algorithms to later compose and play music, and keeps a record encapsulated in the SSFIO 7, of various objects that are pertinent to the description of how the system may compose musical works.
  • the system is also operative in the dynamic state to keep records of extra objects which hold information pertinent to the real-time composition and generation of these works. Many of these objects (the compositional objects 9 for example) are actual instantiations in memory of the descriptions contained in the SSFIO 7. Modification of the descriptions in the SSFIO 7 via the API layer II during the dynamic state, results in those modifications being passed down to the compositional objects 9 so that the real-time composition changes accordingly.
  • Figure 2 shows a breakdown of the SSFIO 7 into its constituent component objects which exist when the system is in its static and dynamic states; the system creates real-time versions of these objects when composing and playing.
  • SSfiles 20 stored each provide information as to "SSObject(s)' 21 representing the different types of object that can be present in the description of a work; these objects may, for example, relate to piece, voice, scale rule, harmony rule, rhythm rule.
  • Each of these objects has a list of "SSFparameters' 22 that describe it; for example, they may relate to tempo, instrument and scale root.
  • the API 6 allows a number of functions to be effected such as "start composing and playing', "change the rules used in the composition', "change the parameters that control how the piece is played' including the configuration of effects etc.
  • the rules and other parameters affecting composition (for example, tempo) within the music engine SKME are defined in memory, specifically within the SSFIO 7, and its real-time instantiation of the compositional objects 9.
  • Use of rules and parameters within the music engine SKME form part of the continual compositional process for other voice objects within the system.
  • Figure 4 illustrates this more general process based on examples of scale and harmony rules shown at (1) and (2) respectively.
  • the scale rule is illustrated at (1) with shaded blocks indicating a non-zero probability of choosing that interval offset from a designated scale root note.
  • the octave Ove, major third M3 and fifth 5 are the most likely choices, followed by M2, 4, M6 and M7; the rest will never be chosen. Sequences that may be generated by the system from this are shown below the blocks, and in this respect the octave has been chosen most often followed by the major third and the fifth.
  • the scale root set in the system as C the resulting sequence of notes output from the system in this example are C,E,C,D,G,A,E,D,C,G,E,B,C,F, as illustrated at (1) of Figure 3.
  • the harmony rule defines how the system may choose the pitches of notes when other notes are playing, that is to say, how those pitches should harmonise together.
  • only the octave and major second are indicated (by shading) to be selected. This means that when the pitch for a voice is chosen, it must be either the same pitch as, or a major second from, all other notes currently being played.
  • rhythm rules applicable to the voice objects VI -V3 in this example give rise to a generated sequence of notes as follows: voice VI starts playing a note, then voice V2 starts playing a note, then voice V3 starts playing a note, and then after all notes have ended, voice V2 starts playing another note, followed by voice VI and then voice V3.
  • voice VI starts playing a note
  • voice V3 starts playing a note
  • voice VI starts playing another note
  • voice VI starts playing another note
  • the pitch for voice V2 must either be the same as (Ove) or a major second above (M2) the fifth. In the case illustrated, it is chosen to be the same, and so the fifth is chosen too.
  • voice V3 starts playing it must harmonise with both voices VI and V2, so the pitch chosen must be the same as, or a major second above that of voices VI and V2.
  • the system chooses voice V3 to be a major second above, therefore giving pitch offset M6 from the scale root.
  • the actual pitches and harmonisation of that sequence is determined by the composer 10 using several items of information, namely: (a) the note-control sub-pattern operational at that moment; (b) the scale, rhythm, harmony and next-note rules depending upon the type of the note-control subsequence; and (c) any piece-level rules which take into account the behaviour of other voices within the piece.
  • the music engine SKME When the music engine SKME is in dynamic (i.e. composing and playing) mode, it typically contains a number of voice compositional objects 9.
  • the composer 12 composes a sequence of notes for each of these and. makes sure they obey the various rules. The process involved is illustrated in the flow diagram of Figure 4.
  • the music engine SKME responds to an external trigger applied at step 51, and the API 6 through step 52 instructs a voice 1 in step 53 to register that it must start a sequence.
  • Voice 1 and the voices 2 to N in step 54 have their own rules, and the composer 10 ensures that the relevant rules are obeyed when utilising any of the voices 1 to N. More particularly, the composer 10 responds in step 55 to the instruction of step 53 for voice 1 to start a sequence, by starting the generative pattern sequence sub-system of Figure 3. This sends note-control sub-sequences to the trigger voice (voice 1 in this example), but the composer 10 makes sure the resulting notes harmonise with the other voices in the piece. The outcome via the conductor 12 in step 56 is played in step 57.
  • the generative pattern sequence triggered will play forever, or until the system is instructed otherwise. If a sequence control sub-pattern is used to define a generative pattern sequence such that the final note control sub-pattern is one which plays silence (rest notes) in an infinite loop, then when this pattern sequence is selected, the voice will become effectively "inactive' until another trigger is detected. Further triggering events for the same generative pattern sequence may sound different as the process is generative, or since the rules in use by the piece or the scale of the trigger voice, its harmony or next note rules may have changed (either via interaction through the API 6 or via internal music engine SKME changes).
  • the sounds used to "render' each note, whether from triggered sequences or generative voices may be played either through the MIDI sounds or the samples of the rendering objects 13, or via software of the synthesiser engine 14 which may add digital signal processing effects such as, for example, filter sweeps, reverberation and chorus.
  • the entire process can be used to generate musical event information that is then fed into, and may thus control, other processing units within the system such as synthesiser related units allowing the triggering of generative sound effects.
  • Voices can also be added which make use of the software synthesiser engine 14 to generate non note-based effects such as sound washes and ambient environmental sounds, such as chimes, wind and other organic sounds.
  • the system operates according to a program which allows the user to modify parameters or create structures representing elements of the composition or sound definitions ("user control information') which can either: (1) affect the current group or solo audio session in progress as soon as the changes can be accommodated; or (2) affect a session exclusive to the user or exclusive to a number of users where the user's input is used to create audio material that can be used independently of the session or added to the session at a later time.
  • user control information' can either: (1) affect the current group or solo audio session in progress as soon as the changes can be accommodated; or (2) affect a session exclusive to the user or exclusive to a number of users where the user's input is used to create audio material that can be used independently of the session or added to the session at a later time.
  • this application might reside either predominantly on the server, or predominantly on the client.
  • This program could be written in a variety of computer languages (for example, client-side Java, Java Servlets, C++ etc.), which can communicate with the server via for example pipes, streams, text blocks, simple messages, MIDI etc.
  • User control information can be either uploaded dynamically to the sever to affect the session on a ongoing basis or instead as a batch-process.
  • the general principle is that one or more remote clients are controlling or influencing (either in real time or not) sounds or music generated by a server.
  • the client control could be direct - e.g. by explicitly instructing the server to generate music in the key of C major, having a particular rhythm - or it could alternatively be implicit - for example where the client is controlling something else on the server such as an image, and the features of the image then control or influence the sound.
  • the control information could include the selection of pre-defined templates on the server which may control or constrain the sound generation process in a multiplicity of predefined ways.
  • the client may also send information which instructs the server to set up, modify or delete such templates. Once a user has gone to the trouble of defining a template, it may then simply be referred to in the future by a short command which instructs the server to select it.
  • the audio session might be rendered immediately via an audio stream (subject to a small latency period), or it might be created in a deferred manner for later delivery (saved to an audio recording within the server file system).
  • the users can influence the shared audio experience through a number of mechanisms, including (but not limited to): (a) voting;
  • WAP phone or PDA used to create personal "musicians' (actually SKME voices all stored on the server along with all other data), sound effects, tunes or tracks. Menus or some other data entry mechanism may be used to design these voices, sound effects, tunes or tracks etc.
  • users can also interact with a composition in progress or design SKME components via a standard touch-tone telephone system.
  • the service When they dial up the telephone server, the service translates touch-tone sounds into data (via the telephone server) and sends this data to the SKME server which interprets this as data to modify compositional parameters.
  • the SKME server interprets this as data to modify compositional parameters.
  • a user can, with a touch-tone telephone, make simple modifications to the SKME compositional processes.
  • the user could also hear the composition in progress on the telephone line. This would happen by the server's streamed or stored audio files (as described below) simply being played down the telephone line to the user. In this way the user does not need a special client-device with the ability to receive and play streams.
  • Rendering may be server-side and audio streamed to the client.
  • one or more servers are used to create one or more generative audio streams, and/or one or more recordings of the audio stream are stored.
  • the audio to be created is dictated by the application logic, which is in turn influenced by the users involved in controlling any one particular server-side output.
  • FIG. 5 illustrates audio generation with rendering at the server and audio streaming to the client.
  • the client 61 interacts with an application 62 and 63 to create music structural components (such as voice rules mentioned above, or more sophisticated musical structures). Communication between the client and server-side in this application takes place via a gateway 64 (such as a WAP server for example). There may be many server-side applications in a one to one correspondence with the number of clients connected to it.
  • the created music structural components are then integrated into a database 65 to be queried by the SKME Server Process 66 to be used as the basis for a time-sequenced collection of multiple users' control information and audio samples from 67 to 68 waiting to be composed and rendered. This rendered audio is streamed to the client starting at the server's current playback point. The client then receives and hears the generated streamed audio from the server as represented at 69.
  • the audio may be sent as a steam of data (or control information) in any other music/audio representation format (such as MIDI data). It would be up to the client-side of the application to convert that data-stream into actual sounds (for example, standard MIDI output device or software synthesiser).
  • audio generation may be with rendering at the server and deferred file delivery to the client.
  • user operations are used to influence the creation of audio by batch processes on the server.
  • the server composition process is deferred, such that as and when suitable server-side CPU power is available, the compositions are rendered and the recorded audio sessions are made available for consumption by the user and/or interested users.
  • the presence of these files is then communicated to the user and/or interested users, via e-mail or some other messaging system.
  • the audio recordings may then be listened to at the convenience of each user. This is illustrated by Figure 6 where the client 71 interacts with an application 72 and 73 to create music structural components (such as SKME voice rules mentioned above, or more sophisticated musical structures).
  • a gateway 74 such as a WAP server for example.
  • server-side applications in a one to one correspondence with the number of clients connected to it.
  • the created music structural components are then integrated into a database 75 of musical compositions to be generated. These are farmed off to numerous "SKME server processes' 76 to generate each client's audio output. When these are completed they are stored as audio files (for example, in MP3 or WAV formats) on a file store 77.
  • the audio may be stored as a sequence of data (or control information) in any other music/audio representation format (such as MIDI data), ready for transmittal to or downloading by the client.
  • a sequence of data or control information
  • any other music/audio representation format such as MIDI data
  • audio generation may be with rendering at the server and audio played to the user via a telephone.
  • the user could also hear the composition in progress on the phone line. This would happen by the server's streamed or stored audio files simply being played down the telephone line to the user. In this way the user does not need a special client-device with the ability to receive and play streams.
  • the server may supply the client with a "rough rendering" of the particular sound or music that has been generated. This can be done relatively rapidly at the server, and, because of the small file size, need not take too much transmission time either. If the user is happy with the sound, he instructs the client to send a message back to the server asking for the sound to be rendered at high resolution.
  • the server complies - either immediately or • when it has time - and sends a message to the client to indicate when the high- resolution file is ready for downloading.
  • An alternative site variant of that approach would be for the server to send control instructions to the client enabling the client itself to do the initial "rough rendering". If the result is acceptable to the user, a message is sent back to the server requesting either control instructions enabling the final high-quality sound to be rendered at the client or, alternatively, a request for the high-quality rendering to be carried out at the server.
  • a user may be allocated an amount of storage space or a number of entries in the database (either for free or for a fee) and further space or entry slots could then be purchased as required.
  • Mail .78 is then sent to the client to inform him/her that the audio generation is completed and ready to be downloaded.
  • the file store can also be accessed by the web-server 79 which the user then logs into to download his completed audio file.
  • the user can log into this from any web terminal 80 where he can then hear the audio.
  • the user could, for example, mail the audio file to a friend, or save it to a CD.
  • a single server option might be appropriate when there is only a small number of compositions being generated at any one time, for example in a system where many users are sharing the same audio experience; a group or community session.
  • a user may choose to purchase copies of segments of the recordings of the session, perhaps as a memento. These copies would be owned by the user and could be distributed to the user via CD-ROM or e-mail, for example.
  • Each generated audio output would be either streamed back to one or more clients, or recorded to backing storage for later consumption by the user and/or interested users where such clients support playback of the output.
  • Output could for example be streamed to multiple clients in the case of the user's component being incorporated in a community composition, or the user could for example, at a time after the interactive session, download the recorded output of his composed component. Alternatively the user could hear (where it is in audio form) his composed component or the community composition over a telephone.
  • the invention has application for group composition sessions where a style is allowed to develop. This may come about through positive feedback when a majority of users note a particular "style' emerging from the work and try to reinforce it. Thus, each server/server-session may embody its own style. Server-sessions may be allowed to interact by passing rule-sets between them. This can act as a primitive form of communication system. A method of assigning fitness to rule-sets could allow spread of rule-sets (which can be seen as style components) from server session to server session. These rule-sets may well operate at different temporal scales, allowing musical components to travel through compositions, much as viruses spread through populations of hosts.
  • the invention may be applied to sever-side sound design, that is to say to provision of a service in which a user defines the elements of a composition, tune, mix or sound, through a WAP interface.
  • a user defines the elements of a composition, tune, mix or sound, through a WAP interface.
  • they submit their editing for batch processing/composition on the server (through a server-side engine). They get SMS mailed when their piece/sound/track/mix is ready and can then dial up a number and listen to it (or download it from a PC etc). This could either have a gaming or creativity slant, and could also be submitted to a server-side real-time session.
  • provision of a server-side real-time session may be through a service in which the customer dials a number and can listen to a composition being created or changed in real time by the listeners themselves.
  • Each keypress can control certain aspects of the group piece, such as, for example, consolidated general voting when the listener likes what is happening, or perhaps even some direct user control over elements they have submitted during server-side sound design as described in the preceding, paragraph. But in general such control would be at a lower level than in server-side sound design, since those creating the elements in server-side sound design are effectively controlling the "space" — so there is a distinct advantage in having a WAP phone and becoming expert at designing the sounds/riffs/tunes etc.
  • GPRS becomes available, users will be able to engage in server-side design and server-side real-time sessions concurrently.
  • Server-side sound design has many applications, for example:
  • the ring tone could either be sent to the phone as fully-rendered audio, or alternatively by means of a series of control instructions (e.g. a midi-file or in OTA format) allowing the phone to render its own tone.
  • the tone could be either polyphonic or monophonic;
  • the client instructions to the server may affect what happens visually within the virtual world as well as what happens aurally.
  • the aural and visual aspects may be linked, for example to increase the realism of the virtual environment.
  • each player in a computer game may be represented by a corresponding computer-generated player who is able to move around within the virtual world.
  • the sounds "heard" by that individual, produced by the server and transmitted to the corresponding client, may depend upon the position of the individual within the virtual world.
  • the sound of an explosion could for example vary according to where the individual is located, and in particular how far away within the virtual world the explosion is.
  • the user may be supplied with differing ambient or background sounds depending upon the location within the world of the corresponding computer- generated individual. More generally, sounds may be generated by the server and transmitted to one or more clients which depend upon the relative position of objects (not necessarily individuals) within the virtual World.
  • sounds may be generated by the server and transmitted to one or more clients which depend upon the relative position of objects (not necessarily individuals) within the virtual World.
  • a user simply uses normal game controls to move objects or individuals around within the virtual world; the server feeds information about the relative virtual positions of the objects into the SKME, and the resultant sounds are generated and/or influenced accordingly without the user needing to control the sounds directly.
  • the audio engine on the server is configured to provide a stream of audio (music and other audio events) to the user that reflects the state of the ongoing user experience within the virtual environment.
  • the audio the user will experience is weighted to reflect the greater significance of local events.

Abstract

A system for automatically generating sound sequences, for example a generative music system, has a server which communicates with a plurality of remote clients. The generative music system resides entirely on the server, with control coming from the remote clients (71). In some embodiments, the sounds or music generated at the server may depend upon the collaborative or competitive real time actions of the clients. The system has numerous applications, including mobile phones and multi-user computer games.

Description

Automated Generation of Sound Sequences
This invention relates to methods and systems for automated generation of sound sequences, and especially (though not exclusively) of sound sequences in the form of music. It relates in the preferred embodiment to a Generative Music System (GMS).
The automated creation of music has a long history, going back at least as far as Mozart's use of musical dice. One of the first musical works generated, by a computer was L. Hiller's Illiac suite. Since that time, of course, the sophistication of computer-generated music or more generally audio sequences has increased substantially.
Systems for creating musical sequences by computer may conveniently be divided up into two areas, which have been called "non-generative" and
"generative". Non-generative systems include deterministic systems which will produce the same sequences every time, along with systems that simply replay (perhaps in a random or other order) pre-composed sections of music.
The vast majority of current systems which produce musical output make use of this type of approach, for example by selecting and playing a particular predefined sequence of notes at random when the key is pressed or a mouse button clicked. Generative Music Systems, on the other hand, may be considerably more complex. Such systems generate musical content, typically note by note, on the basis of a higher-level of musical knowledge. Such systems either explicitly or implicitly are aware of a variety of musical rules which are used to control or influence the generation of the music. In some systems, the rules may operate purely on the individual- notes being generated, without imposing any form of higher order musical structure on the output; in such systems, any musical order that arises will be of an emergent nature. More sophisticated systems may include higher-level rules which can influence the overall musical structure. Generative Music Systems will normally create musical content "on the fly", in other words the musical sequences are built up note by note and phrase by phrase, starting at the beginning and finishing at the end. This means that - in contrast with some of the non-generative systems - the musical content can be generated and played in real time: there is no need for example for the whole of the phrase to be generated before the first few notes of the phrase can be played.
For our present purposes, the essential features of a generative music system are that it generates musical content in a non-deterministic way, based upon a plurality of musical rules (which may either be implicit within the software or which may be explicitly specified by either the program writer or the user of the program). By analogy, a generative sound system produces non-deterministic sound sequences based upon sound-generation rules.
According to a first aspect of the present invention there is provided a system for the composition of audio sequences comprising a server and a generative audio engine on the server controlled by a plurality of remote clients in communication with the server, the server being arranged to transmit a completed composed audio sequence to one or more of the clients. According to a second aspect of the present invention there is provided a method for the composition of audio sequences, comprising composing an audio sequence at a central location, under control of one or more clients at remote locations, and transmitting the completed audio sequence to one or more of the clients. According to a third aspect of the present invention there is provided a generative audio engine arranged to compose audio sequences under the control of instructions received from remote clients, the engine including means for transmitting a completed audio sequence to one or more of the clients.
The invention further extends to a computer game using a system or method as previously defined.
According to another aspect of the present invention there is provided a method or a system in which composition of a sequence of music or other sounds is effected within, or otherwise in conjunction with, one or more servers.
The method and system of the invention are applicable to the use of generative music (or other sound) systems within a server architecture, and in this regard have many aspects by which such an engine can render audio output. The audio output may be streamed in real-time to networked digital devices, or may be composed on the server as a batch-process for subsequent download to such devices.
The server-based audio generation of the present invention may be arranged to deliver the audio by streaming. The generated audio may in this case be streamed from the generative engine on the server, directly to the client device from an interactive session being conducted by the user. Each stream rendered within the server system might go to a single client, or to a number of clients that all share the same stream. This facilitates server-side creation of generative interactive audio streams, which might be delivered uniquely to one client (for a highly personal audio experience), or shared with multiple clients (for shared audio experiences). Users might be charged for this service.
As an alternative, the download of the server-generated audio work may be deferred. In this case, the generated audio may be saved to audio files (for example, as an MP3 or WAV formatted file) on the server, to be made available to client devices after sufficient server computation time has passed. The download, which may be regarded as providing a recorded generative session, may take place on any client device (for example, a personal computer). More especially, the device receiving the download may not necessarily be the original client device from which the request for audio generation was made. This latter feature is of benefit where there is not sufficient local audio processing power on the original client device (for example, a WAP phone) to accept an audio stream from a server, or perhaps to produce the music in real time. As a general matter, users of the deferred download service, like those of the streaming service, might be charged for using the service.
Networked servers may be used according to invention to generate many different streams and/or files simultaneously. The facility for generation of audio over multiple servers is of advantage in the circumstances in which the amount of audio generation required exceeds the processing power of a single server.
The method and system of the invention may include provision for interaction by clients in creating instructions or rules for generating audio, and in general may be adapted to allow individual clients to manipulate or influence directly the audio that is created on the server. Also, the method and system may adapted to allow group interactive sessions, where many clients share and influence the same composition in either a simultaneous manner, or through deferred processing for later listening by group members.
The general context of the invention is thus that of a generative music or other sound system in which data presented or interpreted in a musical or other sound context is integrated and an output reflecting this integration is generated. The generative system may include multiple interdependent musical or other sound agents that generate a stream of coordinated musical or other sound events which can in turn affect the way in which incoming data is interpreted and integrated, and which can act as an input to processing units within and external to the system. The music or other sound may be delivered either in a real-time generated audio stream, or by being written to a storage medium for deferred delivery. In either case the method and system of the invention may be part of operation and structure of a larger method and system for generating musical works, audio, sounds and sound environments in real-time.
A method and system for automated generation of sound sequences, and applications of such method and system, according to the invention will now be described, by way of example, with reference to the accompanying drawings, in which:
Figure 1 is a schematic representation of an embodiment of the present invention;
Figure 2 is illustrative of objects that are involved in a component of the system of Figure 1; Figure 3 is illustrative of operation of the method and system of the invention in relation to scale and harmony rules;
Figure 4 illustrates operation of the method and system of the preferred embodiment in relation to the triggering of note sequences and their integration into a musical work as currently being composed and played; and
Figures 5 and 6 illustrate aspects of operation of the method and system of the preferred embodiment in relation to audio generation with rendering at a server.
The method and system to be described are for automated generation of sound sequences and to integrate data presented or interpreted in a musical context for generating an output reflecting this integration. Operation is within the context of generation of musical works, audio, sounds and sound environments in real- time. More especially, the method and system function in the manner of a "generative music system' operating in real-time to enable user-interaction to be incorporated into the composition on-the-fly. The overall construction of the system is shown in Figure 1 and will now be described.
Referring to Figure 1, the system involves four high-level layers, namely, an applications layer I comprising software components 1 to 5, a layer II formed by an application programmer's interface (API) 6 for interfacing with a music (or, more generally, an audio or compositional) engine SKME (SSEYO Koan Music Engine) that is manifest in objects or components 7 to 14 of a layer III, and a hardware device layer IV comprising hardware components 15 to 19 that interact with the music engine SKME of layer III. Information flow between the software and hardware components of layers I to IV is represented in Figure 1 by arrow-heads on dotted-line interconnections, whereas arrow-heads on solid lines indicate an act of creation; for example, information in the composed- notes buffer 11 is used by the conductor 12 which is created by the soundscape 8. The soundscape is a runtime object, only one of which typically exists within the SKME.
The applications layer I determines the look, feel and physical instantiation of the music engine SKME. Users can interact with the music/audio/compositional engine SKME through web applications 1, or through desktop computer applications 2 such as those marketed by the Applicants under their Registered Trade Mark KOAN" as KOAN PRO and KOAN X; the music engine SKME may itself be such as marketed by the Applicants under the Registered Trade Mark KOAN. Interaction with the engine SKME may also be through applications on other diverse platforms 3 such as, for example through mobile telephones or electronic toys. All applications 1 to 3 ultimately communicate with the music engine SKME via the API 6 which protects the internals of the music engine SKME from the outside world and controls the way in which the applications can interact with it. Typically, the instructions sent to the API 6 from the applications 1 to 3 consist of commands that instruct the music engine SKME to carry out certain tasks, for example starting the composition and playback, and changing the settings of certain parameters (which may affect the way in which the music is composed/played). Depending on the needs of the individual applications, communication with the API 6 may be direct or via an intermediate API. The SKME according to the present invention resides in one or more servers, and the music or sound generation is effected at the server, for transmittal to one or more remote clients. The music engine SKME, which is held in memory within the system, comprises eight main components 7 to 14. Of these, SSFIO (SSEYO file input/output) 7, which as the name suggests is for file input/output, holds a description of the parameters, rules and their settings used by algorithms within the engine, to compose. When the engine SKME is instructed via the API 6 to start composition/playback, a soundscape 8 is created in memory and this is responsible for creating a composer 10, conductor 12 and all the individual compositional objects 9 relating to the description of the piece as recorded in the SSFIO 7. The compositional objects are referred to by the composer 10 (which runs in a background loop) to decide what notes to compose next. The composed notes are stored in a number of buffers 11 along with a time-stamp which specifies when they should be played. The conductor 12 keeps time, by receiving accurate time information from a timer device 19 of level IV. When the current time exceeds the time-stamp of notes in the buffers 11, the relevant notes are removed from the buffers 11 and the information they contain (such as concerning pitch, amplitude, play time, the instrument to be used, etc.) is passed to the appropriate rendering objects 13. The rendering objects 13 determine how to play this information, in particular whether via a MIDI output device 17, or as an audio sample via an audio-out device 18, or via a synthesiser engine 14 which generates complex wave-forms for audio output directly, adding effects as needed.
The hardware devices layer IV includes in addition to the devices 17 to 19, a file system 15 that stores complete descriptions of rules and parameters used for individual compose/playback sessions in the system; each of these descriptions is stored as an 'SSfile', and many of these files may be stored by the file system 15. In addition, a MIDI in device 16 is included in layer IV to allow note and other musical-event information triggered by an external hardware object (such as a musical keyboard) to be passed into the music engine SKME and influence the composition in progress.
The system can be described as having essentially two operative states, one, a "dynamic' state, in which it is composing and the other, a "static' state, in which it is not composing. In the static state the system allows modification of the rules that are used by the algorithms to later compose and play music, and keeps a record encapsulated in the SSFIO 7, of various objects that are pertinent to the description of how the system may compose musical works. The system is also operative in the dynamic state to keep records of extra objects which hold information pertinent to the real-time composition and generation of these works. Many of these objects (the compositional objects 9 for example) are actual instantiations in memory of the descriptions contained in the SSFIO 7. Modification of the descriptions in the SSFIO 7 via the API layer II during the dynamic state, results in those modifications being passed down to the compositional objects 9 so that the real-time composition changes accordingly.
Figure 2 shows a breakdown of the SSFIO 7 into its constituent component objects which exist when the system is in its static and dynamic states; the system creates real-time versions of these objects when composing and playing. In this respect, SSfiles 20 stored each provide information as to "SSObject(s)' 21 representing the different types of object that can be present in the description of a work; these objects may, for example, relate to piece, voice, scale rule, harmony rule, rhythm rule. Each of these objects has a list of "SSFparameters' 22 that describe it; for example, they may relate to tempo, instrument and scale root. When an SSfile 20 is loaded into the music engine SKME, actual instances of these objects 21 and their parameters 22 are created giving rise to "SSFObjec lnstance' 23 and "SSFParameterlnstance' 24 as illustrated in Figure 2.
Referring again to Figure 1, the user interacts with the system through applications 1 to 3 utilising the services of the API 6. The API 6 allows a number of functions to be effected such as "start composing and playing', "change the rules used in the composition', "change the parameters that control how the piece is played' including the configuration of effects etc.
The rules and other parameters affecting composition (for example, tempo) within the music engine SKME are defined in memory, specifically within the SSFIO 7, and its real-time instantiation of the compositional objects 9. Use of rules and parameters within the music engine SKME form part of the continual compositional process for other voice objects within the system. Figure 4 illustrates this more general process based on examples of scale and harmony rules shown at (1) and (2) respectively.
Referring to Figure 3, the scale rule is illustrated at (1) with shaded blocks indicating a non-zero probability of choosing that interval offset from a designated scale root note. The larger the shaded block, the greater the probability of the system choosing that offset. Thus, for this example, the octave Ove, major third M3 and fifth 5 are the most likely choices, followed by M2, 4, M6 and M7; the rest will never be chosen. Sequences that may be generated by the system from this are shown below the blocks, and in this respect the octave has been chosen most often followed by the major third and the fifth. With the scale root set in the system as C, the resulting sequence of notes output from the system in this example are C,E,C,D,G,A,E,D,C,G,E,B,C,F, as illustrated at (1) of Figure 3.
The harmony rule defines how the system may choose the pitches of notes when other notes are playing, that is to say, how those pitches should harmonise together. In the example illustrated at (2) of Figure 3, only the octave and major second are indicated (by shading) to be selected. This means that when the pitch for a voice is chosen, it must be either the same pitch as, or a major second from, all other notes currently being played.
For the purpose of further explanation, consideration will be given to the example represented at (3) of Figure 3 involving three voice objects VI -V3. The rhythm rules applicable to the voice objects VI -V3 in this example, give rise to a generated sequence of notes as follows: voice VI starts playing a note, then voice V2 starts playing a note, then voice V3 starts playing a note, and then after all notes have ended, voice V2 starts playing another note, followed by voice VI and then voice V3. With this scenario, the note from voice V2 must harmonise with that of voice VI and the voice V3 note must harmonise with that of voice V2. If in these circumstances the voice VI is, as illustrated by bold hatching, chosen with a pitch offset of a fifth from the scale root, the pitch for voice V2 must either be the same as (Ove) or a major second above (M2) the fifth. In the case illustrated, it is chosen to be the same, and so the fifth is chosen too. When voice V3 starts playing it must harmonise with both voices VI and V2, so the pitch chosen must be the same as, or a major second above that of voices VI and V2. As illustrated, the system chooses voice V3 to be a major second above, therefore giving pitch offset M6 from the scale root. After voice V3 all notes end, and the next note begins, as illustrated at (4) of Figure 3 with voice V2. This next note by voice V2 is governed by the next- note rule used by voice V2, and the last note played by voice V2. According to this rule, the system chooses pitch offset M2 for voice V2, and then harmonises voices V3 and VI with it by choice of a major second for both of them. With the scale root set in the system to C, the entire generated sequence accordingly follows that indicated at (5) of Figure 3, where "S' denotes a note starting and "E" a note ending.
Thus, when sequences are generated in response to an external trigger, the actual pitches and harmonisation of that sequence is determined by the composer 10 using several items of information, namely: (a) the note-control sub-pattern operational at that moment; (b) the scale, rhythm, harmony and next-note rules depending upon the type of the note-control subsequence; and (c) any piece-level rules which take into account the behaviour of other voices within the piece.
When the music engine SKME is in dynamic (i.e. composing and playing) mode, it typically contains a number of voice compositional objects 9. The composer 12 composes a sequence of notes for each of these and. makes sure they obey the various rules. The process involved is illustrated in the flow diagram of Figure 4.
Referring to Figure 4, the music engine SKME responds to an external trigger applied at step 51, and the API 6 through step 52 instructs a voice 1 in step 53 to register that it must start a sequence. Voice 1 and the voices 2 to N in step 54, have their own rules, and the composer 10 ensures that the relevant rules are obeyed when utilising any of the voices 1 to N. More particularly, the composer 10 responds in step 55 to the instruction of step 53 for voice 1 to start a sequence, by starting the generative pattern sequence sub-system of Figure 3. This sends note-control sub-sequences to the trigger voice (voice 1 in this example), but the composer 10 makes sure the resulting notes harmonise with the other voices in the piece. The outcome via the conductor 12 in step 56 is played in step 57.
The generative pattern sequence triggered will play forever, or until the system is instructed otherwise. If a sequence control sub-pattern is used to define a generative pattern sequence such that the final note control sub-pattern is one which plays silence (rest notes) in an infinite loop, then when this pattern sequence is selected, the voice will become effectively "inactive' until another trigger is detected. Further triggering events for the same generative pattern sequence may sound different as the process is generative, or since the rules in use by the piece or the scale of the trigger voice, its harmony or next note rules may have changed (either via interaction through the API 6 or via internal music engine SKME changes).
The sounds used to "render' each note, whether from triggered sequences or generative voices may be played either through the MIDI sounds or the samples of the rendering objects 13, or via software of the synthesiser engine 14 which may add digital signal processing effects such as, for example, filter sweeps, reverberation and chorus. The entire process can be used to generate musical event information that is then fed into, and may thus control, other processing units within the system such as synthesiser related units allowing the triggering of generative sound effects. Voices can also be added which make use of the software synthesiser engine 14 to generate non note-based effects such as sound washes and ambient environmental sounds, such as chimes, wind and other organic sounds.
The system operates according to a program which allows the user to modify parameters or create structures representing elements of the composition or sound definitions ("user control information') which can either: (1) affect the current group or solo audio session in progress as soon as the changes can be accommodated; or (2) affect a session exclusive to the user or exclusive to a number of users where the user's input is used to create audio material that can be used independently of the session or added to the session at a later time..
Depending on the implementation, the logic of this application might reside either predominantly on the server, or predominantly on the client. This program could be written in a variety of computer languages (for example, client-side Java, Java Servlets, C++ etc.), which can communicate with the server via for example pipes, streams, text blocks, simple messages, MIDI etc.
User control information can be either uploaded dynamically to the sever to affect the session on a ongoing basis or instead as a batch-process. In either case, the general principle is that one or more remote clients are controlling or influencing (either in real time or not) sounds or music generated by a server.
The client control could be direct - e.g. by explicitly instructing the server to generate music in the key of C major, having a particular rhythm - or it could alternatively be implicit - for example where the client is controlling something else on the server such as an image, and the features of the image then control or influence the sound. The control information could include the selection of pre-defined templates on the server which may control or constrain the sound generation process in a multiplicity of predefined ways. The client may also send information which instructs the server to set up, modify or delete such templates. Once a user has gone to the trouble of defining a template, it may then simply be referred to in the future by a short command which instructs the server to select it.
The audio session might be rendered immediately via an audio stream (subject to a small latency period), or it might be created in a deferred manner for later delivery (saved to an audio recording within the server file system).
Where the application allows many users to interact together to influence the audio output, the users can influence the shared audio experience through a number of mechanisms, including (but not limited to): (a) voting;
(b) specific ownership of specific musical objects (or characteristics of objects) — this could allow the application to act as a trading system for buying packs of sound/musical components and then trading them with other users (comparable with POKEMON™) — alternatively, it could allow the application to act as a bulletin board for combining components with those of other users, to form bands;
(c) time-slice based ownership;
(d) payment for the over-riding "privilege' to change a given element; and (e) consolidation of a number of user's inputs for a specific parameter into a single value. Some example applications are:
1. WAP phone or PDA used to create personal "musicians' (actually SKME voices all stored on the server along with all other data), sound effects, tunes or tracks. Menus or some other data entry mechanism may be used to design these voices, sound effects, tunes or tracks etc.
2. Recordings (audio MIDI, parametric data, control data for example) of band sessions created on the server. Users in the band are notified when session recordings are available for download. WAP phone connects to server to mail links to desktop computer (so that recordings can be heard), or to play recordings directly on the phone where that phone supports appropriate playback, e.g. of embedded MP3 content.
Additionally, users can also interact with a composition in progress or design SKME components via a standard touch-tone telephone system. When they dial up the telephone server, the service translates touch-tone sounds into data (via the telephone server) and sends this data to the SKME server which interprets this as data to modify compositional parameters. By this method a user can, with a touch-tone telephone, make simple modifications to the SKME compositional processes. The user could also hear the composition in progress on the telephone line. This would happen by the server's streamed or stored audio files (as described below) simply being played down the telephone line to the user. In this way the user does not need a special client-device with the ability to receive and play streams.
Rendering may be server-side and audio streamed to the client. In this regard, one or more servers are used to create one or more generative audio streams, and/or one or more recordings of the audio stream are stored. The audio to be created is dictated by the application logic, which is in turn influenced by the users involved in controlling any one particular server-side output.
Figure 5 illustrates audio generation with rendering at the server and audio streaming to the client. The client 61 interacts with an application 62 and 63 to create music structural components (such as voice rules mentioned above, or more sophisticated musical structures). Communication between the client and server-side in this application takes place via a gateway 64 (such as a WAP server for example). There may be many server-side applications in a one to one correspondence with the number of clients connected to it. The created music structural components are then integrated into a database 65 to be queried by the SKME Server Process 66 to be used as the basis for a time-sequenced collection of multiple users' control information and audio samples from 67 to 68 waiting to be composed and rendered. This rendered audio is streamed to the client starting at the server's current playback point. The client then receives and hears the generated streamed audio from the server as represented at 69.
Instead of the process rendering the audio to an audio file to be streamed, the audio may be sent as a steam of data (or control information) in any other music/audio representation format (such as MIDI data). It would be up to the client-side of the application to convert that data-stream into actual sounds (for example, standard MIDI output device or software synthesiser).
As an alternative to the above, audio generation may be with rendering at the server and deferred file delivery to the client. In this respect, user operations are used to influence the creation of audio by batch processes on the server. The server composition process is deferred, such that as and when suitable server-side CPU power is available, the compositions are rendered and the recorded audio sessions are made available for consumption by the user and/or interested users. The presence of these files is then communicated to the user and/or interested users, via e-mail or some other messaging system. The audio recordings may then be listened to at the convenience of each user. This is illustrated by Figure 6 where the client 71 interacts with an application 72 and 73 to create music structural components (such as SKME voice rules mentioned above, or more sophisticated musical structures). Communication between the client and server sides of this application takes place via a gateway 74 (such as a WAP server for example). There may be many server-side applications in a one to one correspondence with the number of clients connected to it. The created music structural components are then integrated into a database 75 of musical compositions to be generated. These are farmed off to numerous "SKME server processes' 76 to generate each client's audio output. When these are completed they are stored as audio files (for example, in MP3 or WAV formats) on a file store 77.
Again, instead of the SKME Server Process rendering the audio to an audio file, the audio may be stored as a sequence of data (or control information) in any other music/audio representation format (such as MIDI data), ready for transmittal to or downloading by the client.
Furthermore, audio generation may be with rendering at the server and audio played to the user via a telephone. The user could also hear the composition in progress on the phone line. This would happen by the server's streamed or stored audio files simply being played down the telephone line to the user. In this way the user does not need a special client-device with the ability to receive and play streams.
In a further variation, the server may supply the client with a "rough rendering" of the particular sound or music that has been generated. This can be done relatively rapidly at the server, and, because of the small file size, need not take too much transmission time either. If the user is happy with the sound, he instructs the client to send a message back to the server asking for the sound to be rendered at high resolution. The server complies - either immediately or • when it has time - and sends a message to the client to indicate when the high- resolution file is ready for downloading.
An alternative site variant of that approach would be for the server to send control instructions to the client enabling the client itself to do the initial "rough rendering". If the result is acceptable to the user, a message is sent back to the server requesting either control instructions enabling the final high-quality sound to be rendered at the client or, alternatively, a request for the high-quality rendering to be carried out at the server.
A user may be allocated an amount of storage space or a number of entries in the database (either for free or for a fee) and further space or entry slots could then be purchased as required. Mail .78 is then sent to the client to inform him/her that the audio generation is completed and ready to be downloaded.
The file store can also be accessed by the web-server 79 which the user then logs into to download his completed audio file. The user can log into this from any web terminal 80 where he can then hear the audio. At this time the user could, for example, mail the audio file to a friend, or save it to a CD.
If the file was stored in a representational format such as MIDI (or other music/audio representation or control format) It would be up to the client-side of the application to convert that data-file into actual sounds (for example, by playing the file through standard MIDI output device or software synthesiser).
Where users want to interact with the music engine on an individual basis (that is to say, they do not want to create a joint composition with other users), then potentially a large number of high-capacity servers are needed to accommodate the correspondingly large number of composed audio-streams. A single server option might be appropriate when there is only a small number of compositions being generated at any one time, for example in a system where many users are sharing the same audio experience; a group or community session. A user may choose to purchase copies of segments of the recordings of the session, perhaps as a memento. These copies would be owned by the user and could be distributed to the user via CD-ROM or e-mail, for example.
Each generated audio output would be either streamed back to one or more clients, or recorded to backing storage for later consumption by the user and/or interested users where such clients support playback of the output. Output could for example be streamed to multiple clients in the case of the user's component being incorporated in a community composition, or the user could for example, at a time after the interactive session, download the recorded output of his composed component. Alternatively the user could hear (where it is in audio form) his composed component or the community composition over a telephone.
The invention has application for group composition sessions where a style is allowed to develop. This may come about through positive feedback when a majority of users note a particular "style' emerging from the work and try to reinforce it. Thus, each server/server-session may embody its own style. Server-sessions may be allowed to interact by passing rule-sets between them. This can act as a primitive form of communication system. A method of assigning fitness to rule-sets could allow spread of rule-sets (which can be seen as style components) from server session to server session. These rule-sets may well operate at different temporal scales, allowing musical components to travel through compositions, much as viruses spread through populations of hosts.
The invention may be applied to sever-side sound design, that is to say to provision of a service in which a user defines the elements of a composition, tune, mix or sound, through a WAP interface. When the user has finished editing the settings, they submit their editing for batch processing/composition on the server (through a server-side engine). They get SMS mailed when their piece/sound/track/mix is ready and can then dial up a number and listen to it (or download it from a PC etc). This could either have a gaming or creativity slant, and could also be submitted to a server-side real-time session.
As suggested above, provision of a server-side real-time session may be through a service in which the customer dials a number and can listen to a composition being created or changed in real time by the listeners themselves. Each keypress can control certain aspects of the group piece, such as, for example, consolidated general voting when the listener likes what is happening, or perhaps even some direct user control over elements they have submitted during server-side sound design as described in the preceding, paragraph. But in general such control would be at a lower level than in server-side sound design, since those creating the elements in server-side sound design are effectively controlling the "space" — so there is a distinct advantage in having a WAP phone and becoming expert at designing the sounds/riffs/tunes etc. When GPRS becomes available, users will be able to engage in server-side design and server-side real-time sessions concurrently.
Server-side sound design has many applications, for example:
(a) WAP musician trading/role-playing game - A WAP or other mobile phone is used to create personal "musicians' (actually SKME voices, all stored on the server along with all other data) — menus are used to design these voices, to define sounds to use etc;
(b) trading system for buying packs of sounds and trading them with other users (comparable with POKEMON™);
(c) bulletin-board system for combining these musicians with those of other users, to form bands; (d) MP3 or other "sessions' of band recordings created on the server — the users in the band would be mailed when sessions are available for download, and the WAP or other mobile phone connects to server to mail links to desktop computer, so that files can be heard, or to play files directly on the phone where that phone supports playback of e.g. embedded MP3 content;
(e) users can pay to have sessions cut to CD;
(f) whenever a desktop box is used to access a page which contains the file links, a KOAN Plugin is forced to be downloaded; (g) enhanced on-line instruments for subsequent on-line mixing of sessions or pieces - these would be available only to users who paid for the phone service; users may also pay for increased server priority, longer recordings etc;
(h) the generation of ring tones for mobile phones. As described above, the ring tone could either be sent to the phone as fully-rendered audio, or alternatively by means of a series of control instructions (e.g. a midi-file or in OTA format) allowing the phone to render its own tone. The tone could be either polyphonic or monophonic;
(i) sound-generation at the server being linked to visual displays on the client side. For example, where a number of users are engaged in a multi-user game, within a virtual 3-D world, for example via the Internet, the client instructions to the server may affect what happens visually within the virtual world as well as what happens aurally. The aural and visual aspects may be linked, for example to increase the realism of the virtual environment. For example, each player in a computer game may be represented by a corresponding computer-generated player who is able to move around within the virtual world. The sounds "heard" by that individual, produced by the server and transmitted to the corresponding client, may depend upon the position of the individual within the virtual world. The sound of an explosion could for example vary according to where the individual is located, and in particular how far away within the virtual world the explosion is. Likewise, the user may be supplied with differing ambient or background sounds depending upon the location within the world of the corresponding computer- generated individual. More generally, sounds may be generated by the server and transmitted to one or more clients which depend upon the relative position of objects (not necessarily individuals) within the virtual World. In order to generate or influence the sounds, a user simply uses normal game controls to move objects or individuals around within the virtual world; the server feeds information about the relative virtual positions of the objects into the SKME, and the resultant sounds are generated and/or influenced accordingly without the user needing to control the sounds directly. Generally, then, the audio engine on the server is configured to provide a stream of audio (music and other audio events) to the user that reflects the state of the ongoing user experience within the virtual environment. The audio the user will experience is weighted to reflect the greater significance of local events.

Claims

CLAIMS:
1. A system for the composition of audio sequences comprising a server 5 and a generative audio engine on the server controlled by a plurality of remote clients in communication with the server, the server being arranged to transmit a completed composed audio sequence to one or more of the clients.
2. A system as claimed in claim 1 in which the audio sequence is rendered i 0 on the server and transmitted in rendered form to the client.
3. A system as claimed in claim 1 in which the audio sequence is transmitted in a parametric or encoded format, for example in MIDI format, and is rendered on receipt by the client.
15
4. A system as claimed in any one of the preceding claims in which the clients are mobile phones.
5. A system as claimed in claim 4 in which the audio sequence includes a 0 ring tone.
6. A system as claimed in any one of the preceding claims in which the audio sequence includes a musical composition.
5 7. A system as claimed in any one of the preceding claims in which the audio sequence includes generated non-musical sounds.
8. A system as claimed in any one of the preceding claims in which the audio sequence is streamed in real time to one or more of the clients.
9. A system as claimed in any one of claims 1 to 7 in which the audio sequence is, once generated, held on the server until a client request is received for it to be sent.
10. A system as claimed in claim 7 in which the server informs one or more clients when the completed audio sequence is ready to be sent.
11. A system as claimed in any one of the preceding claims in which the audio engine generates the audio sequence in dependence upon instructions received from a plurality of clients.
12. A system as claimed in any one of the preceding claims in which the audio sequence is generated by the audio engine in real time, with the generation being controlled or influenced by instructions received during generation.
13. A system as claimed in any one of the preceding claims in which the audio engine is controlled by control templates stored at the server, the control templates being changeable by the clients.
14. A system as claimed in any one of claims 1 to 12 in which the audio engine is controlled by control templates stored at the server, the control templates being selectable by the clients.
15. A system as claimed in any one of the preceding claims in which the audio engine is associated at the server with an image, the engine being controlled in accordance with changes in the image effected via instructions received from the clients.
16. A system as claimed in any one of claims 1 to 14 in which the audio engine is associated at the server with a virtual 3 -dimensional world, the engine being controlled in accordance with changes in the virtual world effected via instructions received from the clients.
17. A system as claimed in claim 16 in which the said changes are changes in the relative or absolute positions of objects or individuals within the virtual world.
18. A system as claimed in claim 16 or claim 17 in which the audio sequence is representative of the sounds heard by a virtual individual or object within the virtual world.
19. A system as claimed in any one of the preceding claims when dependent upon claim 11 in which each client's instructions control a specific musical object or characteristic of such an object within the audio engine, for example key, rhythm, voice, reverberation, and so on.
20. A system as claimed in any one of the preceding claims when dependent upon claim 11 in which the audio engine creates the audio sequence in dependence upon votes received from the clients.
21. A system as claimed in any one of the preceding claims when dependent upon claim 11 in which instructions from more than one client are combined into one or more parameter values, the parameter value or values controlling the audio engine.
22. A system as claimed in any one of the preceding claims including a plurality of linked servers, each having a generative music engine.
23. A computer game using a system as claimed in any one of the preceding claims.
24. A method for the composition of audio sequences, comprising composing an audio sequence at a central location, under control of one or more clients at remote locations, and transmitting the completed audio sequence to one or more of the clients.
25. A generative audio engine arranged to compose audio sequences under the control of instructions received from remote clients, the engine including means for transmitting a completed audio sequence to one or more of the clients.
26. A computer game using a method as claimed in claim 24.
PCT/GB2001/001985 2000-05-05 2001-05-04 Automated generation of sound sequences WO2001086628A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2001254939A AU2001254939A1 (en) 2000-05-05 2001-05-04 Automated generation of sound sequences

Applications Claiming Priority (10)

Application Number Priority Date Filing Date Title
GB0010967A GB0010967D0 (en) 2000-05-05 2000-05-05 Automated generation of sound sequences
GB0010969.4 2000-05-05
GB0010969A GB0010969D0 (en) 2000-05-05 2000-05-05 Automated generation of sound sequences
GB0010967.8 2000-05-05
GB0011178A GB0011178D0 (en) 2000-05-09 2000-05-09 Automated generation of sound sequences
GB0011178.1 2000-05-09
GB0022164A GB0022164D0 (en) 2000-09-11 2000-09-11 Automated generation of sound sequences
GB0022164.8 2000-09-11
GB0030833.8 2000-12-18
GB0030833A GB0030833D0 (en) 2000-05-05 2000-12-18 Automated generation of sound sequences

Publications (2)

Publication Number Publication Date
WO2001086628A2 true WO2001086628A2 (en) 2001-11-15
WO2001086628A3 WO2001086628A3 (en) 2002-03-28

Family

ID=27515938

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/GB2001/001985 WO2001086628A2 (en) 2000-05-05 2001-05-04 Automated generation of sound sequences

Country Status (2)

Country Link
AU (1) AU2001254939A1 (en)
WO (1) WO2001086628A2 (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
GB2378306A (en) * 2001-07-31 2003-02-05 Sendo Int Ltd Tone generator for a communication device
WO2004049300A1 (en) * 2002-11-22 2004-06-10 Hutchison Whampoa Three G Ip(Bahamas) Limited Method for generating an audio file on a server upon a request from a mobile phone
US6815600B2 (en) 2002-11-12 2004-11-09 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US6972363B2 (en) 2002-01-04 2005-12-06 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
EP1650743A1 (en) * 2004-10-21 2006-04-26 Yamaha Corporation Electronic musical apparatus system, server-side electronic musical apparatus and client-side electronic musical apparatus
US7076035B2 (en) 2002-01-04 2006-07-11 Medialab Solutions Llc Methods for providing on-hold music using auto-composition
US7169996B2 (en) 2002-11-12 2007-01-30 Medialab Solutions Llc Systems and methods for generating music using data/music data file transmitted/received via a network
CN1838230B (en) * 2005-03-24 2010-05-12 雅马哈株式会社 Music data generation system and method
US8257157B2 (en) 2008-02-04 2012-09-04 Polchin George C Physical data building blocks system for video game interaction
US9818386B2 (en) 1999-10-19 2017-11-14 Medialab Solutions Corp. Interactive digital music recorder and player
WO2019121574A1 (en) * 2017-12-18 2019-06-27 Bytedance Inc. Automated music production
GB2581319A (en) * 2018-12-12 2020-08-19 Bytedance Inc Automated music production
EP3857538A4 (en) * 2018-09-25 2022-06-22 Reactional Music Group AB Real-time music generation engine for interactive systems
US11842710B2 (en) 2021-03-31 2023-12-12 DAACI Limited Generative composition using form atom heuristics

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9065931B2 (en) 2002-11-12 2015-06-23 Medialab Solutions Corp. Systems and methods for portable audio synthesis

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5734119A (en) * 1996-12-19 1998-03-31 Invision Interactive, Inc. Method for streaming transmission of compressed music
US5886274A (en) * 1997-07-11 1999-03-23 Seer Systems, Inc. System and method for generating, distributing, storing and performing musical work files
EP0933906A2 (en) * 1998-01-29 1999-08-04 Yamaha Corporation Network system for ensemble performance by remote terminals
GB2334137A (en) * 1998-02-04 1999-08-11 Ghassan Matar Musical score display device
US5952599A (en) * 1996-12-19 1999-09-14 Interval Research Corporation Interactive music generation system making use of global feature control by non-musicians
WO2001016931A1 (en) * 1999-09-01 2001-03-08 Nokia Corporation Method and arrangement for providing customized audio characteristics to cellular terminals

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5734119A (en) * 1996-12-19 1998-03-31 Invision Interactive, Inc. Method for streaming transmission of compressed music
US5952599A (en) * 1996-12-19 1999-09-14 Interval Research Corporation Interactive music generation system making use of global feature control by non-musicians
US5886274A (en) * 1997-07-11 1999-03-23 Seer Systems, Inc. System and method for generating, distributing, storing and performing musical work files
EP0933906A2 (en) * 1998-01-29 1999-08-04 Yamaha Corporation Network system for ensemble performance by remote terminals
GB2334137A (en) * 1998-02-04 1999-08-11 Ghassan Matar Musical score display device
WO2001016931A1 (en) * 1999-09-01 2001-03-08 Nokia Corporation Method and arrangement for providing customized audio characteristics to cellular terminals

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9818386B2 (en) 1999-10-19 2017-11-14 Medialab Solutions Corp. Interactive digital music recorder and player
GB2378306B (en) * 2001-07-31 2005-05-25 Sendo Int Ltd Tone generation
GB2378306A (en) * 2001-07-31 2003-02-05 Sendo Int Ltd Tone generator for a communication device
US7102069B2 (en) 2002-01-04 2006-09-05 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US6972363B2 (en) 2002-01-04 2005-12-06 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US7076035B2 (en) 2002-01-04 2006-07-11 Medialab Solutions Llc Methods for providing on-hold music using auto-composition
US6815600B2 (en) 2002-11-12 2004-11-09 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US7026534B2 (en) 2002-11-12 2006-04-11 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US6958441B2 (en) 2002-11-12 2005-10-25 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US6977335B2 (en) 2002-11-12 2005-12-20 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US6979767B2 (en) 2002-11-12 2005-12-27 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US7015389B2 (en) 2002-11-12 2006-03-21 Medialab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US7022906B2 (en) 2002-11-12 2006-04-04 Media Lab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US6897368B2 (en) 2002-11-12 2005-05-24 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
US6960714B2 (en) 2002-11-12 2005-11-01 Media Lab Solutions Llc Systems and methods for creating, modifying, interacting with and playing musical compositions
US7169996B2 (en) 2002-11-12 2007-01-30 Medialab Solutions Llc Systems and methods for generating music using data/music data file transmitted/received via a network
US6916978B2 (en) 2002-11-12 2005-07-12 Alain Georges Systems and methods for creating, modifying, interacting with and playing musical compositions
WO2004049300A1 (en) * 2002-11-22 2004-06-10 Hutchison Whampoa Three G Ip(Bahamas) Limited Method for generating an audio file on a server upon a request from a mobile phone
EP1650743A1 (en) * 2004-10-21 2006-04-26 Yamaha Corporation Electronic musical apparatus system, server-side electronic musical apparatus and client-side electronic musical apparatus
US7390954B2 (en) 2004-10-21 2008-06-24 Yamaha Corporation Electronic musical apparatus system, server-side electronic musical apparatus and client-side electronic musical apparatus
CN1838230B (en) * 2005-03-24 2010-05-12 雅马哈株式会社 Music data generation system and method
US8257157B2 (en) 2008-02-04 2012-09-04 Polchin George C Physical data building blocks system for video game interaction
JP2021507309A (en) * 2017-12-18 2021-02-22 バイトダンス・インコーポレイテッドBytedance Inc. Modular automatic music production server
WO2019121577A1 (en) * 2017-12-18 2019-06-27 Bytedance Inc. Automated midi music composition server
WO2019121576A3 (en) * 2017-12-18 2019-08-01 Bytedance Inc. Modular automated music production server
WO2019121574A1 (en) * 2017-12-18 2019-06-27 Bytedance Inc. Automated music production
CN111566724A (en) * 2017-12-18 2020-08-21 字节跳动有限公司 Modular automatic music production server
US11610568B2 (en) 2017-12-18 2023-03-21 Bytedance Inc. Modular automated music production server
JP7041270B2 (en) 2017-12-18 2022-03-23 バイトダンス・インコーポレイテッド Modular automatic music production server
EP3857538A4 (en) * 2018-09-25 2022-06-22 Reactional Music Group AB Real-time music generation engine for interactive systems
GB2581319A (en) * 2018-12-12 2020-08-19 Bytedance Inc Automated music production
GB2581319B (en) * 2018-12-12 2022-05-25 Bytedance Inc Automated music production
US11842710B2 (en) 2021-03-31 2023-12-12 DAACI Limited Generative composition using form atom heuristics
US11887568B2 (en) 2021-03-31 2024-01-30 DAACI Limited Generative composition with defined form atom heuristics

Also Published As

Publication number Publication date
AU2001254939A1 (en) 2001-11-20
WO2001086628A3 (en) 2002-03-28

Similar Documents

Publication Publication Date Title
US6093880A (en) System for prioritizing audio for a virtual environment
US5886274A (en) System and method for generating, distributing, storing and performing musical work files
US6975995B2 (en) Network based music playing/song accompanying service system and method
US7191023B2 (en) Method and apparatus for sound and music mixing on a network
WO2001086628A2 (en) Automated generation of sound sequences
US6835884B2 (en) System, method, and storage media storing a computer program for assisting in composing music with musical template data
US20030167904A1 (en) Player information-providing method, server, program for controlling the server, and storage medium storing the program
JP4382786B2 (en) Audio mixdown device, audio mixdown program
JP2002091440A (en) Performance information converting method and its device and recording medium and sound source device
WO2002077585A1 (en) System and method for music creation and rearrangement
US7884275B2 (en) Music creator for a client-server environment
US7502657B2 (en) Information processing apparatus, information processing method, and storage medium
US20040159216A1 (en) Method and system for creating and performing music electronically via a communications network
JP2007041917A (en) Content distribution system, content-reproducing device, content-providing server, and program used for them
US6658309B1 (en) System for producing sound through blocks and modifiers
JP4075901B2 (en) Music data generation system and program
KR20030030866A (en) Audio signal output method, audio signal generating device, and program
WO2001086625A2 (en) Automated generation of sound sequences
WO2001086630A2 (en) Automated generation of sound sequences
WO2001086626A2 (en) Automated generation of sound sequences
JP2001517814A (en) Sound effect system
WO2001086627A2 (en) Automated generation of sound sequences
JP2003099035A (en) Automatic playing device, information distributing server device, and program used for them
CN1734552B (en) Electronic musical apparatus for reproducing received music content
WO2001086629A2 (en) Automated generation of sound sequences

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
AK Designated states

Kind code of ref document: A3

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A3

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GW ML MR NE SN TD TG

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: COMMUNICATION UNDER RULE 69 EPC (EPO FORM 1205A OF 18.02.2003)

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase in:

Ref country code: JP