Dobrian’s early interactive compositions

As computer processing speed became faster, in the 1980s and 1990s it became clear that a computer program could compose and synthesize sound instantaneously “in real time”, meaning “without noticeable time delay”. Composing in real time is essentially what people often mean when they use the word “improvisation”; a person (or computer) makes up music and performs it at the same time. However, improvisation between two or more musicians involves the important components of listening to the music others are making, reacting in real time to what others do, and interacting with them. Can computer improvisation incorporate these apparently very human traits of listening, reacting, and interacting?

Artificial intelligence is defined by some as having the quality or appearance of intelligent behavior, of successfully imitating human intelligence even if lacking the awareness and understanding that humans have. Similarly, computer perception, cognition, and reaction/interaction can be successfully emulated to the point where we might say the computer exhibits the qualities of those abilities, even if we know it is not actually “hearing” or “interacting” the same way humans do. For this sort of emulation of interaction, we might say that the computer exhibits “interactivity”, the quality of interacting, even if its way of doing so is different from that of a human.

In the 1990s I began to focus less on music in which the computer composed and performed/synthesized its own composition, and more on music that involved some sort of interactivity between the computer and a human performer in real time. In order for the computer to seem to be interactive, its program has to include some rudimentary form of perception and cognition of the musical events produced by the human, and it must also have some degree of autonomy and unpredictability. (If these factors are absent, the computer would be acting in a wholly deterministic way, and could not be said to be in any way truly interactive.)

In the composition There’s Just One Thing You Need To Know (1998) for Disklavier, synthesizer, and interactive computer system, the notes played by a human pianist are transmitted via MIDI to a computer program that responds and contributes musically in various ways.

The piece is composed as a “mini-concerto” for the Disklavier piano, which is accompanied by a Korg Wavestation A/D synthesizer controlled by Max software interacting automatically with the performer in real time. The conceptual and musical theme of the piece is “reflection”. The music composed for the pianist requires — almost without exception — symmetrical hand movement; one hand mirrors precisely the position of the other hand, albeit often with some delay or other rhythmic modification. Even the music played by the computer is at any given moment symmetrical around a given pitch axis. The human and computer performers also act as “mirror images” or “alter egos”, playing inverted versions of the other’s musical material, playing interlocking piano passages in which they share a single musical gesture, and reinforcing ideas presented by the other.

The computer plays three different roles in the piece. In one role, it is a quasi-intelligent accompanist, providing instantaneous synthesizer response to the notes played by the pianist. In a second role, the computer acts as an extension of the pianist’s body—an extra pair of virtual hands—playing the piano at the same time as the pianist, making it possible for the piano to do more than a single pianist could achieve. In a third role, the computer is an improviser, answering the pianist’s notes with musical ideas of its own composed on the spot in response to its perceptions of the pianist’s music.

The title comes from a statement made by composer Morton Feldman. “There’s just one thing you need to know to write music for piano. You’ve got a left hand, and you’ve got a right hand. [gleefully] That’s ‘counterpoint’!”


Another work composed shortly thereafter is Microepiphanies: A Digital Opera (2000), an hour-long multimedia theatrical and musical performance in which the music, sound, lights, and projections are all controlled by computer in response to the actions of the live performers onstage, without there being any offstage technicians running the show.

The performance is conceived as a satire of the tropes and cliches that were commonly found in performances of so-called interactive music of the time. The performers describe their activities to the audience (sometimes deceitfully) as they perform music with various unusual technological devices. Because the technology occasionally seems to malfunction, or functions mysteriously, it’s difficult for the audience to know the true extent to which the computer system is actually interactive. Because it’s made clear that there are in fact no offstage technicians aiding the performance, the apparent musical and interactive sophistication of the computer seems at times magical.


A couple years later I had the opportunity to spend a year living in Seoul, Korea. I was interested to see in what ways interactive computer music could be used in the context of traditional Korean music, not simply to use the sounds of Korean music as a sort of musical exoticism, but rather as a way to find a true symbiosis between two apparently disparate musical worlds, the traditional music of an Asian nation and the modern technological music being practiced in the West.

This project required that I study traditional Korean classical music as seriously as I could, in order to be properly knowledgeable and respectful of that music as I composed the music and designed software for interaction with a live musician. Because I had previously done a series of interactive pieces involving the flute, I chose to work with the Korean bamboo flute known called the daegeum. I was helped by Serin Hong, who was at that time a student of traditional Korean music at Chugye University, specializing in daegeum performance; Serin would play the music I composed for the instrument, and would give me criticism on any passages that were not sufficiently idiomatic or playable.

I eventually wrote a complex interactive computer program and composed a thirteen-minute piece titled Mannam (“Encounter”) (2003) for daegeum and interactive computer system, which was premiered by Serin Hong in the 2003 Seoul International Computer Music Festival.

The computer was programmed to capture the expressive information (pitch and volume fluctuations) from the live daegeum performance; the program used pitch, loudness, and timbre data to shape the computer’s sound synthesis and realtime processing. The computer modifies the sound of the daegum in real time, stores and reconfigures excerpts of the played music, and provides harmonic accompaniment in “intelligent” response to the daegeum notes. The daegeum music is composed in idiomatic style, and leaves the performer considerable opportunity for rubato, ornamentation, and even occasional reordering of phrases, in order to respond to the computer’s performance, which is different every time the piece is played.

The techniques I developed for tracking the performer’s nuances are described in detail in “Strategies for Continuous Pitch and Amplitude Tracking in Realtime Interactive Improvisation Software“, an article I wrote for the 2004 Sound and Music Computing conference in Paris.

Dobrian’s early algorithmic compositions

When I started learning about computer music in 1984, I became interested in the challenge of trying to program the computer to compose music. My first piece of computer music was an experiment in algorithmic composition focusing on automated composition of metric modulations. Since a metric modulation is essentially a mathematical operation, it seemed reasonable that a computer could calculate and perform such tempo changes more readily than humans. So I set out to write a computer program that would compose and synthesize music that included metric modulations.

The problem that I immediately confronted, though, is how to establish a clear sense of tempo and beat for the listener that would be musically interesting. It’s one thing to use standard Western notation to compose metric modulations on paper, but when there’s no paper or notation involved, and no human performers to add inflection to the music, it’s another thing to figure out exactly how to make the beat evident to the listener so that the metric modulation would be clearly audible.

I was obliged to answer the question, usually taken for granted by musicians, “What are the factors that give us a sense of beat?” The most obvious one is that things happen periodically at a regular rate somewhere roughly in the range of walking/running speed, say from 48 to 144 beats per minute. But music consists of more than simply a single constant rhythm, it contains a wide variety of rhythms. So how do we derive the beat? We use other factors that exhibit regularity such as dynamic accent (loudness), timbral accent (choice of instrument, brightness), melodic contour (a repeating shape), harmonic rhythm (implied chord changes), and harmonically-related beat divisions (triplets, sixteenths, quintuplets, etc.), and we take all of those into consideration simultaneously to figure out rates of repetition that might be considered the beat. (There are also stylistic considerations; we may be familiar with a body of works that all belong to the same style or genre, so that we have established cultural cues about where the beat is considered to be in that type of music.)

So I wrote a program in C that composed phrases of music, and that made probabilistic decisions along the way about how to modulate to a new harmonically-related tempo using beat divisions of the same speed in each tempo. It used factors of melodic contour, dynamic accent, and timbral accent to provide a sense of periodicity. (It incidentally also used continuously changing pitch ranges and continuously changing tuning systems, so that the music has a unique inharmonic character.) The program produced a type of musical score for each phrase of music, a time-tagged list of notes that would be used in the Cmusic sound synthesis language to synthesize the notes on a VAX 11/780 (Unix) computer and store the results in a sound file. (The sampling rate was a mere 16 kHz, but that was barely adequate for the sounds being synthesized.)

The composition was titled Degueudoudeloupe (1985), a nonsense word I thought was evocative of the kinds of sounds and rhythms in the piece. The sounds in the piece are string-like, drum-like, and bell-like sounds synthesized using two methods:  Karplus-Strong synthesis and Frequency Modulation synthesis.


In another composition several years later I again used probabilistic computer decision making to compose time-tagged lists of notes, this time intended to be transmitted as MIDI note data to a Yamaha Disklavier computer-controlled piano. The program was written in the C language, and the MIDI data that that program produced was saved as a MIDI file on a Macintosh Plus computer and transmitted to the Disklavier.

The piece is titled Entropy (1991), and was largely inspired by the following passage from The Open Work by Umberto Eco:

Consider the chaotic effect (resulting from a sudden imposition of uniformity) of a strong wind on the innumerable grains of sand that compose a beach: amid this confusion, the action of a human foot on the surface of the beach constitutes a complex interaction of events that leads to the statistically very improbable configuration of a footprint. The organization of events that has produced this configuration, this form, is only temporary: the footprint will soon be swept away by the wind. In other words, a deviation from the general entropy curve (consisting of a decrease in entropy and the establishment of improbable order) will generally tend to be reabsorbed into the universal curve of increasing entropy. And yet, for a moment, the elemental chaos of this system has made room for the appearance of an order…

The composition was conceived as a vehicle with which to explore ideas of information theory and stochasticism in an artistic way. It explores the perception of randomness and order (entropy and negentropy) in musical structure, and demonstrates the use of stochasticism not only as a model for the distribution of sounds in time, but also as a method of variation of a harmonic “order”.

The notes of this piece were all chosen by a computer algorithm written by the composer. The algorithm takes as its input a description of some beginning and ending characteristics of a musical phrase, and outputs the note information necessary to realize a continuous transformation from the beginning to the ending state. Such a transformation can take place over any period of time desired by the composer (in this piece anywhere from 3 to 90 seconds). The input description is stated in terms of relative probabilities of different musical occurrences, thus allowing the composer to describe music which ranges between totally predictable (negentropic) and totally unpredictable (entropic), and which can transform gradually or suddenly from one to the other.