Music Processing - Petra Christian University
Download
Report
Transcript Music Processing - Petra Christian University
Music
Processing
Roger B. Dannenber
Overview
Music Representation
MIDI and Synthesizers
Synthesis Techniques
Music Understanding
Music Representation
Acoustic Level: sound, samples,
spectra
Performance Information: timing,
parameters
Notation Information: parts, clefs,
stem direction
Compositional Structure: notes,
chords, symbolic structure
Performance Information
MIDI bandwidth is 3KB/s, or
180KB/min
More typical: 3KB/minute,
180KB/hour
• Complete Scott Joplin: 1MB
• Output of 50 Composers (400 days of
music):
500MB (1 CD-ROM)
Synthesis of acoustic instruments is
a problem
Music Notation
Compact, symbolic representation
Does not capture performance
information
Expressive “performance” not fully
automated
Compositional Structure
Example: Nyquist (free software!)
(defun melody1 ()
(seq (stretch q (note a4) (note b4)
(note cs5) (note d5))))
(defun counterpoint () …)
(defun composition ()
(sim (melody1) (counterpoint)))
(play (transpose 4 (composition)))
MIDI: Musical Instrument Digital
Interface
Musical Performance Information:
• Piano Keyboard key presses and
releases
• “instrument” selection (by number)
• sustain pedal, switches
• continuous controls: volume pedal,
pitch bend, aftertouch
• very compact
(human gesture < 100Hz bandwidth)
MIDI (cont’d)
Point-to-point connections:
• MIDI IN, OUT, THRU
• Channels
No time stamps
• (almost) everything happens in real
time
Asynchronous serial,
8-bit bytes+start+stop bits,
31.25K baud = 1MHz/32
MIDI Message Formats
Key Up
8 ch
key#
vel
Key Down
9 ch
key#
vel
Polyphonic Aftertouch
A ch
key#
press
Control Change
B ch
ctrl#
value
Program Change
C ch index#
Channel Aftertouch
D ch
press
Pitch Bend
E ch
lo 7
System Exclusive
… DATA …
F 0
F
hi 7
E
Standard MIDI Files
Key point: Must encode timing
information
Interleave time differences with MIDI
data...
<track data> =1 or more <track event>,
<track event> = <delta time> <event>,
<event> = midi data or <meta event>,
<meta event> = FF<type><length><data>
Delta times use variable length
encoding, omit for zero.
Music Synthesis Introduction
Primary issue is control
• No control Digital Audio (start, stop,
...)
• Complete control Digital Audio
(S[0], S[1], S[2], ... )
• Parametric control Synthesis
Music Synthesis Introduction
(cont’d)
What parameters?
• pitch
• loudness
• timbre (e.g. which instrument)
• articulation, expression, vibrato, etc.
• spatial effects (e.g. reverberation)
Why synthesize?
• high-level representation provides
precision of specification and
supports interactivity
Additive Synthesis
n1
Si A j sin j i t
j0
amplitude A[i] and frequency [i]
specified for each partial (sinusoidal
component)
potentially 2n more control samples
than signal samples!
Additive Synthesis (cont’d)
often use piece-wise linear control
envelopes to save space
still difficult to control because of so
many parameters
and parameters do not match
perceptual attributes
Table-Lookup Oscillators
If signal is periodic, store one period
Control parameters: pitch, amplitude, wavefor
Frequency
+
Efficient, but ...
Spectrum is static
Phase
Amplitude
x
(Note that
phase and
frequency are
fixed point or
floating point
numbers)
FM Synthesis
MOD FREQ
A
AMPL
+
A
F
F
Usually use sinusoids
“carrier” and “modulator”
are both at audio
frequencies
If frequencies are simple
ratio (R), output spectrum
is periodic
Output varies from
sinusoid to complex signal
as MOD increases
out = AMPL·sin(2·FREQ·t + MOD sin(2R·FREQ·t))
FM Synthesis (cont’d)
Interesting sounds,
Time-varying spectra, and ...
Low computation requirements
Often uses more than 2 oscillators
… but …
Hard to recreate a specific waveform
No successful analysis procedure
Sample-based Synthesis
Samplers store waveforms for
playback
Sounds are “looped” to extend
duration
Spectrum is static (as in tablelookup), so:
• different samples are used for
different pitches
• simple effects are added: filter,
vibrato,
envelope
Attack amplitude
Loop
Loop again ...
Physical Models
Additive, FM, and sampling:
more-or-less perception-based.
Physical Modeling is source-based:
compute the wave equation, simulate
attached reeds, bows, etc.
Example:
Reed
Bore
Bell
Physical Models (cont’d)
Difficult to control, and ...
Can be very computationally
intensive
… but ...
Produce “characteristic” acoustic
sounds.
Music Understanding
Introduction
Score Following, Computer
Accompaniment
Interactive Performance
Style Recognition
Conclusions